summarylogtreecommitdiffstats
diff options
context:
space:
mode:
authorJerry Lin2018-11-03 16:50:42 +0800
committerJerry Lin2018-11-03 16:50:42 +0800
commit369e1c7ca2382b09bd12aa292803db4b26b23eb4 (patch)
tree529757a06facf61450498e4e032dd60874020349
parent9f9fa15ad814fd6411063537b7c775c7ef12061a (diff)
downloadaur-369e1c7ca2382b09bd12aa292803db4b26b23eb4.tar.gz
Fix undefined functions from console_bridge. Update version to 1.11.21-1.
-rw-r--r--.SRCINFO10
-rw-r--r--PKGBUILD15
-rw-r--r--fix-console-bridge-functions.patch305
3 files changed, 322 insertions, 8 deletions
diff --git a/.SRCINFO b/.SRCINFO
index a60ec45ef62b..26031ca1d9b0 100644
--- a/.SRCINFO
+++ b/.SRCINFO
@@ -1,8 +1,6 @@
-# Generated by mksrcinfo v8
-# Sun May 22 20:54:12 UTC 2016
pkgbase = ros-indigo-rosbag-storage
pkgdesc = ROS - This is a set of tools for recording from and playing back ROS message without relying on the ROS client library.
- pkgver = 1.11.19
+ pkgver = 1.11.21
pkgrel = 1
url = http://www.ros.org/
arch = any
@@ -26,8 +24,10 @@ pkgbase = ros-indigo-rosbag-storage
depends = bzip2
depends = boost
depends = console-bridge
- source = ros-indigo-rosbag-storage-1.11.19-0.tar.gz::https://github.com/ros-gbp/ros_comm-release/archive/release/indigo/rosbag_storage/1.11.19-0.tar.gz
- sha256sums = 70358a5ac70ec401b1c4ee681a3f1e7d7c3c72d874d459a563505a4471bb193b
+ source = ros-indigo-rosbag-storage-1.11.21-0.tar.gz::https://github.com/ros-gbp/ros_comm-release/archive/release/indigo/rosbag_storage/1.11.21-0.tar.gz
+ source = fix-console-bridge-functions.patch
+ sha256sums = 494619a8a8656bc5a0e85c7dcf605c22aa76f2d107e440180baf450a8b9c26e2
+ sha256sums = 714778aab69b69992819217898447628ac92b6efc22c6bc5d5e90a5907683f8e
pkgname = ros-indigo-rosbag-storage
diff --git a/PKGBUILD b/PKGBUILD
index a7d999520109..6fc91f2e01af 100644
--- a/PKGBUILD
+++ b/PKGBUILD
@@ -4,7 +4,7 @@ pkgdesc="ROS - This is a set of tools for recording from and playing back ROS me
url='http://www.ros.org/'
pkgname='ros-indigo-rosbag-storage'
-pkgver='1.11.19'
+pkgver='1.11.21'
_pkgver_patch=0
arch=('any')
pkgrel=1
@@ -40,8 +40,17 @@ depends=(${ros_depends[@]}
# Tarball version (faster download)
_dir="ros_comm-release-release-indigo-rosbag_storage-${pkgver}-${_pkgver_patch}"
-source=("${pkgname}-${pkgver}-${_pkgver_patch}.tar.gz"::"https://github.com/ros-gbp/ros_comm-release/archive/release/indigo/rosbag_storage/${pkgver}-${_pkgver_patch}.tar.gz")
-sha256sums=('70358a5ac70ec401b1c4ee681a3f1e7d7c3c72d874d459a563505a4471bb193b')
+source=(
+ "${pkgname}-${pkgver}-${_pkgver_patch}.tar.gz"::"https://github.com/ros-gbp/ros_comm-release/archive/release/indigo/rosbag_storage/${pkgver}-${_pkgver_patch}.tar.gz"
+ 'fix-console-bridge-functions.patch'
+)
+sha256sums=('494619a8a8656bc5a0e85c7dcf605c22aa76f2d107e440180baf450a8b9c26e2'
+ '714778aab69b69992819217898447628ac92b6efc22c6bc5d5e90a5907683f8e')
+
+prepare () {
+ cd ${srcdir}/${_dir}
+ patch -Np1 -i ${srcdir}/fix-console-bridge-functions.patch
+}
build() {
# Use ROS environment variables
diff --git a/fix-console-bridge-functions.patch b/fix-console-bridge-functions.patch
new file mode 100644
index 000000000000..b0e15e6ce415
--- /dev/null
+++ b/fix-console-bridge-functions.patch
@@ -0,0 +1,305 @@
+diff --git a/include/rosbag/bag.h b/include/rosbag/bag.h
+index 0696d72..f6d688e 100644
+--- a/include/rosbag/bag.h
++++ b/include/rosbag/bag.h
+@@ -564,7 +564,7 @@ void Bag::doWrite(std::string const& topic, ros::Time const& time, T const& msg,
+
+ // Check if we want to stop this chunk
+ uint32_t chunk_size = getChunkOffset();
+- logDebug(" curr_chunk_size=%d (threshold=%d)", chunk_size, chunk_threshold_);
++ CONSOLE_BRIDGE_logDebug(" curr_chunk_size=%d (threshold=%d)", chunk_size, chunk_threshold_);
+ if (chunk_size > chunk_threshold_) {
+ // Empty the outgoing chunk
+ stopWritingChunk();
+@@ -599,7 +599,7 @@ void Bag::writeMessageDataRecord(uint32_t conn_id, ros::Time const& time, T cons
+ seek(0, std::ios::end);
+ file_size_ = file_.getOffset();
+
+- logDebug("Writing MSG_DATA [%llu:%d]: conn=%d sec=%d nsec=%d data_len=%d",
++ CONSOLE_BRIDGE_logDebug("Writing MSG_DATA [%llu:%d]: conn=%d sec=%d nsec=%d data_len=%d",
+ (unsigned long long) file_.getOffset(), getChunkOffset(), conn_id, time.sec, time.nsec, msg_ser_len);
+
+ writeHeader(header);
+diff --git a/src/bag.cpp b/src/bag.cpp
+index 62cd953..b2b9bd2 100644
+--- a/src/bag.cpp
++++ b/src/bag.cpp
+@@ -213,7 +213,7 @@ void Bag::setCompression(CompressionType compression) {
+ void Bag::writeVersion() {
+ string version = string("#ROSBAG V") + VERSION + string("\n");
+
+- logDebug("Writing VERSION [%llu]: %s", (unsigned long long) file_.getOffset(), version.c_str());
++ CONSOLE_BRIDGE_logDebug("Writing VERSION [%llu]: %s", (unsigned long long) file_.getOffset(), version.c_str());
+
+ version_ = 200;
+
+@@ -236,7 +236,7 @@ void Bag::readVersion() {
+
+ version_ = version_major * 100 + version_minor;
+
+- logDebug("Read VERSION: version=%d", version_);
++ CONSOLE_BRIDGE_logDebug("Read VERSION: version=%d", version_);
+ }
+
+ uint32_t Bag::getMajorVersion() const { return version_ / 100; }
+@@ -325,7 +325,7 @@ void Bag::startReadingVersion102() {
+ multiset<IndexEntry> const& index = i->second;
+ IndexEntry const& first_entry = *index.begin();
+
+- logDebug("Reading message definition for connection %d at %llu", i->first, (unsigned long long) first_entry.chunk_pos);
++ CONSOLE_BRIDGE_logDebug("Reading message definition for connection %d at %llu", i->first, (unsigned long long) first_entry.chunk_pos);
+
+ seek(first_entry.chunk_pos);
+
+@@ -339,7 +339,7 @@ void Bag::writeFileHeaderRecord() {
+ connection_count_ = connections_.size();
+ chunk_count_ = chunks_.size();
+
+- logDebug("Writing FILE_HEADER [%llu]: index_pos=%llu connection_count=%d chunk_count=%d",
++ CONSOLE_BRIDGE_logDebug("Writing FILE_HEADER [%llu]: index_pos=%llu connection_count=%d chunk_count=%d",
+ (unsigned long long) file_.getOffset(), (unsigned long long) index_data_pos_, connection_count_, chunk_count_);
+
+ // Write file header record
+@@ -390,7 +390,7 @@ void Bag::readFileHeaderRecord() {
+ readField(fields, CHUNK_COUNT_FIELD_NAME, true, &chunk_count_);
+ }
+
+- logDebug("Read FILE_HEADER: index_pos=%llu connection_count=%d chunk_count=%d",
++ CONSOLE_BRIDGE_logDebug("Read FILE_HEADER: index_pos=%llu connection_count=%d chunk_count=%d",
+ (unsigned long long) index_data_pos_, connection_count_, chunk_count_);
+
+ // Skip the data section (just padding)
+@@ -460,7 +460,7 @@ void Bag::writeChunkHeader(CompressionType compression, uint32_t compressed_size
+ chunk_header.compressed_size = compressed_size;
+ chunk_header.uncompressed_size = uncompressed_size;
+
+- logDebug("Writing CHUNK [%llu]: compression=%s compressed=%d uncompressed=%d",
++ CONSOLE_BRIDGE_logDebug("Writing CHUNK [%llu]: compression=%s compressed=%d uncompressed=%d",
+ (unsigned long long) file_.getOffset(), chunk_header.compression.c_str(), chunk_header.compressed_size, chunk_header.uncompressed_size);
+
+ M_string header;
+@@ -485,7 +485,7 @@ void Bag::readChunkHeader(ChunkHeader& chunk_header) const {
+ readField(fields, COMPRESSION_FIELD_NAME, true, chunk_header.compression);
+ readField(fields, SIZE_FIELD_NAME, true, &chunk_header.uncompressed_size);
+
+- logDebug("Read CHUNK: compression=%s size=%d uncompressed=%d (%f)", chunk_header.compression.c_str(), chunk_header.compressed_size, chunk_header.uncompressed_size, 100 * ((double) chunk_header.compressed_size) / chunk_header.uncompressed_size);
++ CONSOLE_BRIDGE_logDebug("Read CHUNK: compression=%s size=%d uncompressed=%d (%f)", chunk_header.compression.c_str(), chunk_header.compressed_size, chunk_header.uncompressed_size, 100 * ((double) chunk_header.compressed_size) / chunk_header.uncompressed_size);
+ }
+
+ // Index records
+@@ -506,7 +506,7 @@ void Bag::writeIndexRecords() {
+
+ writeDataLength(index_size * 12);
+
+- logDebug("Writing INDEX_DATA: connection=%d ver=%d count=%d", connection_id, INDEX_VERSION, index_size);
++ CONSOLE_BRIDGE_logDebug("Writing INDEX_DATA: connection=%d ver=%d count=%d", connection_id, INDEX_VERSION, index_size);
+
+ // Write the index record data (pairs of timestamp and position in file)
+ foreach(IndexEntry const& e, index) {
+@@ -514,7 +514,7 @@ void Bag::writeIndexRecords() {
+ write((char*) &e.time.nsec, 4);
+ write((char*) &e.offset, 4);
+
+- logDebug(" - %d.%d: %d", e.time.sec, e.time.nsec, e.offset);
++ CONSOLE_BRIDGE_logDebug(" - %d.%d: %d", e.time.sec, e.time.nsec, e.offset);
+ }
+ }
+ }
+@@ -536,7 +536,7 @@ void Bag::readTopicIndexRecord102() {
+ readField(fields, TOPIC_FIELD_NAME, true, topic);
+ readField(fields, COUNT_FIELD_NAME, true, &count);
+
+- logDebug("Read INDEX_DATA: ver=%d topic=%s count=%d", index_version, topic.c_str(), count);
++ CONSOLE_BRIDGE_logDebug("Read INDEX_DATA: ver=%d topic=%s count=%d", index_version, topic.c_str(), count);
+
+ if (index_version != 0)
+ throw BagFormatException((format("Unsupported INDEX_DATA version: %1%") % index_version).str());
+@@ -546,7 +546,7 @@ void Bag::readTopicIndexRecord102() {
+ if (topic_conn_id_iter == topic_connection_ids_.end()) {
+ connection_id = connections_.size();
+
+- logDebug("Creating connection: id=%d topic=%s", connection_id, topic.c_str());
++ CONSOLE_BRIDGE_logDebug("Creating connection: id=%d topic=%s", connection_id, topic.c_str());
+ ConnectionInfo* connection_info = new ConnectionInfo();
+ connection_info->id = connection_id;
+ connection_info->topic = topic;
+@@ -569,11 +569,11 @@ void Bag::readTopicIndexRecord102() {
+ index_entry.time = Time(sec, nsec);
+ index_entry.offset = 0;
+
+- logDebug(" - %d.%d: %llu", sec, nsec, (unsigned long long) index_entry.chunk_pos);
++ CONSOLE_BRIDGE_logDebug(" - %d.%d: %llu", sec, nsec, (unsigned long long) index_entry.chunk_pos);
+
+ if (index_entry.time < ros::TIME_MIN || index_entry.time > ros::TIME_MAX)
+ {
+- logError("Index entry for topic %s contains invalid time.", topic.c_str());
++ CONSOLE_BRIDGE_logError("Index entry for topic %s contains invalid time.", topic.c_str());
+ } else
+ {
+ connection_index.insert(connection_index.end(), index_entry);
+@@ -598,7 +598,7 @@ void Bag::readConnectionIndexRecord200() {
+ readField(fields, CONNECTION_FIELD_NAME, true, &connection_id);
+ readField(fields, COUNT_FIELD_NAME, true, &count);
+
+- logDebug("Read INDEX_DATA: ver=%d connection=%d count=%d", index_version, connection_id, count);
++ CONSOLE_BRIDGE_logDebug("Read INDEX_DATA: ver=%d connection=%d count=%d", index_version, connection_id, count);
+
+ if (index_version != 1)
+ throw BagFormatException((format("Unsupported INDEX_DATA version: %1%") % index_version).str());
+@@ -617,11 +617,11 @@ void Bag::readConnectionIndexRecord200() {
+ read((char*) &index_entry.offset, 4);
+ index_entry.time = Time(sec, nsec);
+
+- logDebug(" - %d.%d: %llu+%d", sec, nsec, (unsigned long long) index_entry.chunk_pos, index_entry.offset);
++ CONSOLE_BRIDGE_logDebug(" - %d.%d: %llu+%d", sec, nsec, (unsigned long long) index_entry.chunk_pos, index_entry.offset);
+
+ if (index_entry.time < ros::TIME_MIN || index_entry.time > ros::TIME_MAX)
+ {
+- logError("Index entry for topic %s contains invalid time. This message will not be loaded.", connections_[connection_id]->topic.c_str());
++ CONSOLE_BRIDGE_logError("Index entry for topic %s contains invalid time. This message will not be loaded.", connections_[connection_id]->topic.c_str());
+ } else
+ {
+ connection_index.insert(connection_index.end(), index_entry);
+@@ -639,7 +639,7 @@ void Bag::writeConnectionRecords() {
+ }
+
+ void Bag::writeConnectionRecord(ConnectionInfo const* connection_info) {
+- logDebug("Writing CONNECTION [%llu:%d]: topic=%s id=%d",
++ CONSOLE_BRIDGE_logDebug("Writing CONNECTION [%llu:%d]: topic=%s id=%d",
+ (unsigned long long) file_.getOffset(), getChunkOffset(), connection_info->topic.c_str(), connection_info->id);
+
+ M_string header;
+@@ -693,7 +693,7 @@ void Bag::readConnectionRecord() {
+ connection_info->md5sum = (*connection_info->header)["md5sum"];
+ connections_[id] = connection_info;
+
+- logDebug("Read CONNECTION: topic=%s id=%d", topic.c_str(), id);
++ CONSOLE_BRIDGE_logDebug("Read CONNECTION: topic=%s id=%d", topic.c_str(), id);
+ }
+ }
+
+@@ -719,7 +719,7 @@ void Bag::readMessageDefinitionRecord102() {
+ if (topic_conn_id_iter == topic_connection_ids_.end()) {
+ uint32_t id = connections_.size();
+
+- logDebug("Creating connection: topic=%s md5sum=%s datatype=%s", topic.c_str(), md5sum.c_str(), datatype.c_str());
++ CONSOLE_BRIDGE_logDebug("Creating connection: topic=%s md5sum=%s datatype=%s", topic.c_str(), md5sum.c_str(), datatype.c_str());
+ connection_info = new ConnectionInfo();
+ connection_info->id = id;
+ connection_info->topic = topic;
+@@ -738,7 +738,7 @@ void Bag::readMessageDefinitionRecord102() {
+ (*connection_info->header)["md5sum"] = connection_info->md5sum;
+ (*connection_info->header)["message_definition"] = connection_info->msg_def;
+
+- logDebug("Read MSG_DEF: topic=%s md5sum=%s datatype=%s", topic.c_str(), md5sum.c_str(), datatype.c_str());
++ CONSOLE_BRIDGE_logDebug("Read MSG_DEF: topic=%s md5sum=%s datatype=%s", topic.c_str(), md5sum.c_str(), datatype.c_str());
+ }
+
+ void Bag::decompressChunk(uint64_t chunk_pos) const {
+@@ -773,7 +773,7 @@ void Bag::decompressChunk(uint64_t chunk_pos) const {
+ }
+
+ void Bag::readMessageDataRecord102(uint64_t offset, ros::Header& header) const {
+- logDebug("readMessageDataRecord: offset=%llu", (unsigned long long) offset);
++ CONSOLE_BRIDGE_logDebug("readMessageDataRecord: offset=%llu", (unsigned long long) offset);
+
+ seek(offset);
+
+@@ -799,7 +799,7 @@ void Bag::decompressRawChunk(ChunkHeader const& chunk_header) const {
+ assert(chunk_header.compression == COMPRESSION_NONE);
+ assert(chunk_header.compressed_size == chunk_header.uncompressed_size);
+
+- logDebug("compressed_size: %d uncompressed_size: %d", chunk_header.compressed_size, chunk_header.uncompressed_size);
++ CONSOLE_BRIDGE_logDebug("compressed_size: %d uncompressed_size: %d", chunk_header.compressed_size, chunk_header.uncompressed_size);
+
+ decompress_buffer_.setSize(chunk_header.compressed_size);
+ file_.read((char*) decompress_buffer_.getData(), chunk_header.compressed_size);
+@@ -812,7 +812,7 @@ void Bag::decompressBz2Chunk(ChunkHeader const& chunk_header) const {
+
+ CompressionType compression = compression::BZ2;
+
+- logDebug("compressed_size: %d uncompressed_size: %d", chunk_header.compressed_size, chunk_header.uncompressed_size);
++ CONSOLE_BRIDGE_logDebug("compressed_size: %d uncompressed_size: %d", chunk_header.compressed_size, chunk_header.uncompressed_size);
+
+ chunk_buffer_.setSize(chunk_header.compressed_size);
+ file_.read((char*) chunk_buffer_.getData(), chunk_header.compressed_size);
+@@ -828,7 +828,7 @@ void Bag::decompressLz4Chunk(ChunkHeader const& chunk_header) const {
+
+ CompressionType compression = compression::LZ4;
+
+- logDebug("lz4 compressed_size: %d uncompressed_size: %d",
++ CONSOLE_BRIDGE_logDebug("lz4 compressed_size: %d uncompressed_size: %d",
+ chunk_header.compressed_size, chunk_header.uncompressed_size);
+
+ chunk_buffer_.setSize(chunk_header.compressed_size);
+@@ -889,7 +889,7 @@ void Bag::writeChunkInfoRecords() {
+ header[END_TIME_FIELD_NAME] = toHeaderString(&chunk_info.end_time);
+ header[COUNT_FIELD_NAME] = toHeaderString(&chunk_connection_count);
+
+- logDebug("Writing CHUNK_INFO [%llu]: ver=%d pos=%llu start=%d.%d end=%d.%d",
++ CONSOLE_BRIDGE_logDebug("Writing CHUNK_INFO [%llu]: ver=%d pos=%llu start=%d.%d end=%d.%d",
+ (unsigned long long) file_.getOffset(), CHUNK_INFO_VERSION, (unsigned long long) chunk_info.pos,
+ chunk_info.start_time.sec, chunk_info.start_time.nsec,
+ chunk_info.end_time.sec, chunk_info.end_time.nsec);
+@@ -906,7 +906,7 @@ void Bag::writeChunkInfoRecords() {
+ write((char*) &connection_id, 4);
+ write((char*) &count, 4);
+
+- logDebug(" - %d: %d", connection_id, count);
++ CONSOLE_BRIDGE_logDebug(" - %d: %d", connection_id, count);
+ }
+ }
+ }
+@@ -935,7 +935,7 @@ void Bag::readChunkInfoRecord() {
+ uint32_t chunk_connection_count = 0;
+ readField(fields, COUNT_FIELD_NAME, true, &chunk_connection_count);
+
+- logDebug("Read CHUNK_INFO: chunk_pos=%llu connection_count=%d start=%d.%d end=%d.%d",
++ CONSOLE_BRIDGE_logDebug("Read CHUNK_INFO: chunk_pos=%llu connection_count=%d start=%d.%d end=%d.%d",
+ (unsigned long long) chunk_info.pos, chunk_connection_count,
+ chunk_info.start_time.sec, chunk_info.start_time.nsec,
+ chunk_info.end_time.sec, chunk_info.end_time.nsec);
+@@ -946,7 +946,7 @@ void Bag::readChunkInfoRecord() {
+ read((char*) &connection_id, 4);
+ read((char*) &connection_count, 4);
+
+- logDebug(" %d: %d messages", connection_id, connection_count);
++ CONSOLE_BRIDGE_logDebug(" %d: %d messages", connection_id, connection_count);
+
+ chunk_info.connection_counts[connection_id] = connection_count;
+ }
+@@ -1028,7 +1028,7 @@ void Bag::readMessageDataHeaderFromBuffer(Buffer& buffer, uint32_t offset, ros::
+ total_bytes_read = 0;
+ uint8_t op = 0xFF;
+ do {
+- logDebug("reading header from buffer: offset=%d", offset);
++ CONSOLE_BRIDGE_logDebug("reading header from buffer: offset=%d", offset);
+ uint32_t bytes_read;
+ readHeaderFromBuffer(*current_buffer_, offset, header, data_size, bytes_read);
+
+diff --git a/src/bz2_stream.cpp b/src/bz2_stream.cpp
+index 900cd04..d158d1d 100644
+--- a/src/bz2_stream.cpp
++++ b/src/bz2_stream.cpp
+@@ -115,7 +115,7 @@ void BZ2Stream::read(void* ptr, size_t size) {
+ case BZ_OK: return;
+ case BZ_STREAM_END:
+ if (getUnused() || getUnusedLength() > 0)
+- logError("unused data already available");
++ CONSOLE_BRIDGE_logError("unused data already available");
+ else {
+ char* unused;
+ int nUnused;
+diff --git a/src/lz4_stream.cpp b/src/lz4_stream.cpp
+index 9502664..41ec0df 100644
+--- a/src/lz4_stream.cpp
++++ b/src/lz4_stream.cpp
+@@ -155,7 +155,7 @@ void LZ4Stream::read(void* ptr, size_t size) {
+ case ROSLZ4_OK: break;
+ case ROSLZ4_STREAM_END:
+ if (getUnused() || getUnusedLength() > 0)
+- logError("unused data already available");
++ CONSOLE_BRIDGE_logError("unused data already available");
+ else {
+ setUnused(lz4s_.input_next);
+ setUnusedLength(lz4s_.input_left);