diff options
author | Jerry Lin | 2018-11-03 16:50:42 +0800 |
---|---|---|
committer | Jerry Lin | 2018-11-03 16:50:42 +0800 |
commit | 369e1c7ca2382b09bd12aa292803db4b26b23eb4 (patch) | |
tree | 529757a06facf61450498e4e032dd60874020349 | |
parent | 9f9fa15ad814fd6411063537b7c775c7ef12061a (diff) | |
download | aur-ros-indigo-rosbag-storage.tar.gz |
Fix undefined functions from console_bridge. Update version to 1.11.21-1.
-rw-r--r-- | .SRCINFO | 10 | ||||
-rw-r--r-- | PKGBUILD | 15 | ||||
-rw-r--r-- | fix-console-bridge-functions.patch | 305 |
3 files changed, 322 insertions, 8 deletions
@@ -1,8 +1,6 @@ -# Generated by mksrcinfo v8 -# Sun May 22 20:54:12 UTC 2016 pkgbase = ros-indigo-rosbag-storage pkgdesc = ROS - This is a set of tools for recording from and playing back ROS message without relying on the ROS client library. - pkgver = 1.11.19 + pkgver = 1.11.21 pkgrel = 1 url = http://www.ros.org/ arch = any @@ -26,8 +24,10 @@ pkgbase = ros-indigo-rosbag-storage depends = bzip2 depends = boost depends = console-bridge - source = ros-indigo-rosbag-storage-1.11.19-0.tar.gz::https://github.com/ros-gbp/ros_comm-release/archive/release/indigo/rosbag_storage/1.11.19-0.tar.gz - sha256sums = 70358a5ac70ec401b1c4ee681a3f1e7d7c3c72d874d459a563505a4471bb193b + source = ros-indigo-rosbag-storage-1.11.21-0.tar.gz::https://github.com/ros-gbp/ros_comm-release/archive/release/indigo/rosbag_storage/1.11.21-0.tar.gz + source = fix-console-bridge-functions.patch + sha256sums = 494619a8a8656bc5a0e85c7dcf605c22aa76f2d107e440180baf450a8b9c26e2 + sha256sums = 714778aab69b69992819217898447628ac92b6efc22c6bc5d5e90a5907683f8e pkgname = ros-indigo-rosbag-storage @@ -4,7 +4,7 @@ pkgdesc="ROS - This is a set of tools for recording from and playing back ROS me url='http://www.ros.org/' pkgname='ros-indigo-rosbag-storage' -pkgver='1.11.19' +pkgver='1.11.21' _pkgver_patch=0 arch=('any') pkgrel=1 @@ -40,8 +40,17 @@ depends=(${ros_depends[@]} # Tarball version (faster download) _dir="ros_comm-release-release-indigo-rosbag_storage-${pkgver}-${_pkgver_patch}" -source=("${pkgname}-${pkgver}-${_pkgver_patch}.tar.gz"::"https://github.com/ros-gbp/ros_comm-release/archive/release/indigo/rosbag_storage/${pkgver}-${_pkgver_patch}.tar.gz") -sha256sums=('70358a5ac70ec401b1c4ee681a3f1e7d7c3c72d874d459a563505a4471bb193b') +source=( + "${pkgname}-${pkgver}-${_pkgver_patch}.tar.gz"::"https://github.com/ros-gbp/ros_comm-release/archive/release/indigo/rosbag_storage/${pkgver}-${_pkgver_patch}.tar.gz" + 'fix-console-bridge-functions.patch' +) +sha256sums=('494619a8a8656bc5a0e85c7dcf605c22aa76f2d107e440180baf450a8b9c26e2' + '714778aab69b69992819217898447628ac92b6efc22c6bc5d5e90a5907683f8e') + +prepare () { + cd ${srcdir}/${_dir} + patch -Np1 -i ${srcdir}/fix-console-bridge-functions.patch +} build() { # Use ROS environment variables diff --git a/fix-console-bridge-functions.patch b/fix-console-bridge-functions.patch new file mode 100644 index 000000000000..b0e15e6ce415 --- /dev/null +++ b/fix-console-bridge-functions.patch @@ -0,0 +1,305 @@ +diff --git a/include/rosbag/bag.h b/include/rosbag/bag.h +index 0696d72..f6d688e 100644 +--- a/include/rosbag/bag.h ++++ b/include/rosbag/bag.h +@@ -564,7 +564,7 @@ void Bag::doWrite(std::string const& topic, ros::Time const& time, T const& msg, + + // Check if we want to stop this chunk + uint32_t chunk_size = getChunkOffset(); +- logDebug(" curr_chunk_size=%d (threshold=%d)", chunk_size, chunk_threshold_); ++ CONSOLE_BRIDGE_logDebug(" curr_chunk_size=%d (threshold=%d)", chunk_size, chunk_threshold_); + if (chunk_size > chunk_threshold_) { + // Empty the outgoing chunk + stopWritingChunk(); +@@ -599,7 +599,7 @@ void Bag::writeMessageDataRecord(uint32_t conn_id, ros::Time const& time, T cons + seek(0, std::ios::end); + file_size_ = file_.getOffset(); + +- logDebug("Writing MSG_DATA [%llu:%d]: conn=%d sec=%d nsec=%d data_len=%d", ++ CONSOLE_BRIDGE_logDebug("Writing MSG_DATA [%llu:%d]: conn=%d sec=%d nsec=%d data_len=%d", + (unsigned long long) file_.getOffset(), getChunkOffset(), conn_id, time.sec, time.nsec, msg_ser_len); + + writeHeader(header); +diff --git a/src/bag.cpp b/src/bag.cpp +index 62cd953..b2b9bd2 100644 +--- a/src/bag.cpp ++++ b/src/bag.cpp +@@ -213,7 +213,7 @@ void Bag::setCompression(CompressionType compression) { + void Bag::writeVersion() { + string version = string("#ROSBAG V") + VERSION + string("\n"); + +- logDebug("Writing VERSION [%llu]: %s", (unsigned long long) file_.getOffset(), version.c_str()); ++ CONSOLE_BRIDGE_logDebug("Writing VERSION [%llu]: %s", (unsigned long long) file_.getOffset(), version.c_str()); + + version_ = 200; + +@@ -236,7 +236,7 @@ void Bag::readVersion() { + + version_ = version_major * 100 + version_minor; + +- logDebug("Read VERSION: version=%d", version_); ++ CONSOLE_BRIDGE_logDebug("Read VERSION: version=%d", version_); + } + + uint32_t Bag::getMajorVersion() const { return version_ / 100; } +@@ -325,7 +325,7 @@ void Bag::startReadingVersion102() { + multiset<IndexEntry> const& index = i->second; + IndexEntry const& first_entry = *index.begin(); + +- logDebug("Reading message definition for connection %d at %llu", i->first, (unsigned long long) first_entry.chunk_pos); ++ CONSOLE_BRIDGE_logDebug("Reading message definition for connection %d at %llu", i->first, (unsigned long long) first_entry.chunk_pos); + + seek(first_entry.chunk_pos); + +@@ -339,7 +339,7 @@ void Bag::writeFileHeaderRecord() { + connection_count_ = connections_.size(); + chunk_count_ = chunks_.size(); + +- logDebug("Writing FILE_HEADER [%llu]: index_pos=%llu connection_count=%d chunk_count=%d", ++ CONSOLE_BRIDGE_logDebug("Writing FILE_HEADER [%llu]: index_pos=%llu connection_count=%d chunk_count=%d", + (unsigned long long) file_.getOffset(), (unsigned long long) index_data_pos_, connection_count_, chunk_count_); + + // Write file header record +@@ -390,7 +390,7 @@ void Bag::readFileHeaderRecord() { + readField(fields, CHUNK_COUNT_FIELD_NAME, true, &chunk_count_); + } + +- logDebug("Read FILE_HEADER: index_pos=%llu connection_count=%d chunk_count=%d", ++ CONSOLE_BRIDGE_logDebug("Read FILE_HEADER: index_pos=%llu connection_count=%d chunk_count=%d", + (unsigned long long) index_data_pos_, connection_count_, chunk_count_); + + // Skip the data section (just padding) +@@ -460,7 +460,7 @@ void Bag::writeChunkHeader(CompressionType compression, uint32_t compressed_size + chunk_header.compressed_size = compressed_size; + chunk_header.uncompressed_size = uncompressed_size; + +- logDebug("Writing CHUNK [%llu]: compression=%s compressed=%d uncompressed=%d", ++ CONSOLE_BRIDGE_logDebug("Writing CHUNK [%llu]: compression=%s compressed=%d uncompressed=%d", + (unsigned long long) file_.getOffset(), chunk_header.compression.c_str(), chunk_header.compressed_size, chunk_header.uncompressed_size); + + M_string header; +@@ -485,7 +485,7 @@ void Bag::readChunkHeader(ChunkHeader& chunk_header) const { + readField(fields, COMPRESSION_FIELD_NAME, true, chunk_header.compression); + readField(fields, SIZE_FIELD_NAME, true, &chunk_header.uncompressed_size); + +- logDebug("Read CHUNK: compression=%s size=%d uncompressed=%d (%f)", chunk_header.compression.c_str(), chunk_header.compressed_size, chunk_header.uncompressed_size, 100 * ((double) chunk_header.compressed_size) / chunk_header.uncompressed_size); ++ CONSOLE_BRIDGE_logDebug("Read CHUNK: compression=%s size=%d uncompressed=%d (%f)", chunk_header.compression.c_str(), chunk_header.compressed_size, chunk_header.uncompressed_size, 100 * ((double) chunk_header.compressed_size) / chunk_header.uncompressed_size); + } + + // Index records +@@ -506,7 +506,7 @@ void Bag::writeIndexRecords() { + + writeDataLength(index_size * 12); + +- logDebug("Writing INDEX_DATA: connection=%d ver=%d count=%d", connection_id, INDEX_VERSION, index_size); ++ CONSOLE_BRIDGE_logDebug("Writing INDEX_DATA: connection=%d ver=%d count=%d", connection_id, INDEX_VERSION, index_size); + + // Write the index record data (pairs of timestamp and position in file) + foreach(IndexEntry const& e, index) { +@@ -514,7 +514,7 @@ void Bag::writeIndexRecords() { + write((char*) &e.time.nsec, 4); + write((char*) &e.offset, 4); + +- logDebug(" - %d.%d: %d", e.time.sec, e.time.nsec, e.offset); ++ CONSOLE_BRIDGE_logDebug(" - %d.%d: %d", e.time.sec, e.time.nsec, e.offset); + } + } + } +@@ -536,7 +536,7 @@ void Bag::readTopicIndexRecord102() { + readField(fields, TOPIC_FIELD_NAME, true, topic); + readField(fields, COUNT_FIELD_NAME, true, &count); + +- logDebug("Read INDEX_DATA: ver=%d topic=%s count=%d", index_version, topic.c_str(), count); ++ CONSOLE_BRIDGE_logDebug("Read INDEX_DATA: ver=%d topic=%s count=%d", index_version, topic.c_str(), count); + + if (index_version != 0) + throw BagFormatException((format("Unsupported INDEX_DATA version: %1%") % index_version).str()); +@@ -546,7 +546,7 @@ void Bag::readTopicIndexRecord102() { + if (topic_conn_id_iter == topic_connection_ids_.end()) { + connection_id = connections_.size(); + +- logDebug("Creating connection: id=%d topic=%s", connection_id, topic.c_str()); ++ CONSOLE_BRIDGE_logDebug("Creating connection: id=%d topic=%s", connection_id, topic.c_str()); + ConnectionInfo* connection_info = new ConnectionInfo(); + connection_info->id = connection_id; + connection_info->topic = topic; +@@ -569,11 +569,11 @@ void Bag::readTopicIndexRecord102() { + index_entry.time = Time(sec, nsec); + index_entry.offset = 0; + +- logDebug(" - %d.%d: %llu", sec, nsec, (unsigned long long) index_entry.chunk_pos); ++ CONSOLE_BRIDGE_logDebug(" - %d.%d: %llu", sec, nsec, (unsigned long long) index_entry.chunk_pos); + + if (index_entry.time < ros::TIME_MIN || index_entry.time > ros::TIME_MAX) + { +- logError("Index entry for topic %s contains invalid time.", topic.c_str()); ++ CONSOLE_BRIDGE_logError("Index entry for topic %s contains invalid time.", topic.c_str()); + } else + { + connection_index.insert(connection_index.end(), index_entry); +@@ -598,7 +598,7 @@ void Bag::readConnectionIndexRecord200() { + readField(fields, CONNECTION_FIELD_NAME, true, &connection_id); + readField(fields, COUNT_FIELD_NAME, true, &count); + +- logDebug("Read INDEX_DATA: ver=%d connection=%d count=%d", index_version, connection_id, count); ++ CONSOLE_BRIDGE_logDebug("Read INDEX_DATA: ver=%d connection=%d count=%d", index_version, connection_id, count); + + if (index_version != 1) + throw BagFormatException((format("Unsupported INDEX_DATA version: %1%") % index_version).str()); +@@ -617,11 +617,11 @@ void Bag::readConnectionIndexRecord200() { + read((char*) &index_entry.offset, 4); + index_entry.time = Time(sec, nsec); + +- logDebug(" - %d.%d: %llu+%d", sec, nsec, (unsigned long long) index_entry.chunk_pos, index_entry.offset); ++ CONSOLE_BRIDGE_logDebug(" - %d.%d: %llu+%d", sec, nsec, (unsigned long long) index_entry.chunk_pos, index_entry.offset); + + if (index_entry.time < ros::TIME_MIN || index_entry.time > ros::TIME_MAX) + { +- logError("Index entry for topic %s contains invalid time. This message will not be loaded.", connections_[connection_id]->topic.c_str()); ++ CONSOLE_BRIDGE_logError("Index entry for topic %s contains invalid time. This message will not be loaded.", connections_[connection_id]->topic.c_str()); + } else + { + connection_index.insert(connection_index.end(), index_entry); +@@ -639,7 +639,7 @@ void Bag::writeConnectionRecords() { + } + + void Bag::writeConnectionRecord(ConnectionInfo const* connection_info) { +- logDebug("Writing CONNECTION [%llu:%d]: topic=%s id=%d", ++ CONSOLE_BRIDGE_logDebug("Writing CONNECTION [%llu:%d]: topic=%s id=%d", + (unsigned long long) file_.getOffset(), getChunkOffset(), connection_info->topic.c_str(), connection_info->id); + + M_string header; +@@ -693,7 +693,7 @@ void Bag::readConnectionRecord() { + connection_info->md5sum = (*connection_info->header)["md5sum"]; + connections_[id] = connection_info; + +- logDebug("Read CONNECTION: topic=%s id=%d", topic.c_str(), id); ++ CONSOLE_BRIDGE_logDebug("Read CONNECTION: topic=%s id=%d", topic.c_str(), id); + } + } + +@@ -719,7 +719,7 @@ void Bag::readMessageDefinitionRecord102() { + if (topic_conn_id_iter == topic_connection_ids_.end()) { + uint32_t id = connections_.size(); + +- logDebug("Creating connection: topic=%s md5sum=%s datatype=%s", topic.c_str(), md5sum.c_str(), datatype.c_str()); ++ CONSOLE_BRIDGE_logDebug("Creating connection: topic=%s md5sum=%s datatype=%s", topic.c_str(), md5sum.c_str(), datatype.c_str()); + connection_info = new ConnectionInfo(); + connection_info->id = id; + connection_info->topic = topic; +@@ -738,7 +738,7 @@ void Bag::readMessageDefinitionRecord102() { + (*connection_info->header)["md5sum"] = connection_info->md5sum; + (*connection_info->header)["message_definition"] = connection_info->msg_def; + +- logDebug("Read MSG_DEF: topic=%s md5sum=%s datatype=%s", topic.c_str(), md5sum.c_str(), datatype.c_str()); ++ CONSOLE_BRIDGE_logDebug("Read MSG_DEF: topic=%s md5sum=%s datatype=%s", topic.c_str(), md5sum.c_str(), datatype.c_str()); + } + + void Bag::decompressChunk(uint64_t chunk_pos) const { +@@ -773,7 +773,7 @@ void Bag::decompressChunk(uint64_t chunk_pos) const { + } + + void Bag::readMessageDataRecord102(uint64_t offset, ros::Header& header) const { +- logDebug("readMessageDataRecord: offset=%llu", (unsigned long long) offset); ++ CONSOLE_BRIDGE_logDebug("readMessageDataRecord: offset=%llu", (unsigned long long) offset); + + seek(offset); + +@@ -799,7 +799,7 @@ void Bag::decompressRawChunk(ChunkHeader const& chunk_header) const { + assert(chunk_header.compression == COMPRESSION_NONE); + assert(chunk_header.compressed_size == chunk_header.uncompressed_size); + +- logDebug("compressed_size: %d uncompressed_size: %d", chunk_header.compressed_size, chunk_header.uncompressed_size); ++ CONSOLE_BRIDGE_logDebug("compressed_size: %d uncompressed_size: %d", chunk_header.compressed_size, chunk_header.uncompressed_size); + + decompress_buffer_.setSize(chunk_header.compressed_size); + file_.read((char*) decompress_buffer_.getData(), chunk_header.compressed_size); +@@ -812,7 +812,7 @@ void Bag::decompressBz2Chunk(ChunkHeader const& chunk_header) const { + + CompressionType compression = compression::BZ2; + +- logDebug("compressed_size: %d uncompressed_size: %d", chunk_header.compressed_size, chunk_header.uncompressed_size); ++ CONSOLE_BRIDGE_logDebug("compressed_size: %d uncompressed_size: %d", chunk_header.compressed_size, chunk_header.uncompressed_size); + + chunk_buffer_.setSize(chunk_header.compressed_size); + file_.read((char*) chunk_buffer_.getData(), chunk_header.compressed_size); +@@ -828,7 +828,7 @@ void Bag::decompressLz4Chunk(ChunkHeader const& chunk_header) const { + + CompressionType compression = compression::LZ4; + +- logDebug("lz4 compressed_size: %d uncompressed_size: %d", ++ CONSOLE_BRIDGE_logDebug("lz4 compressed_size: %d uncompressed_size: %d", + chunk_header.compressed_size, chunk_header.uncompressed_size); + + chunk_buffer_.setSize(chunk_header.compressed_size); +@@ -889,7 +889,7 @@ void Bag::writeChunkInfoRecords() { + header[END_TIME_FIELD_NAME] = toHeaderString(&chunk_info.end_time); + header[COUNT_FIELD_NAME] = toHeaderString(&chunk_connection_count); + +- logDebug("Writing CHUNK_INFO [%llu]: ver=%d pos=%llu start=%d.%d end=%d.%d", ++ CONSOLE_BRIDGE_logDebug("Writing CHUNK_INFO [%llu]: ver=%d pos=%llu start=%d.%d end=%d.%d", + (unsigned long long) file_.getOffset(), CHUNK_INFO_VERSION, (unsigned long long) chunk_info.pos, + chunk_info.start_time.sec, chunk_info.start_time.nsec, + chunk_info.end_time.sec, chunk_info.end_time.nsec); +@@ -906,7 +906,7 @@ void Bag::writeChunkInfoRecords() { + write((char*) &connection_id, 4); + write((char*) &count, 4); + +- logDebug(" - %d: %d", connection_id, count); ++ CONSOLE_BRIDGE_logDebug(" - %d: %d", connection_id, count); + } + } + } +@@ -935,7 +935,7 @@ void Bag::readChunkInfoRecord() { + uint32_t chunk_connection_count = 0; + readField(fields, COUNT_FIELD_NAME, true, &chunk_connection_count); + +- logDebug("Read CHUNK_INFO: chunk_pos=%llu connection_count=%d start=%d.%d end=%d.%d", ++ CONSOLE_BRIDGE_logDebug("Read CHUNK_INFO: chunk_pos=%llu connection_count=%d start=%d.%d end=%d.%d", + (unsigned long long) chunk_info.pos, chunk_connection_count, + chunk_info.start_time.sec, chunk_info.start_time.nsec, + chunk_info.end_time.sec, chunk_info.end_time.nsec); +@@ -946,7 +946,7 @@ void Bag::readChunkInfoRecord() { + read((char*) &connection_id, 4); + read((char*) &connection_count, 4); + +- logDebug(" %d: %d messages", connection_id, connection_count); ++ CONSOLE_BRIDGE_logDebug(" %d: %d messages", connection_id, connection_count); + + chunk_info.connection_counts[connection_id] = connection_count; + } +@@ -1028,7 +1028,7 @@ void Bag::readMessageDataHeaderFromBuffer(Buffer& buffer, uint32_t offset, ros:: + total_bytes_read = 0; + uint8_t op = 0xFF; + do { +- logDebug("reading header from buffer: offset=%d", offset); ++ CONSOLE_BRIDGE_logDebug("reading header from buffer: offset=%d", offset); + uint32_t bytes_read; + readHeaderFromBuffer(*current_buffer_, offset, header, data_size, bytes_read); + +diff --git a/src/bz2_stream.cpp b/src/bz2_stream.cpp +index 900cd04..d158d1d 100644 +--- a/src/bz2_stream.cpp ++++ b/src/bz2_stream.cpp +@@ -115,7 +115,7 @@ void BZ2Stream::read(void* ptr, size_t size) { + case BZ_OK: return; + case BZ_STREAM_END: + if (getUnused() || getUnusedLength() > 0) +- logError("unused data already available"); ++ CONSOLE_BRIDGE_logError("unused data already available"); + else { + char* unused; + int nUnused; +diff --git a/src/lz4_stream.cpp b/src/lz4_stream.cpp +index 9502664..41ec0df 100644 +--- a/src/lz4_stream.cpp ++++ b/src/lz4_stream.cpp +@@ -155,7 +155,7 @@ void LZ4Stream::read(void* ptr, size_t size) { + case ROSLZ4_OK: break; + case ROSLZ4_STREAM_END: + if (getUnused() || getUnusedLength() > 0) +- logError("unused data already available"); ++ CONSOLE_BRIDGE_logError("unused data already available"); + else { + setUnused(lz4s_.input_next); + setUnusedLength(lz4s_.input_left); |