X-Git-Url: https://sigrok.org/gitweb/?p=pulseview.git;a=blobdiff_plain;f=pv%2Fdata%2Fsegment.cpp;h=905b79274efeb6ee5ddebd7045348c5da85d9db3;hp=88015c3639860597accf8bf23eb21ed98d499e53;hb=0c5fe73e5b66bf31d2dcfa2b2846a86cd2fdf0c4;hpb=c063290ac7189bdd15221450f598504f43286b43 diff --git a/pv/data/segment.cpp b/pv/data/segment.cpp index 88015c36..905b7927 100644 --- a/pv/data/segment.cpp +++ b/pv/data/segment.cpp @@ -31,7 +31,7 @@ using std::recursive_mutex; namespace pv { namespace data { -const uint64_t Segment::MaxChunkSize = 10*1024*1024; /* 10MiB */ +const uint64_t Segment::MaxChunkSize = 10 * 1024 * 1024; /* 10MiB */ Segment::Segment(uint64_t samplerate, unsigned int unit_size) : sample_count_(0), @@ -135,35 +135,38 @@ void Segment::append_samples(void* data, uint64_t samples) { lock_guard lock(mutex_); - if (unused_samples_ >= samples) { - // All samples fit into the current chunk - memcpy(current_chunk_ + (used_samples_ * unit_size_), - data, (samples * unit_size_)); - used_samples_ += samples; - unused_samples_ -= samples; - } else { - // Only a part of the samples fit, split data up between chunks - memcpy(current_chunk_ + (used_samples_ * unit_size_), - data, (unused_samples_ * unit_size_)); - const uint64_t remaining_samples = samples - unused_samples_; - - // If we're out of memory, this will throw std::bad_alloc - current_chunk_ = new uint8_t[chunk_size_]; - data_chunks_.push_back(current_chunk_); - memcpy(current_chunk_, (uint8_t*)data + (unused_samples_ * unit_size_), - (remaining_samples * unit_size_)); - - used_samples_ = remaining_samples; - unused_samples_ = (chunk_size_ / unit_size_) - remaining_samples; - } - - if (unused_samples_ == 0) { - // If we're out of memory, this will throw std::bad_alloc - current_chunk_ = new uint8_t[chunk_size_]; - data_chunks_.push_back(current_chunk_); - used_samples_ = 0; - unused_samples_ = chunk_size_ / unit_size_; - } + const uint8_t* data_byte_ptr = (uint8_t*)data; + uint64_t remaining_samples = samples; + uint64_t data_offset = 0; + + do { + uint64_t copy_count = 0; + + if (remaining_samples <= unused_samples_) { + // All samples fit into the current chunk + copy_count = remaining_samples; + } else { + // Only a part of the samples fit, fill up current chunk + copy_count = unused_samples_; + } + + const uint8_t* dest = &(current_chunk_[used_samples_ * unit_size_]); + const uint8_t* src = &(data_byte_ptr[data_offset]); + memcpy((void*)dest, (void*)src, (copy_count * unit_size_)); + + used_samples_ += copy_count; + unused_samples_ -= copy_count; + remaining_samples -= copy_count; + data_offset += (copy_count * unit_size_); + + if (unused_samples_ == 0) { + // If we're out of memory, this will throw std::bad_alloc + current_chunk_ = new uint8_t[chunk_size_]; + data_chunks_.push_back(current_chunk_); + used_samples_ = 0; + unused_samples_ = chunk_size_ / unit_size_; + } + } while (remaining_samples > 0); sample_count_ += samples; } @@ -219,8 +222,6 @@ SegmentRawDataIterator* Segment::begin_raw_sample_iteration(uint64_t start) void Segment::continue_raw_sample_iteration(SegmentRawDataIterator* it, uint64_t increase) { - lock_guard lock(mutex_); - // Fail gracefully if we are asked to deliver data we don't have if (it->sample_index > sample_count_) return;