summaryrefslogtreecommitdiff
path: root/chromium/media/base/audio_buffer.cc
diff options
context:
space:
mode:
Diffstat (limited to 'chromium/media/base/audio_buffer.cc')
-rw-r--r--chromium/media/base/audio_buffer.cc254
1 files changed, 254 insertions, 0 deletions
diff --git a/chromium/media/base/audio_buffer.cc b/chromium/media/base/audio_buffer.cc
new file mode 100644
index 00000000000..b2cdd8c41a7
--- /dev/null
+++ b/chromium/media/base/audio_buffer.cc
@@ -0,0 +1,254 @@
+// Copyright 2013 The Chromium Authors. All rights reserved.
+// Use of this source code is governed by a BSD-style license that can be
+// found in the LICENSE file.
+
+#include "media/base/audio_buffer.h"
+
+#include "base/logging.h"
+#include "media/base/audio_bus.h"
+#include "media/base/buffers.h"
+#include "media/base/limits.h"
+
+namespace media {
+
+// Alignment of each channel's data; this must match what ffmpeg expects
+// (which may be 0, 16, or 32, depending on the processor). Selecting 32 in
+// order to work on all processors.
+enum { kChannelAlignment = 32 };
+
+AudioBuffer::AudioBuffer(SampleFormat sample_format,
+ int channel_count,
+ int frame_count,
+ bool create_buffer,
+ const uint8* const* data,
+ const base::TimeDelta timestamp,
+ const base::TimeDelta duration)
+ : sample_format_(sample_format),
+ channel_count_(channel_count),
+ adjusted_frame_count_(frame_count),
+ trim_start_(0),
+ end_of_stream_(!create_buffer && data == NULL && frame_count == 0),
+ timestamp_(timestamp),
+ duration_(duration) {
+ CHECK_GE(channel_count, 0);
+ CHECK_LE(channel_count, limits::kMaxChannels);
+ CHECK_GE(frame_count, 0);
+ int bytes_per_channel = SampleFormatToBytesPerChannel(sample_format);
+ DCHECK_LE(bytes_per_channel, kChannelAlignment);
+ int data_size = frame_count * bytes_per_channel;
+
+ // Empty buffer?
+ if (!create_buffer)
+ return;
+
+ if (sample_format == kSampleFormatPlanarF32 ||
+ sample_format == kSampleFormatPlanarS16) {
+ // Planar data, so need to allocate buffer for each channel.
+ // Determine per channel data size, taking into account alignment.
+ int block_size_per_channel =
+ (data_size + kChannelAlignment - 1) & ~(kChannelAlignment - 1);
+ DCHECK_GE(block_size_per_channel, data_size);
+
+ // Allocate a contiguous buffer for all the channel data.
+ data_.reset(static_cast<uint8*>(base::AlignedAlloc(
+ channel_count * block_size_per_channel, kChannelAlignment)));
+ channel_data_.reserve(channel_count);
+
+ // Copy each channel's data into the appropriate spot.
+ for (int i = 0; i < channel_count; ++i) {
+ channel_data_.push_back(data_.get() + i * block_size_per_channel);
+ if (data)
+ memcpy(channel_data_[i], data[i], data_size);
+ }
+ return;
+ }
+
+ // Remaining formats are interleaved data.
+ DCHECK(sample_format_ == kSampleFormatU8 ||
+ sample_format_ == kSampleFormatS16 ||
+ sample_format_ == kSampleFormatS32 ||
+ sample_format_ == kSampleFormatF32) << sample_format_;
+ // Allocate our own buffer and copy the supplied data into it. Buffer must
+ // contain the data for all channels.
+ data_size *= channel_count;
+ data_.reset(
+ static_cast<uint8*>(base::AlignedAlloc(data_size, kChannelAlignment)));
+ if (data)
+ memcpy(data_.get(), data[0], data_size);
+}
+
+AudioBuffer::~AudioBuffer() {}
+
+// static
+scoped_refptr<AudioBuffer> AudioBuffer::CopyFrom(
+ SampleFormat sample_format,
+ int channel_count,
+ int frame_count,
+ const uint8* const* data,
+ const base::TimeDelta timestamp,
+ const base::TimeDelta duration) {
+ // If you hit this CHECK you likely have a bug in a demuxer. Go fix it.
+ CHECK_GT(frame_count, 0); // Otherwise looks like an EOF buffer.
+ CHECK(data[0]);
+ return make_scoped_refptr(new AudioBuffer(sample_format,
+ channel_count,
+ frame_count,
+ true,
+ data,
+ timestamp,
+ duration));
+}
+
+// static
+scoped_refptr<AudioBuffer> AudioBuffer::CreateBuffer(SampleFormat sample_format,
+ int channel_count,
+ int frame_count) {
+ CHECK_GT(frame_count, 0); // Otherwise looks like an EOF buffer.
+ return make_scoped_refptr(new AudioBuffer(sample_format,
+ channel_count,
+ frame_count,
+ true,
+ NULL,
+ kNoTimestamp(),
+ kNoTimestamp()));
+}
+
+// static
+scoped_refptr<AudioBuffer> AudioBuffer::CreateEmptyBuffer(
+ int channel_count,
+ int frame_count,
+ const base::TimeDelta timestamp,
+ const base::TimeDelta duration) {
+ CHECK_GT(frame_count, 0); // Otherwise looks like an EOF buffer.
+ // Since data == NULL, format doesn't matter.
+ return make_scoped_refptr(new AudioBuffer(kSampleFormatF32,
+ channel_count,
+ frame_count,
+ false,
+ NULL,
+ timestamp,
+ duration));
+}
+
+// static
+scoped_refptr<AudioBuffer> AudioBuffer::CreateEOSBuffer() {
+ return make_scoped_refptr(new AudioBuffer(
+ kUnknownSampleFormat, 1, 0, false, NULL, kNoTimestamp(), kNoTimestamp()));
+}
+
+// Convert int16 values in the range [kint16min, kint16max] to [-1.0, 1.0].
+static inline float ConvertS16ToFloat(int16 value) {
+ return value * (value < 0 ? -1.0f / kint16min : 1.0f / kint16max);
+}
+
+void AudioBuffer::ReadFrames(int frames_to_copy,
+ int source_frame_offset,
+ int dest_frame_offset,
+ AudioBus* dest) {
+ // Deinterleave each channel (if necessary) and convert to 32bit
+ // floating-point with nominal range -1.0 -> +1.0 (if necessary).
+
+ // |dest| must have the same number of channels, and the number of frames
+ // specified must be in range.
+ DCHECK(!end_of_stream());
+ DCHECK_EQ(dest->channels(), channel_count_);
+ DCHECK_LE(source_frame_offset + frames_to_copy, adjusted_frame_count_);
+ DCHECK_LE(dest_frame_offset + frames_to_copy, dest->frames());
+
+ // Move the start past any frames that have been trimmed.
+ source_frame_offset += trim_start_;
+
+ if (!data_) {
+ // Special case for an empty buffer.
+ dest->ZeroFramesPartial(dest_frame_offset, frames_to_copy);
+ return;
+ }
+
+ if (sample_format_ == kSampleFormatPlanarF32) {
+ // Format is planar float32. Copy the data from each channel as a block.
+ for (int ch = 0; ch < channel_count_; ++ch) {
+ const float* source_data =
+ reinterpret_cast<const float*>(channel_data_[ch]) +
+ source_frame_offset;
+ memcpy(dest->channel(ch) + dest_frame_offset,
+ source_data,
+ sizeof(float) * frames_to_copy);
+ }
+ return;
+ }
+
+ if (sample_format_ == kSampleFormatPlanarS16) {
+ // Format is planar signed16. Convert each value into float and insert into
+ // output channel data.
+ for (int ch = 0; ch < channel_count_; ++ch) {
+ const int16* source_data =
+ reinterpret_cast<const int16*>(channel_data_[ch]) +
+ source_frame_offset;
+ float* dest_data = dest->channel(ch) + dest_frame_offset;
+ for (int i = 0; i < frames_to_copy; ++i) {
+ dest_data[i] = ConvertS16ToFloat(source_data[i]);
+ }
+ }
+ return;
+ }
+
+ if (sample_format_ == kSampleFormatF32) {
+ // Format is interleaved float32. Copy the data into each channel.
+ const float* source_data = reinterpret_cast<const float*>(data_.get()) +
+ source_frame_offset * channel_count_;
+ for (int ch = 0; ch < channel_count_; ++ch) {
+ float* dest_data = dest->channel(ch) + dest_frame_offset;
+ for (int i = 0, offset = ch; i < frames_to_copy;
+ ++i, offset += channel_count_) {
+ dest_data[i] = source_data[offset];
+ }
+ }
+ return;
+ }
+
+ // Remaining formats are integer interleaved data. Use the deinterleaving code
+ // in AudioBus to copy the data.
+ DCHECK(sample_format_ == kSampleFormatU8 ||
+ sample_format_ == kSampleFormatS16 ||
+ sample_format_ == kSampleFormatS32);
+ int bytes_per_channel = SampleFormatToBytesPerChannel(sample_format_);
+ int frame_size = channel_count_ * bytes_per_channel;
+ const uint8* source_data = data_.get() + source_frame_offset * frame_size;
+ dest->FromInterleavedPartial(
+ source_data, dest_frame_offset, frames_to_copy, bytes_per_channel);
+}
+
+void AudioBuffer::TrimStart(int frames_to_trim) {
+ CHECK_GE(frames_to_trim, 0);
+ CHECK_LE(frames_to_trim, adjusted_frame_count_);
+
+ // Adjust timestamp_ and duration_ to reflect the smaller number of frames.
+ double offset = static_cast<double>(duration_.InMicroseconds()) *
+ frames_to_trim / adjusted_frame_count_;
+ base::TimeDelta offset_as_time =
+ base::TimeDelta::FromMicroseconds(static_cast<int64>(offset));
+ timestamp_ += offset_as_time;
+ duration_ -= offset_as_time;
+
+ // Finally adjust the number of frames in this buffer and where the start
+ // really is.
+ adjusted_frame_count_ -= frames_to_trim;
+ trim_start_ += frames_to_trim;
+}
+
+void AudioBuffer::TrimEnd(int frames_to_trim) {
+ CHECK_GE(frames_to_trim, 0);
+ CHECK_LE(frames_to_trim, adjusted_frame_count_);
+
+ // Adjust duration_ only to reflect the smaller number of frames.
+ double offset = static_cast<double>(duration_.InMicroseconds()) *
+ frames_to_trim / adjusted_frame_count_;
+ base::TimeDelta offset_as_time =
+ base::TimeDelta::FromMicroseconds(static_cast<int64>(offset));
+ duration_ -= offset_as_time;
+
+ // Finally adjust the number of frames in this buffer.
+ adjusted_frame_count_ -= frames_to_trim;
+}
+
+} // namespace media