blob: 0c339cd851a6055d49ae7f4a73a6ec45227a2caa [file] [log] [blame]
// Copyright 2014 The Chromium Authors. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file.
#include "content/renderer/media/speech_recognition_audio_sink.h"
#include <stddef.h>
#include <utility>
#include "base/logging.h"
#include "base/memory/shared_memory.h"
#include "base/time/time.h"
#include "content/renderer/media/media_stream_audio_source.h"
#include "media/base/audio_fifo.h"
#include "media/base/audio_parameters.h"
namespace content {
SpeechRecognitionAudioSink::SpeechRecognitionAudioSink(
const blink::WebMediaStreamTrack& track,
const media::AudioParameters& params,
const base::SharedMemoryHandle memory,
std::unique_ptr<base::SyncSocket> socket,
const OnStoppedCB& on_stopped_cb)
: track_(track),
shared_memory_(memory, false),
socket_(std::move(socket)),
output_params_(params),
track_stopped_(false),
buffer_index_(0),
on_stopped_cb_(on_stopped_cb) {
DCHECK(socket_.get());
DCHECK(main_render_thread_checker_.CalledOnValidThread());
DCHECK(params.IsValid());
DCHECK(IsSupportedTrack(track));
const size_t kSharedMemorySize = sizeof(media::AudioInputBufferParameters) +
media::AudioBus::CalculateMemorySize(params);
CHECK(shared_memory_.Map(kSharedMemorySize));
media::AudioInputBuffer* buffer =
static_cast<media::AudioInputBuffer*>(shared_memory_.memory());
// The peer must manage their own counter and reset it to 0.
DCHECK_EQ(0U, buffer->params.size);
output_bus_ = media::AudioBus::WrapMemory(params, buffer->audio);
// Connect this audio sink to the track
MediaStreamAudioSink::AddToAudioTrack(this, track_);
}
SpeechRecognitionAudioSink::~SpeechRecognitionAudioSink() {
DCHECK(main_render_thread_checker_.CalledOnValidThread());
if (audio_converter_.get())
audio_converter_->RemoveInput(this);
// Notify the track before this sink goes away.
if (!track_stopped_)
MediaStreamAudioSink::RemoveFromAudioTrack(this, track_);
}
// static
bool SpeechRecognitionAudioSink::IsSupportedTrack(
const blink::WebMediaStreamTrack& track) {
MediaStreamAudioSource* native_source =
MediaStreamAudioSource::From(track.Source());
if (!native_source)
return false;
const StreamDeviceInfo& device_info = native_source->device_info();
// Purposely only support tracks from an audio device. Dissallow WebAudio.
return (device_info.device.type == content::MEDIA_DEVICE_AUDIO_CAPTURE);
}
void SpeechRecognitionAudioSink::OnSetFormat(
const media::AudioParameters& input_params) {
DCHECK(input_params.IsValid());
DCHECK_LE(
input_params.frames_per_buffer() * 1000 / input_params.sample_rate(),
output_params_.frames_per_buffer() * 1000 / output_params_.sample_rate());
// Detach the thread here because it will be a new capture thread
// calling OnSetFormat() and OnData() if the source is restarted.
capture_thread_checker_.DetachFromThread();
input_params_ = input_params;
fifo_buffer_size_ =
std::ceil(output_params_.frames_per_buffer() *
static_cast<double>(input_params_.sample_rate()) /
output_params_.sample_rate());
DCHECK_GE(fifo_buffer_size_, input_params_.frames_per_buffer());
// Allows for some delays on the peer.
static const int kNumberOfBuffersInFifo = 2;
int frames_in_fifo = kNumberOfBuffersInFifo * fifo_buffer_size_;
fifo_.reset(new media::AudioFifo(input_params.channels(), frames_in_fifo));
// Create the audio converter with |disable_fifo| as false so that the
// converter will request input_params.frames_per_buffer() each time.
// This will not increase the complexity as there is only one client to
// the converter.
audio_converter_.reset(
new media::AudioConverter(input_params, output_params_, false));
audio_converter_->AddInput(this);
}
void SpeechRecognitionAudioSink::OnReadyStateChanged(
blink::WebMediaStreamSource::ReadyState state) {
DCHECK(main_render_thread_checker_.CalledOnValidThread());
DCHECK(!track_stopped_);
if (state == blink::WebMediaStreamSource::kReadyStateEnded) {
track_stopped_ = true;
if (!on_stopped_cb_.is_null())
on_stopped_cb_.Run();
}
}
void SpeechRecognitionAudioSink::OnData(
const media::AudioBus& audio_bus,
base::TimeTicks estimated_capture_time) {
DCHECK(capture_thread_checker_.CalledOnValidThread());
DCHECK_EQ(audio_bus.frames(), input_params_.frames_per_buffer());
DCHECK_EQ(audio_bus.channels(), input_params_.channels());
if (fifo_->frames() + audio_bus.frames() > fifo_->max_frames()) {
// This would indicate a serious issue with the browser process or the
// SyncSocket and/or SharedMemory. We drop any previous buffers and try to
// recover by resuming where the peer left of.
DLOG(ERROR) << "Audio FIFO overflow";
fifo_->Clear();
buffer_index_ = GetAudioInputBuffer()->params.size;
}
fifo_->Push(&audio_bus);
// Wait for FIFO to have at least |fifo_buffer_size_| frames ready.
if (fifo_->frames() < fifo_buffer_size_)
return;
// Make sure the previous output buffer was consumed by the peer before we
// send the next buffer.
// The peer must write to it (incrementing by 1) once the the buffer was
// consumed. This is intentional not to block this audio capturing thread.
if (buffer_index_ != GetAudioInputBuffer()->params.size) {
DVLOG(1) << "Buffer synchronization lag";
return;
}
audio_converter_->Convert(output_bus_.get());
// Notify peer to consume buffer |buffer_index_| on |output_bus_|.
const size_t bytes_sent =
socket_->Send(&buffer_index_, sizeof(buffer_index_));
if (bytes_sent != sizeof(buffer_index_)) {
// The send ocasionally fails if the user changes their input audio device.
DVLOG(1) << "Failed sending buffer index to peer";
// We have discarded this buffer, but could still recover on the next one.
return;
}
// Count the sent buffer. We expect the peer to do the same on their end.
++buffer_index_;
}
double SpeechRecognitionAudioSink::ProvideInput(media::AudioBus* audio_bus,
uint32_t frames_delayed) {
DCHECK(capture_thread_checker_.CalledOnValidThread());
if (fifo_->frames() >= audio_bus->frames())
fifo_->Consume(audio_bus, 0, audio_bus->frames());
else
audio_bus->Zero();
// Return volume greater than zero to indicate we have more data.
return 1.0;
}
media::AudioInputBuffer*
SpeechRecognitionAudioSink::GetAudioInputBuffer() const {
DCHECK(capture_thread_checker_.CalledOnValidThread());
DCHECK(shared_memory_.memory());
return static_cast<media::AudioInputBuffer*>(shared_memory_.memory());
}
} // namespace content