|  | // Copyright (c) 2012 The Chromium Authors. All rights reserved. | 
|  | // Use of this source code is governed by a BSD-style license that can be | 
|  | // found in the LICENSE file. | 
|  |  | 
|  | #include "content/browser/speech/endpointer/endpointer.h" | 
|  |  | 
|  | #include "base/time/time.h" | 
|  | #include "content/browser/speech/audio_buffer.h" | 
|  |  | 
|  | namespace { | 
|  | const int64_t kMicrosecondsPerSecond = base::Time::kMicrosecondsPerSecond; | 
|  | const int kFrameRate = 50;  // 1 frame = 20ms of audio. | 
|  | } | 
|  |  | 
|  | namespace content { | 
|  |  | 
|  | Endpointer::Endpointer(int sample_rate) | 
|  | : speech_input_possibly_complete_silence_length_us_(-1), | 
|  | speech_input_complete_silence_length_us_(-1), | 
|  | audio_frame_time_us_(0), | 
|  | sample_rate_(sample_rate), | 
|  | frame_size_(0) { | 
|  | Reset(); | 
|  |  | 
|  | frame_size_ = static_cast<int>(sample_rate / static_cast<float>(kFrameRate)); | 
|  |  | 
|  | speech_input_minimum_length_us_ = | 
|  | static_cast<int64_t>(1.7 * kMicrosecondsPerSecond); | 
|  | speech_input_complete_silence_length_us_ = | 
|  | static_cast<int64_t>(0.5 * kMicrosecondsPerSecond); | 
|  | long_speech_input_complete_silence_length_us_ = -1; | 
|  | long_speech_length_us_ = -1; | 
|  | speech_input_possibly_complete_silence_length_us_ = | 
|  | 1 * kMicrosecondsPerSecond; | 
|  |  | 
|  | // Set the default configuration for Push To Talk mode. | 
|  | EnergyEndpointerParams ep_config; | 
|  | ep_config.set_frame_period(1.0f / static_cast<float>(kFrameRate)); | 
|  | ep_config.set_frame_duration(1.0f / static_cast<float>(kFrameRate)); | 
|  | ep_config.set_endpoint_margin(0.2f); | 
|  | ep_config.set_onset_window(0.15f); | 
|  | ep_config.set_speech_on_window(0.4f); | 
|  | ep_config.set_offset_window(0.15f); | 
|  | ep_config.set_onset_detect_dur(0.09f); | 
|  | ep_config.set_onset_confirm_dur(0.075f); | 
|  | ep_config.set_on_maintain_dur(0.10f); | 
|  | ep_config.set_offset_confirm_dur(0.12f); | 
|  | ep_config.set_decision_threshold(1000.0f); | 
|  | ep_config.set_min_decision_threshold(50.0f); | 
|  | ep_config.set_fast_update_dur(0.2f); | 
|  | ep_config.set_sample_rate(static_cast<float>(sample_rate)); | 
|  | ep_config.set_min_fundamental_frequency(57.143f); | 
|  | ep_config.set_max_fundamental_frequency(400.0f); | 
|  | ep_config.set_contamination_rejection_period(0.25f); | 
|  | energy_endpointer_.Init(ep_config); | 
|  | } | 
|  |  | 
|  | void Endpointer::Reset() { | 
|  | old_ep_status_ = EP_PRE_SPEECH; | 
|  | waiting_for_speech_possibly_complete_timeout_ = false; | 
|  | waiting_for_speech_complete_timeout_ = false; | 
|  | speech_previously_detected_ = false; | 
|  | speech_input_complete_ = false; | 
|  | audio_frame_time_us_ = 0; // Reset time for packets sent to endpointer. | 
|  | speech_end_time_us_ = -1; | 
|  | speech_start_time_us_ = -1; | 
|  | } | 
|  |  | 
|  | void Endpointer::StartSession() { | 
|  | Reset(); | 
|  | energy_endpointer_.StartSession(); | 
|  | } | 
|  |  | 
|  | void Endpointer::EndSession() { | 
|  | energy_endpointer_.EndSession(); | 
|  | } | 
|  |  | 
|  | void Endpointer::SetEnvironmentEstimationMode() { | 
|  | Reset(); | 
|  | energy_endpointer_.SetEnvironmentEstimationMode(); | 
|  | } | 
|  |  | 
|  | void Endpointer::SetUserInputMode() { | 
|  | energy_endpointer_.SetUserInputMode(); | 
|  | } | 
|  |  | 
|  | EpStatus Endpointer::Status(int64_t* time) { | 
|  | return energy_endpointer_.Status(time); | 
|  | } | 
|  |  | 
|  | EpStatus Endpointer::ProcessAudio(const AudioChunk& raw_audio, float* rms_out) { | 
|  | const int16_t* audio_data = raw_audio.SamplesData16(); | 
|  | const int num_samples = raw_audio.NumSamples(); | 
|  | EpStatus ep_status = EP_PRE_SPEECH; | 
|  |  | 
|  | // Process the input data in blocks of frame_size_, dropping any incomplete | 
|  | // frames at the end (which is ok since typically the caller will be recording | 
|  | // audio in multiples of our frame size). | 
|  | int sample_index = 0; | 
|  | while (sample_index + frame_size_ <= num_samples) { | 
|  | // Have the endpointer process the frame. | 
|  | energy_endpointer_.ProcessAudioFrame(audio_frame_time_us_, | 
|  | audio_data + sample_index, | 
|  | frame_size_, | 
|  | rms_out); | 
|  | sample_index += frame_size_; | 
|  | audio_frame_time_us_ += | 
|  | (frame_size_ * kMicrosecondsPerSecond) / sample_rate_; | 
|  |  | 
|  | // Get the status of the endpointer. | 
|  | int64_t ep_time; | 
|  | ep_status = energy_endpointer_.Status(&ep_time); | 
|  |  | 
|  | // Handle state changes. | 
|  | if ((EP_SPEECH_PRESENT == ep_status) && | 
|  | (EP_POSSIBLE_ONSET == old_ep_status_)) { | 
|  | speech_end_time_us_ = -1; | 
|  | waiting_for_speech_possibly_complete_timeout_ = false; | 
|  | waiting_for_speech_complete_timeout_ = false; | 
|  | // Trigger SpeechInputDidStart event on first detection. | 
|  | if (false == speech_previously_detected_) { | 
|  | speech_previously_detected_ = true; | 
|  | speech_start_time_us_ = ep_time; | 
|  | } | 
|  | } | 
|  | if ((EP_PRE_SPEECH == ep_status) && | 
|  | (EP_POSSIBLE_OFFSET == old_ep_status_)) { | 
|  | speech_end_time_us_ = ep_time; | 
|  | waiting_for_speech_possibly_complete_timeout_ = true; | 
|  | waiting_for_speech_complete_timeout_ = true; | 
|  | } | 
|  | if (ep_time > speech_input_minimum_length_us_) { | 
|  | // Speech possibly complete timeout. | 
|  | if ((waiting_for_speech_possibly_complete_timeout_) && | 
|  | (ep_time - speech_end_time_us_ > | 
|  | speech_input_possibly_complete_silence_length_us_)) { | 
|  | waiting_for_speech_possibly_complete_timeout_ = false; | 
|  | } | 
|  | if (waiting_for_speech_complete_timeout_) { | 
|  | // The length of the silence timeout period can be held constant, or it | 
|  | // can be changed after a fixed amount of time from the beginning of | 
|  | // speech. | 
|  | bool has_stepped_silence = | 
|  | (long_speech_length_us_ > 0) && | 
|  | (long_speech_input_complete_silence_length_us_ > 0); | 
|  | int64_t requested_silence_length; | 
|  | if (has_stepped_silence && | 
|  | (ep_time - speech_start_time_us_) > long_speech_length_us_) { | 
|  | requested_silence_length = | 
|  | long_speech_input_complete_silence_length_us_; | 
|  | } else { | 
|  | requested_silence_length = | 
|  | speech_input_complete_silence_length_us_; | 
|  | } | 
|  |  | 
|  | // Speech complete timeout. | 
|  | if ((ep_time - speech_end_time_us_) > requested_silence_length) { | 
|  | waiting_for_speech_complete_timeout_ = false; | 
|  | speech_input_complete_ = true; | 
|  | } | 
|  | } | 
|  | } | 
|  | old_ep_status_ = ep_status; | 
|  | } | 
|  | return ep_status; | 
|  | } | 
|  |  | 
|  | }  // namespace content |