| /* |
| * Copyright (C) 2012 Google Inc. All rights reserved. |
| * |
| * Redistribution and use in source and binary forms, with or without |
| * modification, are permitted provided that the following conditions |
| * are met: |
| * * Redistributions of source code must retain the above copyright |
| * notice, this list of conditions and the following disclaimer. |
| * * Redistributions in binary form must reproduce the above copyright |
| * notice, this list of conditions and the following disclaimer in the |
| * documentation and/or other materials provided with the distribution. |
| * |
| * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS ``AS IS'' AND ANY |
| * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE |
| * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR |
| * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR |
| * CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, |
| * EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, |
| * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR |
| * PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY |
| * OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
| * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE |
| * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
| */ |
| |
| #include "modules/speech/SpeechRecognition.h" |
| |
| #include "bindings/core/v8/ExceptionState.h" |
| #include "core/dom/Document.h" |
| #include "core/dom/ExceptionCode.h" |
| #include "core/page/Page.h" |
| #include "modules/mediastream/MediaStreamTrack.h" |
| #include "modules/speech/SpeechRecognitionController.h" |
| #include "modules/speech/SpeechRecognitionError.h" |
| #include "modules/speech/SpeechRecognitionEvent.h" |
| |
| namespace blink { |
| |
| SpeechRecognition* SpeechRecognition::create(ExecutionContext* context) |
| { |
| ASSERT(context && context->isDocument()); |
| Document* document = toDocument(context); |
| ASSERT(document); |
| SpeechRecognition* speechRecognition = new SpeechRecognition(document->page(), context); |
| speechRecognition->suspendIfNeeded(); |
| return speechRecognition; |
| } |
| |
| void SpeechRecognition::start(ExceptionState& exceptionState) |
| { |
| if (!m_controller) |
| return; |
| |
| if (m_started) { |
| exceptionState.throwDOMException(InvalidStateError, "recognition has already started."); |
| return; |
| } |
| |
| m_finalResults.clear(); |
| m_controller->start(this, m_grammars, m_lang, m_continuous, m_interimResults, m_maxAlternatives, m_audioTrack); |
| m_started = true; |
| } |
| |
| void SpeechRecognition::stopFunction() |
| { |
| if (!m_controller) |
| return; |
| |
| if (m_started && !m_stopping) { |
| m_stopping = true; |
| m_controller->stop(this); |
| } |
| } |
| |
| void SpeechRecognition::abort() |
| { |
| if (!m_controller) |
| return; |
| |
| if (m_started && !m_stopping) { |
| m_stopping = true; |
| m_controller->abort(this); |
| } |
| } |
| |
| void SpeechRecognition::didStartAudio() |
| { |
| dispatchEvent(Event::create(EventTypeNames::audiostart)); |
| } |
| |
| void SpeechRecognition::didStartSound() |
| { |
| dispatchEvent(Event::create(EventTypeNames::soundstart)); |
| } |
| |
| void SpeechRecognition::didStartSpeech() |
| { |
| dispatchEvent(Event::create(EventTypeNames::speechstart)); |
| } |
| |
| void SpeechRecognition::didEndSpeech() |
| { |
| dispatchEvent(Event::create(EventTypeNames::speechend)); |
| } |
| |
| void SpeechRecognition::didEndSound() |
| { |
| dispatchEvent(Event::create(EventTypeNames::soundend)); |
| } |
| |
| void SpeechRecognition::didEndAudio() |
| { |
| dispatchEvent(Event::create(EventTypeNames::audioend)); |
| } |
| |
| void SpeechRecognition::didReceiveResults(const HeapVector<Member<SpeechRecognitionResult>>& newFinalResults, const HeapVector<Member<SpeechRecognitionResult>>& currentInterimResults) |
| { |
| size_t resultIndex = m_finalResults.size(); |
| |
| for (size_t i = 0; i < newFinalResults.size(); ++i) |
| m_finalResults.append(newFinalResults[i]); |
| |
| HeapVector<Member<SpeechRecognitionResult>> results = m_finalResults; |
| for (size_t i = 0; i < currentInterimResults.size(); ++i) |
| results.append(currentInterimResults[i]); |
| |
| dispatchEvent(SpeechRecognitionEvent::createResult(resultIndex, results)); |
| } |
| |
| void SpeechRecognition::didReceiveNoMatch(SpeechRecognitionResult* result) |
| { |
| dispatchEvent(SpeechRecognitionEvent::createNoMatch(result)); |
| } |
| |
| void SpeechRecognition::didReceiveError(SpeechRecognitionError* error) |
| { |
| dispatchEvent(error); |
| m_started = false; |
| } |
| |
| void SpeechRecognition::didStart() |
| { |
| dispatchEvent(Event::create(EventTypeNames::start)); |
| } |
| |
| void SpeechRecognition::didEnd() |
| { |
| m_started = false; |
| m_stopping = false; |
| if (!m_stoppedByActiveDOMObject) |
| dispatchEvent(Event::create(EventTypeNames::end)); |
| } |
| |
| const AtomicString& SpeechRecognition::interfaceName() const |
| { |
| return EventTargetNames::SpeechRecognition; |
| } |
| |
| ExecutionContext* SpeechRecognition::getExecutionContext() const |
| { |
| return ActiveDOMObject::getExecutionContext(); |
| } |
| |
| void SpeechRecognition::stop() |
| { |
| m_stoppedByActiveDOMObject = true; |
| if (hasPendingActivity()) |
| abort(); |
| } |
| |
| bool SpeechRecognition::hasPendingActivity() const |
| { |
| return m_started; |
| } |
| |
| SpeechRecognition::SpeechRecognition(Page* page, ExecutionContext* context) |
| : PageLifecycleObserver(page) |
| , ActiveScriptWrappable(this) |
| , ActiveDOMObject(context) |
| , m_grammars(SpeechGrammarList::create()) // FIXME: The spec is not clear on the default value for the grammars attribute. |
| , m_audioTrack(nullptr) |
| , m_continuous(false) |
| , m_interimResults(false) |
| , m_maxAlternatives(1) |
| , m_controller(SpeechRecognitionController::from(page)) |
| , m_stoppedByActiveDOMObject(false) |
| , m_started(false) |
| , m_stopping(false) |
| { |
| // FIXME: Need to hook up with Page to get notified when the visibility changes. |
| } |
| |
| SpeechRecognition::~SpeechRecognition() |
| { |
| } |
| |
| void SpeechRecognition::contextDestroyed() |
| { |
| m_controller = nullptr; |
| PageLifecycleObserver::contextDestroyed(); |
| } |
| |
| DEFINE_TRACE(SpeechRecognition) |
| { |
| visitor->trace(m_grammars); |
| visitor->trace(m_audioTrack); |
| visitor->trace(m_controller); |
| visitor->trace(m_finalResults); |
| EventTargetWithInlineData::trace(visitor); |
| PageLifecycleObserver::trace(visitor); |
| ActiveDOMObject::trace(visitor); |
| } |
| |
| } // namespace blink |