blob: 628ebbccdce194e389633e5c042f2e4aa04b5f91 [file] [log] [blame]
<html><body>
<style>
body, h1, h2, h3, div, span, p, pre, a {
margin: 0;
padding: 0;
border: 0;
font-weight: inherit;
font-style: inherit;
font-size: 100%;
font-family: inherit;
vertical-align: baseline;
}
body {
font-size: 13px;
padding: 1em;
}
h1 {
font-size: 26px;
margin-bottom: 1em;
}
h2 {
font-size: 24px;
margin-bottom: 1em;
}
h3 {
font-size: 20px;
margin-bottom: 1em;
margin-top: 1em;
}
pre, code {
line-height: 1.5;
font-family: Monaco, 'DejaVu Sans Mono', 'Bitstream Vera Sans Mono', 'Lucida Console', monospace;
}
pre {
margin-top: 0.5em;
}
h1, h2, h3, p {
font-family: Arial, sans serif;
}
h1, h2, h3 {
border-bottom: solid #CCC 1px;
}
.toc_element {
margin-top: 0.5em;
}
.firstline {
margin-left: 2 em;
}
.method {
margin-top: 1em;
border: solid 1px #CCC;
padding: 1em;
background: #EEE;
}
.details {
font-weight: bold;
font-size: 14px;
}
</style>
<h1><a href="speech_v1beta1.html">Google Cloud Speech API</a> . <a href="speech_v1beta1.speech.html">speech</a></h1>
<h2>Instance Methods</h2>
<p class="toc_element">
<code><a href="#asyncrecognize">asyncrecognize(body, x__xgafv=None)</a></code></p>
<p class="firstline">Performs asynchronous speech recognition: receive results via the</p>
<p class="toc_element">
<code><a href="#syncrecognize">syncrecognize(body, x__xgafv=None)</a></code></p>
<p class="firstline">Performs synchronous speech recognition: receive results after all audio</p>
<h3>Method Details</h3>
<div class="method">
<code class="details" id="asyncrecognize">asyncrecognize(body, x__xgafv=None)</code>
<pre>Performs asynchronous speech recognition: receive results via the
[google.longrunning.Operations]
(/speech/reference/rest/v1beta1/operations#Operation)
interface. Returns either an
`Operation.error` or an `Operation.response` which contains
an `AsyncRecognizeResponse` message.
Args:
body: object, The request body. (required)
The object takes the form of:
{ # `AsyncRecognizeRequest` is the top-level message sent by the client for
# the `AsyncRecognize` method.
"audio": { # Contains audio data in the encoding specified in the `RecognitionConfig`. # [Required] The audio data to be recognized.
# Either `content` or `uri` must be supplied. Supplying both or neither
# returns google.rpc.Code.INVALID_ARGUMENT. See
# [audio limits](https://cloud.google.com/speech/limits#content).
"content": "A String", # The audio data bytes encoded as specified in
# `RecognitionConfig`. Note: as with all bytes fields, protobuffers use a
# pure binary representation, whereas JSON representations use base64.
"uri": "A String", # URI that points to a file that contains audio data bytes as specified in
# `RecognitionConfig`. Currently, only Google Cloud Storage URIs are
# supported, which must be specified in the following format:
# `gs://bucket_name/object_name` (other URI formats return
# google.rpc.Code.INVALID_ARGUMENT). For more information, see
# [Request URIs](https://cloud.google.com/storage/docs/reference-uris).
},
"config": { # The `RecognitionConfig` message provides information to the recognizer # [Required] The `config` message provides information to the recognizer
# that specifies how to process the request.
# that specifies how to process the request.
"languageCode": "A String", # [Optional] The language of the supplied audio as a BCP-47 language tag.
# Example: "en-GB" https://www.rfc-editor.org/rfc/bcp/bcp47.txt
# If omitted, defaults to "en-US". See
# [Language Support](https://cloud.google.com/speech/docs/languages)
# for a list of the currently supported language codes.
"speechContext": { # Provides "hints" to the speech recognizer to favor specific words and phrases # [Optional] A means to provide context to assist the speech recognition.
# in the results.
"phrases": [ # [Optional] A list of strings containing words and phrases "hints" so that
# the speech recognition is more likely to recognize them. This can be used
# to improve the accuracy for specific words and phrases, for example, if
# specific commands are typically spoken by the user. This can also be used
# to add additional words to the vocabulary of the recognizer. See
# [usage limits](https://cloud.google.com/speech/limits#content).
"A String",
],
},
"encoding": "A String", # [Required] Encoding of audio data sent in all `RecognitionAudio` messages.
"maxAlternatives": 42, # [Optional] Maximum number of recognition hypotheses to be returned.
# Specifically, the maximum number of `SpeechRecognitionAlternative` messages
# within each `SpeechRecognitionResult`.
# The server may return fewer than `max_alternatives`.
# Valid values are `0`-`30`. A value of `0` or `1` will return a maximum of
# `1`. If omitted, defaults to `1`.
"profanityFilter": True or False, # [Optional] If set to `true`, the server will attempt to filter out
# profanities, replacing all but the initial character in each filtered word
# with asterisks, e.g. "f***". If set to `false` or omitted, profanities
# won't be filtered out.
"sampleRate": 42, # [Required] Sample rate in Hertz of the audio data sent in all
# `RecognitionAudio` messages. Valid values are: 8000-48000.
# 16000 is optimal. For best results, set the sampling rate of the audio
# source to 16000 Hz. If that's not possible, use the native sample rate of
# the audio source (instead of re-sampling).
},
}
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
2 - v2 error format
Returns:
An object of the form:
{ # This resource represents a long-running operation that is the result of a
# network API call.
"metadata": { # Service-specific metadata associated with the operation. It typically
# contains progress information and common metadata such as create time.
# Some services might not provide such metadata. Any method that returns a
# long-running operation should document the metadata type, if any.
"a_key": "", # Properties of the object. Contains field @type with type URL.
},
"done": True or False, # If the value is `false`, it means the operation is still in progress.
# If true, the operation is completed, and either `error` or `response` is
# available.
"response": { # The normal response of the operation in case of success. If the original
# method returns no data on success, such as `Delete`, the response is
# `google.protobuf.Empty`. If the original method is standard
# `Get`/`Create`/`Update`, the response should be the resource. For other
# methods, the response should have the type `XxxResponse`, where `Xxx`
# is the original method name. For example, if the original method name
# is `TakeSnapshot()`, the inferred response type is
# `TakeSnapshotResponse`.
"a_key": "", # Properties of the object. Contains field @type with type URL.
},
"name": "A String", # The server-assigned name, which is only unique within the same service that
# originally returns it. If you use the default HTTP mapping, the
# `name` should have the format of `operations/some/unique/name`.
"error": { # The `Status` type defines a logical error model that is suitable for different # The error result of the operation in case of failure or cancellation.
# programming environments, including REST APIs and RPC APIs. It is used by
# [gRPC](https://github.com/grpc). The error model is designed to be:
#
# - Simple to use and understand for most users
# - Flexible enough to meet unexpected needs
#
# # Overview
#
# The `Status` message contains three pieces of data: error code, error message,
# and error details. The error code should be an enum value of
# google.rpc.Code, but it may accept additional error codes if needed. The
# error message should be a developer-facing English message that helps
# developers *understand* and *resolve* the error. If a localized user-facing
# error message is needed, put the localized message in the error details or
# localize it in the client. The optional error details may contain arbitrary
# information about the error. There is a predefined set of error detail types
# in the package `google.rpc` which can be used for common error conditions.
#
# # Language mapping
#
# The `Status` message is the logical representation of the error model, but it
# is not necessarily the actual wire format. When the `Status` message is
# exposed in different client libraries and different wire protocols, it can be
# mapped differently. For example, it will likely be mapped to some exceptions
# in Java, but more likely mapped to some error codes in C.
#
# # Other uses
#
# The error model and the `Status` message can be used in a variety of
# environments, either with or without APIs, to provide a
# consistent developer experience across different environments.
#
# Example uses of this error model include:
#
# - Partial errors. If a service needs to return partial errors to the client,
# it may embed the `Status` in the normal response to indicate the partial
# errors.
#
# - Workflow errors. A typical workflow has multiple steps. Each step may
# have a `Status` message for error reporting purpose.
#
# - Batch operations. If a client uses batch request and batch response, the
# `Status` message should be used directly inside batch response, one for
# each error sub-response.
#
# - Asynchronous operations. If an API call embeds asynchronous operation
# results in its response, the status of those operations should be
# represented directly using the `Status` message.
#
# - Logging. If some API errors are stored in logs, the message `Status` could
# be used directly after any stripping needed for security/privacy reasons.
"message": "A String", # A developer-facing error message, which should be in English. Any
# user-facing error message should be localized and sent in the
# google.rpc.Status.details field, or localized by the client.
"code": 42, # The status code, which should be an enum value of google.rpc.Code.
"details": [ # A list of messages that carry the error details. There will be a
# common set of message types for APIs to use.
{
"a_key": "", # Properties of the object. Contains field @type with type URL.
},
],
},
}</pre>
</div>
<div class="method">
<code class="details" id="syncrecognize">syncrecognize(body, x__xgafv=None)</code>
<pre>Performs synchronous speech recognition: receive results after all audio
has been sent and processed.
Args:
body: object, The request body. (required)
The object takes the form of:
{ # `SyncRecognizeRequest` is the top-level message sent by the client for
# the `SyncRecognize` method.
"audio": { # Contains audio data in the encoding specified in the `RecognitionConfig`. # [Required] The audio data to be recognized.
# Either `content` or `uri` must be supplied. Supplying both or neither
# returns google.rpc.Code.INVALID_ARGUMENT. See
# [audio limits](https://cloud.google.com/speech/limits#content).
"content": "A String", # The audio data bytes encoded as specified in
# `RecognitionConfig`. Note: as with all bytes fields, protobuffers use a
# pure binary representation, whereas JSON representations use base64.
"uri": "A String", # URI that points to a file that contains audio data bytes as specified in
# `RecognitionConfig`. Currently, only Google Cloud Storage URIs are
# supported, which must be specified in the following format:
# `gs://bucket_name/object_name` (other URI formats return
# google.rpc.Code.INVALID_ARGUMENT). For more information, see
# [Request URIs](https://cloud.google.com/storage/docs/reference-uris).
},
"config": { # The `RecognitionConfig` message provides information to the recognizer # [Required] The `config` message provides information to the recognizer
# that specifies how to process the request.
# that specifies how to process the request.
"languageCode": "A String", # [Optional] The language of the supplied audio as a BCP-47 language tag.
# Example: "en-GB" https://www.rfc-editor.org/rfc/bcp/bcp47.txt
# If omitted, defaults to "en-US". See
# [Language Support](https://cloud.google.com/speech/docs/languages)
# for a list of the currently supported language codes.
"speechContext": { # Provides "hints" to the speech recognizer to favor specific words and phrases # [Optional] A means to provide context to assist the speech recognition.
# in the results.
"phrases": [ # [Optional] A list of strings containing words and phrases "hints" so that
# the speech recognition is more likely to recognize them. This can be used
# to improve the accuracy for specific words and phrases, for example, if
# specific commands are typically spoken by the user. This can also be used
# to add additional words to the vocabulary of the recognizer. See
# [usage limits](https://cloud.google.com/speech/limits#content).
"A String",
],
},
"encoding": "A String", # [Required] Encoding of audio data sent in all `RecognitionAudio` messages.
"maxAlternatives": 42, # [Optional] Maximum number of recognition hypotheses to be returned.
# Specifically, the maximum number of `SpeechRecognitionAlternative` messages
# within each `SpeechRecognitionResult`.
# The server may return fewer than `max_alternatives`.
# Valid values are `0`-`30`. A value of `0` or `1` will return a maximum of
# `1`. If omitted, defaults to `1`.
"profanityFilter": True or False, # [Optional] If set to `true`, the server will attempt to filter out
# profanities, replacing all but the initial character in each filtered word
# with asterisks, e.g. "f***". If set to `false` or omitted, profanities
# won't be filtered out.
"sampleRate": 42, # [Required] Sample rate in Hertz of the audio data sent in all
# `RecognitionAudio` messages. Valid values are: 8000-48000.
# 16000 is optimal. For best results, set the sampling rate of the audio
# source to 16000 Hz. If that's not possible, use the native sample rate of
# the audio source (instead of re-sampling).
},
}
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
2 - v2 error format
Returns:
An object of the form:
{ # `SyncRecognizeResponse` is the only message returned to the client by
# `SyncRecognize`. It contains the result as zero or more sequential
# `SpeechRecognitionResult` messages.
"results": [ # [Output-only] Sequential list of transcription results corresponding to
# sequential portions of audio.
{ # A speech recognition result corresponding to a portion of the audio.
"alternatives": [ # [Output-only] May contain one or more recognition hypotheses (up to the
# maximum specified in `max_alternatives`).
{ # Alternative hypotheses (a.k.a. n-best list).
"confidence": 3.14, # [Output-only] The confidence estimate between 0.0 and 1.0. A higher number
# means the system is more confident that the recognition is correct.
# This field is typically provided only for the top hypothesis, and only for
# `is_final=true` results.
# The default of 0.0 is a sentinel value indicating confidence was not set.
"transcript": "A String", # [Output-only] Transcript text representing the words that the user spoke.
},
],
},
],
}</pre>
</div>
</body></html>