On January 1, 2020 this library will no longer support Python 2 on the latest released version. Previously released library versions will continue to be available. For more information please visit Python 2 support on Google Cloud.

Client for Cloud Speech-to-Text API

class google.cloud.speech_v1p1beta1.SpeechClient(transport=None, channel=None, credentials=None, client_config=None, client_info=None, client_options=None)[source]

Service that implements Google Cloud Speech API.

Constructor.

Parameters
  • (Union[SpeechGrpcTransport, (transport) – Callable[[~.Credentials, type], ~.SpeechGrpcTransport]): A transport instance, responsible for actually making the API calls. The default transport uses the gRPC protocol. This argument may also be a callable which returns a transport instance. Callables will be sent the credentials as the first argument and the default transport class as the second argument.

  • channel (grpc.Channel) – DEPRECATED. A Channel instance through which to make calls. This argument is mutually exclusive with credentials; providing both will raise an exception.

  • credentials (google.auth.credentials.Credentials) – The authorization credentials to attach to requests. These credentials identify this application to the service. If none are specified, the client will attempt to ascertain the credentials from the environment. This argument is mutually exclusive with providing a transport instance to transport; doing so will raise an exception.

  • client_config (dict) – DEPRECATED. A dictionary of call options for each method. If not specified, the default configuration is used.

  • client_info (google.api_core.gapic_v1.client_info.ClientInfo) – The client info used to send a user-agent string along with API requests. If None, then default info will be used. Generally, you only need to set this if you’re developing your own client library.

  • client_options (Union[dict, google.api_core.client_options.ClientOptions]) – Client options used to set user options on the client. API Endpoint should be set through client_options.

enums = <module 'google.cloud.speech_v1p1beta1.gapic.enums' from '/tmpfs/src/github/python-speech/google/cloud/speech_v1p1beta1/gapic/enums.py'>
classmethod from_service_account_file(filename, *args, **kwargs)[source]

Creates an instance of this client using the provided credentials file.

Parameters
  • filename (str) – The path to the service account private key json file.

  • args – Additional arguments to pass to the constructor.

  • kwargs – Additional arguments to pass to the constructor.

Returns

The constructed client.

Return type

SpeechClient

classmethod from_service_account_json(filename, *args, **kwargs)

Creates an instance of this client using the provided credentials file.

Parameters
  • filename (str) – The path to the service account private key json file.

  • args – Additional arguments to pass to the constructor.

  • kwargs – Additional arguments to pass to the constructor.

Returns

The constructed client.

Return type

SpeechClient

long_running_recognize(config, audio, retry=<object object>, timeout=<object object>, metadata=None)[source]

Performs asynchronous speech recognition: receive results via the google.longrunning.Operations interface. Returns either an Operation.error or an Operation.response which contains a LongRunningRecognizeResponse message. For more information on asynchronous speech recognition, see the how-to.

Example

>>> from google.cloud import speech_v1p1beta1
>>> from google.cloud.speech_v1p1beta1 import enums
>>>
>>> client = speech_v1p1beta1.SpeechClient()
>>>
>>> encoding = enums.RecognitionConfig.AudioEncoding.FLAC
>>> sample_rate_hertz = 44100
>>> language_code = 'en-US'
>>> config = {'encoding': encoding, 'sample_rate_hertz': sample_rate_hertz, 'language_code': language_code}
>>> uri = 'gs://bucket_name/file_name.flac'
>>> audio = {'uri': uri}
>>>
>>> response = client.long_running_recognize(config, audio)
>>>
>>> def callback(operation_future):
...     # Handle result.
...     result = operation_future.result()
>>>
>>> response.add_done_callback(callback)
>>>
>>> # Handle metadata.
>>> metadata = response.metadata()
Parameters
  • config (Union[dict, RecognitionConfig]) –

    Required. Provides information to the recognizer that specifies how to process the request.

    If a dict is provided, it must be of the same form as the protobuf message RecognitionConfig

  • audio (Union[dict, RecognitionAudio]) –

    Required. The audio data to be recognized.

    If a dict is provided, it must be of the same form as the protobuf message RecognitionAudio

  • retry (Optional[google.api_core.retry.Retry]) – A retry object used to retry requests. If None is specified, requests will be retried using a default configuration.

  • timeout (Optional[float]) – The amount of time, in seconds, to wait for the request to complete. Note that if retry is specified, the timeout applies to each individual attempt.

  • metadata (Optional[Sequence[Tuple[str, str]]]) – Additional metadata that is provided to the method.

Returns

A _OperationFuture instance.

Raises
recognize(config, audio, retry=<object object>, timeout=<object object>, metadata=None)[source]

Performs synchronous speech recognition: receive results after all audio has been sent and processed.

Example

>>> from google.cloud import speech_v1p1beta1
>>> from google.cloud.speech_v1p1beta1 import enums
>>>
>>> client = speech_v1p1beta1.SpeechClient()
>>>
>>> encoding = enums.RecognitionConfig.AudioEncoding.FLAC
>>> sample_rate_hertz = 44100
>>> language_code = 'en-US'
>>> config = {'encoding': encoding, 'sample_rate_hertz': sample_rate_hertz, 'language_code': language_code}
>>> uri = 'gs://bucket_name/file_name.flac'
>>> audio = {'uri': uri}
>>>
>>> response = client.recognize(config, audio)
Parameters
  • config (Union[dict, RecognitionConfig]) –

    Required. Provides information to the recognizer that specifies how to process the request.

    If a dict is provided, it must be of the same form as the protobuf message RecognitionConfig

  • audio (Union[dict, RecognitionAudio]) –

    Required. The audio data to be recognized.

    If a dict is provided, it must be of the same form as the protobuf message RecognitionAudio

  • retry (Optional[google.api_core.retry.Retry]) – A retry object used to retry requests. If None is specified, requests will be retried using a default configuration.

  • timeout (Optional[float]) – The amount of time, in seconds, to wait for the request to complete. Note that if retry is specified, the timeout applies to each individual attempt.

  • metadata (Optional[Sequence[Tuple[str, str]]]) – Additional metadata that is provided to the method.

Returns

A RecognizeResponse instance.

Raises
streaming_recognize(config, requests, retry=<object object>, timeout=<object object>)

Perform bi-directional speech recognition.

This method allows you to receive results while sending audio; it is only available via. gRPC (not REST).

Warning

This method is EXPERIMENTAL. Its interface might change in the future.

Example

>>> from google.cloud.speech_v1 import enums
>>> from google.cloud.speech_v1 import SpeechClient
>>> from google.cloud.speech_v1 import types
>>> client = SpeechClient()
>>> config = types.StreamingRecognitionConfig(
...     config=types.RecognitionConfig(
...         encoding=enums.RecognitionConfig.AudioEncoding.FLAC,
...     ),
... )
>>> request = types.StreamingRecognizeRequest(audio_content=b'...')
>>> requests = [request]
>>> for element in client.streaming_recognize(config, requests):
...     # process element
...     pass
Parameters
  • config (StreamingRecognitionConfig) – The configuration to use for the stream.

  • requests (Iterable[StreamingRecognizeRequest]) – The input objects.

  • retry (Optional[google.api_core.retry.Retry]) – A retry object used to retry requests. If None is specified, requests will not be retried.

  • timeout (Optional[float]) – The amount of time, in seconds, to wait for the request to complete. Note that if retry is specified, the timeout applies to each individual attempt.

Returns

Iterable[StreamingRecognizeResponse]

Raises