// Copyright 2019 Google LLC
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// https://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
'use strict';
const gapicConfig = require('./speech_client_config.json');
const gax = require('google-gax');
const path = require('path');
const protobuf = require('protobufjs');
const VERSION = require('../../package.json').version;
/**
* Service that implements Google Cloud Speech API.
*
* @class
* @memberof v1p1beta1
*/
class SpeechClient {
/**
* Construct an instance of SpeechClient.
*
* @param {object} [options] - The configuration object. See the subsequent
* parameters for more details.
* @param {object} [options.credentials] - Credentials object.
* @param {string} [options.credentials.client_email]
* @param {string} [options.credentials.private_key]
* @param {string} [options.email] - Account email address. Required when
* using a .pem or .p12 keyFilename.
* @param {string} [options.keyFilename] - Full path to the a .json, .pem, or
* .p12 key downloaded from the Google Developers Console. If you provide
* a path to a JSON file, the projectId option below is not necessary.
* NOTE: .pem and .p12 require you to specify options.email as well.
* @param {number} [options.port] - The port on which to connect to
* the remote host.
* @param {string} [options.projectId] - The project ID from the Google
* Developer's Console, e.g. 'grape-spaceship-123'. We will also check
* the environment variable GCLOUD_PROJECT for your project ID. If your
* app is running in an environment which supports
* {@link https://developers.google.com/identity/protocols/application-default-credentials Application Default Credentials},
* your project ID will be detected automatically.
* @param {function} [options.promise] - Custom promise module to use instead
* of native Promises.
* @param {string} [options.apiEndpoint] - The domain name of the
* API remote host.
*/
constructor(opts) {
opts = opts || {};
this._descriptors = {};
const servicePath =
opts.servicePath || opts.apiEndpoint || this.constructor.servicePath;
// Ensure that options include the service address and port.
opts = Object.assign(
{
clientConfig: {},
port: this.constructor.port,
servicePath,
},
opts
);
// Create a `gaxGrpc` object, with any grpc-specific options
// sent to the client.
opts.scopes = this.constructor.scopes;
const gaxGrpc = new gax.GrpcClient(opts);
// Save the auth object to the client, for use by other methods.
this.auth = gaxGrpc.auth;
// Determine the client header string.
const clientHeader = [
`gl-node/${process.versions.node}`,
`grpc/${gaxGrpc.grpcVersion}`,
`gax/${gax.version}`,
`gapic/${VERSION}`,
];
if (opts.libName && opts.libVersion) {
clientHeader.push(`${opts.libName}/${opts.libVersion}`);
}
// Load the applicable protos.
const protos = gaxGrpc.loadProto(
path.join(__dirname, '..', '..', 'protos'),
['google/cloud/speech/v1p1beta1/cloud_speech.proto']
);
// Some of the methods on this service provide streaming responses.
// Provide descriptors for these.
this._descriptors.stream = {
streamingRecognize: new gax.StreamDescriptor(
gax.StreamType.BIDI_STREAMING
),
};
let protoFilesRoot = new gax.GoogleProtoFilesRoot();
protoFilesRoot = protobuf.loadSync(
path.join(
__dirname,
'..',
'..',
'protos',
'google/cloud/speech/v1p1beta1/cloud_speech.proto'
),
protoFilesRoot
);
// This API contains "long-running operations", which return a
// an Operation object that allows for tracking of the operation,
// rather than holding a request open.
this.operationsClient = new gax.lro({
auth: gaxGrpc.auth,
grpc: gaxGrpc.grpc,
}).operationsClient(opts);
const longRunningRecognizeResponse = protoFilesRoot.lookup(
'google.cloud.speech.v1p1beta1.LongRunningRecognizeResponse'
);
const longRunningRecognizeMetadata = protoFilesRoot.lookup(
'google.cloud.speech.v1p1beta1.LongRunningRecognizeMetadata'
);
this._descriptors.longrunning = {
longRunningRecognize: new gax.LongrunningDescriptor(
this.operationsClient,
longRunningRecognizeResponse.decode.bind(longRunningRecognizeResponse),
longRunningRecognizeMetadata.decode.bind(longRunningRecognizeMetadata)
),
};
// Put together the default options sent with requests.
const defaults = gaxGrpc.constructSettings(
'google.cloud.speech.v1p1beta1.Speech',
gapicConfig,
opts.clientConfig,
{'x-goog-api-client': clientHeader.join(' ')}
);
// Set up a dictionary of "inner API calls"; the core implementation
// of calling the API is handled in `google-gax`, with this code
// merely providing the destination and request information.
this._innerApiCalls = {};
// Put together the "service stub" for
// google.cloud.speech.v1p1beta1.Speech.
const speechStub = gaxGrpc.createStub(
protos.google.cloud.speech.v1p1beta1.Speech,
opts
);
// Iterate over each of the methods that the service provides
// and create an API call method for each.
const speechStubMethods = [
'recognize',
'longRunningRecognize',
'streamingRecognize',
];
for (const methodName of speechStubMethods) {
this._innerApiCalls[methodName] = gax.createApiCall(
speechStub.then(
stub =>
function() {
const args = Array.prototype.slice.call(arguments, 0);
return stub[methodName].apply(stub, args);
},
err =>
function() {
throw err;
}
),
defaults[methodName],
this._descriptors.stream[methodName] ||
this._descriptors.longrunning[methodName]
);
}
}
/**
* The DNS address for this API service.
*/
static get servicePath() {
return 'speech.googleapis.com';
}
/**
* The DNS address for this API service - same as servicePath(),
* exists for compatibility reasons.
*/
static get apiEndpoint() {
return 'speech.googleapis.com';
}
/**
* The port for this API service.
*/
static get port() {
return 443;
}
/**
* The scopes needed to make gRPC calls for every method defined
* in this service.
*/
static get scopes() {
return ['https://www.googleapis.com/auth/cloud-platform'];
}
/**
* Return the project ID used by this class.
* @param {function(Error, string)} callback - the callback to
* be called with the current project Id.
*/
getProjectId(callback) {
return this.auth.getProjectId(callback);
}
// -------------------
// -- Service calls --
// -------------------
/**
* Performs synchronous speech recognition: receive results after all audio
* has been sent and processed.
*
* @param {Object} request
* The request object that will be sent.
* @param {Object} request.config
* *Required* Provides information to the recognizer that specifies how to
* process the request.
*
* This object should have the same structure as [RecognitionConfig]{@link google.cloud.speech.v1p1beta1.RecognitionConfig}
* @param {Object} request.audio
* *Required* The audio data to be recognized.
*
* This object should have the same structure as [RecognitionAudio]{@link google.cloud.speech.v1p1beta1.RecognitionAudio}
* @param {Object} [options]
* Optional parameters. You can override the default settings for this call, e.g, timeout,
* retries, paginations, etc. See [gax.CallOptions]{@link https://googleapis.github.io/gax-nodejs/interfaces/CallOptions.html} for the details.
* @param {function(?Error, ?Object)} [callback]
* The function which will be called with the result of the API call.
*
* The second parameter to the callback is an object representing [RecognizeResponse]{@link google.cloud.speech.v1p1beta1.RecognizeResponse}.
* @returns {Promise} - The promise which resolves to an array.
* The first element of the array is an object representing [RecognizeResponse]{@link google.cloud.speech.v1p1beta1.RecognizeResponse}.
* The promise has a method named "cancel" which cancels the ongoing API call.
*
* @example
*
* const speech = require('@google-cloud/speech');
*
* const client = new speech.v1p1beta1.SpeechClient({
* // optional auth parameters.
* });
*
* const encoding = 'FLAC';
* const sampleRateHertz = 44100;
* const languageCode = 'en-US';
* const config = {
* encoding: encoding,
* sampleRateHertz: sampleRateHertz,
* languageCode: languageCode,
* };
* const uri = 'gs://bucket_name/file_name.flac';
* const audio = {
* uri: uri,
* };
* const request = {
* config: config,
* audio: audio,
* };
* client.recognize(request)
* .then(responses => {
* const response = responses[0];
* // doThingsWith(response)
* })
* .catch(err => {
* console.error(err);
* });
*/
recognize(request, options, callback) {
if (options instanceof Function && callback === undefined) {
callback = options;
options = {};
}
request = request || {};
options = options || {};
return this._innerApiCalls.recognize(request, options, callback);
}
/**
* Performs asynchronous speech recognition: receive results via the
* google.longrunning.Operations interface. Returns either an
* `Operation.error` or an `Operation.response` which contains
* a `LongRunningRecognizeResponse` message.
*
* @param {Object} request
* The request object that will be sent.
* @param {Object} request.config
* *Required* Provides information to the recognizer that specifies how to
* process the request.
*
* This object should have the same structure as [RecognitionConfig]{@link google.cloud.speech.v1p1beta1.RecognitionConfig}
* @param {Object} request.audio
* *Required* The audio data to be recognized.
*
* This object should have the same structure as [RecognitionAudio]{@link google.cloud.speech.v1p1beta1.RecognitionAudio}
* @param {Object} [options]
* Optional parameters. You can override the default settings for this call, e.g, timeout,
* retries, paginations, etc. See [gax.CallOptions]{@link https://googleapis.github.io/gax-nodejs/interfaces/CallOptions.html} for the details.
* @param {function(?Error, ?Object)} [callback]
* The function which will be called with the result of the API call.
*
* The second parameter to the callback is a [gax.Operation]{@link https://googleapis.github.io/gax-nodejs/classes/Operation.html} object.
* @returns {Promise} - The promise which resolves to an array.
* The first element of the array is a [gax.Operation]{@link https://googleapis.github.io/gax-nodejs/classes/Operation.html} object.
* The promise has a method named "cancel" which cancels the ongoing API call.
*
* @example
*
* const speech = require('@google-cloud/speech');
*
* const client = new speech.v1p1beta1.SpeechClient({
* // optional auth parameters.
* });
*
* const encoding = 'FLAC';
* const sampleRateHertz = 44100;
* const languageCode = 'en-US';
* const config = {
* encoding: encoding,
* sampleRateHertz: sampleRateHertz,
* languageCode: languageCode,
* };
* const uri = 'gs://bucket_name/file_name.flac';
* const audio = {
* uri: uri,
* };
* const request = {
* config: config,
* audio: audio,
* };
*
* // Handle the operation using the promise pattern.
* client.longRunningRecognize(request)
* .then(responses => {
* const [operation, initialApiResponse] = responses;
*
* // Operation#promise starts polling for the completion of the LRO.
* return operation.promise();
* })
* .then(responses => {
* const result = responses[0];
* const metadata = responses[1];
* const finalApiResponse = responses[2];
* })
* .catch(err => {
* console.error(err);
* });
*
* const encoding = 'FLAC';
* const sampleRateHertz = 44100;
* const languageCode = 'en-US';
* const config = {
* encoding: encoding,
* sampleRateHertz: sampleRateHertz,
* languageCode: languageCode,
* };
* const uri = 'gs://bucket_name/file_name.flac';
* const audio = {
* uri: uri,
* };
* const request = {
* config: config,
* audio: audio,
* };
*
* // Handle the operation using the event emitter pattern.
* client.longRunningRecognize(request)
* .then(responses => {
* const [operation, initialApiResponse] = responses;
*
* // Adding a listener for the "complete" event starts polling for the
* // completion of the operation.
* operation.on('complete', (result, metadata, finalApiResponse) => {
* // doSomethingWith(result);
* });
*
* // Adding a listener for the "progress" event causes the callback to be
* // called on any change in metadata when the operation is polled.
* operation.on('progress', (metadata, apiResponse) => {
* // doSomethingWith(metadata)
* });
*
* // Adding a listener for the "error" event handles any errors found during polling.
* operation.on('error', err => {
* // throw(err);
* });
* })
* .catch(err => {
* console.error(err);
* });
*
* const encoding = 'FLAC';
* const sampleRateHertz = 44100;
* const languageCode = 'en-US';
* const config = {
* encoding: encoding,
* sampleRateHertz: sampleRateHertz,
* languageCode: languageCode,
* };
* const uri = 'gs://bucket_name/file_name.flac';
* const audio = {
* uri: uri,
* };
* const request = {
* config: config,
* audio: audio,
* };
*
* // Handle the operation using the await pattern.
* const [operation] = await client.longRunningRecognize(request);
*
* const [response] = await operation.promise();
*/
longRunningRecognize(request, options, callback) {
if (options instanceof Function && callback === undefined) {
callback = options;
options = {};
}
request = request || {};
options = options || {};
return this._innerApiCalls.longRunningRecognize(request, options, callback);
}
/**
* Performs bidirectional streaming speech recognition: receive results while
* sending audio. This method is only available via the gRPC API (not REST).
*
* @param {Object} [options]
* Optional parameters. You can override the default settings for this call, e.g, timeout,
* retries, paginations, etc. See [gax.CallOptions]{@link https://googleapis.github.io/gax-nodejs/interfaces/CallOptions.html} for the details.
* @returns {Stream}
* An object stream which is both readable and writable. It accepts objects
* representing [StreamingRecognizeRequest]{@link google.cloud.speech.v1p1beta1.StreamingRecognizeRequest} for write() method, and
* will emit objects representing [StreamingRecognizeResponse]{@link google.cloud.speech.v1p1beta1.StreamingRecognizeResponse} on 'data' event asynchronously.
*
* @example
*
* const speech = require('@google-cloud/speech');
*
* const client = new speech.v1p1beta1.SpeechClient({
* // optional auth parameters.
* });
*
* const stream = client.streamingRecognize().on('data', response => {
* // doThingsWith(response)
* });
* const request = {};
* // Write request objects.
* stream.write(request);
*/
streamingRecognize(options) {
options = options || {};
return this._innerApiCalls.streamingRecognize(options);
}
}
module.exports = SpeechClient;