public static final class StreamingRecognitionResult.Builder extends GeneratedMessageV3.Builder<StreamingRecognitionResult.Builder> implements StreamingRecognitionResultOrBuilder
Contains a speech recognition result corresponding to a portion of the audio that is currently being processed or an indication that this is the end of the single requested utterance. While end-user audio is being processed, Dialogflow sends a series of results. Each result may contain a `transcript` value. A transcript represents a portion of the utterance. While the recognizer is processing audio, transcript values may be interim values or finalized values. Once a transcript is finalized, the `is_final` value is set to true and processing continues for the next transcript. If `StreamingDetectIntentRequest.query_input.audio_config.single_utterance` was true, and the recognizer has completed processing audio, the `message_type` value is set to `END_OF_SINGLE_UTTERANCE and the following (last) result contains the last finalized transcript. The complete end-user utterance is determined by concatenating the finalized transcript values received for the series of results. In the following example, single utterance is enabled. In the case where single utterance is not enabled, result 7 would not occur. ``` Num | transcript | message_type | is_final --- | ----------------------- | ----------------------- | -------- 1 | "tube" | TRANSCRIPT | false 2 | "to be a" | TRANSCRIPT | false 3 | "to be" | TRANSCRIPT | false 4 | "to be or not to be" | TRANSCRIPT | true 5 | "that's" | TRANSCRIPT | false 6 | "that is | TRANSCRIPT | false 7 | unset | END_OF_SINGLE_UTTERANCE | unset 8 | " that is the question" | TRANSCRIPT | true ``` Concatenating the finalized transcripts with `is_final` set to true, the complete utterance becomes "to be or not to be that is the question".Protobuf type
google.cloud.dialogflow.v2beta1.StreamingRecognitionResult
Modifier and Type | Method and Description |
---|---|
StreamingRecognitionResult.Builder |
addAllSpeechWordInfo(Iterable<? extends SpeechWordInfo> values)
Word-specific information for the words recognized by Speech in
[transcript][google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.transcript].
|
StreamingRecognitionResult.Builder |
addRepeatedField(Descriptors.FieldDescriptor field,
Object value) |
StreamingRecognitionResult.Builder |
addSpeechWordInfo(int index,
SpeechWordInfo.Builder builderForValue)
Word-specific information for the words recognized by Speech in
[transcript][google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.transcript].
|
StreamingRecognitionResult.Builder |
addSpeechWordInfo(int index,
SpeechWordInfo value)
Word-specific information for the words recognized by Speech in
[transcript][google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.transcript].
|
StreamingRecognitionResult.Builder |
addSpeechWordInfo(SpeechWordInfo.Builder builderForValue)
Word-specific information for the words recognized by Speech in
[transcript][google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.transcript].
|
StreamingRecognitionResult.Builder |
addSpeechWordInfo(SpeechWordInfo value)
Word-specific information for the words recognized by Speech in
[transcript][google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.transcript].
|
SpeechWordInfo.Builder |
addSpeechWordInfoBuilder()
Word-specific information for the words recognized by Speech in
[transcript][google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.transcript].
|
SpeechWordInfo.Builder |
addSpeechWordInfoBuilder(int index)
Word-specific information for the words recognized by Speech in
[transcript][google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.transcript].
|
StreamingRecognitionResult |
build() |
StreamingRecognitionResult |
buildPartial() |
StreamingRecognitionResult.Builder |
clear() |
StreamingRecognitionResult.Builder |
clearConfidence()
The Speech confidence between 0.0 and 1.0 for the current portion of audio.
|
StreamingRecognitionResult.Builder |
clearDtmfDigits()
DTMF digits.
|
StreamingRecognitionResult.Builder |
clearField(Descriptors.FieldDescriptor field) |
StreamingRecognitionResult.Builder |
clearIsFinal()
If `false`, the `StreamingRecognitionResult` represents an
interim result that may change.
|
StreamingRecognitionResult.Builder |
clearLanguageCode()
Detected language code for the transcript.
|
StreamingRecognitionResult.Builder |
clearMessageType()
Type of the result message.
|
StreamingRecognitionResult.Builder |
clearOneof(Descriptors.OneofDescriptor oneof) |
StreamingRecognitionResult.Builder |
clearSpeechEndOffset()
Time offset of the end of this Speech recognition result relative to the
beginning of the audio.
|
StreamingRecognitionResult.Builder |
clearSpeechWordInfo()
Word-specific information for the words recognized by Speech in
[transcript][google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.transcript].
|
StreamingRecognitionResult.Builder |
clearStability()
An estimate of the likelihood that the speech recognizer will
not change its guess about this interim recognition result:
* If the value is unspecified or 0.0, Dialogflow didn't compute the
stability.
|
StreamingRecognitionResult.Builder |
clearTranscript()
Transcript text representing the words that the user spoke.
|
StreamingRecognitionResult.Builder |
clone() |
float |
getConfidence()
The Speech confidence between 0.0 and 1.0 for the current portion of audio.
|
StreamingRecognitionResult |
getDefaultInstanceForType() |
static Descriptors.Descriptor |
getDescriptor() |
Descriptors.Descriptor |
getDescriptorForType() |
TelephonyDtmfEvents |
getDtmfDigits()
DTMF digits.
|
TelephonyDtmfEvents.Builder |
getDtmfDigitsBuilder()
DTMF digits.
|
TelephonyDtmfEventsOrBuilder |
getDtmfDigitsOrBuilder()
DTMF digits.
|
boolean |
getIsFinal()
If `false`, the `StreamingRecognitionResult` represents an
interim result that may change.
|
String |
getLanguageCode()
Detected language code for the transcript.
|
ByteString |
getLanguageCodeBytes()
Detected language code for the transcript.
|
StreamingRecognitionResult.MessageType |
getMessageType()
Type of the result message.
|
int |
getMessageTypeValue()
Type of the result message.
|
Duration |
getSpeechEndOffset()
Time offset of the end of this Speech recognition result relative to the
beginning of the audio.
|
Duration.Builder |
getSpeechEndOffsetBuilder()
Time offset of the end of this Speech recognition result relative to the
beginning of the audio.
|
DurationOrBuilder |
getSpeechEndOffsetOrBuilder()
Time offset of the end of this Speech recognition result relative to the
beginning of the audio.
|
SpeechWordInfo |
getSpeechWordInfo(int index)
Word-specific information for the words recognized by Speech in
[transcript][google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.transcript].
|
SpeechWordInfo.Builder |
getSpeechWordInfoBuilder(int index)
Word-specific information for the words recognized by Speech in
[transcript][google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.transcript].
|
List<SpeechWordInfo.Builder> |
getSpeechWordInfoBuilderList()
Word-specific information for the words recognized by Speech in
[transcript][google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.transcript].
|
int |
getSpeechWordInfoCount()
Word-specific information for the words recognized by Speech in
[transcript][google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.transcript].
|
List<SpeechWordInfo> |
getSpeechWordInfoList()
Word-specific information for the words recognized by Speech in
[transcript][google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.transcript].
|
SpeechWordInfoOrBuilder |
getSpeechWordInfoOrBuilder(int index)
Word-specific information for the words recognized by Speech in
[transcript][google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.transcript].
|
List<? extends SpeechWordInfoOrBuilder> |
getSpeechWordInfoOrBuilderList()
Word-specific information for the words recognized by Speech in
[transcript][google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.transcript].
|
float |
getStability()
An estimate of the likelihood that the speech recognizer will
not change its guess about this interim recognition result:
* If the value is unspecified or 0.0, Dialogflow didn't compute the
stability.
|
String |
getTranscript()
Transcript text representing the words that the user spoke.
|
ByteString |
getTranscriptBytes()
Transcript text representing the words that the user spoke.
|
boolean |
hasDtmfDigits()
DTMF digits.
|
boolean |
hasSpeechEndOffset()
Time offset of the end of this Speech recognition result relative to the
beginning of the audio.
|
protected GeneratedMessageV3.FieldAccessorTable |
internalGetFieldAccessorTable() |
boolean |
isInitialized() |
StreamingRecognitionResult.Builder |
mergeDtmfDigits(TelephonyDtmfEvents value)
DTMF digits.
|
StreamingRecognitionResult.Builder |
mergeFrom(CodedInputStream input,
ExtensionRegistryLite extensionRegistry) |
StreamingRecognitionResult.Builder |
mergeFrom(Message other) |
StreamingRecognitionResult.Builder |
mergeFrom(StreamingRecognitionResult other) |
StreamingRecognitionResult.Builder |
mergeSpeechEndOffset(Duration value)
Time offset of the end of this Speech recognition result relative to the
beginning of the audio.
|
StreamingRecognitionResult.Builder |
mergeUnknownFields(UnknownFieldSet unknownFields) |
StreamingRecognitionResult.Builder |
removeSpeechWordInfo(int index)
Word-specific information for the words recognized by Speech in
[transcript][google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.transcript].
|
StreamingRecognitionResult.Builder |
setConfidence(float value)
The Speech confidence between 0.0 and 1.0 for the current portion of audio.
|
StreamingRecognitionResult.Builder |
setDtmfDigits(TelephonyDtmfEvents.Builder builderForValue)
DTMF digits.
|
StreamingRecognitionResult.Builder |
setDtmfDigits(TelephonyDtmfEvents value)
DTMF digits.
|
StreamingRecognitionResult.Builder |
setField(Descriptors.FieldDescriptor field,
Object value) |
StreamingRecognitionResult.Builder |
setIsFinal(boolean value)
If `false`, the `StreamingRecognitionResult` represents an
interim result that may change.
|
StreamingRecognitionResult.Builder |
setLanguageCode(String value)
Detected language code for the transcript.
|
StreamingRecognitionResult.Builder |
setLanguageCodeBytes(ByteString value)
Detected language code for the transcript.
|
StreamingRecognitionResult.Builder |
setMessageType(StreamingRecognitionResult.MessageType value)
Type of the result message.
|
StreamingRecognitionResult.Builder |
setMessageTypeValue(int value)
Type of the result message.
|
StreamingRecognitionResult.Builder |
setRepeatedField(Descriptors.FieldDescriptor field,
int index,
Object value) |
StreamingRecognitionResult.Builder |
setSpeechEndOffset(Duration.Builder builderForValue)
Time offset of the end of this Speech recognition result relative to the
beginning of the audio.
|
StreamingRecognitionResult.Builder |
setSpeechEndOffset(Duration value)
Time offset of the end of this Speech recognition result relative to the
beginning of the audio.
|
StreamingRecognitionResult.Builder |
setSpeechWordInfo(int index,
SpeechWordInfo.Builder builderForValue)
Word-specific information for the words recognized by Speech in
[transcript][google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.transcript].
|
StreamingRecognitionResult.Builder |
setSpeechWordInfo(int index,
SpeechWordInfo value)
Word-specific information for the words recognized by Speech in
[transcript][google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.transcript].
|
StreamingRecognitionResult.Builder |
setStability(float value)
An estimate of the likelihood that the speech recognizer will
not change its guess about this interim recognition result:
* If the value is unspecified or 0.0, Dialogflow didn't compute the
stability.
|
StreamingRecognitionResult.Builder |
setTranscript(String value)
Transcript text representing the words that the user spoke.
|
StreamingRecognitionResult.Builder |
setTranscriptBytes(ByteString value)
Transcript text representing the words that the user spoke.
|
StreamingRecognitionResult.Builder |
setUnknownFields(UnknownFieldSet unknownFields) |
getAllFields, getField, getFieldBuilder, getOneofFieldDescriptor, getParentForChildren, getRepeatedField, getRepeatedFieldBuilder, getRepeatedFieldCount, getUnknownFields, getUnknownFieldSetBuilder, hasField, hasOneof, internalGetMapField, internalGetMutableMapField, isClean, markClean, mergeUnknownLengthDelimitedField, mergeUnknownVarintField, newBuilderForField, onBuilt, onChanged, parseUnknownField, setUnknownFieldSetBuilder, setUnknownFieldsProto3
findInitializationErrors, getInitializationErrorString, internalMergeFrom, mergeFrom, mergeFrom, mergeFrom, mergeFrom, mergeFrom, mergeFrom, mergeFrom, mergeFrom, mergeFrom, newUninitializedMessageException, toString
addAll, addAll, mergeDelimitedFrom, mergeDelimitedFrom, mergeFrom, newUninitializedMessageException
equals, finalize, getClass, hashCode, notify, notifyAll, wait, wait, wait
findInitializationErrors, getAllFields, getField, getInitializationErrorString, getOneofFieldDescriptor, getRepeatedField, getRepeatedFieldCount, getUnknownFields, hasField, hasOneof
mergeDelimitedFrom, mergeDelimitedFrom
mergeFrom
public static final Descriptors.Descriptor getDescriptor()
protected GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable()
internalGetFieldAccessorTable
in class GeneratedMessageV3.Builder<StreamingRecognitionResult.Builder>
public StreamingRecognitionResult.Builder clear()
clear
in interface Message.Builder
clear
in interface MessageLite.Builder
clear
in class GeneratedMessageV3.Builder<StreamingRecognitionResult.Builder>
public Descriptors.Descriptor getDescriptorForType()
getDescriptorForType
in interface Message.Builder
getDescriptorForType
in interface MessageOrBuilder
getDescriptorForType
in class GeneratedMessageV3.Builder<StreamingRecognitionResult.Builder>
public StreamingRecognitionResult getDefaultInstanceForType()
getDefaultInstanceForType
in interface MessageLiteOrBuilder
getDefaultInstanceForType
in interface MessageOrBuilder
public StreamingRecognitionResult build()
build
in interface Message.Builder
build
in interface MessageLite.Builder
public StreamingRecognitionResult buildPartial()
buildPartial
in interface Message.Builder
buildPartial
in interface MessageLite.Builder
public StreamingRecognitionResult.Builder clone()
clone
in interface Message.Builder
clone
in interface MessageLite.Builder
clone
in class GeneratedMessageV3.Builder<StreamingRecognitionResult.Builder>
public StreamingRecognitionResult.Builder setField(Descriptors.FieldDescriptor field, Object value)
setField
in interface Message.Builder
setField
in class GeneratedMessageV3.Builder<StreamingRecognitionResult.Builder>
public StreamingRecognitionResult.Builder clearField(Descriptors.FieldDescriptor field)
clearField
in interface Message.Builder
clearField
in class GeneratedMessageV3.Builder<StreamingRecognitionResult.Builder>
public StreamingRecognitionResult.Builder clearOneof(Descriptors.OneofDescriptor oneof)
clearOneof
in interface Message.Builder
clearOneof
in class GeneratedMessageV3.Builder<StreamingRecognitionResult.Builder>
public StreamingRecognitionResult.Builder setRepeatedField(Descriptors.FieldDescriptor field, int index, Object value)
setRepeatedField
in interface Message.Builder
setRepeatedField
in class GeneratedMessageV3.Builder<StreamingRecognitionResult.Builder>
public StreamingRecognitionResult.Builder addRepeatedField(Descriptors.FieldDescriptor field, Object value)
addRepeatedField
in interface Message.Builder
addRepeatedField
in class GeneratedMessageV3.Builder<StreamingRecognitionResult.Builder>
public StreamingRecognitionResult.Builder mergeFrom(Message other)
mergeFrom
in interface Message.Builder
mergeFrom
in class AbstractMessage.Builder<StreamingRecognitionResult.Builder>
public StreamingRecognitionResult.Builder mergeFrom(StreamingRecognitionResult other)
public final boolean isInitialized()
isInitialized
in interface MessageLiteOrBuilder
isInitialized
in class GeneratedMessageV3.Builder<StreamingRecognitionResult.Builder>
public StreamingRecognitionResult.Builder mergeFrom(CodedInputStream input, ExtensionRegistryLite extensionRegistry) throws IOException
mergeFrom
in interface Message.Builder
mergeFrom
in interface MessageLite.Builder
mergeFrom
in class AbstractMessage.Builder<StreamingRecognitionResult.Builder>
IOException
public int getMessageTypeValue()
Type of the result message.
.google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.MessageType message_type = 1;
getMessageTypeValue
in interface StreamingRecognitionResultOrBuilder
public StreamingRecognitionResult.Builder setMessageTypeValue(int value)
Type of the result message.
.google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.MessageType message_type = 1;
value
- The enum numeric value on the wire for messageType to set.public StreamingRecognitionResult.MessageType getMessageType()
Type of the result message.
.google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.MessageType message_type = 1;
getMessageType
in interface StreamingRecognitionResultOrBuilder
public StreamingRecognitionResult.Builder setMessageType(StreamingRecognitionResult.MessageType value)
Type of the result message.
.google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.MessageType message_type = 1;
value
- The messageType to set.public StreamingRecognitionResult.Builder clearMessageType()
Type of the result message.
.google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.MessageType message_type = 1;
public String getTranscript()
Transcript text representing the words that the user spoke. Populated if and only if `message_type` = `TRANSCRIPT`.
string transcript = 2;
getTranscript
in interface StreamingRecognitionResultOrBuilder
public ByteString getTranscriptBytes()
Transcript text representing the words that the user spoke. Populated if and only if `message_type` = `TRANSCRIPT`.
string transcript = 2;
getTranscriptBytes
in interface StreamingRecognitionResultOrBuilder
public StreamingRecognitionResult.Builder setTranscript(String value)
Transcript text representing the words that the user spoke. Populated if and only if `message_type` = `TRANSCRIPT`.
string transcript = 2;
value
- The transcript to set.public StreamingRecognitionResult.Builder clearTranscript()
Transcript text representing the words that the user spoke. Populated if and only if `message_type` = `TRANSCRIPT`.
string transcript = 2;
public StreamingRecognitionResult.Builder setTranscriptBytes(ByteString value)
Transcript text representing the words that the user spoke. Populated if and only if `message_type` = `TRANSCRIPT`.
string transcript = 2;
value
- The bytes for transcript to set.public boolean getIsFinal()
If `false`, the `StreamingRecognitionResult` represents an interim result that may change. If `true`, the recognizer will not return any further hypotheses about this piece of the audio. May only be populated for `message_type` = `TRANSCRIPT`.
bool is_final = 3;
getIsFinal
in interface StreamingRecognitionResultOrBuilder
public StreamingRecognitionResult.Builder setIsFinal(boolean value)
If `false`, the `StreamingRecognitionResult` represents an interim result that may change. If `true`, the recognizer will not return any further hypotheses about this piece of the audio. May only be populated for `message_type` = `TRANSCRIPT`.
bool is_final = 3;
value
- The isFinal to set.public StreamingRecognitionResult.Builder clearIsFinal()
If `false`, the `StreamingRecognitionResult` represents an interim result that may change. If `true`, the recognizer will not return any further hypotheses about this piece of the audio. May only be populated for `message_type` = `TRANSCRIPT`.
bool is_final = 3;
public float getConfidence()
The Speech confidence between 0.0 and 1.0 for the current portion of audio. A higher number indicates an estimated greater likelihood that the recognized words are correct. The default of 0.0 is a sentinel value indicating that confidence was not set. This field is typically only provided if `is_final` is true and you should not rely on it being accurate or even set.
float confidence = 4;
getConfidence
in interface StreamingRecognitionResultOrBuilder
public StreamingRecognitionResult.Builder setConfidence(float value)
The Speech confidence between 0.0 and 1.0 for the current portion of audio. A higher number indicates an estimated greater likelihood that the recognized words are correct. The default of 0.0 is a sentinel value indicating that confidence was not set. This field is typically only provided if `is_final` is true and you should not rely on it being accurate or even set.
float confidence = 4;
value
- The confidence to set.public StreamingRecognitionResult.Builder clearConfidence()
The Speech confidence between 0.0 and 1.0 for the current portion of audio. A higher number indicates an estimated greater likelihood that the recognized words are correct. The default of 0.0 is a sentinel value indicating that confidence was not set. This field is typically only provided if `is_final` is true and you should not rely on it being accurate or even set.
float confidence = 4;
public float getStability()
An estimate of the likelihood that the speech recognizer will not change its guess about this interim recognition result: * If the value is unspecified or 0.0, Dialogflow didn't compute the stability. In particular, Dialogflow will only provide stability for `TRANSCRIPT` results with `is_final = false`. * Otherwise, the value is in (0.0, 1.0] where 0.0 means completely unstable and 1.0 means completely stable.
float stability = 6;
getStability
in interface StreamingRecognitionResultOrBuilder
public StreamingRecognitionResult.Builder setStability(float value)
An estimate of the likelihood that the speech recognizer will not change its guess about this interim recognition result: * If the value is unspecified or 0.0, Dialogflow didn't compute the stability. In particular, Dialogflow will only provide stability for `TRANSCRIPT` results with `is_final = false`. * Otherwise, the value is in (0.0, 1.0] where 0.0 means completely unstable and 1.0 means completely stable.
float stability = 6;
value
- The stability to set.public StreamingRecognitionResult.Builder clearStability()
An estimate of the likelihood that the speech recognizer will not change its guess about this interim recognition result: * If the value is unspecified or 0.0, Dialogflow didn't compute the stability. In particular, Dialogflow will only provide stability for `TRANSCRIPT` results with `is_final = false`. * Otherwise, the value is in (0.0, 1.0] where 0.0 means completely unstable and 1.0 means completely stable.
float stability = 6;
public List<SpeechWordInfo> getSpeechWordInfoList()
Word-specific information for the words recognized by Speech in [transcript][google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.transcript]. Populated if and only if `message_type` = `TRANSCRIPT` and [InputAudioConfig.enable_word_info] is set.
repeated .google.cloud.dialogflow.v2beta1.SpeechWordInfo speech_word_info = 7;
getSpeechWordInfoList
in interface StreamingRecognitionResultOrBuilder
public int getSpeechWordInfoCount()
Word-specific information for the words recognized by Speech in [transcript][google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.transcript]. Populated if and only if `message_type` = `TRANSCRIPT` and [InputAudioConfig.enable_word_info] is set.
repeated .google.cloud.dialogflow.v2beta1.SpeechWordInfo speech_word_info = 7;
getSpeechWordInfoCount
in interface StreamingRecognitionResultOrBuilder
public SpeechWordInfo getSpeechWordInfo(int index)
Word-specific information for the words recognized by Speech in [transcript][google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.transcript]. Populated if and only if `message_type` = `TRANSCRIPT` and [InputAudioConfig.enable_word_info] is set.
repeated .google.cloud.dialogflow.v2beta1.SpeechWordInfo speech_word_info = 7;
getSpeechWordInfo
in interface StreamingRecognitionResultOrBuilder
public StreamingRecognitionResult.Builder setSpeechWordInfo(int index, SpeechWordInfo value)
Word-specific information for the words recognized by Speech in [transcript][google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.transcript]. Populated if and only if `message_type` = `TRANSCRIPT` and [InputAudioConfig.enable_word_info] is set.
repeated .google.cloud.dialogflow.v2beta1.SpeechWordInfo speech_word_info = 7;
public StreamingRecognitionResult.Builder setSpeechWordInfo(int index, SpeechWordInfo.Builder builderForValue)
Word-specific information for the words recognized by Speech in [transcript][google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.transcript]. Populated if and only if `message_type` = `TRANSCRIPT` and [InputAudioConfig.enable_word_info] is set.
repeated .google.cloud.dialogflow.v2beta1.SpeechWordInfo speech_word_info = 7;
public StreamingRecognitionResult.Builder addSpeechWordInfo(SpeechWordInfo value)
Word-specific information for the words recognized by Speech in [transcript][google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.transcript]. Populated if and only if `message_type` = `TRANSCRIPT` and [InputAudioConfig.enable_word_info] is set.
repeated .google.cloud.dialogflow.v2beta1.SpeechWordInfo speech_word_info = 7;
public StreamingRecognitionResult.Builder addSpeechWordInfo(int index, SpeechWordInfo value)
Word-specific information for the words recognized by Speech in [transcript][google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.transcript]. Populated if and only if `message_type` = `TRANSCRIPT` and [InputAudioConfig.enable_word_info] is set.
repeated .google.cloud.dialogflow.v2beta1.SpeechWordInfo speech_word_info = 7;
public StreamingRecognitionResult.Builder addSpeechWordInfo(SpeechWordInfo.Builder builderForValue)
Word-specific information for the words recognized by Speech in [transcript][google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.transcript]. Populated if and only if `message_type` = `TRANSCRIPT` and [InputAudioConfig.enable_word_info] is set.
repeated .google.cloud.dialogflow.v2beta1.SpeechWordInfo speech_word_info = 7;
public StreamingRecognitionResult.Builder addSpeechWordInfo(int index, SpeechWordInfo.Builder builderForValue)
Word-specific information for the words recognized by Speech in [transcript][google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.transcript]. Populated if and only if `message_type` = `TRANSCRIPT` and [InputAudioConfig.enable_word_info] is set.
repeated .google.cloud.dialogflow.v2beta1.SpeechWordInfo speech_word_info = 7;
public StreamingRecognitionResult.Builder addAllSpeechWordInfo(Iterable<? extends SpeechWordInfo> values)
Word-specific information for the words recognized by Speech in [transcript][google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.transcript]. Populated if and only if `message_type` = `TRANSCRIPT` and [InputAudioConfig.enable_word_info] is set.
repeated .google.cloud.dialogflow.v2beta1.SpeechWordInfo speech_word_info = 7;
public StreamingRecognitionResult.Builder clearSpeechWordInfo()
Word-specific information for the words recognized by Speech in [transcript][google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.transcript]. Populated if and only if `message_type` = `TRANSCRIPT` and [InputAudioConfig.enable_word_info] is set.
repeated .google.cloud.dialogflow.v2beta1.SpeechWordInfo speech_word_info = 7;
public StreamingRecognitionResult.Builder removeSpeechWordInfo(int index)
Word-specific information for the words recognized by Speech in [transcript][google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.transcript]. Populated if and only if `message_type` = `TRANSCRIPT` and [InputAudioConfig.enable_word_info] is set.
repeated .google.cloud.dialogflow.v2beta1.SpeechWordInfo speech_word_info = 7;
public SpeechWordInfo.Builder getSpeechWordInfoBuilder(int index)
Word-specific information for the words recognized by Speech in [transcript][google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.transcript]. Populated if and only if `message_type` = `TRANSCRIPT` and [InputAudioConfig.enable_word_info] is set.
repeated .google.cloud.dialogflow.v2beta1.SpeechWordInfo speech_word_info = 7;
public SpeechWordInfoOrBuilder getSpeechWordInfoOrBuilder(int index)
Word-specific information for the words recognized by Speech in [transcript][google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.transcript]. Populated if and only if `message_type` = `TRANSCRIPT` and [InputAudioConfig.enable_word_info] is set.
repeated .google.cloud.dialogflow.v2beta1.SpeechWordInfo speech_word_info = 7;
getSpeechWordInfoOrBuilder
in interface StreamingRecognitionResultOrBuilder
public List<? extends SpeechWordInfoOrBuilder> getSpeechWordInfoOrBuilderList()
Word-specific information for the words recognized by Speech in [transcript][google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.transcript]. Populated if and only if `message_type` = `TRANSCRIPT` and [InputAudioConfig.enable_word_info] is set.
repeated .google.cloud.dialogflow.v2beta1.SpeechWordInfo speech_word_info = 7;
getSpeechWordInfoOrBuilderList
in interface StreamingRecognitionResultOrBuilder
public SpeechWordInfo.Builder addSpeechWordInfoBuilder()
Word-specific information for the words recognized by Speech in [transcript][google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.transcript]. Populated if and only if `message_type` = `TRANSCRIPT` and [InputAudioConfig.enable_word_info] is set.
repeated .google.cloud.dialogflow.v2beta1.SpeechWordInfo speech_word_info = 7;
public SpeechWordInfo.Builder addSpeechWordInfoBuilder(int index)
Word-specific information for the words recognized by Speech in [transcript][google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.transcript]. Populated if and only if `message_type` = `TRANSCRIPT` and [InputAudioConfig.enable_word_info] is set.
repeated .google.cloud.dialogflow.v2beta1.SpeechWordInfo speech_word_info = 7;
public List<SpeechWordInfo.Builder> getSpeechWordInfoBuilderList()
Word-specific information for the words recognized by Speech in [transcript][google.cloud.dialogflow.v2beta1.StreamingRecognitionResult.transcript]. Populated if and only if `message_type` = `TRANSCRIPT` and [InputAudioConfig.enable_word_info] is set.
repeated .google.cloud.dialogflow.v2beta1.SpeechWordInfo speech_word_info = 7;
public boolean hasSpeechEndOffset()
Time offset of the end of this Speech recognition result relative to the beginning of the audio. Only populated for `message_type` = `TRANSCRIPT`.
.google.protobuf.Duration speech_end_offset = 8;
hasSpeechEndOffset
in interface StreamingRecognitionResultOrBuilder
public Duration getSpeechEndOffset()
Time offset of the end of this Speech recognition result relative to the beginning of the audio. Only populated for `message_type` = `TRANSCRIPT`.
.google.protobuf.Duration speech_end_offset = 8;
getSpeechEndOffset
in interface StreamingRecognitionResultOrBuilder
public StreamingRecognitionResult.Builder setSpeechEndOffset(Duration value)
Time offset of the end of this Speech recognition result relative to the beginning of the audio. Only populated for `message_type` = `TRANSCRIPT`.
.google.protobuf.Duration speech_end_offset = 8;
public StreamingRecognitionResult.Builder setSpeechEndOffset(Duration.Builder builderForValue)
Time offset of the end of this Speech recognition result relative to the beginning of the audio. Only populated for `message_type` = `TRANSCRIPT`.
.google.protobuf.Duration speech_end_offset = 8;
public StreamingRecognitionResult.Builder mergeSpeechEndOffset(Duration value)
Time offset of the end of this Speech recognition result relative to the beginning of the audio. Only populated for `message_type` = `TRANSCRIPT`.
.google.protobuf.Duration speech_end_offset = 8;
public StreamingRecognitionResult.Builder clearSpeechEndOffset()
Time offset of the end of this Speech recognition result relative to the beginning of the audio. Only populated for `message_type` = `TRANSCRIPT`.
.google.protobuf.Duration speech_end_offset = 8;
public Duration.Builder getSpeechEndOffsetBuilder()
Time offset of the end of this Speech recognition result relative to the beginning of the audio. Only populated for `message_type` = `TRANSCRIPT`.
.google.protobuf.Duration speech_end_offset = 8;
public DurationOrBuilder getSpeechEndOffsetOrBuilder()
Time offset of the end of this Speech recognition result relative to the beginning of the audio. Only populated for `message_type` = `TRANSCRIPT`.
.google.protobuf.Duration speech_end_offset = 8;
getSpeechEndOffsetOrBuilder
in interface StreamingRecognitionResultOrBuilder
public String getLanguageCode()
Detected language code for the transcript.
string language_code = 10;
getLanguageCode
in interface StreamingRecognitionResultOrBuilder
public ByteString getLanguageCodeBytes()
Detected language code for the transcript.
string language_code = 10;
getLanguageCodeBytes
in interface StreamingRecognitionResultOrBuilder
public StreamingRecognitionResult.Builder setLanguageCode(String value)
Detected language code for the transcript.
string language_code = 10;
value
- The languageCode to set.public StreamingRecognitionResult.Builder clearLanguageCode()
Detected language code for the transcript.
string language_code = 10;
public StreamingRecognitionResult.Builder setLanguageCodeBytes(ByteString value)
Detected language code for the transcript.
string language_code = 10;
value
- The bytes for languageCode to set.public boolean hasDtmfDigits()
DTMF digits. Populated if and only if `message_type` = `DTMF_DIGITS`.
.google.cloud.dialogflow.v2beta1.TelephonyDtmfEvents dtmf_digits = 5;
hasDtmfDigits
in interface StreamingRecognitionResultOrBuilder
public TelephonyDtmfEvents getDtmfDigits()
DTMF digits. Populated if and only if `message_type` = `DTMF_DIGITS`.
.google.cloud.dialogflow.v2beta1.TelephonyDtmfEvents dtmf_digits = 5;
getDtmfDigits
in interface StreamingRecognitionResultOrBuilder
public StreamingRecognitionResult.Builder setDtmfDigits(TelephonyDtmfEvents value)
DTMF digits. Populated if and only if `message_type` = `DTMF_DIGITS`.
.google.cloud.dialogflow.v2beta1.TelephonyDtmfEvents dtmf_digits = 5;
public StreamingRecognitionResult.Builder setDtmfDigits(TelephonyDtmfEvents.Builder builderForValue)
DTMF digits. Populated if and only if `message_type` = `DTMF_DIGITS`.
.google.cloud.dialogflow.v2beta1.TelephonyDtmfEvents dtmf_digits = 5;
public StreamingRecognitionResult.Builder mergeDtmfDigits(TelephonyDtmfEvents value)
DTMF digits. Populated if and only if `message_type` = `DTMF_DIGITS`.
.google.cloud.dialogflow.v2beta1.TelephonyDtmfEvents dtmf_digits = 5;
public StreamingRecognitionResult.Builder clearDtmfDigits()
DTMF digits. Populated if and only if `message_type` = `DTMF_DIGITS`.
.google.cloud.dialogflow.v2beta1.TelephonyDtmfEvents dtmf_digits = 5;
public TelephonyDtmfEvents.Builder getDtmfDigitsBuilder()
DTMF digits. Populated if and only if `message_type` = `DTMF_DIGITS`.
.google.cloud.dialogflow.v2beta1.TelephonyDtmfEvents dtmf_digits = 5;
public TelephonyDtmfEventsOrBuilder getDtmfDigitsOrBuilder()
DTMF digits. Populated if and only if `message_type` = `DTMF_DIGITS`.
.google.cloud.dialogflow.v2beta1.TelephonyDtmfEvents dtmf_digits = 5;
getDtmfDigitsOrBuilder
in interface StreamingRecognitionResultOrBuilder
public final StreamingRecognitionResult.Builder setUnknownFields(UnknownFieldSet unknownFields)
setUnknownFields
in interface Message.Builder
setUnknownFields
in class GeneratedMessageV3.Builder<StreamingRecognitionResult.Builder>
public final StreamingRecognitionResult.Builder mergeUnknownFields(UnknownFieldSet unknownFields)
mergeUnknownFields
in interface Message.Builder
mergeUnknownFields
in class GeneratedMessageV3.Builder<StreamingRecognitionResult.Builder>
Copyright © 2022 Google LLC. All rights reserved.