All Downloads are FREE. Search and download functionalities are using the official Maven repository.

com.google.cloud.dialogflow.v2.InputAudioConfig Maven / Gradle / Ivy

There is a newer version: 4.61.0
Show newest version
/*
 * Copyright 2020 Google LLC
 *
 * Licensed under the Apache License, Version 2.0 (the "License");
 * you may not use this file except in compliance with the License.
 * You may obtain a copy of the License at
 *
 *     https://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */
// Generated by the protocol buffer compiler.  DO NOT EDIT!
// source: google/cloud/dialogflow/v2/audio_config.proto

package com.google.cloud.dialogflow.v2;

/**
 *
 *
 * 
 * Instructs the speech recognizer how to process the audio content.
 * 
* * Protobuf type {@code google.cloud.dialogflow.v2.InputAudioConfig} */ public final class InputAudioConfig extends com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:google.cloud.dialogflow.v2.InputAudioConfig) InputAudioConfigOrBuilder { private static final long serialVersionUID = 0L; // Use InputAudioConfig.newBuilder() to construct. private InputAudioConfig(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private InputAudioConfig() { audioEncoding_ = 0; languageCode_ = ""; phraseHints_ = com.google.protobuf.LazyStringArrayList.EMPTY; speechContexts_ = java.util.Collections.emptyList(); model_ = ""; modelVariant_ = 0; } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance(UnusedPrivateParameter unused) { return new InputAudioConfig(); } @java.lang.Override public final com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private InputAudioConfig( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { int rawValue = input.readEnum(); audioEncoding_ = rawValue; break; } case 16: { sampleRateHertz_ = input.readInt32(); break; } case 26: { java.lang.String s = input.readStringRequireUtf8(); languageCode_ = s; break; } case 34: { java.lang.String s = input.readStringRequireUtf8(); if (!((mutable_bitField0_ & 0x00000001) != 0)) { phraseHints_ = new com.google.protobuf.LazyStringArrayList(); mutable_bitField0_ |= 0x00000001; } phraseHints_.add(s); break; } case 58: { java.lang.String s = input.readStringRequireUtf8(); model_ = s; break; } case 64: { singleUtterance_ = input.readBool(); break; } case 80: { int rawValue = input.readEnum(); modelVariant_ = rawValue; break; } case 90: { if (!((mutable_bitField0_ & 0x00000002) != 0)) { speechContexts_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000002; } speechContexts_.add( input.readMessage( com.google.cloud.dialogflow.v2.SpeechContext.parser(), extensionRegistry)); break; } case 104: { enableWordInfo_ = input.readBool(); break; } case 112: { disableNoSpeechRecognizedEvent_ = input.readBool(); break; } default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException(e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { phraseHints_ = phraseHints_.getUnmodifiableView(); } if (((mutable_bitField0_ & 0x00000002) != 0)) { speechContexts_ = java.util.Collections.unmodifiableList(speechContexts_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return com.google.cloud.dialogflow.v2.AudioConfigProto .internal_static_google_cloud_dialogflow_v2_InputAudioConfig_descriptor; } @java.lang.Override protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return com.google.cloud.dialogflow.v2.AudioConfigProto .internal_static_google_cloud_dialogflow_v2_InputAudioConfig_fieldAccessorTable .ensureFieldAccessorsInitialized( com.google.cloud.dialogflow.v2.InputAudioConfig.class, com.google.cloud.dialogflow.v2.InputAudioConfig.Builder.class); } public static final int AUDIO_ENCODING_FIELD_NUMBER = 1; private int audioEncoding_; /** * * *
   * Required. Audio encoding of the audio content to process.
   * 
* * .google.cloud.dialogflow.v2.AudioEncoding audio_encoding = 1; * * @return The enum numeric value on the wire for audioEncoding. */ @java.lang.Override public int getAudioEncodingValue() { return audioEncoding_; } /** * * *
   * Required. Audio encoding of the audio content to process.
   * 
* * .google.cloud.dialogflow.v2.AudioEncoding audio_encoding = 1; * * @return The audioEncoding. */ @java.lang.Override public com.google.cloud.dialogflow.v2.AudioEncoding getAudioEncoding() { @SuppressWarnings("deprecation") com.google.cloud.dialogflow.v2.AudioEncoding result = com.google.cloud.dialogflow.v2.AudioEncoding.valueOf(audioEncoding_); return result == null ? com.google.cloud.dialogflow.v2.AudioEncoding.UNRECOGNIZED : result; } public static final int SAMPLE_RATE_HERTZ_FIELD_NUMBER = 2; private int sampleRateHertz_; /** * * *
   * Required. Sample rate (in Hertz) of the audio content sent in the query.
   * Refer to
   * [Cloud Speech API
   * documentation](https://cloud.google.com/speech-to-text/docs/basics) for
   * more details.
   * 
* * int32 sample_rate_hertz = 2; * * @return The sampleRateHertz. */ @java.lang.Override public int getSampleRateHertz() { return sampleRateHertz_; } public static final int LANGUAGE_CODE_FIELD_NUMBER = 3; private volatile java.lang.Object languageCode_; /** * * *
   * Required. The language of the supplied audio. Dialogflow does not do
   * translations. See [Language
   * Support](https://cloud.google.com/dialogflow/docs/reference/language)
   * for a list of the currently supported language codes. Note that queries in
   * the same session do not necessarily need to specify the same language.
   * 
* * string language_code = 3; * * @return The languageCode. */ @java.lang.Override public java.lang.String getLanguageCode() { java.lang.Object ref = languageCode_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); languageCode_ = s; return s; } } /** * * *
   * Required. The language of the supplied audio. Dialogflow does not do
   * translations. See [Language
   * Support](https://cloud.google.com/dialogflow/docs/reference/language)
   * for a list of the currently supported language codes. Note that queries in
   * the same session do not necessarily need to specify the same language.
   * 
* * string language_code = 3; * * @return The bytes for languageCode. */ @java.lang.Override public com.google.protobuf.ByteString getLanguageCodeBytes() { java.lang.Object ref = languageCode_; if (ref instanceof java.lang.String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8((java.lang.String) ref); languageCode_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } public static final int ENABLE_WORD_INFO_FIELD_NUMBER = 13; private boolean enableWordInfo_; /** * * *
   * If `true`, Dialogflow returns [SpeechWordInfo][google.cloud.dialogflow.v2.SpeechWordInfo] in
   * [StreamingRecognitionResult][google.cloud.dialogflow.v2.StreamingRecognitionResult] with information about the recognized speech
   * words, e.g. start and end time offsets. If false or unspecified, Speech
   * doesn't return any word-level information.
   * 
* * bool enable_word_info = 13; * * @return The enableWordInfo. */ @java.lang.Override public boolean getEnableWordInfo() { return enableWordInfo_; } public static final int PHRASE_HINTS_FIELD_NUMBER = 4; private com.google.protobuf.LazyStringList phraseHints_; /** * * *
   * A list of strings containing words and phrases that the speech
   * recognizer should recognize with higher likelihood.
   * See [the Cloud Speech
   * documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
   * for more details.
   * This field is deprecated. Please use [speech_contexts]() instead. If you
   * specify both [phrase_hints]() and [speech_contexts](), Dialogflow will
   * treat the [phrase_hints]() as a single additional [SpeechContext]().
   * 
* * repeated string phrase_hints = 4 [deprecated = true]; * * @return A list containing the phraseHints. */ @java.lang.Deprecated public com.google.protobuf.ProtocolStringList getPhraseHintsList() { return phraseHints_; } /** * * *
   * A list of strings containing words and phrases that the speech
   * recognizer should recognize with higher likelihood.
   * See [the Cloud Speech
   * documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
   * for more details.
   * This field is deprecated. Please use [speech_contexts]() instead. If you
   * specify both [phrase_hints]() and [speech_contexts](), Dialogflow will
   * treat the [phrase_hints]() as a single additional [SpeechContext]().
   * 
* * repeated string phrase_hints = 4 [deprecated = true]; * * @return The count of phraseHints. */ @java.lang.Deprecated public int getPhraseHintsCount() { return phraseHints_.size(); } /** * * *
   * A list of strings containing words and phrases that the speech
   * recognizer should recognize with higher likelihood.
   * See [the Cloud Speech
   * documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
   * for more details.
   * This field is deprecated. Please use [speech_contexts]() instead. If you
   * specify both [phrase_hints]() and [speech_contexts](), Dialogflow will
   * treat the [phrase_hints]() as a single additional [SpeechContext]().
   * 
* * repeated string phrase_hints = 4 [deprecated = true]; * * @param index The index of the element to return. * @return The phraseHints at the given index. */ @java.lang.Deprecated public java.lang.String getPhraseHints(int index) { return phraseHints_.get(index); } /** * * *
   * A list of strings containing words and phrases that the speech
   * recognizer should recognize with higher likelihood.
   * See [the Cloud Speech
   * documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
   * for more details.
   * This field is deprecated. Please use [speech_contexts]() instead. If you
   * specify both [phrase_hints]() and [speech_contexts](), Dialogflow will
   * treat the [phrase_hints]() as a single additional [SpeechContext]().
   * 
* * repeated string phrase_hints = 4 [deprecated = true]; * * @param index The index of the value to return. * @return The bytes of the phraseHints at the given index. */ @java.lang.Deprecated public com.google.protobuf.ByteString getPhraseHintsBytes(int index) { return phraseHints_.getByteString(index); } public static final int SPEECH_CONTEXTS_FIELD_NUMBER = 11; private java.util.List speechContexts_; /** * * *
   * Context information to assist speech recognition.
   * See [the Cloud Speech
   * documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
   * for more details.
   * 
* * repeated .google.cloud.dialogflow.v2.SpeechContext speech_contexts = 11; */ @java.lang.Override public java.util.List getSpeechContextsList() { return speechContexts_; } /** * * *
   * Context information to assist speech recognition.
   * See [the Cloud Speech
   * documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
   * for more details.
   * 
* * repeated .google.cloud.dialogflow.v2.SpeechContext speech_contexts = 11; */ @java.lang.Override public java.util.List getSpeechContextsOrBuilderList() { return speechContexts_; } /** * * *
   * Context information to assist speech recognition.
   * See [the Cloud Speech
   * documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
   * for more details.
   * 
* * repeated .google.cloud.dialogflow.v2.SpeechContext speech_contexts = 11; */ @java.lang.Override public int getSpeechContextsCount() { return speechContexts_.size(); } /** * * *
   * Context information to assist speech recognition.
   * See [the Cloud Speech
   * documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
   * for more details.
   * 
* * repeated .google.cloud.dialogflow.v2.SpeechContext speech_contexts = 11; */ @java.lang.Override public com.google.cloud.dialogflow.v2.SpeechContext getSpeechContexts(int index) { return speechContexts_.get(index); } /** * * *
   * Context information to assist speech recognition.
   * See [the Cloud Speech
   * documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
   * for more details.
   * 
* * repeated .google.cloud.dialogflow.v2.SpeechContext speech_contexts = 11; */ @java.lang.Override public com.google.cloud.dialogflow.v2.SpeechContextOrBuilder getSpeechContextsOrBuilder( int index) { return speechContexts_.get(index); } public static final int MODEL_FIELD_NUMBER = 7; private volatile java.lang.Object model_; /** * * *
   * Which Speech model to select for the given request. Select the
   * model best suited to your domain to get best results. If a model is not
   * explicitly specified, then we auto-select a model based on the parameters
   * in the InputAudioConfig.
   * If enhanced speech model is enabled for the agent and an enhanced
   * version of the specified model for the language does not exist, then the
   * speech is recognized using the standard version of the specified model.
   * Refer to
   * [Cloud Speech API
   * documentation](https://cloud.google.com/speech-to-text/docs/basics#select-model)
   * for more details.
   * 
* * string model = 7; * * @return The model. */ @java.lang.Override public java.lang.String getModel() { java.lang.Object ref = model_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); model_ = s; return s; } } /** * * *
   * Which Speech model to select for the given request. Select the
   * model best suited to your domain to get best results. If a model is not
   * explicitly specified, then we auto-select a model based on the parameters
   * in the InputAudioConfig.
   * If enhanced speech model is enabled for the agent and an enhanced
   * version of the specified model for the language does not exist, then the
   * speech is recognized using the standard version of the specified model.
   * Refer to
   * [Cloud Speech API
   * documentation](https://cloud.google.com/speech-to-text/docs/basics#select-model)
   * for more details.
   * 
* * string model = 7; * * @return The bytes for model. */ @java.lang.Override public com.google.protobuf.ByteString getModelBytes() { java.lang.Object ref = model_; if (ref instanceof java.lang.String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8((java.lang.String) ref); model_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } public static final int MODEL_VARIANT_FIELD_NUMBER = 10; private int modelVariant_; /** * * *
   * Which variant of the [Speech model][google.cloud.dialogflow.v2.InputAudioConfig.model] to use.
   * 
* * .google.cloud.dialogflow.v2.SpeechModelVariant model_variant = 10; * * @return The enum numeric value on the wire for modelVariant. */ @java.lang.Override public int getModelVariantValue() { return modelVariant_; } /** * * *
   * Which variant of the [Speech model][google.cloud.dialogflow.v2.InputAudioConfig.model] to use.
   * 
* * .google.cloud.dialogflow.v2.SpeechModelVariant model_variant = 10; * * @return The modelVariant. */ @java.lang.Override public com.google.cloud.dialogflow.v2.SpeechModelVariant getModelVariant() { @SuppressWarnings("deprecation") com.google.cloud.dialogflow.v2.SpeechModelVariant result = com.google.cloud.dialogflow.v2.SpeechModelVariant.valueOf(modelVariant_); return result == null ? com.google.cloud.dialogflow.v2.SpeechModelVariant.UNRECOGNIZED : result; } public static final int SINGLE_UTTERANCE_FIELD_NUMBER = 8; private boolean singleUtterance_; /** * * *
   * If `false` (default), recognition does not cease until the
   * client closes the stream.
   * If `true`, the recognizer will detect a single spoken utterance in input
   * audio. Recognition ceases when it detects the audio's voice has
   * stopped or paused. In this case, once a detected intent is received, the
   * client should close the stream and start a new request with a new stream as
   * needed.
   * Note: This setting is relevant only for streaming methods.
   * Note: When specified, InputAudioConfig.single_utterance takes precedence
   * over StreamingDetectIntentRequest.single_utterance.
   * 
* * bool single_utterance = 8; * * @return The singleUtterance. */ @java.lang.Override public boolean getSingleUtterance() { return singleUtterance_; } public static final int DISABLE_NO_SPEECH_RECOGNIZED_EVENT_FIELD_NUMBER = 14; private boolean disableNoSpeechRecognizedEvent_; /** * * *
   * Only used in [Participants.AnalyzeContent][google.cloud.dialogflow.v2.Participants.AnalyzeContent] and
   * [Participants.StreamingAnalyzeContent][google.cloud.dialogflow.v2.Participants.StreamingAnalyzeContent].
   * If `false` and recognition doesn't return any result, trigger
   * `NO_SPEECH_RECOGNIZED` event to Dialogflow agent.
   * 
* * bool disable_no_speech_recognized_event = 14; * * @return The disableNoSpeechRecognizedEvent. */ @java.lang.Override public boolean getDisableNoSpeechRecognizedEvent() { return disableNoSpeechRecognizedEvent_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (audioEncoding_ != com.google.cloud.dialogflow.v2.AudioEncoding.AUDIO_ENCODING_UNSPECIFIED.getNumber()) { output.writeEnum(1, audioEncoding_); } if (sampleRateHertz_ != 0) { output.writeInt32(2, sampleRateHertz_); } if (!getLanguageCodeBytes().isEmpty()) { com.google.protobuf.GeneratedMessageV3.writeString(output, 3, languageCode_); } for (int i = 0; i < phraseHints_.size(); i++) { com.google.protobuf.GeneratedMessageV3.writeString(output, 4, phraseHints_.getRaw(i)); } if (!getModelBytes().isEmpty()) { com.google.protobuf.GeneratedMessageV3.writeString(output, 7, model_); } if (singleUtterance_ != false) { output.writeBool(8, singleUtterance_); } if (modelVariant_ != com.google.cloud.dialogflow.v2.SpeechModelVariant.SPEECH_MODEL_VARIANT_UNSPECIFIED .getNumber()) { output.writeEnum(10, modelVariant_); } for (int i = 0; i < speechContexts_.size(); i++) { output.writeMessage(11, speechContexts_.get(i)); } if (enableWordInfo_ != false) { output.writeBool(13, enableWordInfo_); } if (disableNoSpeechRecognizedEvent_ != false) { output.writeBool(14, disableNoSpeechRecognizedEvent_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (audioEncoding_ != com.google.cloud.dialogflow.v2.AudioEncoding.AUDIO_ENCODING_UNSPECIFIED.getNumber()) { size += com.google.protobuf.CodedOutputStream.computeEnumSize(1, audioEncoding_); } if (sampleRateHertz_ != 0) { size += com.google.protobuf.CodedOutputStream.computeInt32Size(2, sampleRateHertz_); } if (!getLanguageCodeBytes().isEmpty()) { size += com.google.protobuf.GeneratedMessageV3.computeStringSize(3, languageCode_); } { int dataSize = 0; for (int i = 0; i < phraseHints_.size(); i++) { dataSize += computeStringSizeNoTag(phraseHints_.getRaw(i)); } size += dataSize; size += 1 * getPhraseHintsList().size(); } if (!getModelBytes().isEmpty()) { size += com.google.protobuf.GeneratedMessageV3.computeStringSize(7, model_); } if (singleUtterance_ != false) { size += com.google.protobuf.CodedOutputStream.computeBoolSize(8, singleUtterance_); } if (modelVariant_ != com.google.cloud.dialogflow.v2.SpeechModelVariant.SPEECH_MODEL_VARIANT_UNSPECIFIED .getNumber()) { size += com.google.protobuf.CodedOutputStream.computeEnumSize(10, modelVariant_); } for (int i = 0; i < speechContexts_.size(); i++) { size += com.google.protobuf.CodedOutputStream.computeMessageSize(11, speechContexts_.get(i)); } if (enableWordInfo_ != false) { size += com.google.protobuf.CodedOutputStream.computeBoolSize(13, enableWordInfo_); } if (disableNoSpeechRecognizedEvent_ != false) { size += com.google.protobuf.CodedOutputStream.computeBoolSize( 14, disableNoSpeechRecognizedEvent_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof com.google.cloud.dialogflow.v2.InputAudioConfig)) { return super.equals(obj); } com.google.cloud.dialogflow.v2.InputAudioConfig other = (com.google.cloud.dialogflow.v2.InputAudioConfig) obj; if (audioEncoding_ != other.audioEncoding_) return false; if (getSampleRateHertz() != other.getSampleRateHertz()) return false; if (!getLanguageCode().equals(other.getLanguageCode())) return false; if (getEnableWordInfo() != other.getEnableWordInfo()) return false; if (!getPhraseHintsList().equals(other.getPhraseHintsList())) return false; if (!getSpeechContextsList().equals(other.getSpeechContextsList())) return false; if (!getModel().equals(other.getModel())) return false; if (modelVariant_ != other.modelVariant_) return false; if (getSingleUtterance() != other.getSingleUtterance()) return false; if (getDisableNoSpeechRecognizedEvent() != other.getDisableNoSpeechRecognizedEvent()) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (37 * hash) + AUDIO_ENCODING_FIELD_NUMBER; hash = (53 * hash) + audioEncoding_; hash = (37 * hash) + SAMPLE_RATE_HERTZ_FIELD_NUMBER; hash = (53 * hash) + getSampleRateHertz(); hash = (37 * hash) + LANGUAGE_CODE_FIELD_NUMBER; hash = (53 * hash) + getLanguageCode().hashCode(); hash = (37 * hash) + ENABLE_WORD_INFO_FIELD_NUMBER; hash = (53 * hash) + com.google.protobuf.Internal.hashBoolean(getEnableWordInfo()); if (getPhraseHintsCount() > 0) { hash = (37 * hash) + PHRASE_HINTS_FIELD_NUMBER; hash = (53 * hash) + getPhraseHintsList().hashCode(); } if (getSpeechContextsCount() > 0) { hash = (37 * hash) + SPEECH_CONTEXTS_FIELD_NUMBER; hash = (53 * hash) + getSpeechContextsList().hashCode(); } hash = (37 * hash) + MODEL_FIELD_NUMBER; hash = (53 * hash) + getModel().hashCode(); hash = (37 * hash) + MODEL_VARIANT_FIELD_NUMBER; hash = (53 * hash) + modelVariant_; hash = (37 * hash) + SINGLE_UTTERANCE_FIELD_NUMBER; hash = (53 * hash) + com.google.protobuf.Internal.hashBoolean(getSingleUtterance()); hash = (37 * hash) + DISABLE_NO_SPEECH_RECOGNIZED_EVENT_FIELD_NUMBER; hash = (53 * hash) + com.google.protobuf.Internal.hashBoolean(getDisableNoSpeechRecognizedEvent()); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static com.google.cloud.dialogflow.v2.InputAudioConfig parseFrom(java.nio.ByteBuffer data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static com.google.cloud.dialogflow.v2.InputAudioConfig parseFrom( java.nio.ByteBuffer data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static com.google.cloud.dialogflow.v2.InputAudioConfig parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static com.google.cloud.dialogflow.v2.InputAudioConfig parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static com.google.cloud.dialogflow.v2.InputAudioConfig parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static com.google.cloud.dialogflow.v2.InputAudioConfig parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static com.google.cloud.dialogflow.v2.InputAudioConfig parseFrom(java.io.InputStream input) throws java.io.IOException { return com.google.protobuf.GeneratedMessageV3.parseWithIOException(PARSER, input); } public static com.google.cloud.dialogflow.v2.InputAudioConfig parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return com.google.protobuf.GeneratedMessageV3.parseWithIOException( PARSER, input, extensionRegistry); } public static com.google.cloud.dialogflow.v2.InputAudioConfig parseDelimitedFrom( java.io.InputStream input) throws java.io.IOException { return com.google.protobuf.GeneratedMessageV3.parseDelimitedWithIOException(PARSER, input); } public static com.google.cloud.dialogflow.v2.InputAudioConfig parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return com.google.protobuf.GeneratedMessageV3.parseDelimitedWithIOException( PARSER, input, extensionRegistry); } public static com.google.cloud.dialogflow.v2.InputAudioConfig parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return com.google.protobuf.GeneratedMessageV3.parseWithIOException(PARSER, input); } public static com.google.cloud.dialogflow.v2.InputAudioConfig parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return com.google.protobuf.GeneratedMessageV3.parseWithIOException( PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(com.google.cloud.dialogflow.v2.InputAudioConfig prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType(com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * * *
   * Instructs the speech recognizer how to process the audio content.
   * 
* * Protobuf type {@code google.cloud.dialogflow.v2.InputAudioConfig} */ public static final class Builder extends com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:google.cloud.dialogflow.v2.InputAudioConfig) com.google.cloud.dialogflow.v2.InputAudioConfigOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return com.google.cloud.dialogflow.v2.AudioConfigProto .internal_static_google_cloud_dialogflow_v2_InputAudioConfig_descriptor; } @java.lang.Override protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return com.google.cloud.dialogflow.v2.AudioConfigProto .internal_static_google_cloud_dialogflow_v2_InputAudioConfig_fieldAccessorTable .ensureFieldAccessorsInitialized( com.google.cloud.dialogflow.v2.InputAudioConfig.class, com.google.cloud.dialogflow.v2.InputAudioConfig.Builder.class); } // Construct using com.google.cloud.dialogflow.v2.InputAudioConfig.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder(com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders) { getSpeechContextsFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); audioEncoding_ = 0; sampleRateHertz_ = 0; languageCode_ = ""; enableWordInfo_ = false; phraseHints_ = com.google.protobuf.LazyStringArrayList.EMPTY; bitField0_ = (bitField0_ & ~0x00000001); if (speechContextsBuilder_ == null) { speechContexts_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000002); } else { speechContextsBuilder_.clear(); } model_ = ""; modelVariant_ = 0; singleUtterance_ = false; disableNoSpeechRecognizedEvent_ = false; return this; } @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return com.google.cloud.dialogflow.v2.AudioConfigProto .internal_static_google_cloud_dialogflow_v2_InputAudioConfig_descriptor; } @java.lang.Override public com.google.cloud.dialogflow.v2.InputAudioConfig getDefaultInstanceForType() { return com.google.cloud.dialogflow.v2.InputAudioConfig.getDefaultInstance(); } @java.lang.Override public com.google.cloud.dialogflow.v2.InputAudioConfig build() { com.google.cloud.dialogflow.v2.InputAudioConfig result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public com.google.cloud.dialogflow.v2.InputAudioConfig buildPartial() { com.google.cloud.dialogflow.v2.InputAudioConfig result = new com.google.cloud.dialogflow.v2.InputAudioConfig(this); int from_bitField0_ = bitField0_; result.audioEncoding_ = audioEncoding_; result.sampleRateHertz_ = sampleRateHertz_; result.languageCode_ = languageCode_; result.enableWordInfo_ = enableWordInfo_; if (((bitField0_ & 0x00000001) != 0)) { phraseHints_ = phraseHints_.getUnmodifiableView(); bitField0_ = (bitField0_ & ~0x00000001); } result.phraseHints_ = phraseHints_; if (speechContextsBuilder_ == null) { if (((bitField0_ & 0x00000002) != 0)) { speechContexts_ = java.util.Collections.unmodifiableList(speechContexts_); bitField0_ = (bitField0_ & ~0x00000002); } result.speechContexts_ = speechContexts_; } else { result.speechContexts_ = speechContextsBuilder_.build(); } result.model_ = model_; result.modelVariant_ = modelVariant_; result.singleUtterance_ = singleUtterance_; result.disableNoSpeechRecognizedEvent_ = disableNoSpeechRecognizedEvent_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField(com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof(com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof com.google.cloud.dialogflow.v2.InputAudioConfig) { return mergeFrom((com.google.cloud.dialogflow.v2.InputAudioConfig) other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(com.google.cloud.dialogflow.v2.InputAudioConfig other) { if (other == com.google.cloud.dialogflow.v2.InputAudioConfig.getDefaultInstance()) return this; if (other.audioEncoding_ != 0) { setAudioEncodingValue(other.getAudioEncodingValue()); } if (other.getSampleRateHertz() != 0) { setSampleRateHertz(other.getSampleRateHertz()); } if (!other.getLanguageCode().isEmpty()) { languageCode_ = other.languageCode_; onChanged(); } if (other.getEnableWordInfo() != false) { setEnableWordInfo(other.getEnableWordInfo()); } if (!other.phraseHints_.isEmpty()) { if (phraseHints_.isEmpty()) { phraseHints_ = other.phraseHints_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensurePhraseHintsIsMutable(); phraseHints_.addAll(other.phraseHints_); } onChanged(); } if (speechContextsBuilder_ == null) { if (!other.speechContexts_.isEmpty()) { if (speechContexts_.isEmpty()) { speechContexts_ = other.speechContexts_; bitField0_ = (bitField0_ & ~0x00000002); } else { ensureSpeechContextsIsMutable(); speechContexts_.addAll(other.speechContexts_); } onChanged(); } } else { if (!other.speechContexts_.isEmpty()) { if (speechContextsBuilder_.isEmpty()) { speechContextsBuilder_.dispose(); speechContextsBuilder_ = null; speechContexts_ = other.speechContexts_; bitField0_ = (bitField0_ & ~0x00000002); speechContextsBuilder_ = com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getSpeechContextsFieldBuilder() : null; } else { speechContextsBuilder_.addAllMessages(other.speechContexts_); } } } if (!other.getModel().isEmpty()) { model_ = other.model_; onChanged(); } if (other.modelVariant_ != 0) { setModelVariantValue(other.getModelVariantValue()); } if (other.getSingleUtterance() != false) { setSingleUtterance(other.getSingleUtterance()); } if (other.getDisableNoSpeechRecognizedEvent() != false) { setDisableNoSpeechRecognizedEvent(other.getDisableNoSpeechRecognizedEvent()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { com.google.cloud.dialogflow.v2.InputAudioConfig parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (com.google.cloud.dialogflow.v2.InputAudioConfig) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private int audioEncoding_ = 0; /** * * *
     * Required. Audio encoding of the audio content to process.
     * 
* * .google.cloud.dialogflow.v2.AudioEncoding audio_encoding = 1; * * @return The enum numeric value on the wire for audioEncoding. */ @java.lang.Override public int getAudioEncodingValue() { return audioEncoding_; } /** * * *
     * Required. Audio encoding of the audio content to process.
     * 
* * .google.cloud.dialogflow.v2.AudioEncoding audio_encoding = 1; * * @param value The enum numeric value on the wire for audioEncoding to set. * @return This builder for chaining. */ public Builder setAudioEncodingValue(int value) { audioEncoding_ = value; onChanged(); return this; } /** * * *
     * Required. Audio encoding of the audio content to process.
     * 
* * .google.cloud.dialogflow.v2.AudioEncoding audio_encoding = 1; * * @return The audioEncoding. */ @java.lang.Override public com.google.cloud.dialogflow.v2.AudioEncoding getAudioEncoding() { @SuppressWarnings("deprecation") com.google.cloud.dialogflow.v2.AudioEncoding result = com.google.cloud.dialogflow.v2.AudioEncoding.valueOf(audioEncoding_); return result == null ? com.google.cloud.dialogflow.v2.AudioEncoding.UNRECOGNIZED : result; } /** * * *
     * Required. Audio encoding of the audio content to process.
     * 
* * .google.cloud.dialogflow.v2.AudioEncoding audio_encoding = 1; * * @param value The audioEncoding to set. * @return This builder for chaining. */ public Builder setAudioEncoding(com.google.cloud.dialogflow.v2.AudioEncoding value) { if (value == null) { throw new NullPointerException(); } audioEncoding_ = value.getNumber(); onChanged(); return this; } /** * * *
     * Required. Audio encoding of the audio content to process.
     * 
* * .google.cloud.dialogflow.v2.AudioEncoding audio_encoding = 1; * * @return This builder for chaining. */ public Builder clearAudioEncoding() { audioEncoding_ = 0; onChanged(); return this; } private int sampleRateHertz_; /** * * *
     * Required. Sample rate (in Hertz) of the audio content sent in the query.
     * Refer to
     * [Cloud Speech API
     * documentation](https://cloud.google.com/speech-to-text/docs/basics) for
     * more details.
     * 
* * int32 sample_rate_hertz = 2; * * @return The sampleRateHertz. */ @java.lang.Override public int getSampleRateHertz() { return sampleRateHertz_; } /** * * *
     * Required. Sample rate (in Hertz) of the audio content sent in the query.
     * Refer to
     * [Cloud Speech API
     * documentation](https://cloud.google.com/speech-to-text/docs/basics) for
     * more details.
     * 
* * int32 sample_rate_hertz = 2; * * @param value The sampleRateHertz to set. * @return This builder for chaining. */ public Builder setSampleRateHertz(int value) { sampleRateHertz_ = value; onChanged(); return this; } /** * * *
     * Required. Sample rate (in Hertz) of the audio content sent in the query.
     * Refer to
     * [Cloud Speech API
     * documentation](https://cloud.google.com/speech-to-text/docs/basics) for
     * more details.
     * 
* * int32 sample_rate_hertz = 2; * * @return This builder for chaining. */ public Builder clearSampleRateHertz() { sampleRateHertz_ = 0; onChanged(); return this; } private java.lang.Object languageCode_ = ""; /** * * *
     * Required. The language of the supplied audio. Dialogflow does not do
     * translations. See [Language
     * Support](https://cloud.google.com/dialogflow/docs/reference/language)
     * for a list of the currently supported language codes. Note that queries in
     * the same session do not necessarily need to specify the same language.
     * 
* * string language_code = 3; * * @return The languageCode. */ public java.lang.String getLanguageCode() { java.lang.Object ref = languageCode_; if (!(ref instanceof java.lang.String)) { com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); languageCode_ = s; return s; } else { return (java.lang.String) ref; } } /** * * *
     * Required. The language of the supplied audio. Dialogflow does not do
     * translations. See [Language
     * Support](https://cloud.google.com/dialogflow/docs/reference/language)
     * for a list of the currently supported language codes. Note that queries in
     * the same session do not necessarily need to specify the same language.
     * 
* * string language_code = 3; * * @return The bytes for languageCode. */ public com.google.protobuf.ByteString getLanguageCodeBytes() { java.lang.Object ref = languageCode_; if (ref instanceof String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8((java.lang.String) ref); languageCode_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } /** * * *
     * Required. The language of the supplied audio. Dialogflow does not do
     * translations. See [Language
     * Support](https://cloud.google.com/dialogflow/docs/reference/language)
     * for a list of the currently supported language codes. Note that queries in
     * the same session do not necessarily need to specify the same language.
     * 
* * string language_code = 3; * * @param value The languageCode to set. * @return This builder for chaining. */ public Builder setLanguageCode(java.lang.String value) { if (value == null) { throw new NullPointerException(); } languageCode_ = value; onChanged(); return this; } /** * * *
     * Required. The language of the supplied audio. Dialogflow does not do
     * translations. See [Language
     * Support](https://cloud.google.com/dialogflow/docs/reference/language)
     * for a list of the currently supported language codes. Note that queries in
     * the same session do not necessarily need to specify the same language.
     * 
* * string language_code = 3; * * @return This builder for chaining. */ public Builder clearLanguageCode() { languageCode_ = getDefaultInstance().getLanguageCode(); onChanged(); return this; } /** * * *
     * Required. The language of the supplied audio. Dialogflow does not do
     * translations. See [Language
     * Support](https://cloud.google.com/dialogflow/docs/reference/language)
     * for a list of the currently supported language codes. Note that queries in
     * the same session do not necessarily need to specify the same language.
     * 
* * string language_code = 3; * * @param value The bytes for languageCode to set. * @return This builder for chaining. */ public Builder setLanguageCodeBytes(com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } checkByteStringIsUtf8(value); languageCode_ = value; onChanged(); return this; } private boolean enableWordInfo_; /** * * *
     * If `true`, Dialogflow returns [SpeechWordInfo][google.cloud.dialogflow.v2.SpeechWordInfo] in
     * [StreamingRecognitionResult][google.cloud.dialogflow.v2.StreamingRecognitionResult] with information about the recognized speech
     * words, e.g. start and end time offsets. If false or unspecified, Speech
     * doesn't return any word-level information.
     * 
* * bool enable_word_info = 13; * * @return The enableWordInfo. */ @java.lang.Override public boolean getEnableWordInfo() { return enableWordInfo_; } /** * * *
     * If `true`, Dialogflow returns [SpeechWordInfo][google.cloud.dialogflow.v2.SpeechWordInfo] in
     * [StreamingRecognitionResult][google.cloud.dialogflow.v2.StreamingRecognitionResult] with information about the recognized speech
     * words, e.g. start and end time offsets. If false or unspecified, Speech
     * doesn't return any word-level information.
     * 
* * bool enable_word_info = 13; * * @param value The enableWordInfo to set. * @return This builder for chaining. */ public Builder setEnableWordInfo(boolean value) { enableWordInfo_ = value; onChanged(); return this; } /** * * *
     * If `true`, Dialogflow returns [SpeechWordInfo][google.cloud.dialogflow.v2.SpeechWordInfo] in
     * [StreamingRecognitionResult][google.cloud.dialogflow.v2.StreamingRecognitionResult] with information about the recognized speech
     * words, e.g. start and end time offsets. If false or unspecified, Speech
     * doesn't return any word-level information.
     * 
* * bool enable_word_info = 13; * * @return This builder for chaining. */ public Builder clearEnableWordInfo() { enableWordInfo_ = false; onChanged(); return this; } private com.google.protobuf.LazyStringList phraseHints_ = com.google.protobuf.LazyStringArrayList.EMPTY; private void ensurePhraseHintsIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { phraseHints_ = new com.google.protobuf.LazyStringArrayList(phraseHints_); bitField0_ |= 0x00000001; } } /** * * *
     * A list of strings containing words and phrases that the speech
     * recognizer should recognize with higher likelihood.
     * See [the Cloud Speech
     * documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
     * for more details.
     * This field is deprecated. Please use [speech_contexts]() instead. If you
     * specify both [phrase_hints]() and [speech_contexts](), Dialogflow will
     * treat the [phrase_hints]() as a single additional [SpeechContext]().
     * 
* * repeated string phrase_hints = 4 [deprecated = true]; * * @return A list containing the phraseHints. */ @java.lang.Deprecated public com.google.protobuf.ProtocolStringList getPhraseHintsList() { return phraseHints_.getUnmodifiableView(); } /** * * *
     * A list of strings containing words and phrases that the speech
     * recognizer should recognize with higher likelihood.
     * See [the Cloud Speech
     * documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
     * for more details.
     * This field is deprecated. Please use [speech_contexts]() instead. If you
     * specify both [phrase_hints]() and [speech_contexts](), Dialogflow will
     * treat the [phrase_hints]() as a single additional [SpeechContext]().
     * 
* * repeated string phrase_hints = 4 [deprecated = true]; * * @return The count of phraseHints. */ @java.lang.Deprecated public int getPhraseHintsCount() { return phraseHints_.size(); } /** * * *
     * A list of strings containing words and phrases that the speech
     * recognizer should recognize with higher likelihood.
     * See [the Cloud Speech
     * documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
     * for more details.
     * This field is deprecated. Please use [speech_contexts]() instead. If you
     * specify both [phrase_hints]() and [speech_contexts](), Dialogflow will
     * treat the [phrase_hints]() as a single additional [SpeechContext]().
     * 
* * repeated string phrase_hints = 4 [deprecated = true]; * * @param index The index of the element to return. * @return The phraseHints at the given index. */ @java.lang.Deprecated public java.lang.String getPhraseHints(int index) { return phraseHints_.get(index); } /** * * *
     * A list of strings containing words and phrases that the speech
     * recognizer should recognize with higher likelihood.
     * See [the Cloud Speech
     * documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
     * for more details.
     * This field is deprecated. Please use [speech_contexts]() instead. If you
     * specify both [phrase_hints]() and [speech_contexts](), Dialogflow will
     * treat the [phrase_hints]() as a single additional [SpeechContext]().
     * 
* * repeated string phrase_hints = 4 [deprecated = true]; * * @param index The index of the value to return. * @return The bytes of the phraseHints at the given index. */ @java.lang.Deprecated public com.google.protobuf.ByteString getPhraseHintsBytes(int index) { return phraseHints_.getByteString(index); } /** * * *
     * A list of strings containing words and phrases that the speech
     * recognizer should recognize with higher likelihood.
     * See [the Cloud Speech
     * documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
     * for more details.
     * This field is deprecated. Please use [speech_contexts]() instead. If you
     * specify both [phrase_hints]() and [speech_contexts](), Dialogflow will
     * treat the [phrase_hints]() as a single additional [SpeechContext]().
     * 
* * repeated string phrase_hints = 4 [deprecated = true]; * * @param index The index to set the value at. * @param value The phraseHints to set. * @return This builder for chaining. */ @java.lang.Deprecated public Builder setPhraseHints(int index, java.lang.String value) { if (value == null) { throw new NullPointerException(); } ensurePhraseHintsIsMutable(); phraseHints_.set(index, value); onChanged(); return this; } /** * * *
     * A list of strings containing words and phrases that the speech
     * recognizer should recognize with higher likelihood.
     * See [the Cloud Speech
     * documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
     * for more details.
     * This field is deprecated. Please use [speech_contexts]() instead. If you
     * specify both [phrase_hints]() and [speech_contexts](), Dialogflow will
     * treat the [phrase_hints]() as a single additional [SpeechContext]().
     * 
* * repeated string phrase_hints = 4 [deprecated = true]; * * @param value The phraseHints to add. * @return This builder for chaining. */ @java.lang.Deprecated public Builder addPhraseHints(java.lang.String value) { if (value == null) { throw new NullPointerException(); } ensurePhraseHintsIsMutable(); phraseHints_.add(value); onChanged(); return this; } /** * * *
     * A list of strings containing words and phrases that the speech
     * recognizer should recognize with higher likelihood.
     * See [the Cloud Speech
     * documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
     * for more details.
     * This field is deprecated. Please use [speech_contexts]() instead. If you
     * specify both [phrase_hints]() and [speech_contexts](), Dialogflow will
     * treat the [phrase_hints]() as a single additional [SpeechContext]().
     * 
* * repeated string phrase_hints = 4 [deprecated = true]; * * @param values The phraseHints to add. * @return This builder for chaining. */ @java.lang.Deprecated public Builder addAllPhraseHints(java.lang.Iterable values) { ensurePhraseHintsIsMutable(); com.google.protobuf.AbstractMessageLite.Builder.addAll(values, phraseHints_); onChanged(); return this; } /** * * *
     * A list of strings containing words and phrases that the speech
     * recognizer should recognize with higher likelihood.
     * See [the Cloud Speech
     * documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
     * for more details.
     * This field is deprecated. Please use [speech_contexts]() instead. If you
     * specify both [phrase_hints]() and [speech_contexts](), Dialogflow will
     * treat the [phrase_hints]() as a single additional [SpeechContext]().
     * 
* * repeated string phrase_hints = 4 [deprecated = true]; * * @return This builder for chaining. */ @java.lang.Deprecated public Builder clearPhraseHints() { phraseHints_ = com.google.protobuf.LazyStringArrayList.EMPTY; bitField0_ = (bitField0_ & ~0x00000001); onChanged(); return this; } /** * * *
     * A list of strings containing words and phrases that the speech
     * recognizer should recognize with higher likelihood.
     * See [the Cloud Speech
     * documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
     * for more details.
     * This field is deprecated. Please use [speech_contexts]() instead. If you
     * specify both [phrase_hints]() and [speech_contexts](), Dialogflow will
     * treat the [phrase_hints]() as a single additional [SpeechContext]().
     * 
* * repeated string phrase_hints = 4 [deprecated = true]; * * @param value The bytes of the phraseHints to add. * @return This builder for chaining. */ @java.lang.Deprecated public Builder addPhraseHintsBytes(com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } checkByteStringIsUtf8(value); ensurePhraseHintsIsMutable(); phraseHints_.add(value); onChanged(); return this; } private java.util.List speechContexts_ = java.util.Collections.emptyList(); private void ensureSpeechContextsIsMutable() { if (!((bitField0_ & 0x00000002) != 0)) { speechContexts_ = new java.util.ArrayList(speechContexts_); bitField0_ |= 0x00000002; } } private com.google.protobuf.RepeatedFieldBuilderV3< com.google.cloud.dialogflow.v2.SpeechContext, com.google.cloud.dialogflow.v2.SpeechContext.Builder, com.google.cloud.dialogflow.v2.SpeechContextOrBuilder> speechContextsBuilder_; /** * * *
     * Context information to assist speech recognition.
     * See [the Cloud Speech
     * documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
     * for more details.
     * 
* * repeated .google.cloud.dialogflow.v2.SpeechContext speech_contexts = 11; */ public java.util.List getSpeechContextsList() { if (speechContextsBuilder_ == null) { return java.util.Collections.unmodifiableList(speechContexts_); } else { return speechContextsBuilder_.getMessageList(); } } /** * * *
     * Context information to assist speech recognition.
     * See [the Cloud Speech
     * documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
     * for more details.
     * 
* * repeated .google.cloud.dialogflow.v2.SpeechContext speech_contexts = 11; */ public int getSpeechContextsCount() { if (speechContextsBuilder_ == null) { return speechContexts_.size(); } else { return speechContextsBuilder_.getCount(); } } /** * * *
     * Context information to assist speech recognition.
     * See [the Cloud Speech
     * documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
     * for more details.
     * 
* * repeated .google.cloud.dialogflow.v2.SpeechContext speech_contexts = 11; */ public com.google.cloud.dialogflow.v2.SpeechContext getSpeechContexts(int index) { if (speechContextsBuilder_ == null) { return speechContexts_.get(index); } else { return speechContextsBuilder_.getMessage(index); } } /** * * *
     * Context information to assist speech recognition.
     * See [the Cloud Speech
     * documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
     * for more details.
     * 
* * repeated .google.cloud.dialogflow.v2.SpeechContext speech_contexts = 11; */ public Builder setSpeechContexts( int index, com.google.cloud.dialogflow.v2.SpeechContext value) { if (speechContextsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureSpeechContextsIsMutable(); speechContexts_.set(index, value); onChanged(); } else { speechContextsBuilder_.setMessage(index, value); } return this; } /** * * *
     * Context information to assist speech recognition.
     * See [the Cloud Speech
     * documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
     * for more details.
     * 
* * repeated .google.cloud.dialogflow.v2.SpeechContext speech_contexts = 11; */ public Builder setSpeechContexts( int index, com.google.cloud.dialogflow.v2.SpeechContext.Builder builderForValue) { if (speechContextsBuilder_ == null) { ensureSpeechContextsIsMutable(); speechContexts_.set(index, builderForValue.build()); onChanged(); } else { speechContextsBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * * *
     * Context information to assist speech recognition.
     * See [the Cloud Speech
     * documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
     * for more details.
     * 
* * repeated .google.cloud.dialogflow.v2.SpeechContext speech_contexts = 11; */ public Builder addSpeechContexts(com.google.cloud.dialogflow.v2.SpeechContext value) { if (speechContextsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureSpeechContextsIsMutable(); speechContexts_.add(value); onChanged(); } else { speechContextsBuilder_.addMessage(value); } return this; } /** * * *
     * Context information to assist speech recognition.
     * See [the Cloud Speech
     * documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
     * for more details.
     * 
* * repeated .google.cloud.dialogflow.v2.SpeechContext speech_contexts = 11; */ public Builder addSpeechContexts( int index, com.google.cloud.dialogflow.v2.SpeechContext value) { if (speechContextsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureSpeechContextsIsMutable(); speechContexts_.add(index, value); onChanged(); } else { speechContextsBuilder_.addMessage(index, value); } return this; } /** * * *
     * Context information to assist speech recognition.
     * See [the Cloud Speech
     * documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
     * for more details.
     * 
* * repeated .google.cloud.dialogflow.v2.SpeechContext speech_contexts = 11; */ public Builder addSpeechContexts( com.google.cloud.dialogflow.v2.SpeechContext.Builder builderForValue) { if (speechContextsBuilder_ == null) { ensureSpeechContextsIsMutable(); speechContexts_.add(builderForValue.build()); onChanged(); } else { speechContextsBuilder_.addMessage(builderForValue.build()); } return this; } /** * * *
     * Context information to assist speech recognition.
     * See [the Cloud Speech
     * documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
     * for more details.
     * 
* * repeated .google.cloud.dialogflow.v2.SpeechContext speech_contexts = 11; */ public Builder addSpeechContexts( int index, com.google.cloud.dialogflow.v2.SpeechContext.Builder builderForValue) { if (speechContextsBuilder_ == null) { ensureSpeechContextsIsMutable(); speechContexts_.add(index, builderForValue.build()); onChanged(); } else { speechContextsBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * * *
     * Context information to assist speech recognition.
     * See [the Cloud Speech
     * documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
     * for more details.
     * 
* * repeated .google.cloud.dialogflow.v2.SpeechContext speech_contexts = 11; */ public Builder addAllSpeechContexts( java.lang.Iterable values) { if (speechContextsBuilder_ == null) { ensureSpeechContextsIsMutable(); com.google.protobuf.AbstractMessageLite.Builder.addAll(values, speechContexts_); onChanged(); } else { speechContextsBuilder_.addAllMessages(values); } return this; } /** * * *
     * Context information to assist speech recognition.
     * See [the Cloud Speech
     * documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
     * for more details.
     * 
* * repeated .google.cloud.dialogflow.v2.SpeechContext speech_contexts = 11; */ public Builder clearSpeechContexts() { if (speechContextsBuilder_ == null) { speechContexts_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000002); onChanged(); } else { speechContextsBuilder_.clear(); } return this; } /** * * *
     * Context information to assist speech recognition.
     * See [the Cloud Speech
     * documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
     * for more details.
     * 
* * repeated .google.cloud.dialogflow.v2.SpeechContext speech_contexts = 11; */ public Builder removeSpeechContexts(int index) { if (speechContextsBuilder_ == null) { ensureSpeechContextsIsMutable(); speechContexts_.remove(index); onChanged(); } else { speechContextsBuilder_.remove(index); } return this; } /** * * *
     * Context information to assist speech recognition.
     * See [the Cloud Speech
     * documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
     * for more details.
     * 
* * repeated .google.cloud.dialogflow.v2.SpeechContext speech_contexts = 11; */ public com.google.cloud.dialogflow.v2.SpeechContext.Builder getSpeechContextsBuilder( int index) { return getSpeechContextsFieldBuilder().getBuilder(index); } /** * * *
     * Context information to assist speech recognition.
     * See [the Cloud Speech
     * documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
     * for more details.
     * 
* * repeated .google.cloud.dialogflow.v2.SpeechContext speech_contexts = 11; */ public com.google.cloud.dialogflow.v2.SpeechContextOrBuilder getSpeechContextsOrBuilder( int index) { if (speechContextsBuilder_ == null) { return speechContexts_.get(index); } else { return speechContextsBuilder_.getMessageOrBuilder(index); } } /** * * *
     * Context information to assist speech recognition.
     * See [the Cloud Speech
     * documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
     * for more details.
     * 
* * repeated .google.cloud.dialogflow.v2.SpeechContext speech_contexts = 11; */ public java.util.List getSpeechContextsOrBuilderList() { if (speechContextsBuilder_ != null) { return speechContextsBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(speechContexts_); } } /** * * *
     * Context information to assist speech recognition.
     * See [the Cloud Speech
     * documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
     * for more details.
     * 
* * repeated .google.cloud.dialogflow.v2.SpeechContext speech_contexts = 11; */ public com.google.cloud.dialogflow.v2.SpeechContext.Builder addSpeechContextsBuilder() { return getSpeechContextsFieldBuilder() .addBuilder(com.google.cloud.dialogflow.v2.SpeechContext.getDefaultInstance()); } /** * * *
     * Context information to assist speech recognition.
     * See [the Cloud Speech
     * documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
     * for more details.
     * 
* * repeated .google.cloud.dialogflow.v2.SpeechContext speech_contexts = 11; */ public com.google.cloud.dialogflow.v2.SpeechContext.Builder addSpeechContextsBuilder( int index) { return getSpeechContextsFieldBuilder() .addBuilder(index, com.google.cloud.dialogflow.v2.SpeechContext.getDefaultInstance()); } /** * * *
     * Context information to assist speech recognition.
     * See [the Cloud Speech
     * documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
     * for more details.
     * 
* * repeated .google.cloud.dialogflow.v2.SpeechContext speech_contexts = 11; */ public java.util.List getSpeechContextsBuilderList() { return getSpeechContextsFieldBuilder().getBuilderList(); } private com.google.protobuf.RepeatedFieldBuilderV3< com.google.cloud.dialogflow.v2.SpeechContext, com.google.cloud.dialogflow.v2.SpeechContext.Builder, com.google.cloud.dialogflow.v2.SpeechContextOrBuilder> getSpeechContextsFieldBuilder() { if (speechContextsBuilder_ == null) { speechContextsBuilder_ = new com.google.protobuf.RepeatedFieldBuilderV3< com.google.cloud.dialogflow.v2.SpeechContext, com.google.cloud.dialogflow.v2.SpeechContext.Builder, com.google.cloud.dialogflow.v2.SpeechContextOrBuilder>( speechContexts_, ((bitField0_ & 0x00000002) != 0), getParentForChildren(), isClean()); speechContexts_ = null; } return speechContextsBuilder_; } private java.lang.Object model_ = ""; /** * * *
     * Which Speech model to select for the given request. Select the
     * model best suited to your domain to get best results. If a model is not
     * explicitly specified, then we auto-select a model based on the parameters
     * in the InputAudioConfig.
     * If enhanced speech model is enabled for the agent and an enhanced
     * version of the specified model for the language does not exist, then the
     * speech is recognized using the standard version of the specified model.
     * Refer to
     * [Cloud Speech API
     * documentation](https://cloud.google.com/speech-to-text/docs/basics#select-model)
     * for more details.
     * 
* * string model = 7; * * @return The model. */ public java.lang.String getModel() { java.lang.Object ref = model_; if (!(ref instanceof java.lang.String)) { com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); model_ = s; return s; } else { return (java.lang.String) ref; } } /** * * *
     * Which Speech model to select for the given request. Select the
     * model best suited to your domain to get best results. If a model is not
     * explicitly specified, then we auto-select a model based on the parameters
     * in the InputAudioConfig.
     * If enhanced speech model is enabled for the agent and an enhanced
     * version of the specified model for the language does not exist, then the
     * speech is recognized using the standard version of the specified model.
     * Refer to
     * [Cloud Speech API
     * documentation](https://cloud.google.com/speech-to-text/docs/basics#select-model)
     * for more details.
     * 
* * string model = 7; * * @return The bytes for model. */ public com.google.protobuf.ByteString getModelBytes() { java.lang.Object ref = model_; if (ref instanceof String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8((java.lang.String) ref); model_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } /** * * *
     * Which Speech model to select for the given request. Select the
     * model best suited to your domain to get best results. If a model is not
     * explicitly specified, then we auto-select a model based on the parameters
     * in the InputAudioConfig.
     * If enhanced speech model is enabled for the agent and an enhanced
     * version of the specified model for the language does not exist, then the
     * speech is recognized using the standard version of the specified model.
     * Refer to
     * [Cloud Speech API
     * documentation](https://cloud.google.com/speech-to-text/docs/basics#select-model)
     * for more details.
     * 
* * string model = 7; * * @param value The model to set. * @return This builder for chaining. */ public Builder setModel(java.lang.String value) { if (value == null) { throw new NullPointerException(); } model_ = value; onChanged(); return this; } /** * * *
     * Which Speech model to select for the given request. Select the
     * model best suited to your domain to get best results. If a model is not
     * explicitly specified, then we auto-select a model based on the parameters
     * in the InputAudioConfig.
     * If enhanced speech model is enabled for the agent and an enhanced
     * version of the specified model for the language does not exist, then the
     * speech is recognized using the standard version of the specified model.
     * Refer to
     * [Cloud Speech API
     * documentation](https://cloud.google.com/speech-to-text/docs/basics#select-model)
     * for more details.
     * 
* * string model = 7; * * @return This builder for chaining. */ public Builder clearModel() { model_ = getDefaultInstance().getModel(); onChanged(); return this; } /** * * *
     * Which Speech model to select for the given request. Select the
     * model best suited to your domain to get best results. If a model is not
     * explicitly specified, then we auto-select a model based on the parameters
     * in the InputAudioConfig.
     * If enhanced speech model is enabled for the agent and an enhanced
     * version of the specified model for the language does not exist, then the
     * speech is recognized using the standard version of the specified model.
     * Refer to
     * [Cloud Speech API
     * documentation](https://cloud.google.com/speech-to-text/docs/basics#select-model)
     * for more details.
     * 
* * string model = 7; * * @param value The bytes for model to set. * @return This builder for chaining. */ public Builder setModelBytes(com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } checkByteStringIsUtf8(value); model_ = value; onChanged(); return this; } private int modelVariant_ = 0; /** * * *
     * Which variant of the [Speech model][google.cloud.dialogflow.v2.InputAudioConfig.model] to use.
     * 
* * .google.cloud.dialogflow.v2.SpeechModelVariant model_variant = 10; * * @return The enum numeric value on the wire for modelVariant. */ @java.lang.Override public int getModelVariantValue() { return modelVariant_; } /** * * *
     * Which variant of the [Speech model][google.cloud.dialogflow.v2.InputAudioConfig.model] to use.
     * 
* * .google.cloud.dialogflow.v2.SpeechModelVariant model_variant = 10; * * @param value The enum numeric value on the wire for modelVariant to set. * @return This builder for chaining. */ public Builder setModelVariantValue(int value) { modelVariant_ = value; onChanged(); return this; } /** * * *
     * Which variant of the [Speech model][google.cloud.dialogflow.v2.InputAudioConfig.model] to use.
     * 
* * .google.cloud.dialogflow.v2.SpeechModelVariant model_variant = 10; * * @return The modelVariant. */ @java.lang.Override public com.google.cloud.dialogflow.v2.SpeechModelVariant getModelVariant() { @SuppressWarnings("deprecation") com.google.cloud.dialogflow.v2.SpeechModelVariant result = com.google.cloud.dialogflow.v2.SpeechModelVariant.valueOf(modelVariant_); return result == null ? com.google.cloud.dialogflow.v2.SpeechModelVariant.UNRECOGNIZED : result; } /** * * *
     * Which variant of the [Speech model][google.cloud.dialogflow.v2.InputAudioConfig.model] to use.
     * 
* * .google.cloud.dialogflow.v2.SpeechModelVariant model_variant = 10; * * @param value The modelVariant to set. * @return This builder for chaining. */ public Builder setModelVariant(com.google.cloud.dialogflow.v2.SpeechModelVariant value) { if (value == null) { throw new NullPointerException(); } modelVariant_ = value.getNumber(); onChanged(); return this; } /** * * *
     * Which variant of the [Speech model][google.cloud.dialogflow.v2.InputAudioConfig.model] to use.
     * 
* * .google.cloud.dialogflow.v2.SpeechModelVariant model_variant = 10; * * @return This builder for chaining. */ public Builder clearModelVariant() { modelVariant_ = 0; onChanged(); return this; } private boolean singleUtterance_; /** * * *
     * If `false` (default), recognition does not cease until the
     * client closes the stream.
     * If `true`, the recognizer will detect a single spoken utterance in input
     * audio. Recognition ceases when it detects the audio's voice has
     * stopped or paused. In this case, once a detected intent is received, the
     * client should close the stream and start a new request with a new stream as
     * needed.
     * Note: This setting is relevant only for streaming methods.
     * Note: When specified, InputAudioConfig.single_utterance takes precedence
     * over StreamingDetectIntentRequest.single_utterance.
     * 
* * bool single_utterance = 8; * * @return The singleUtterance. */ @java.lang.Override public boolean getSingleUtterance() { return singleUtterance_; } /** * * *
     * If `false` (default), recognition does not cease until the
     * client closes the stream.
     * If `true`, the recognizer will detect a single spoken utterance in input
     * audio. Recognition ceases when it detects the audio's voice has
     * stopped or paused. In this case, once a detected intent is received, the
     * client should close the stream and start a new request with a new stream as
     * needed.
     * Note: This setting is relevant only for streaming methods.
     * Note: When specified, InputAudioConfig.single_utterance takes precedence
     * over StreamingDetectIntentRequest.single_utterance.
     * 
* * bool single_utterance = 8; * * @param value The singleUtterance to set. * @return This builder for chaining. */ public Builder setSingleUtterance(boolean value) { singleUtterance_ = value; onChanged(); return this; } /** * * *
     * If `false` (default), recognition does not cease until the
     * client closes the stream.
     * If `true`, the recognizer will detect a single spoken utterance in input
     * audio. Recognition ceases when it detects the audio's voice has
     * stopped or paused. In this case, once a detected intent is received, the
     * client should close the stream and start a new request with a new stream as
     * needed.
     * Note: This setting is relevant only for streaming methods.
     * Note: When specified, InputAudioConfig.single_utterance takes precedence
     * over StreamingDetectIntentRequest.single_utterance.
     * 
* * bool single_utterance = 8; * * @return This builder for chaining. */ public Builder clearSingleUtterance() { singleUtterance_ = false; onChanged(); return this; } private boolean disableNoSpeechRecognizedEvent_; /** * * *
     * Only used in [Participants.AnalyzeContent][google.cloud.dialogflow.v2.Participants.AnalyzeContent] and
     * [Participants.StreamingAnalyzeContent][google.cloud.dialogflow.v2.Participants.StreamingAnalyzeContent].
     * If `false` and recognition doesn't return any result, trigger
     * `NO_SPEECH_RECOGNIZED` event to Dialogflow agent.
     * 
* * bool disable_no_speech_recognized_event = 14; * * @return The disableNoSpeechRecognizedEvent. */ @java.lang.Override public boolean getDisableNoSpeechRecognizedEvent() { return disableNoSpeechRecognizedEvent_; } /** * * *
     * Only used in [Participants.AnalyzeContent][google.cloud.dialogflow.v2.Participants.AnalyzeContent] and
     * [Participants.StreamingAnalyzeContent][google.cloud.dialogflow.v2.Participants.StreamingAnalyzeContent].
     * If `false` and recognition doesn't return any result, trigger
     * `NO_SPEECH_RECOGNIZED` event to Dialogflow agent.
     * 
* * bool disable_no_speech_recognized_event = 14; * * @param value The disableNoSpeechRecognizedEvent to set. * @return This builder for chaining. */ public Builder setDisableNoSpeechRecognizedEvent(boolean value) { disableNoSpeechRecognizedEvent_ = value; onChanged(); return this; } /** * * *
     * Only used in [Participants.AnalyzeContent][google.cloud.dialogflow.v2.Participants.AnalyzeContent] and
     * [Participants.StreamingAnalyzeContent][google.cloud.dialogflow.v2.Participants.StreamingAnalyzeContent].
     * If `false` and recognition doesn't return any result, trigger
     * `NO_SPEECH_RECOGNIZED` event to Dialogflow agent.
     * 
* * bool disable_no_speech_recognized_event = 14; * * @return This builder for chaining. */ public Builder clearDisableNoSpeechRecognizedEvent() { disableNoSpeechRecognizedEvent_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields(final com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:google.cloud.dialogflow.v2.InputAudioConfig) } // @@protoc_insertion_point(class_scope:google.cloud.dialogflow.v2.InputAudioConfig) private static final com.google.cloud.dialogflow.v2.InputAudioConfig DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new com.google.cloud.dialogflow.v2.InputAudioConfig(); } public static com.google.cloud.dialogflow.v2.InputAudioConfig getDefaultInstance() { return DEFAULT_INSTANCE; } private static final com.google.protobuf.Parser PARSER = new com.google.protobuf.AbstractParser() { @java.lang.Override public InputAudioConfig parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return new InputAudioConfig(input, extensionRegistry); } }; public static com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public com.google.cloud.dialogflow.v2.InputAudioConfig getDefaultInstanceForType() { return DEFAULT_INSTANCE; } }




© 2015 - 2025 Weber Informatics LLC | Privacy Policy