com.google.cloud.dialogflow.v2.InputAudioConfigOrBuilder Maven / Gradle / Ivy
Go to download
Show more of this group Show more artifacts with this name
Show all versions of proto-google-cloud-dialogflow-v2 Show documentation
Show all versions of proto-google-cloud-dialogflow-v2 Show documentation
PROTO library for proto-google-cloud-dialogflow-v2
/*
* Copyright 2020 Google LLC
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* https://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
// Generated by the protocol buffer compiler. DO NOT EDIT!
// source: google/cloud/dialogflow/v2/audio_config.proto
package com.google.cloud.dialogflow.v2;
public interface InputAudioConfigOrBuilder
extends
// @@protoc_insertion_point(interface_extends:google.cloud.dialogflow.v2.InputAudioConfig)
com.google.protobuf.MessageOrBuilder {
/**
*
*
*
* Required. Audio encoding of the audio content to process.
*
*
* .google.cloud.dialogflow.v2.AudioEncoding audio_encoding = 1;
*
* @return The enum numeric value on the wire for audioEncoding.
*/
int getAudioEncodingValue();
/**
*
*
*
* Required. Audio encoding of the audio content to process.
*
*
* .google.cloud.dialogflow.v2.AudioEncoding audio_encoding = 1;
*
* @return The audioEncoding.
*/
com.google.cloud.dialogflow.v2.AudioEncoding getAudioEncoding();
/**
*
*
*
* Required. Sample rate (in Hertz) of the audio content sent in the query.
* Refer to
* [Cloud Speech API
* documentation](https://cloud.google.com/speech-to-text/docs/basics) for
* more details.
*
*
* int32 sample_rate_hertz = 2;
*
* @return The sampleRateHertz.
*/
int getSampleRateHertz();
/**
*
*
*
* Required. The language of the supplied audio. Dialogflow does not do
* translations. See [Language
* Support](https://cloud.google.com/dialogflow/docs/reference/language)
* for a list of the currently supported language codes. Note that queries in
* the same session do not necessarily need to specify the same language.
*
*
* string language_code = 3;
*
* @return The languageCode.
*/
java.lang.String getLanguageCode();
/**
*
*
*
* Required. The language of the supplied audio. Dialogflow does not do
* translations. See [Language
* Support](https://cloud.google.com/dialogflow/docs/reference/language)
* for a list of the currently supported language codes. Note that queries in
* the same session do not necessarily need to specify the same language.
*
*
* string language_code = 3;
*
* @return The bytes for languageCode.
*/
com.google.protobuf.ByteString getLanguageCodeBytes();
/**
*
*
*
* If `true`, Dialogflow returns [SpeechWordInfo][google.cloud.dialogflow.v2.SpeechWordInfo] in
* [StreamingRecognitionResult][google.cloud.dialogflow.v2.StreamingRecognitionResult] with information about the recognized speech
* words, e.g. start and end time offsets. If false or unspecified, Speech
* doesn't return any word-level information.
*
*
* bool enable_word_info = 13;
*
* @return The enableWordInfo.
*/
boolean getEnableWordInfo();
/**
*
*
*
* A list of strings containing words and phrases that the speech
* recognizer should recognize with higher likelihood.
* See [the Cloud Speech
* documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
* for more details.
* This field is deprecated. Please use [speech_contexts]() instead. If you
* specify both [phrase_hints]() and [speech_contexts](), Dialogflow will
* treat the [phrase_hints]() as a single additional [SpeechContext]().
*
*
* repeated string phrase_hints = 4 [deprecated = true];
*
* @return A list containing the phraseHints.
*/
@java.lang.Deprecated
java.util.List getPhraseHintsList();
/**
*
*
*
* A list of strings containing words and phrases that the speech
* recognizer should recognize with higher likelihood.
* See [the Cloud Speech
* documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
* for more details.
* This field is deprecated. Please use [speech_contexts]() instead. If you
* specify both [phrase_hints]() and [speech_contexts](), Dialogflow will
* treat the [phrase_hints]() as a single additional [SpeechContext]().
*
*
* repeated string phrase_hints = 4 [deprecated = true];
*
* @return The count of phraseHints.
*/
@java.lang.Deprecated
int getPhraseHintsCount();
/**
*
*
*
* A list of strings containing words and phrases that the speech
* recognizer should recognize with higher likelihood.
* See [the Cloud Speech
* documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
* for more details.
* This field is deprecated. Please use [speech_contexts]() instead. If you
* specify both [phrase_hints]() and [speech_contexts](), Dialogflow will
* treat the [phrase_hints]() as a single additional [SpeechContext]().
*
*
* repeated string phrase_hints = 4 [deprecated = true];
*
* @param index The index of the element to return.
* @return The phraseHints at the given index.
*/
@java.lang.Deprecated
java.lang.String getPhraseHints(int index);
/**
*
*
*
* A list of strings containing words and phrases that the speech
* recognizer should recognize with higher likelihood.
* See [the Cloud Speech
* documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
* for more details.
* This field is deprecated. Please use [speech_contexts]() instead. If you
* specify both [phrase_hints]() and [speech_contexts](), Dialogflow will
* treat the [phrase_hints]() as a single additional [SpeechContext]().
*
*
* repeated string phrase_hints = 4 [deprecated = true];
*
* @param index The index of the value to return.
* @return The bytes of the phraseHints at the given index.
*/
@java.lang.Deprecated
com.google.protobuf.ByteString getPhraseHintsBytes(int index);
/**
*
*
*
* Context information to assist speech recognition.
* See [the Cloud Speech
* documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
* for more details.
*
*
* repeated .google.cloud.dialogflow.v2.SpeechContext speech_contexts = 11;
*/
java.util.List getSpeechContextsList();
/**
*
*
*
* Context information to assist speech recognition.
* See [the Cloud Speech
* documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
* for more details.
*
*
* repeated .google.cloud.dialogflow.v2.SpeechContext speech_contexts = 11;
*/
com.google.cloud.dialogflow.v2.SpeechContext getSpeechContexts(int index);
/**
*
*
*
* Context information to assist speech recognition.
* See [the Cloud Speech
* documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
* for more details.
*
*
* repeated .google.cloud.dialogflow.v2.SpeechContext speech_contexts = 11;
*/
int getSpeechContextsCount();
/**
*
*
*
* Context information to assist speech recognition.
* See [the Cloud Speech
* documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
* for more details.
*
*
* repeated .google.cloud.dialogflow.v2.SpeechContext speech_contexts = 11;
*/
java.util.List extends com.google.cloud.dialogflow.v2.SpeechContextOrBuilder>
getSpeechContextsOrBuilderList();
/**
*
*
*
* Context information to assist speech recognition.
* See [the Cloud Speech
* documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
* for more details.
*
*
* repeated .google.cloud.dialogflow.v2.SpeechContext speech_contexts = 11;
*/
com.google.cloud.dialogflow.v2.SpeechContextOrBuilder getSpeechContextsOrBuilder(int index);
/**
*
*
*
* Which Speech model to select for the given request. Select the
* model best suited to your domain to get best results. If a model is not
* explicitly specified, then we auto-select a model based on the parameters
* in the InputAudioConfig.
* If enhanced speech model is enabled for the agent and an enhanced
* version of the specified model for the language does not exist, then the
* speech is recognized using the standard version of the specified model.
* Refer to
* [Cloud Speech API
* documentation](https://cloud.google.com/speech-to-text/docs/basics#select-model)
* for more details.
*
*
* string model = 7;
*
* @return The model.
*/
java.lang.String getModel();
/**
*
*
*
* Which Speech model to select for the given request. Select the
* model best suited to your domain to get best results. If a model is not
* explicitly specified, then we auto-select a model based on the parameters
* in the InputAudioConfig.
* If enhanced speech model is enabled for the agent and an enhanced
* version of the specified model for the language does not exist, then the
* speech is recognized using the standard version of the specified model.
* Refer to
* [Cloud Speech API
* documentation](https://cloud.google.com/speech-to-text/docs/basics#select-model)
* for more details.
*
*
* string model = 7;
*
* @return The bytes for model.
*/
com.google.protobuf.ByteString getModelBytes();
/**
*
*
*
* Which variant of the [Speech model][google.cloud.dialogflow.v2.InputAudioConfig.model] to use.
*
*
* .google.cloud.dialogflow.v2.SpeechModelVariant model_variant = 10;
*
* @return The enum numeric value on the wire for modelVariant.
*/
int getModelVariantValue();
/**
*
*
*
* Which variant of the [Speech model][google.cloud.dialogflow.v2.InputAudioConfig.model] to use.
*
*
* .google.cloud.dialogflow.v2.SpeechModelVariant model_variant = 10;
*
* @return The modelVariant.
*/
com.google.cloud.dialogflow.v2.SpeechModelVariant getModelVariant();
/**
*
*
*
* If `false` (default), recognition does not cease until the
* client closes the stream.
* If `true`, the recognizer will detect a single spoken utterance in input
* audio. Recognition ceases when it detects the audio's voice has
* stopped or paused. In this case, once a detected intent is received, the
* client should close the stream and start a new request with a new stream as
* needed.
* Note: This setting is relevant only for streaming methods.
* Note: When specified, InputAudioConfig.single_utterance takes precedence
* over StreamingDetectIntentRequest.single_utterance.
*
*
* bool single_utterance = 8;
*
* @return The singleUtterance.
*/
boolean getSingleUtterance();
/**
*
*
*
* Only used in [Participants.AnalyzeContent][google.cloud.dialogflow.v2.Participants.AnalyzeContent] and
* [Participants.StreamingAnalyzeContent][google.cloud.dialogflow.v2.Participants.StreamingAnalyzeContent].
* If `false` and recognition doesn't return any result, trigger
* `NO_SPEECH_RECOGNIZED` event to Dialogflow agent.
*
*
* bool disable_no_speech_recognized_event = 14;
*
* @return The disableNoSpeechRecognizedEvent.
*/
boolean getDisableNoSpeechRecognizedEvent();
}
© 2015 - 2025 Weber Informatics LLC | Privacy Policy