All Downloads are FREE. Search and download functionalities are using the official Maven repository.

com.ibm.watson.speech_to_text.v1.model.SpeechRecognitionResult Maven / Gradle / Ivy

There is a newer version: 13.0.0
Show newest version
/*
 * (C) Copyright IBM Corp. 2018, 2023.
 *
 * Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with
 * the License. You may obtain a copy of the License at
 *
 * http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on
 * an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the
 * specific language governing permissions and limitations under the License.
 */
package com.ibm.watson.speech_to_text.v1.model;

import com.google.gson.annotations.SerializedName;
import com.ibm.cloud.sdk.core.service.model.GenericModel;
import java.util.List;
import java.util.Map;

/** Component results for a speech recognition request. */
public class SpeechRecognitionResult extends GenericModel {

  /**
   * If the `split_transcript_at_phrase_end` parameter is `true`, describes the reason for the
   * split: * `end_of_data` - The end of the input audio stream. * `full_stop` - A full semantic
   * stop, such as for the conclusion of a grammatical sentence. The insertion of splits is
   * influenced by the base language model and biased by custom language models and grammars. *
   * `reset` - The amount of audio that is currently being processed exceeds the two-minute maximum.
   * The service splits the transcript to avoid excessive memory use. * `silence` - A pause or
   * silence that is at least as long as the pause interval.
   */
  public interface EndOfUtterance {
    /** end_of_data. */
    String END_OF_DATA = "end_of_data";
    /** full_stop. */
    String FULL_STOP = "full_stop";
    /** reset. */
    String RESET = "reset";
    /** silence. */
    String SILENCE = "silence";
  }

  @SerializedName("final")
  protected Boolean xFinal;

  protected List alternatives;

  @SerializedName("keywords_result")
  protected Map> keywordsResult;

  @SerializedName("word_alternatives")
  protected List wordAlternatives;

  @SerializedName("end_of_utterance")
  protected String endOfUtterance;

  protected SpeechRecognitionResult() {}

  /**
   * Gets the xFinal.
   *
   * 

An indication of whether the transcription results are final: * If `true`, the results for * this utterance are final. They are guaranteed not to be updated further. * If `false`, the * results are interim. They can be updated with further interim results until final results are * eventually sent. * *

**Note:** Because `final` is a reserved word in Java and Swift, the field is renamed * `xFinal` in Java and is escaped with back quotes in Swift. * * @return the xFinal */ public Boolean isXFinal() { return xFinal; } /** * Gets the alternatives. * *

An array of alternative transcripts. The `alternatives` array can include additional * requested output such as word confidence or timestamps. * * @return the alternatives */ public List getAlternatives() { return alternatives; } /** * Gets the keywordsResult. * *

A dictionary (or associative array) whose keys are the strings specified for `keywords` if * both that parameter and `keywords_threshold` are specified. The value for each key is an array * of matches spotted in the audio for that keyword. Each match is described by a `KeywordResult` * object. A keyword for which no matches are found is omitted from the dictionary. The dictionary * is omitted entirely if no matches are found for any keywords. * * @return the keywordsResult */ public Map> getKeywordsResult() { return keywordsResult; } /** * Gets the wordAlternatives. * *

An array of alternative hypotheses found for words of the input audio if a * `word_alternatives_threshold` is specified. * * @return the wordAlternatives */ public List getWordAlternatives() { return wordAlternatives; } /** * Gets the endOfUtterance. * *

If the `split_transcript_at_phrase_end` parameter is `true`, describes the reason for the * split: * `end_of_data` - The end of the input audio stream. * `full_stop` - A full semantic * stop, such as for the conclusion of a grammatical sentence. The insertion of splits is * influenced by the base language model and biased by custom language models and grammars. * * `reset` - The amount of audio that is currently being processed exceeds the two-minute maximum. * The service splits the transcript to avoid excessive memory use. * `silence` - A pause or * silence that is at least as long as the pause interval. * * @return the endOfUtterance */ public String getEndOfUtterance() { return endOfUtterance; } }





© 2015 - 2024 Weber Informatics LLC | Privacy Policy