V-Blaze and V-Cloud Online Help

Other / Advanced

Table 1. Advanced Parameters





false (default), true

Specifies that the input file is encoded in base 64.


The value specified is the name or number of the billing account that a transcription request should be associated with.


false (default), true, language_model

The lid parameter enables you to use the ASR engine's Language Identification Module to identify the language spoken in the input audio. When lid identifies the language, the appropriate language model is selected based on the spoken language detected in the audio. For example, if Spanish is detected, then the resulting transcript will be in Spanish. Additionally, if you require an alternate model, specify it using lid=language_model.

  • lid=true - automatically selects the language identification model based on the LID and language models that are available.

  • lid=language_model - the alternate language model.

  • lid=language_model,language_model - for dual-channel audio, this specifies the alternate language models to use for channel 0 and channel 1 respectively.

  • lid=language_model:notext - use this to not decode when a language is detected, just that a language was detected. If this option is specified, the JSON will not contain a model metadata element in each utterance.

  • lid=false - lid is not used.

The following parameters provide additional options when using the lid tag:

  • lidmaxtime (default 20.0 seconds) - maximum audio duration (seconds) to analyze. For example, if lidmaxtime=20, the ASR engine will analyze 20 seconds of audio at most. This tag has no effect when lidutt=true.

  • lidthreshold (default 0.7) - specifies the required confidence level before lid will stop analyzing audio.

  • lidoffset=N - Delays the start of LID until specified (N) seconds into audio. If there is not enough audio left after offset, this will process preceding utterances in reverse. This option is only available in V‑Blaze 7.1+ and V‑Cloud 1.6+.