Performance

The solution has various configurations that trade off accuracy and performance. In this page:

  • Realtime streams (RTS) is the number of concurrent streams that can be serviced by a single accelerator using default settings.
  • Compute latency 99th-percentile (CL99) is the 99th-percentile compute latency, which measures how long it takes for a model to make a prediction for one audio frame. Note that CL99 increases with the number of concurrent streams.
  • User-perceived latency (UPL) is the time difference between when the user finishes saying a word and when it is returned as a transcript by the system.
  • WER is the Word Error Rate, a measure of the accuracy of the model. Lower is better.
  • HF Leaderboard WER is the WER of the model on the Huggingface Open ASR Leaderboard. WER is averaged across 9 test datasets.

The WERs in the following section are for models trained on 13k hours of open-source data described at the bottom of this page.

The UPL were computed by streaming librispeech-dev-clean audio live to an FPGA backend server on-site. Please refer to this document for more details on latencies.

Without state resets

The solution supports decoding with a beam search (default beam width=4) with an n-gram language model for improved accuracy. The solution supports greedy decoding for higher throughput.

ModelParametersDecodingRTSCL99 at max RTSmedian UPLHF Leaderboard WER
base85Mgreedy200025 ms147 ms12.31%
base85Mbeam, width=4130080 ms-11.64%
large196Mgreedy80025 ms-11.41%
large196Mbeam, width=450040 ms158 ms10.95%

State resets

State resets is a technique that improves the accuracy on long utterances (over 60s) by resetting the model's hidden state after a fixed duration. This reduces the number of real-time streams that can be supported by around 25%:

ModelParametersDecodingRTSCL99 at max RTSmedian UPLHF Leaderboard WER
base85Mgreedy160045 ms147 ms12.14%
base85Mbeam, width=4120050 ms-11.38%
large196Mgreedy65055 ms-11.13%
large196Mbeam, width=440060 ms158 ms10.61%

Note that most of the data in the Huggingface leaderboard is less than 60s long so the impact of state resets is not reflected in the leaderboard WER.

Since the UPL numbers were computed from librispeech-dev-clean, the effect of state resets is not reflected in measured latencies.

13k hour dataset

The models above were trained on 13k hrs of open-source training data consisting of:

  • LibriSpeech-960h
  • Common Voice 17.0
  • 961 hours from MLS
  • Peoples' Speech: A ~9389 hour subset filtered for transcription quality
  • 155 hrs of AMI

This data has a maximum_duration of 20s.