Back to overview
Blog

Whisper Deployment Decisions: Part I — Evaluating Latency, Costs, and Performance Metrics

Updated
16 Sep 2025
Published
5 Nov 2023
Reading time
2 min
Tags
 Whisper Deployment Decisions: Part I — Evaluating Latency, Costs, and Performance Metrics
Share this on:
Whisper Deployment Decisions: Part I — Evaluating Latency, Costs, and Performance Metrics
1:50

Deploying any Machine Learning model into production requires a thorough consideration of three major factors:

  • Performance metrics
  • Latency
  • Deployment cost

While numerous articles cover training, fine-tuning, and explaining the paper behind Whisper, there is a scarcity of resources focusing on running Whisper in a production environment. Embarking on a two-part blog series, we delve into the practicalities of implementing OpenAI Whisper in a production environment. In the first part, we navigate the tradeoffs between model sizes and GPUs, shedding light on optimal choices. The sequel takes a closer look at the transformative effects of tools and techniques like JAX, ONNX, and KernlAI on these metrics.

Using HuggingFace's Whisper implementation, we benchmarked multilingual models across different batch sizes (1,2,4,8,and 16) on CPUs and GPUs (T4,V100,and A100) to evaluate inference speed. All the benchmarks were done on the test split of the HuggingFace dataset: librispeech_asr.

Key Findings:

  • As the size of the Whisper model increases, the inference time becomes slower since larger models have more parameters.
  • Running Whisper on CPUs is noticeably slower compared to using GPUs.
  • Irrespective of the model size, the inference time is the quickest on A100.

To sum it up: The T4 GPU emerges as the optimal choice for supporting any Whisper model (except Whisper large-v2) in online (Batch-size = 1) and batch settings. It offers a cost-effective solution compared to the P100 and A100 GPUs. Despite the P100’s superior speed in batch settings compared to the T4, the tradeoff of higher costs makes it a less economical choice.

Read the full article on our Medium account.

About the author

ML6

ML6 is an AI consulting and engineering company with expertise in data, cloud, and applied machine learning. The team helps organizations bring scalable and reliable AI solutions into production, turning cutting-edge technology into real business impact.

The answers you've been looking for

Frequently asked questions