Whisper Large v2 with Key-Value-Cache enabled in ONNX fp16 format

Description

This repo contains the ONNX files for the ONNX conversion of Whisper Large v2 done by Esperanto Technologies. The model is in the fp16 format and has the KVC enabled.

How to download ONNX model and weight files

The easiest way to obtain the model is to clone this whole repo. Alternatively you can download the files is using the huggingface-hub Python library.

pip3 install huggingface-hub>=0.17.1

Then you can download any individual model file to the current directory, at high speed, with a command like this:

huggingface-cli download Esperanto/whisper-large-v2-kvc-fp16-onnx --local-dir whisper-large-v2-kvc-fp16-onnx --local-dir-use-symlinks False

For more documentation on downloading with huggingface-cli, please see: HF -> Hub Python Library -> Download files -> Download from the CLI.

How to run from Python code using ONNXRuntime

This model can easily be ran in a CPU using ONNXRuntime.

Scripts about how to run these models will be provided soon.

Downloads last month
4
Inference API
Unable to determine this model's library. Check the docs .

Model tree for Esperanto/whisper-large-v2-kvc-fp16-onnx

Quantized
(3)
this model

Collection including Esperanto/whisper-large-v2-kvc-fp16-onnx