File size: 6,197 Bytes
2c67b83 e2e3aa3 2c67b83 e2e3aa3 2c67b83 e2e3aa3 465468b e2e3aa3 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 |
---
language: is
datasets:
- language-and-voice-lab/samromur_asr
- language-and-voice-lab/samromur_children
- language-and-voice-lab/malromur_asr
- language-and-voice-lab/althingi_asr
tags:
- audio
- automatic-speech-recognition
- icelandic
- whisper
- whisper-large
- iceland
- reykjavik
- samromur
license: cc-by-4.0
model-index:
- name: whisper-large-icelandic-10k-steps-1000h
results:
- task:
name: Automatic Speech Recognition
type: automatic-speech-recognition
dataset:
name: Samrómur (Test)
type: language-and-voice-lab/samromur_asr
split: test
args:
language: is
metrics:
- name: WER
type: wer
value: 11.879
- task:
name: Automatic Speech Recognition
type: automatic-speech-recognition
dataset:
name: Samrómur (Dev)
type: language-and-voice-lab/samromur_asr
split: validation
args:
language: is
metrics:
- name: WER
type: wer
value: 10.849
- task:
name: Automatic Speech Recognition
type: automatic-speech-recognition
dataset:
name: Samrómur Children (Test)
type: language-and-voice-lab/samromur_children
split: test
args:
language: is
metrics:
- name: WER
type: wer
value: 12.325
- task:
name: Automatic Speech Recognition
type: automatic-speech-recognition
dataset:
name: Samrómur Children (Dev)
type: language-and-voice-lab/samromur_children
split: validation
args:
language: is
metrics:
- name: WER
type: wer
value: 8.078
- task:
name: Automatic Speech Recognition
type: automatic-speech-recognition
dataset:
name: Malrómur (Test)
type: language-and-voice-lab/malromur_asr
split: test
args:
language: is
metrics:
- name: WER
type: wer
value: 10.132
- task:
name: Automatic Speech Recognition
type: automatic-speech-recognition
dataset:
name: Malrómur (Dev)
type: language-and-voice-lab/malromur_asr
split: validation
args:
language: is
metrics:
- name: WER
type: wer
value: 10.157
- task:
name: Automatic Speech Recognition
type: automatic-speech-recognition
dataset:
name: Althingi (Test)
type: language-and-voice-lab/althingi_asr
split: test
args:
language: is
metrics:
- name: WER
type: wer
value: 11.750
- task:
name: Automatic Speech Recognition
type: automatic-speech-recognition
dataset:
name: Althingi (Dev)
type: language-and-voice-lab/althingi_asr
split: validation
args:
language: is
metrics:
- name: WER
type: wer
value: 11.141
---
# whisper-large-icelandic-10k-steps-1000h
The "whisper-large-icelandic-10k-steps-1000h" is an acoustic model suitable for Automatic Speech Recognition in Icelandic. It is the result of fine-tuning the model "openai/whisper-large" with around 1000 hours of Icelandic data developed by the [Language and Voice Laboratory](https://huggingface.co./language-and-voice-lab). Most of the data is available at public repositories such as [LDC](https://www.ldc.upenn.edu/), [OpenSLR](https://openslr.org/) or [Clarin.is](https://clarin.is/)
The specific list of corpora used to fine-tune the model is:
- [Samrómur 21.05 (114h34m)](http://www.openslr.org/112/)
- [Samrómur Children (127h25m)](https://catalog.ldc.upenn.edu/LDC2022S11)
- [Malrómur (119hh03m)](https://clarin.is/en/resources/malromur/)
- [Althingi Parliamentary Speech (514h29m)](https://catalog.ldc.upenn.edu/LDC2021S01)
- L2-Speakers Data (125h55m) **Unpublished material**
The fine-tuning process was performed during March (2023) in the servers of the Language and Voice Laboratory (https://lvl.ru.is/) at Reykjavík University (Iceland) by Carlos Daniel Hernández Mena.
# Evaluation
```python
import torch
from transformers import WhisperForConditionalGeneration, WhisperProcessor
#Load the processor and model.
MODEL_NAME="carlosdanielhernandezmena/whisper-large-icelandic-10k-steps-1000h"
processor = WhisperProcessor.from_pretrained(MODEL_NAME)
model = WhisperForConditionalGeneration.from_pretrained(MODEL_NAME).to("cuda")
#Load the dataset
from datasets import load_dataset, load_metric, Audio
ds=load_dataset("language-and-voice-lab/samromur_children",split='test')
#Downsample to 16kHz
ds = ds.cast_column("audio", Audio(sampling_rate=16_000))
#Process the dataset
def map_to_pred(batch):
audio = batch["audio"]
input_features = processor(audio["array"], sampling_rate=audio["sampling_rate"], return_tensors="pt").input_features
batch["reference"] = processor.tokenizer._normalize(batch['normalized_text'])
with torch.no_grad():
predicted_ids = model.generate(input_features.to("cuda"))[0]
transcription = processor.decode(predicted_ids)
batch["prediction"] = processor.tokenizer._normalize(transcription)
return batch
#Do the evaluation
result = ds.map(map_to_pred)
#Compute the overall WER now.
from evaluate import load
wer = load("wer")
WER=100 * wer.compute(references=result["reference"], predictions=result["prediction"])
print(WER)
```
**Test Result**: 12.325364793542379
# BibTeX entry and citation info
*When publishing results based on these models please refer to:*
```bibtex
@misc{mena2023whisperlarge10kicelandic,
title={Acoustic Model in Icelandic: whisper-large-icelandic-10k-steps-1000h.},
author={Hernandez Mena, Carlos Daniel},
url={https://huggingface.co./carlosdanielhernandezmena/whisper-large-icelandic-10k-steps-1000h},
year={2023}
}
```
# Acknowledgements
Thanks to Jón Guðnason, head of the Language and Voice Lab for providing computational power to make this model possible. We also want to thank to the "Language Technology Programme for Icelandic 2019-2023" which is managed and coordinated by Almannarómur, and it is funded by the Icelandic Ministry of Education, Science and Culture.
Special thanks to Björn Ingi Stefánsson for setting up the configuration of the server where this model was trained.
|