cwkeam commited on
Commit
eaa3a8c
1 Parent(s): a48f800

update readme to working code

Browse files
Files changed (1) hide show
  1. README.md +5 -4
README.md CHANGED
@@ -45,18 +45,19 @@ For more information on how the model was trained, please take a look at the [of
45
  To transcribe audio files the model can be used as a standalone acoustic model as follows:
46
 
47
  ```python
48
- import torch
49
  import torchaudio
 
50
  from transformers import MCTCTForCTC, MCTCTProcessor
51
 
52
- model = MCTCTForCTC.from_pretrained("speechbrain/mctct-large")
53
- processor = MCTCTProcessor.from_pretrained("speechbrain/mctct-large")
54
 
55
  # load dummy dataset and read soundfiles
56
  ds = load_dataset("patrickvonplaten/librispeech_asr_dummy", "clean", split="validation")
57
 
58
  # tokenize
59
- input_features = processor(ds[0]["audio"]["array"], return_tensors="pt", padding="longest").input_features # Batch size 1
60
 
61
  # retrieve logits
62
  logits = model(input_features).logits
 
45
  To transcribe audio files the model can be used as a standalone acoustic model as follows:
46
 
47
  ```python
48
+ import torch
49
  import torchaudio
50
+ from datasets import load_dataset
51
  from transformers import MCTCTForCTC, MCTCTProcessor
52
 
53
+ model = MCTCTForCTC.from_pretrained("cwkeam/mctct-large")
54
+ processor = MCTCTProcessor.from_pretrained("cwkeam/mctct-large")
55
 
56
  # load dummy dataset and read soundfiles
57
  ds = load_dataset("patrickvonplaten/librispeech_asr_dummy", "clean", split="validation")
58
 
59
  # tokenize
60
+ input_features = processor(ds[0]["audio"]["array"], return_tensors="pt").input_features
61
 
62
  # retrieve logits
63
  logits = model(input_features).logits