bofenghuang

45 models • 3 total models in database
Sort by:

whisper-large-v3-french-distil-dec16

license:mit
1,665
11

vigostral-7b-chat

NaNK
license:apache-2.0
1,340
28

vigogne-2-7b-instruct

NaNK
llama
775
25

vigogne-2-7b-chat

NaNK
llama
680
24

vigogne-2-13b-instruct

NaNK
llama
676
15

vigogne-33b-instruct

NaNK
llama
674
5

vigogne-13b-instruct

NaNK
llama
667
13

vigogne-7b-instruct

NaNK
llama
666
23

vigogne-7b-chat

NaNK
llama
664
4

vigogne-13b-chat

NaNK
llama
664
1

vigogne-2-13b-chat

NaNK
llama
657
0

vigogne-2-70b-chat

NaNK
llama
653
6

whisper-large-v3-french

license:mit
632
34

asr-wav2vec2-ctc-french

license:apache-2.0
501
13

Whisper Small Cv11 French

This model is a fine-tuned version of openai/whisper-small, trained on the mozilla-foundation/commonvoice110 fr dataset. When using the model make sure that your speech input is also sampled at 16Khz. This model also predicts casing and punctuation. Below are the WERs of the pre-trained models on the Common Voice 9.0, Multilingual LibriSpeech, Voxpopuli and Fleurs. These results are reported in the original paper. | Model | Common Voice 9.0 | MLS | VoxPopuli | Fleurs | | --- | :---: | :---: | :---: | :---: | | openai/whisper-small | 22.7 | 16.2 | 15.7 | 15.0 | | openai/whisper-medium | 16.0 | 8.9 | 12.2 | 8.7 | | openai/whisper-large | 14.7 | 8.9 | 11.0 | 7.7 | | openai/whisper-large-v2 | 13.9 | 7.3 | 11.4 | 8.3 | Below are the WERs of the fine-tuned models on the Common Voice 11.0, Multilingual LibriSpeech, Voxpopuli, and Fleurs. Note that these evaluation datasets have been filtered and preprocessed to only contain French alphabet characters and are removed of punctuation outside of apostrophe. The results in the table are reported as `WER (greedy search) / WER (beam search with beam width 5)`. | Model | Common Voice 11.0 | MLS | VoxPopuli | Fleurs | | --- | :---: | :---: | :---: | :---: | | bofenghuang/whisper-small-cv11-french | 11.76 / 10.99 | 9.65 / 8.91 | 14.45 / 13.66 | 10.76 / 9.83 | | bofenghuang/whisper-medium-cv11-french | 9.03 / 8.54 | 6.34 / 5.86 | 11.64 / 11.35 | 7.13 / 6.85 | | bofenghuang/whisper-medium-french | 9.03 / 8.73 | 4.60 / 4.44 | 9.53 / 9.46 | 6.33 / 5.94 | | bofenghuang/whisper-large-v2-cv11-french | 8.05 / 7.67 | 5.56 / 5.28 | 11.50 / 10.69 | 5.42 / 5.05 | | bofenghuang/whisper-large-v2-french | 8.15 / 7.83 | 4.20 / 4.03 | 9.10 / 8.66 | 5.22 / 4.98 |

license:apache-2.0
178
6

whisper-large-v3-distil-it-v0.2

license:mit
141
1

Whisper Medium French

This model is a fine-tuned version of openai/whisper-medium, trained on a composite dataset comprising of over 2200 hours of French speech audio, using the train and the validation splits of Common Voice 11.0, Multilingual LibriSpeech, Voxpopuli, Fleurs, Multilingual TEDx, MediaSpeech, and African Accented French. When using the model make sure that your speech input is sampled at 16Khz. This model doesn't predict casing or punctuation. Below are the WERs of the pre-trained models on the Common Voice 9.0, Multilingual LibriSpeech, Voxpopuli and Fleurs. These results are reported in the original paper. | Model | Common Voice 9.0 | MLS | VoxPopuli | Fleurs | | --- | :---: | :---: | :---: | :---: | | openai/whisper-small | 22.7 | 16.2 | 15.7 | 15.0 | | openai/whisper-medium | 16.0 | 8.9 | 12.2 | 8.7 | | openai/whisper-large | 14.7 | 8.9 | 11.0 | 7.7 | | openai/whisper-large-v2 | 13.9 | 7.3 | 11.4 | 8.3 | Below are the WERs of the fine-tuned models on the Common Voice 11.0, Multilingual LibriSpeech, Voxpopuli, and Fleurs. Note that these evaluation datasets have been filtered and preprocessed to only contain French alphabet characters and are removed of punctuation outside of apostrophe. The results in the table are reported as `WER (greedy search) / WER (beam search with beam width 5)`. | Model | Common Voice 11.0 | MLS | VoxPopuli | Fleurs | | --- | :---: | :---: | :---: | :---: | | bofenghuang/whisper-small-cv11-french | 11.76 / 10.99 | 9.65 / 8.91 | 14.45 / 13.66 | 10.76 / 9.83 | | bofenghuang/whisper-medium-cv11-french | 9.03 / 8.54 | 6.34 / 5.86 | 11.64 / 11.35 | 7.13 / 6.85 | | bofenghuang/whisper-medium-french | 9.03 / 8.73 | 4.60 / 4.44 | 9.53 / 9.46 | 6.33 / 5.94 | | bofenghuang/whisper-large-v2-cv11-french | 8.05 / 7.67 | 5.56 / 5.28 | 11.50 / 10.69 | 5.42 / 5.05 | | bofenghuang/whisper-large-v2-french | 8.15 / 7.83 | 4.20 / 4.03 | 9.10 / 8.66 | 5.22 / 4.98 |

license:apache-2.0
54
20

phonemizer-wav2vec2-ctc-french

license:mit
44
0

whisper-medium-cv11-german

license:apache-2.0
32
3

asr-wav2vec2-xls-r-1b-ctc-french

NaNK
license:apache-2.0
31
0

whisper-large-v2-french

license:apache-2.0
29
14

Whisper Medium Cv11 French

This model is a fine-tuned version of openai/whisper-medium, trained on the mozilla-foundation/commonvoice110 fr dataset. When using the model make sure that your speech input is also sampled at 16Khz. This model also predicts casing and punctuation. Below are the WERs of the pre-trained models on the Common Voice 9.0, Multilingual LibriSpeech, Voxpopuli and Fleurs. These results are reported in the original paper. | Model | Common Voice 9.0 | MLS | VoxPopuli | Fleurs | | --- | :---: | :---: | :---: | :---: | | openai/whisper-small | 22.7 | 16.2 | 15.7 | 15.0 | | openai/whisper-medium | 16.0 | 8.9 | 12.2 | 8.7 | | openai/whisper-large | 14.7 | 8.9 | 11.0 | 7.7 | | openai/whisper-large-v2 | 13.9 | 7.3 | 11.4 | 8.3 | Below are the WERs of the fine-tuned models on the Common Voice 11.0, Multilingual LibriSpeech, Voxpopuli, and Fleurs. Note that these evaluation datasets have been filtered and preprocessed to only contain French alphabet characters and are removed of punctuation outside of apostrophe. The results in the table are reported as `WER (greedy search) / WER (beam search with beam width 5)`. | Model | Common Voice 11.0 | MLS | VoxPopuli | Fleurs | | --- | :---: | :---: | :---: | :---: | | bofenghuang/whisper-small-cv11-french | 11.76 / 10.99 | 9.65 / 8.91 | 14.45 / 13.66 | 10.76 / 9.83 | | bofenghuang/whisper-medium-cv11-french | 9.03 / 8.54 | 6.34 / 5.86 | 11.64 / 11.35 | 7.13 / 6.85 | | bofenghuang/whisper-medium-french | 9.03 / 8.73 | 4.60 / 4.44 | 9.53 / 9.46 | 6.33 / 5.94 | | bofenghuang/whisper-large-v2-cv11-french | 8.05 / 7.67 | 5.56 / 5.28 | 11.50 / 10.69 | 5.42 / 5.05 | | bofenghuang/whisper-large-v2-french | 8.15 / 7.83 | 4.20 / 4.03 | 9.10 / 8.66 | 5.22 / 4.98 |

license:apache-2.0
28
3

whisper-large-v3-distil-fr-v0.2

license:mit
26
2

whisper-large-v2-cv11-german

license:apache-2.0
24
17

whisper-large-v2-cv11-french-ct2

license:apache-2.0
24
0

whisper-small-cv11-german

license:apache-2.0
23
7

stt_fr_fastconformer_hybrid_large

license:apache-2.0
23
1

whisper-large-v2-cv11-french

license:apache-2.0
15
5

whisper-large-v2-cv11-german-ct2

license:apache-2.0
9
0

whisper-large-v3-french-distil-dec8

license:mit
8
4

vigogne-mpt-7b-instruct

NaNK
8
0

whisper-large-v3-french-distil-dec2

license:mit
7
1

whisper-large-v3-distil-multi4-v0.2

license:mit
7
1

whisper-large-v3-french-distil-dec4

license:mit
5
0

parakeet-tdt-0.6b-v3-hybrid

Extend nvidia/parakeet-tdt-0.6b-v3 from TDT to hybrid TDT-CTC: - Kept encoder and TDT decoder, reinitialized CTC decoder with the same vocab of 8192 tokens - Can be used for pure CTC or hybrid CTC-RNNT finetuning Sanity check seen below passed, getting the same transcriptions using TDT and gibberish with reinitialized CTC:

NaNK
license:cc-by-4.0
4
0

vigogne-falcon-7b-chat

NaNK
license:apache-2.0
3
1

vigogne-stablelm-3b-4e1t-chat

NaNK
license:apache-2.0
3
1

Meta-Llama-3-8B

NaNK
llama
3
1

flan-t5-large-dialogsum-fr

license:apache-2.0
2
2

deprecated-whisper-large-v2-cv11-french-punct-plus

license:apache-2.0
2
1

whisper-large-v3-distil-multi7-v0.2

license:mit
1
3

wav2vec2-xls-r-1b-cv9-fr

NaNK
license:apache-2.0
1
1

vigogne-falcon-7b-instruct

NaNK
license:apache-2.0
1
1

vigogne-bloom-7b1-instruct

NaNK
0
4

vigogne-opt-6.7b-instruct

NaNK
0
2