badrex

20 models • 1 total models in database
Sort by:

Ethio-ASR-multilingual-1B

NaNK
license:cc-by-nc-4.0
344
1

Ethio-ASR-tigrinya

license:cc-by-4.0
335
0

mms-300m-arabic-dialect-identifier

NaNK
license:cc-by-4.0
323
8

w2v-bert-2.0-tigrinya-asr

NaNK
license:cc-by-4.0
299
0

w2v-bert-2.0-kinyarwanda-asr

This model is a fine-tuned version of Wav2Vec2-BERT 2.0 for Kinyarwanda automatic speech recognition (ASR). It was trained on the 1000 hours dataset from the Kinyarwanda ASR hackthon on Kaggle (Track B), dataset covering Health, Government, Finance, Education, and Agriculture domains. The model is robust and the in-domain WER is below 8.4%. - Developed by: Badr al-Absi - Model type: Speech Recognition (ASR) - Language: Kinyarwanda (rw) - License: CC-BY-4.0 - Finetuned from: facebook/w2v-bert-2.0 Examples 🚀 | | Audio | Human Transcription | ASR Transcription | |----------|--------|----------------|----------------| | 1 | | Umugore wambaye umupira w'akazi mpuzankano iri mu ibara ry'umuhondo handitseho amagambo yandikishije ibara ry'ubururu. Afite igikoresho cy'itumanaho gikoreshwa mu guhamagara no kwandika ubutumwa bugufi. | umugore wambaye umupira w'akazi impuzankano iri mu ibara ry'umuhondo handitseho amagambo yandikishije ibara ry'ubururu afite igikoresho cy'itumanaho gikoreshwa mu guhamagara no kwandika ubutumwa bugufi | | 2 | | Igikoresho cyifashishwa mu kwiga imibare ndetse kiba kirimo ibindi bikoresho byinshi harimo amarati atatu ndetse n'irati imwe ndende n'ikaramu na kompa, ibigibi byahawe abanyeshuri biga mu myaka ya mbere n'iya kabiri kugira ngo bajye babyifashisha bari kwiga imibare. | igikoresho cyifashishwa mu kwiga imibare ndetse kiba kirimo ibindi bikoresho byinshi harimo amarati atatu ndetse n'irati imwe ndende n'ikaramu na kompa ibi ngibi byahawe abanyeshuri biga mu myaka ya mbere n'iya kabiri kugira ngo bajye babyifashisha bari kwiga imibare | | 3 | | Iyi ni Kizimyamwoto iri mu ibara ry'umutuku. Hejuru hakaba hariho amabara y'umuhondo ku ruhande hakaba hariho akantu kameze nk'isaha, hasi hakaba hariho akabara gasa n'ubururu kari amagambo menshi mu rurimi rw'icyongereza hasi yako hakaba hari n'akandi kari mu ibara ry' umuhondo handikishijemo amagambo y'icyongereza, hasi yako hakaba hari n' utundi tuntu tw' utubokisi tw' umweru harimo utuntu tujyiye dushushanyije hakaba hariho n' inyajwi bi na si. | iyinzuzinyamwoto iri mu ibara ry'umutuku hejuru hakaba hariho ahariho amabara y'umuhondo ku ruhande hakaba hariho akantu kameze nk'isaha hasi hakaba hariho akabara gatoya k'ubururu kariho amagambo yandikishije mu rurimi rw'icyongereza hasi yako hakaba hari n'akandi kari mu ibara ry'umuhondo wandikishijemo amagambo y'icyongereza hasi yako hakaba hariho utundi tutu tw'tuboisi hariho amaotw'utubogisi tw'umweru harimo utuntu tugiye dushushanyije hakaba hariho n'inyajwi bi na si | - Repository: https://huggingface.co/badrex/w2v-bert-2.0-kinyarwanda-asr - Dataset: Kinyarwanda ASR Track B The model can be used directly for automatic speech recognition of Kinyarwanda audio: This model can be used as a foundation for: - building voice assistants for Kinyarwanda speakers - transcription services for Kinyarwanda content - accessibility tools for Kinyarwanda-speaking communities - research in low-resource speech recognition - transcribing languages other than Kinyarwanda - real-time applications without proper latency testing - high-stakes applications without domain-specific validation - Domain bias: primarily trained on formal speech from specific domains (Health, Government, Finance, Education, Agriculture) - Accent variation: may not perform well on dialects or accents not represented in training data - Audio quality: performance may degrade on noisy or low-quality audio - Technical terms: may struggle with specialized vocabulary outside training domains The model was fine-tuned on the Kinyarwanda ASR hackthon - Track B dataset: - Size: ~1000 hours of transcribed Kinyarwanda speech - Domains: Health, Government, Finance, Education, Agriculture - Source: Digital Umuganda (Gates Foundation funded) - License: CC-BY-4.0 - Base model: Wav2Vec2-BERT 2.0 - Architecture: transformer-based with convolutional feature extractor - Parameters: ~600M (inherited from base model) - Objective: connectionist temporal classification (CTC) For questions or issues, please contact via the Hugging Face model repository.

NaNK
license:cc-by-4.0
291
2

Ethio-ASR-multilingual-94M

license:cc-by-nc-sa-4.0
277
0

Ethio-ASR-multilingual-300M

license:cc-by-nc-4.0
267
0

Ethio-ASR-multilingual-600M

license:cc-by-4.0
238
0

Ethio-ASR-amharic

license:cc-by-4.0
186
0

Ethio-ASR-oromo

license:cc-by-4.0
180
0

Ethio-ASR-walaytta

license:cc-by-4.0
172
0

Ethio-ASR-sidaama

license:cc-by-4.0
172
0

w2v-bert-2.0-swahili-asr

NaNK
license:cc-by-4.0
119
0

w2v-bert-2.0-zulu-asr

This model is a fine-tuned version of Wav2Vec2-BERT 2.0 for Zulu automatic speech recognition (ASR). It was trained on the 250 hours of transcribed Zulu speech. The ASR model is robust and the in-domain WER is below 16.3%. - Developed by: Badr al-Absi - Model type: Speech Recognition (ASR) - Language: Zulu (zu) - License: CC-BY-4.0 - Finetuned from: facebook/w2v-bert-2.0 Examples 🚀 | | Audio | Human Transcription | ASR Transcription | |----------|--------|----------------|----------------| | 1 | | Yenza isinqumo ngezilimo uzozitshala kumaphi amasimu uphinde idwebe imephu njengereferensi yakho. | yenza isinqumo ngezilimo ozozitshala kumaphi amasimu uphinde igwebe imephu njengereference yakho | | 2 | | Emdlalweni wokugcina ngokumelene IFrance, wayengumuntu ongasetshenziswanga esikhundleni njengoba i-Argentina inqobe ngo-4-2 nge-penalty ukuze ithole isiqu sayo sesithathu seNdebe Yomhlaba. | emdlalweni wokugqina ngokumelene i-france wayengumuntu ongasetshenziswanga esikhundleni njengoba i-argentina incobe ngo-4-2 ngephelnathi ukuze ithole isiqu sayo sesithathu sendebe yomhlaba | | 3 | | Amadolobhana angaphandle angaphezu kwamamitha ambalwa, Reneging cishe 140m, amamitha angu-459.3, ngaphezu kogu lolwandle. Le ndawo iningi emahlathini ama-dune asogwini, ikakhulukazi eceleni kwe-zindunduma zasogwini nasedolobheni lase-Meerensee. | amadolobhana angaphandle angaphezu kwamamitha ambalwa reneging cishe 140m amamitha angu 4593 ngaphezu kogulolwandle le ndawo iningi emahlabathini amedum esogwini ikakhulukazi eceleni kwezindunduma zasogwini nasedolobheni lasemerins | The model can be used directly for automatic speech recognition of Zulu audio: This model can be used as a foundation for: - building voice assistants for Zulu speakers - transcription services for Zulu content - accessibility tools for Zulu-speaking communities - research in low-resource speech recognition - Base model: Wav2Vec2-BERT 2.0 - Architecture: transformer-based with convolutional feature extractor - Parameters: ~600M (inherited from base model) - Objective: connectionist temporal classification (CTC) Funding The development of this model was supported by CLEAR Global and Gates Foundation. For questions or issues, please contact via the Hugging Face model repository in the community discussion section.

NaNK
license:cc-by-4.0
105
0

w2v-bert-2.0-ethiopian-asr

license:cc-by-4.0
46
0

w2v-bert-2.0-amharic-asr

NaNK
license:cc-by-4.0
37
0

w2v-bert-2.0-oromo-asr

NaNK
license:cc-by-4.0
24
0

w2v-bert-2.0-sidama-asr

NaNK
license:cc-by-4.0
19
0

w2v-bert-2.0-wolaytta-asr

license:cc-by-4.0
19
0

w2v-bert-2.0-kikuyu-asr

This model is a fine-tuned version of Wav2Vec2-BERT 2.0 for Kikuyu automatic speech recognition (ASR). It was trained on the 100+ hours of transcribed speech, covering Health, Government, Finance, Education, and Agriculture domains. The in-domain WER for this ASR model is below 25.0%. - Developed by: Badr al-Absi - Model type: Speech Recognition (ASR) - Language: Kikuyu (kik) - License: CC-BY-4.0 - Finetuned from: facebook/w2v-bert-2.0 - Base model: Wav2Vec2-BERT 2.0 - Architecture: transformer-based with convolutional feature extractor - Parameters: ~600M (inherited from base model) - Objective: connectionist temporal classification (CTC) Funding The development of this model was supported by CLEAR Global and Gates Foundation. For questions or issues, please contact via the Hugging Face model repository in the community discussion section.

NaNK
license:cc-by-4.0
17
0