haoranxu

19 models • 1 total models in database
Sort by:

ALMA-7B-R

NaNK
llama
8,656
14

ALMA-13B-Pretrain

NaNK
llama
8,163
10

X-ALMA-13B-Pretrain

NaNK
llama
2,008
11

ALMA-13B

NaNK
llama
1,098
37

ALMA-7B

NaNK
llama
850
25

ALMA-7B-Pretrain

NaNK
llama
806
4

X-ALMA-13B-Group4

NaNK
llama
149
0

X-ALMA-13B-Group1

NaNK
llama
107
2

X-ALMA-13B-Group2

NaNK
llama
99
1

X-ALMA-13B-Group6

NaNK
llama
91
4

X-ALMA-13B-Group5

NaNK
llama
71
1

X-ALMA-13B-Group8

NaNK
llama
44
1

X-ALMA-13B-Group7

NaNK
llama
42
0

X-ALMA-13B-Group3

NaNK
llama
29
0

X ALMA

X-ALMA builds upon ALMA-R by expanding support from 6 to 50 languages. It utilizes a plug-and-play architecture with language-specific modules, complemented by a carefully designed training recipe. This release includes the the complete X-ALMA model that contains the X-ALMA pre-trained base model and all its language-specific modules. X-ALMA supports 50 languages: en,da,nl,de,is,no,sv,af,ca,ro,gl,it,pt,es,bg,mk,sr,uk,ru,id,ms,th,vi,mg,fr,hu,el,cs,pl,lt,lv,ka,zh,ja,ko,fi,et,gu,hi,mr,ne,ur,az,kk,ky,tr,uz,ar,he,fa, ensuring their high-performance in translation, regardless of thir resource level. All X-ALMA checkpoints are released at huggingface: | Models | Model Link | Description | |:-------------:|:---------------:|:---------------:| | X-ALMA | haoranxu/X-ALMA) | X-ALMA model with all its modules | | X-ALMA-13B-Pretrain | haoranxu/X-ALMA-13B-Pretrain | X-ALMA 13B multilingual pre-trained base model | | X-ALMA-Group1 | haoranxu/X-ALMA-13B-Group1 | X-ALMA group1 specific module and the merged model | | X-ALMA-Group2 | haoranxu/X-ALMA-13B-Group2 | X-ALMA group2 specific module and the merged model | | X-ALMA-Group3 | haoranxu/X-ALMA-13B-Group3 | X-ALMA group3 specific module and the merged model | | X-ALMA-Group4 | haoranxu/X-ALMA-13B-Group4 | X-ALMA group4 specific module and the merged model | | X-ALMA-Group5 | haoranxu/X-ALMA-13B-Group5 | X-ALMA group5 specific module and the merged model | | X-ALMA-Group6 | haoranxu/X-ALMA-13B-Group6 | X-ALMA group6 specific module and the merged model | | X-ALMA-Group7 | haoranxu/X-ALMA-13B-Group7 | X-ALMA group7 specific module and the merged model | | X-ALMA-Group8 | haoranxu/X-ALMA-13B-Group8 | X-ALMA group8 specific module and the merged model | A quick start: There are three ways to load X-ALMA for translation. An example of translating "我爱机器翻译。" into English (X-ALMA should also able to do multilingual open-ended QA). The first way: loading the merged model where the language-specific module has been merged into the base model (Recommended): The second way: loading the base model and language-specific module (Recommended): The third way: loading the base model with all language-specific modules like MoE: (Require large GPU memory)

NaNK
license:mit
17
19

Llama-3-Instruct-8B-CPO

NaNK
llama
1
1

ALMA-13B-Pretrain-LoRA

NaNK
license:mit
0
7

ALMA-7B-Pretrain-LoRA

NaNK
license:mit
0
3

Mistral7B-ALMA-Pretrain

NaNK
license:mit
0
1