unicamp-dl

37 models • 1 total models in database
Sort by:

ptt5-base-portuguese-vocab

Introduction PTT5 is a T5 model pretrained in the BrWac corpus, a large collection of web pages in Portuguese, improving T5's performance on Portuguese sentence similarity and entailment tasks. It's available in three sizes (small, base and large) and two vocabularies (Google's T5 original and ours, trained on Portuguese Wikipedia). For further information or requests, please go to PTT5 repository. Available models | Model | Size | #Params | Vocabulary | | :-: | :-: | :-: | :-: | | unicamp-dl/ptt5-small-t5-vocab | small | 60M | Google's T5 | | unicamp-dl/ptt5-base-t5-vocab | base | 220M | Google's T5 | | unicamp-dl/ptt5-large-t5-vocab | large | 740M | Google's T5 | | unicamp-dl/ptt5-small-portuguese-vocab | small | 60M | Portuguese | | unicamp-dl/ptt5-base-portuguese-vocab (Recommended) | base | 220M | Portuguese | | unicamp-dl/ptt5-large-portuguese-vocab | large | 740M | Portuguese | @article{ptt52020, title={PTT5: Pretraining and validating the T5 model on Brazilian Portuguese data}, author={Carmo, Diedre and Piau, Marcos and Campiotti, Israel and Nogueira, Rodrigo and Lotufo, Roberto}, journal={arXiv preprint arXiv:2008.09144}, year={2020} }

license:mit
9,149
44

translation-en-pt-t5

8,202
21

InRanker-base

804
5

InRanker-small

519
6

translation-pt-en-t5

455
26

ptt5-v2-base

license:apache-2.0
182
2

ptt5-small-portuguese-vocab

license:mit
171
5

mMiniLM-L6-v2-mmarco-v2

license:mit
111
5

mMiniLM-L6-v2-pt-v2

license:mit
90
3

ptt5-v2-3b

NaNK
license:apache-2.0
70
0

mt5-base-mmarco-v2

license:mit
61
4

ptt5-base-en-pt-msmarco-100k-v2

license:mit
55
0

ptt5-large-portuguese-vocab

license:mit
52
10

MMiniLM L6 V2 En Pt Msmarco V2

mMiniLM-L6-v2 Reranker finetuned on mMARCO Introduction mMiniLM-L6-v2-en-pt-msmarco-v2 is a multilingual miniLM-based model finetuned on a bilingual version of MS MARCO passage dataset. This bilingual dataset version is formed by the original MS MARCO dataset (in English) and a Portuguese translated version. In the v2 version, the Portuguese dataset was translated using Google Translate. Further information about the dataset or the translation method can be found on our mMARCO: A Multilingual...

license:mit
39
4

ptt5-base-t5-vocab

license:mit
31
2

mMiniLM-L6-v2-pt-msmarco-v1

license:mit
28
1

ptt5-small-t5-vocab

license:mit
26
0

mMiniLM-L6-v2-en-msmarco

license:mit
24
1

monoptt5-small

NaNK
license:apache-2.0
24
1

ptt5-large-t5-vocab

license:mit
17
2

InRanker-3B

NaNK
11
4

mt5-13b-mmarco-100k

NaNK
license:apache-2.0
10
4

ptt5-v2-small

NaNK
license:apache-2.0
9
1

mt5-base-mmarco-v1

license:mit
8
0

mt5-base-en-msmarco

license:mit
4
0

mt5-base-en-pt-msmarco-v2

license:mit
3
1

mMiniLM-L6-v2-en-pt-msmarco-v1

license:mit
2
1

mMiniLM-L6-v2-mmarco-v1

license:mit
2
1

monoptt5-large

NaNK
license:apache-2.0
2
1

mt5-base-en-pt-msmarco-v1

license:mit
2
0

ptt5-base-pt-msmarco-100k-v1

license:mit
2
0

ptt5-base-pt-msmarco-10k-v2

license:mit
2
0

monoptt5-base

NaNK
license:apache-2.0
2
0

ptt5-base-pt-msmarco-10k-v1

license:mit
1
1

ptt5-v2-large

NaNK
license:apache-2.0
1
1

ptt5-base-pt-msmarco-100k-v2

license:mit
1
0

mt5-3B-mmarco-en-pt

NaNK
license:apache-2.0
0
1