ltg

61 models • 1 total models in database
Sort by:

gpt-bert-babylm-base

license:mit
17,628
10

norbert4-base

license:apache-2.0
4,529
6

norbert3-xs

license:apache-2.0
1,284
4

norbert4-xsmall

license:apache-2.0
673
2

norbert3-large

license:apache-2.0
396
6

norbert4-large

license:apache-2.0
373
2

norbert3-small

license:apache-2.0
339
2

norbert4-xlarge

license:apache-2.0
317
2

norbert4-small

license:apache-2.0
314
1

norbert3-base

license:apache-2.0
301
7

Deberta Xxlarge Fixed

This is deberta-v2-xxlarge updated to implement the `AutoModelForCausalLM` class, enabling it to generate text. This implementation is based on our paper "BERTs are Generative In-Context Learners". This repository also fixes three bugs in the original HF implementation of DeBERTa: 1. We fixed the incorrect name of the output embedding weights in the checkpoint file; 2. We fixed the implementation of the enhanced mask decoder (EMD), based on the original GitHub repository; 3. We clamp the positional embeddings so that they work with long sequence lengths. If you find DeBERTa useful for your work, please cite the following paper:

license:mit
187
14

norbert3-fine-absa

155
1

norbert2

license:cc-by-4.0
120
7

norbert3-base_sentence-sentiment

license:cc-by-4.0
104
2

norbert

license:cc-by-4.0
98
7

ltg-bert-bnc

license:cc-by-4.0
69
2

nort5-base-en-no-translation

license:cc-by-4.0
61
1

norbert3-large_sentence-sentiment

license:cc-by-4.0
37
1

norbert3-coarse-absa

36
0

flan-t5-definition-en-xl

license:cc-by-sa-4.0
26
1

mt0-definition-ru-xl-axolotl24st

license:cc-by-sa-4.0
23
0

nort5-base

license:apache-2.0
20
2

gpt-bert-babylm-small

license:mit
18
1

norbert3-fine-absa-full

15
0

nort5-large

license:apache-2.0
14
7

aya-definition-fi-axolotl24st

This model is a version of CohereLabs/aya-101, fine-tuned on datasets of Finnish usage examples and definitions. It generates definitions of Finnish words in context. Its input is the usage example and the instruction question ". Mitä tarkoittaa \ ?" - Github repository: MultilingualDefGen - Paper: EMNLP 2025 Findings The model is intended for research purposes, as a source of contextualized dictionary-like lexical definitions. The fine-tuning datasets were limited to Finnish. Although the original model is multilingual, we did not evaluate its ability to generate definitions in other languages. Generated definitions can contain all sorts of biases and stereotypes, stemming from the underlying language model and raw dictionary data.

NaNK
license:cc-by-sa-4.0
14
0

aya-definition-fi-axolotl24st_dbnary

This model is a version of CohereLabs/aya-101, fine-tuned on datasets of Finnish usage examples and definitions. It generates definitions of Finnish words in context. Its input is the usage example and the instruction question ". Mitä tarkoittaa \ ?" - Github repository: MultilingualDefGen - Paper: EMNLP 2025 Findings The model is intended for research purposes, as a source of contextualized dictionary-like lexical definitions. The fine-tuning datasets were limited to Finnish. Although the original model is multilingual, we did not evaluate its ability to generate definitions in other languages. Generated definitions can contain all sorts of biases and stereotypes, stemming from the underlying language model and raw dictionary data.

NaNK
license:cc-by-sa-4.0
14
0

nort5-xs

license:apache-2.0
13
1

mt0-definition-de-xl-dbnary

This model is a version of bigscience/mt0-xl, fine-tuned on datasets of German usage examples and definitions. It generates definitions of German words in context. Its input is the usage example and the instruction question ". Was ist die Definition von \ ?" - Github repository: MultilingualDefGen - Paper: EMNLP 2025 Findings The model is intended for research purposes, as a source of contextualized dictionary-like lexical definitions. The fine-tuning datasets were limited to German. Although the original model is multilingual, we did not evaluate its ability to generate definitions in other languages. Generated definitions can contain all sorts of biases and stereotypes, stemming from the underlying language model and raw dictionary data.

license:cc-by-sa-4.0
13
0

norbert3-coarse-absa-full

12
0

nort5-small

license:apache-2.0
12
0

aya-definition-ru-axolotl24st

NaNK
license:cc-by-sa-4.0
12
0

mt0-definition-fi-xl-axolotl24st

license:cc-by-sa-4.0
11
0

mt0-definition-ru-xl-axolotl24st_dbnary

This model is a version of bigscience/mt0-xl, fine-tuned on datasets of Russian usage examples and definitions. It generates definitions of Russian words in context. Its input is the usage example and the instruction question "Что такое ?" - Github repository: MultilingualDefGen - Paper: EMNLP 2025 Findings The model is intended for research purposes, as a source of contextualized dictionary-like lexical definitions. The fine-tuning datasets were limited to Russian. Although the original model is multilingual, we did not evaluate its ability to generate definitions in other languages. Generated definitions can contain all sorts of biases and stereotypes, stemming from the underlying language model and raw dictionary data.

license:cc-by-sa-4.0
11
0

tower-definition-ru-axolotl24st_dbnary

This model is a version of Unbabel/TowerInstruct-7B-v0.2, fine-tuned on datasets of Russian usage examples and definitions. It generates definitions of Russian words in context. Its input is the usage example and the instruction question "Что такое \ ?" - Github repository: MultilingualDefGen - Paper: EMNLP 2025 Findings The model is intended for research purposes, as a source of contextualized dictionary-like lexical definitions. The fine-tuning datasets were limited to Russian. Although the original model is multilingual, we did not evaluate its ability to generate definitions in other languages. Generated definitions can contain all sorts of biases and stereotypes, stemming from the underlying language model and raw dictionary data.

NaNK
license:cc-by-sa-4.0
11
0

tower-definition-de-dbnary

This model is a version of Unbabel/TowerInstruct-7B-v0.2, fine-tuned on datasets of German usage examples and definitions. It generates definitions of German words in context. Its input is the usage example and the instruction question ". Was ist die Definition von \ ?" - Github repository: MultilingualDefGen - Paper: EMNLP 2025 Findings The model is intended for research purposes, as a source of contextualized dictionary-like lexical definitions. The fine-tuning datasets were limited to German. Although the original model is multilingual, we did not evaluate its ability to generate definitions in other languages. Generated definitions can contain all sorts of biases and stereotypes, stemming from the underlying language model and raw dictionary data.

NaNK
license:cc-by-sa-4.0
10
0

tower-definition-ru-axolotl24st

This model is a version of Unbabel/TowerInstruct-7B-v0.2, fine-tuned on datasets of Russian usage examples and definitions. It generates definitions of Russian words in context. Its input is the usage example and the instruction question "Что такое ?" - Github repository: MultilingualDefGen - Paper: EMNLP 2025 Findings The model is intended for research purposes, as a source of contextualized dictionary-like lexical definitions. The fine-tuning datasets were limited to Russian. Although the original model is multilingual, we did not evaluate its ability to generate definitions in other languages. Generated definitions can contain all sorts of biases and stereotypes, stemming from the underlying language model and raw dictionary data.

NaNK
license:cc-by-sa-4.0
10
0

aya-definition-ru-axolotl24st_dbnary

NaNK
license:cc-by-sa-4.0
8
0

flan-t5-definition-en-base

license:cc-by-sa-4.0
7
0

aya-definition-de-dbnary

NaNK
license:cc-by-sa-4.0
7
0

mt0-definition-fi-xl-axolotl24st_dbnary

license:cc-by-sa-4.0
7
0

tower-definition-fi-axolotl24st_dbnary

This model is a version of Unbabel/TowerInstruct-7B-v0.2, fine-tuned on datasets of Finnish usage examples and definitions. It generates definitions of Finnish words in context. Its input is the usage example and the instruction question ". Mitä tarkoittaa ?" - Github repository: MultilingualDefGen - Paper: EMNLP 2025 Findings The model is intended for research purposes, as a source of contextualized dictionary-like lexical definitions. The fine-tuning datasets were limited to Finnish. Although the original model is multilingual, we did not evaluate its ability to generate definitions in other languages. Generated definitions can contain all sorts of biases and stereotypes, stemming from the underlying language model and raw dictionary data.

NaNK
license:cc-by-sa-4.0
7
0

tower-definition-fi-axolotl24st

This model is a version of Unbabel/TowerInstruct-7B-v0.2, fine-tuned on datasets of Finnish usage examples and definitions. It generates definitions of Finnish words in context. Its input is the usage example and the instruction question ". Mitä tarkoittaa ?" - Github repository: MultilingualDefGen - Paper: EMNLP 2025 Findings The model is intended for research purposes, as a source of contextualized dictionary-like lexical definitions. The fine-tuning datasets were limited to Finnish. Although the original model is multilingual, we did not evaluate its ability to generate definitions in other languages. Generated definitions can contain all sorts of biases and stereotypes, stemming from the underlying language model and raw dictionary data.

NaNK
license:cc-by-sa-4.0
7
0

SLIDE-translation

6
0

flan-t5-definition-en-large

license:cc-by-sa-4.0
5
1

ltg-bert-babylm

license:cc-by-4.0
4
2

norbert3-large_TSA

license:cc-by-4.0
4
1

SLIDE-base

NaNK
license:apache-2.0
4
0

nort5-large-en-no-translation

license:cc-by-4.0
3
4

mt0-definition-ru-xl

license:cc-by-sa-4.0
3
0

bnc-bert-span-0.25x

2
0

bnc-bert-span-2x

2
0

bnc-bert-span-document

2
0

bnc-bert-span-order

2
0

bnc-bert-subword

2
0

bnc-bert-word

2
0

mt0-definition-en-xl

license:cc-by-sa-4.0
2
0

mt0-definition-no-xl

license:cc-by-sa-4.0
2
0

SLIDE-small

NaNK
license:apache-2.0
2
0

SLIDE-x-small

NaNK
license:apache-2.0
1
0

ssa-perin

license:apache-2.0
0
1