ltg
gpt-bert-babylm-base
norbert4-base
norbert3-xs
norbert4-xsmall
norbert3-large
norbert4-large
norbert3-small
norbert4-xlarge
norbert4-small
norbert3-base
Deberta Xxlarge Fixed
This is deberta-v2-xxlarge updated to implement the `AutoModelForCausalLM` class, enabling it to generate text. This implementation is based on our paper "BERTs are Generative In-Context Learners". This repository also fixes three bugs in the original HF implementation of DeBERTa: 1. We fixed the incorrect name of the output embedding weights in the checkpoint file; 2. We fixed the implementation of the enhanced mask decoder (EMD), based on the original GitHub repository; 3. We clamp the positional embeddings so that they work with long sequence lengths. If you find DeBERTa useful for your work, please cite the following paper:
norbert3-fine-absa
norbert2
norbert3-base_sentence-sentiment
norbert
ltg-bert-bnc
nort5-base-en-no-translation
norbert3-large_sentence-sentiment
norbert3-coarse-absa
flan-t5-definition-en-xl
mt0-definition-ru-xl-axolotl24st
nort5-base
gpt-bert-babylm-small
norbert3-fine-absa-full
nort5-large
aya-definition-fi-axolotl24st
This model is a version of CohereLabs/aya-101, fine-tuned on datasets of Finnish usage examples and definitions. It generates definitions of Finnish words in context. Its input is the usage example and the instruction question ". Mitä tarkoittaa \ ?" - Github repository: MultilingualDefGen - Paper: EMNLP 2025 Findings The model is intended for research purposes, as a source of contextualized dictionary-like lexical definitions. The fine-tuning datasets were limited to Finnish. Although the original model is multilingual, we did not evaluate its ability to generate definitions in other languages. Generated definitions can contain all sorts of biases and stereotypes, stemming from the underlying language model and raw dictionary data.
aya-definition-fi-axolotl24st_dbnary
This model is a version of CohereLabs/aya-101, fine-tuned on datasets of Finnish usage examples and definitions. It generates definitions of Finnish words in context. Its input is the usage example and the instruction question ". Mitä tarkoittaa \ ?" - Github repository: MultilingualDefGen - Paper: EMNLP 2025 Findings The model is intended for research purposes, as a source of contextualized dictionary-like lexical definitions. The fine-tuning datasets were limited to Finnish. Although the original model is multilingual, we did not evaluate its ability to generate definitions in other languages. Generated definitions can contain all sorts of biases and stereotypes, stemming from the underlying language model and raw dictionary data.
nort5-xs
mt0-definition-de-xl-dbnary
This model is a version of bigscience/mt0-xl, fine-tuned on datasets of German usage examples and definitions. It generates definitions of German words in context. Its input is the usage example and the instruction question ". Was ist die Definition von \ ?" - Github repository: MultilingualDefGen - Paper: EMNLP 2025 Findings The model is intended for research purposes, as a source of contextualized dictionary-like lexical definitions. The fine-tuning datasets were limited to German. Although the original model is multilingual, we did not evaluate its ability to generate definitions in other languages. Generated definitions can contain all sorts of biases and stereotypes, stemming from the underlying language model and raw dictionary data.
norbert3-coarse-absa-full
nort5-small
aya-definition-ru-axolotl24st
mt0-definition-fi-xl-axolotl24st
mt0-definition-ru-xl-axolotl24st_dbnary
This model is a version of bigscience/mt0-xl, fine-tuned on datasets of Russian usage examples and definitions. It generates definitions of Russian words in context. Its input is the usage example and the instruction question "Что такое ?" - Github repository: MultilingualDefGen - Paper: EMNLP 2025 Findings The model is intended for research purposes, as a source of contextualized dictionary-like lexical definitions. The fine-tuning datasets were limited to Russian. Although the original model is multilingual, we did not evaluate its ability to generate definitions in other languages. Generated definitions can contain all sorts of biases and stereotypes, stemming from the underlying language model and raw dictionary data.
tower-definition-ru-axolotl24st_dbnary
This model is a version of Unbabel/TowerInstruct-7B-v0.2, fine-tuned on datasets of Russian usage examples and definitions. It generates definitions of Russian words in context. Its input is the usage example and the instruction question "Что такое \ ?" - Github repository: MultilingualDefGen - Paper: EMNLP 2025 Findings The model is intended for research purposes, as a source of contextualized dictionary-like lexical definitions. The fine-tuning datasets were limited to Russian. Although the original model is multilingual, we did not evaluate its ability to generate definitions in other languages. Generated definitions can contain all sorts of biases and stereotypes, stemming from the underlying language model and raw dictionary data.
tower-definition-de-dbnary
This model is a version of Unbabel/TowerInstruct-7B-v0.2, fine-tuned on datasets of German usage examples and definitions. It generates definitions of German words in context. Its input is the usage example and the instruction question ". Was ist die Definition von \ ?" - Github repository: MultilingualDefGen - Paper: EMNLP 2025 Findings The model is intended for research purposes, as a source of contextualized dictionary-like lexical definitions. The fine-tuning datasets were limited to German. Although the original model is multilingual, we did not evaluate its ability to generate definitions in other languages. Generated definitions can contain all sorts of biases and stereotypes, stemming from the underlying language model and raw dictionary data.
tower-definition-ru-axolotl24st
This model is a version of Unbabel/TowerInstruct-7B-v0.2, fine-tuned on datasets of Russian usage examples and definitions. It generates definitions of Russian words in context. Its input is the usage example and the instruction question "Что такое ?" - Github repository: MultilingualDefGen - Paper: EMNLP 2025 Findings The model is intended for research purposes, as a source of contextualized dictionary-like lexical definitions. The fine-tuning datasets were limited to Russian. Although the original model is multilingual, we did not evaluate its ability to generate definitions in other languages. Generated definitions can contain all sorts of biases and stereotypes, stemming from the underlying language model and raw dictionary data.
aya-definition-ru-axolotl24st_dbnary
flan-t5-definition-en-base
aya-definition-de-dbnary
mt0-definition-fi-xl-axolotl24st_dbnary
tower-definition-fi-axolotl24st_dbnary
This model is a version of Unbabel/TowerInstruct-7B-v0.2, fine-tuned on datasets of Finnish usage examples and definitions. It generates definitions of Finnish words in context. Its input is the usage example and the instruction question ". Mitä tarkoittaa ?" - Github repository: MultilingualDefGen - Paper: EMNLP 2025 Findings The model is intended for research purposes, as a source of contextualized dictionary-like lexical definitions. The fine-tuning datasets were limited to Finnish. Although the original model is multilingual, we did not evaluate its ability to generate definitions in other languages. Generated definitions can contain all sorts of biases and stereotypes, stemming from the underlying language model and raw dictionary data.
tower-definition-fi-axolotl24st
This model is a version of Unbabel/TowerInstruct-7B-v0.2, fine-tuned on datasets of Finnish usage examples and definitions. It generates definitions of Finnish words in context. Its input is the usage example and the instruction question ". Mitä tarkoittaa ?" - Github repository: MultilingualDefGen - Paper: EMNLP 2025 Findings The model is intended for research purposes, as a source of contextualized dictionary-like lexical definitions. The fine-tuning datasets were limited to Finnish. Although the original model is multilingual, we did not evaluate its ability to generate definitions in other languages. Generated definitions can contain all sorts of biases and stereotypes, stemming from the underlying language model and raw dictionary data.