occiglot

10 models • 1 total models in database
Sort by:

occiglot-7b-it-en-instruct

NaNK
license:apache-2.0
628
5

occiglot-7b-fr-en

NaNK
license:apache-2.0
172
3

occiglot-7b-eu5-instruct

NaNK
license:apache-2.0
112
10

occiglot-7b-de-en-instruct

NaNK
license:apache-2.0
61
24

occiglot-7b-eu5

NaNK
license:apache-2.0
50
27

occiglot-7b-es-en-instruct

Occiglot-7B-ES-EN-Instruct is a the instruct version of occiglot-7b-es-en, a generative language model with 7B parameters supporting the Spanish and English and trained by the Occiglot Research Collective. It was trained on 160M tokens of additional multilingual and code instructions. Note that the model was not safety aligned and might generate problematic outputs. This is the first release of an ongoing open research project for multilingual language models. If you want to train a model for your own language or are working on evaluations, please contact us or join our Discord server. We are open for collaborations! - Instruction tuned from: occiglot-7b-es-en - Model type: Causal decoder-only transformer language model - Languages: English, Spanish, and code. - License: Apache 2.0 - Compute resources: DFKI cluster - Contributors: Manuel Brack, Patrick Schramowski, Pedro Ortiz, Malte Ostendorff, Fabio Barth, Georg Rehm, Kristian Kersting - Research labs: Occiglot with support from SAINT and SLT - Contact: Discord The model was trained using the chatml instruction template. You can use the transformers chat template feature for interaction. Since the generation relies on some randomness, we set a seed for reproducibility: The training data was split evenly amongst Spanish and English based on the total number of tokens. Spanish - Mentor-ES - Squad-es - OASST-2 (Spanish subset) - Aya-Dataset (Spanish subset) - Full instruction fine-tuning on 8xH100. - 0.6 - 4 training epochs (depending on dataset sampling). - Framework: axolotl - Precision: bf16 - Optimizer: AdamW - Global batch size: 128 (with 8192 context length) - Cosine Annealing with Warmup Preliminary evaluation results can be found below. Please note that the non-English results are based on partially machine-translated datasets and English prompts (Belebele and Okapi framework) and thus should be interpreted with caution, e.g., biased towards English model performance. Currently, we are working on more suitable benchmarks for Spanish, French, German, and Italian. | | avg | arcchallenge | belebele | hellaswag | mmlu | truthfulqa | |:---------------------------|---------:|----------------:|-----------:|------------:|---------:|-------------:| | Occiglot-7b-eu5 | 0.516895 | 0.508109 | 0.675556 | 0.718963 | 0.402064 | 0.279782 | | Occiglot-7b-eu5-instruct | 0.537799 | 0.53632 | 0.691111 | 0.731918 | 0.405198 | 0.32445 | | Occiglot-7b-es-en | 0.483388 | 0.482949 | 0.606889 | 0.653902 | 0.398922 | 0.274277 | | Occiglot-7b-es-en-instruct | 0.504023 | 0.494576 | 0.65 | 0.670847 | 0.406176 | 0.298513 | | Lince-mistral-7b-it-es | 0.543427 | 0.540222 | 0.745111 | 0.692931 | 0.426241 | 0.312629 | | Mistral-7b-v0.1 | 0.547111 | 0.528937 | 0.768444 | 0.682516 | 0.448253 | 0.307403 | | Mistral-7b-instruct-v0.2 | 0.56713 | 0.547228 | 0.741111 | 0.69455 | 0.422501 | 0.430262 | | | avg | arcchallenge | belebele | hellaswag | mmlu | truthfulqa | |:---------------------------|---------:|----------------:|-----------:|------------:|---------:|-------------:| | Occiglot-7b-eu5 | 0.59657 | 0.530717 | 0.726667 | 0.789882 | 0.531904 | 0.403678 | | Occiglot-7b-eu5-instruct | 0.617905 | 0.558874 | 0.746667 | 0.799841 | 0.535109 | 0.449 | | Occiglot-7b-es-en | 0.593609 | 0.543515 | 0.697778 | 0.788289 | 0.548355 | 0.390109 | | Occiglot-7b-es-en-instruct | 0.615707 | 0.552048 | 0.736667 | 0.797451 | 0.557328 | 0.435042 | | Leo-mistral-hessianai-7b | 0.600949 | 0.522184 | 0.736667 | 0.777833 | 0.538812 | 0.429248 | | Mistral-7b-v0.1 | 0.668385 | 0.612628 | 0.844444 | 0.834097 | 0.624555 | 0.426201 | | Mistral-7b-instruct-v0.2 | 0.713657 | 0.637372 | 0.824444 | 0.846345 | 0.59201 | 0.668116 | | | avg | arcchallengees | belebelees | hellaswages | mmlues | truthfulqaes | |:---------------------------|---------:|-------------------:|--------------:|---------------:|----------:|----------------:| | Occiglot-7b-eu5 | 0.533194 | 0.508547 | 0.676667 | 0.725411 | 0.499325 | 0.25602 | | Occiglot-7b-eu5-instruct | 0.548155 | 0.535043 | 0.68 | 0.737039 | 0.503525 | 0.285171 | | Occiglot-7b-es-en | 0.527264 | 0.529915 | 0.627778 | 0.72253 | 0.512749 | 0.243346 | | Occiglot-7b-es-en-instruct | 0.5396 | 0.545299 | 0.636667 | 0.734372 | 0.524374 | 0.257288 | | Lince-mistral-7b-it-es | 0.547212 | 0.52906 | 0.721111 | 0.687967 | 0.512749 | 0.285171 | | Mistral-7b-v0.1 | 0.554817 | 0.528205 | 0.747778 | 0.672712 | 0.544023 | 0.281369 | | Mistral-7b-instruct-v0.2 | 0.568575 | 0.54188 | 0.73 | 0.685406 | 0.511699 | 0.373891 | The pre-trained model training was supported by a compute grant at the 42 supercomputer which is a central component in the development of hessian AI, the AI Innovation Lab (funded by the Hessian Ministry of Higher Education, Research and the Art (HMWK) & the Hessian Ministry of the Interior, for Security and Homeland Security (HMinD)) and the AI Service Centers (funded by the German Federal Ministry for Economic Affairs and Climate Action (BMWK)). The curation of the training data is partially funded by the German Federal Ministry for Economic Affairs and Climate Action (BMWK) through the project OpenGPT-X (project no. 68GX21007D). - https://huggingface.co/collections/occiglot/occiglot-eu5-7b-v01-65dbed502a6348b052695e01 - https://huggingface.co/NikolayKozloff/occiglot-7b-es-en-GGUF Open LLM Leaderboard Evaluation Results Detailed results can be found here | Metric |Value| |-------------------|----:| |Avg. |12.37| |IFEval (0-Shot) |34.85| |BBH (3-Shot) |17.24| |MATH Lvl 5 (4-Shot)| 1.89| |GPQA (0-shot) | 1.23| |MuSR (0-shot) | 4.45| |MMLU-PRO (5-shot) |14.56|

NaNK
license:apache-2.0
6
2

occiglot-7b-it-en

NaNK
license:apache-2.0
5
5

occiglot-7b-fr-en-instruct

NaNK
license:apache-2.0
5
3

occiglot-7b-es-en

NaNK
license:apache-2.0
1
4

occiglot-7b-de-en

NaNK
license:apache-2.0
0
7