Lyte

9 models • 1 total models in database
Sort by:

Gemma-3-1B-Moroccan-Instruct

- Developed by: Lyte - License: Apache-2.0 - Base model: `unsloth/gemma-3-1b-it-unsloth-bnb-4bit` - Dataset: `Lyte/Moroccan-QA-Extended` (with additional English Questions -> Moroccan Darija Answers) - Language: Moroccan Arabic (Darija) - Q80: gemma-3-1b-moroccan-instruct-q80.gguf - Q4KM: gemma-3-1b-moroccan-instruct-q4km.gguf Here is an example of the model's output in LM Studio, answering a question about Newton's law of universal gravitation in Moroccan Darija. - Max Length: 1024 tokens - Epochs: 3 - Total Steps: 843 - Batch size: 2 (per device) - Gradient Accumulation: 4 (Total effective batch size: 16) - Learning rate: 2e-4 - Optimizer: 8-bit AdamW - Scheduler: Linear - Weight decay: 0.01 - Seed: 3407 - Num of Examples: 4,495 - Trainable Parameters: 52.18M (4.96%) - Training Time: ~1 hour on a single GPU. This was the first test finetune run, not a final production model. Training was done using Unsloth for speedup and Hugging Face TRL for supervised finetuning. - Training Loss: From 2.171600 to 0.9392 (at final step 843) - Evaluation Loss: From 2.198849 to 1.5074 (at final step 800) Training converged without issues. The loss metrics show expected early-stage improvement, but this checkpoint is experimental and requires further tuning and validation before use. - Experimental model — not yet optimized or fully Moroccan-Darija-aligned. - Performance outside Moroccan Arabic QA tasks may be limited. - Further finetuning and evaluation are needed before production use. - Developed by: Lyte - License: apache-2.0 - Finetuned from model : unsloth/gemma-3-1b-it-unsloth-bnb-4bit This gemma3text model was trained 2x faster with Unsloth and Huggingface's TRL library.

NaNK
license:apache-2.0
348
1

Nanochat-Moroccan-Instruct-0.7B

NaNK
license:cc-by-sa-4.0
95
1

Llama-3.2-3B-Overthinker

Language model designed for English with an Apache 2.0 license.

NaNK
llama
21
21

QuadConnect2.5-1.5B-v0.1.0b

NaNK
12
1

QuadConnect2.5-0.5B-v0.0.9b

NaNK
10
0

Whisper-Small-Darija

- Developed by: Lyte - License: apache-2.0 - Finetuned from model : unsloth/whisper-small This whisper model was trained 2x faster with Unsloth and Huggingface's TRL library.

license:apache-2.0
7
0

QuadConnect2.5-0.5B-v0.1.1b

NaNK
5
0

Whisper-Tiny-Darija

3
0

whisper-base-darija

2
0