olabs-ai
TFFT-20241101_213900-Llama-3.2-1B
qLeap_instruct_v02
- Developed by: olabs-ai - License: apache-2.0 - Finetuned from model : unsloth/Llama-3.2-1B-bnb-4bit This llama model was trained 2x faster with Unsloth and Huggingface's TRL library.
unsloth-Llama-3.2-1B-Instruct-bnb-4bit-GGUF
qLeap_v06_instruct
qLeap_base_v01
unsloth-cpt-hindi-v01
qLeap_v05_instruct
TFFT-20241101_221234-Llama-3.2-1B
qLeap_model_v0_8bit_Q8_1730963323
qLeap_v04
- Developed by: olabs-ai - License: apache-2.0 - Finetuned from model : unsloth/Llama-3.2-1B-bnb-4bit This llama model was trained 2x faster with Unsloth and Huggingface's TRL library.
qLeap_v07_instruct
qLeap_instruct_v04
- Developed by: olabs-ai - License: apache-2.0 - Finetuned from model : unsloth/Llama-3.2-1B-bnb-4bit This llama model was trained 2x faster with Unsloth and Huggingface's TRL library.
TFFT-20241102_123621-Llama-3.2-1B-Instruct
qLeap_model_v0_16bit_GGUF_1730963323
unsloth-Llama-3.2-1B-bnb-4bit
qLeap_v04_instruct
- Developed by: olabs-ai - License: apache-2.0 - Finetuned from model : unsloth/Llama-3.2-1B-bnb-4bit This llama model was trained 2x faster with Unsloth and Huggingface's TRL library.
qLeap_base_v02
qLeap_model_v0_q4_k_m_16bit
qLeap_model_v0_q5_k_m_16bit
rohitx11
qLeap_instruct_v3
reflection_model
--- language: en tags: - text-generation - causal-lm - fine-tuning - unsupervised --- The `olabs-ai/reflectionmodel` is a fine-tuned language model based on Meta-Llama-3.1-8B-Instruct. It has been further fine-tuned using LoRA (Low-Rank Adaptation) for improved performance in specific tasks. This model is designed for text generation and can be used for various applications like conversational agents, content creation, and more. - Base Model: Meta-Llama-3.1-8B-Instruct - Fine-Tuning Method: LoRA - Architecture: LlamaForCausalLM - Number of Parameters: 8 Billion (Base Model) - Training Data: [Details about the training data used for fine-tuning, if available] To use this model, you need to have the `transformers` and `unsloth` libraries installed. You can load the model and tokenizer as follows: ```python from transformers import AutoConfig, AutoModel, AutoTokenizer from unsloth import FastLanguageModel Load base model configuration basemodelname = "olabs-ai/Meta-Llama-3.1-8B-Instruct" baseconfig = AutoConfig.frompretrained(basemodelname) basemodel = AutoModel.frompretrained(basemodelname, config=baseconfig) tokenizer = AutoTokenizer.frompretrained(basemodelname) Load LoRA adapter adapterconfigpath = "pathtoyouradapterconfig.json" adapterweightspath = "pathtoyouradapterweights" Use FastLanguageModel to apply LoRA adapter model = FastLanguageModel.frompretrained( modelname=basemodelname, adapterweights=adapterweightspath, config=adapterconfigpath ) Set inference mode for LoRA FastLanguageModel.forinference(model) Prepare inputs customprompt = "What is a famous tall tower in Paris?" inputs = tokenizer([customprompt], returntensors="pt").to("cuda") from transformers import TextStreamer textstreamer = TextStreamer(tokenizer) Generate outputs outputs = model.generate(inputs, streamer=textstreamer, maxnewtokens=1000)