vonjack
whisper-large-v3-gguf
bge-m3-gguf
Qwen-LLaMAfied-HFTok-7B-Chat
MobileLLM-125M-HF
granite-docling-258M-gguf
gemma2-2b-merged
This is a merge of pre-trained language models created using mergekit. This model was merged using the TIES merge method using google/gemma-2-2b as a base. The following models were included in the merge: google/gemma-2-2b-it The following YAML configuration was used to produce this model:
hyper-sd-v1_5-segmind-vegart-gguf
Phi-3.5-mini-instruct-GGUF
SmolLM2-1.7B-Merged
This is a merge of pre-trained language models created using mergekit. This model was merged using the TIES merge method using HuggingFaceTB/SmolLM2-1.7B as a base. The following models were included in the merge: HuggingFaceTB/SmolLM2-1.7B-Instruct The following YAML configuration was used to produce this model:
Qwen2.5-Coder-1.5B-Merged
This is a merge of pre-trained language models created using mergekit. This model was merged using the TIES merge method using Qwen/Qwen2.5-Coder-1.5B as a base. The following models were included in the merge: Qwen/Qwen2.5-Coder-1.5B-Instruct The following YAML configuration was used to produce this model:
SmolLM2-135M-Merged
SmolLM2-360M-Merged
Phi-3.5-mini-instruct-hermes-fc-json
Phi-3.5-mini is a lightweight, state-of-the-art open model built upon datasets used for Phi-3 - synthetic data and filtered publicly available websites - with a focus on very high-quality, reasoning dense data. The model belongs to the Phi-3 model family and supports 128K token context length. The model underwent a rigorous enhancement process, incorporating both supervised fine-tuning, proximal policy optimization, and direct preference optimization to ensure precise instruction adherence and robust safety measures. This dataset is the compilation of structured output and function calling data used in the Hermes 2 Pro series of models. This repository contains a structured output dataset with function-calling conversations, json-mode, agentic json-mode and structured extraction samples, designed to train LLM models in performing function calls and returning structured output based on natural language instructions. The dataset features various conversational scenarios where AI agents are required to interpret queries and execute appropriate single or multiple function calls. The synthetic data generation was led by @interstellarninja in collaboration with @NousResearch, @teknium, @THEODOROS and many others who provided guidance. Hermes Function-calling Standard enables creation of LLM agents that are capable of executing API calls directly from user instructions. For instance, when asked to "find a flight from New York to Los Angeles for next Friday," a function-calling agent can interpret the request, generate the necessary function call (e.g., `searchflights`), and return the results. These agents significantly enhance the utility of AI by enabling direct interactions with APIs, making them invaluable in digital assistants across various domains. For a complete useage guide of models trained on this data, see our github repo: https://github.com/NousResearch/Hermes-Function-Calling
Qwen2.5-Coder-0.5B-Merged
This is a merge of pre-trained language models created using mergekit. This model was merged using the TIES merge method using Qwen/Qwen2.5-Coder-0.5B as a base. The following models were included in the merge: Qwen/Qwen2.5-Coder-0.5B-Instruct The following YAML configuration was used to produce this model: