vonjack

19 models • 7 total models in database
Sort by:

whisper-large-v3-gguf

license:apache-2.0
1,120
17

bge-m3-gguf

license:mit
727
18

Qwen-LLaMAfied-HFTok-7B-Chat

NaNK
llama
716
24

MobileLLM-125M-HF

llama
357
1

granite-docling-258M-gguf

license:apache-2.0
120
1

gemma2-2b-merged

This is a merge of pre-trained language models created using mergekit. This model was merged using the TIES merge method using google/gemma-2-2b as a base. The following models were included in the merge: google/gemma-2-2b-it The following YAML configuration was used to produce this model:

NaNK
41
0

hyper-sd-v1_5-segmind-vegart-gguf

license:apache-2.0
19
3

Phi-3.5-mini-instruct-GGUF

license:apache-2.0
12
0

SmolLM2-1.7B-Merged

This is a merge of pre-trained language models created using mergekit. This model was merged using the TIES merge method using HuggingFaceTB/SmolLM2-1.7B as a base. The following models were included in the merge: HuggingFaceTB/SmolLM2-1.7B-Instruct The following YAML configuration was used to produce this model:

NaNK
llama
12
0

Qwen2.5-Coder-1.5B-Merged

This is a merge of pre-trained language models created using mergekit. This model was merged using the TIES merge method using Qwen/Qwen2.5-Coder-1.5B as a base. The following models were included in the merge: Qwen/Qwen2.5-Coder-1.5B-Instruct The following YAML configuration was used to produce this model:

NaNK
10
1

SmolLM2-135M-Merged

NaNK
llama
10
0

SmolLM2-360M-Merged

NaNK
llama
10
0

Phi-3.5-mini-instruct-hermes-fc-json

Phi-3.5-mini is a lightweight, state-of-the-art open model built upon datasets used for Phi-3 - synthetic data and filtered publicly available websites - with a focus on very high-quality, reasoning dense data. The model belongs to the Phi-3 model family and supports 128K token context length. The model underwent a rigorous enhancement process, incorporating both supervised fine-tuning, proximal policy optimization, and direct preference optimization to ensure precise instruction adherence and robust safety measures. This dataset is the compilation of structured output and function calling data used in the Hermes 2 Pro series of models. This repository contains a structured output dataset with function-calling conversations, json-mode, agentic json-mode and structured extraction samples, designed to train LLM models in performing function calls and returning structured output based on natural language instructions. The dataset features various conversational scenarios where AI agents are required to interpret queries and execute appropriate single or multiple function calls. The synthetic data generation was led by @interstellarninja in collaboration with @NousResearch, @teknium, @THEODOROS and many others who provided guidance. Hermes Function-calling Standard enables creation of LLM agents that are capable of executing API calls directly from user instructions. For instance, when asked to "find a flight from New York to Los Angeles for next Friday," a function-calling agent can interpret the request, generate the necessary function call (e.g., `searchflights`), and return the results. These agents significantly enhance the utility of AI by enabling direct interactions with APIs, making them invaluable in digital assistants across various domains. For a complete useage guide of models trained on this data, see our github repo: https://github.com/NousResearch/Hermes-Function-Calling

NaNK
license:apache-2.0
8
1

Qwen2.5-Coder-0.5B-Merged

This is a merge of pre-trained language models created using mergekit. This model was merged using the TIES merge method using Qwen/Qwen2.5-Coder-0.5B as a base. The following models were included in the merge: Qwen/Qwen2.5-Coder-0.5B-Instruct The following YAML configuration was used to produce this model:

NaNK
4
0

opus-mt-mul-en-big

mit
3
3

Sensei-7B-V2-gguf

NaNK
license:apache-2.0
3
0

Phi-3-mini-4k-instruct-LLaMAfied

llama
2
11

Hermes-2-Pro-BakLLaVA-Mistral-7B

NaNK
license:apache-2.0
1
15

Nous-Hermes-2-Pro-Xtuner-LLaVA-v1_1-Llama-3-8B

NaNK
llama
1
2