prithivMLmods

500 models • 115 total models in database
Sort by:

open-deepfake-detection

license:apache-2.0
92,092
7

Common-Voice-Gender-Detection

license:apache-2.0
58,771
9

Deep-Fake-Detector-v2-Model

The Deep-Fake-Detector-v2-Model is a state-of-the-art deep learning model designed to detect deepfake images. It leverages the Vision Transformer (ViT) architecture, specifically the `google/vit-base-patch16-224-in21k` model, fine-tuned on a dataset of real and deepfake images. The model is trained to classify images as either "Realism" or "Deepfake" with high accuracy, making it a powerful tool for detecting manipulated media. Update : The previous model checkpoint was obtained using a smaller classification dataset. Although it performed well in evaluation scores, its real-time performance was average due to limited variations in the training set. The new update includes a larger dataset to improve the detection of fake images. | Repository | Link | |------------|------| | Deep Fake Detector v2 Model | GitHub Repository | Key Features - Architecture: Vision Transformer (ViT) - `google/vit-base-patch16-224-in21k`. - Input: RGB images resized to 224x224 pixels. - Output: Binary classification ("Realism" or "Deepfake"). - Training Dataset: A curated dataset of real and deepfake images. - Fine-Tuning: The model is fine-tuned using Hugging Face's `Trainer` API with advanced data augmentation techniques. - Performance: Achieves high accuracy and F1 score on validation and test datasets. Model Architecture The model is based on the Vision Transformer (ViT), which treats images as sequences of patches and applies a transformer encoder to learn spatial relationships. Key components include: - Patch Embedding: Divides the input image into fixed-size patches (16x16 pixels). - Transformer Encoder: Processes patch embeddings using multi-head self-attention mechanisms. - Classification Head: A fully connected layer for binary classification. Training Details - Optimizer: AdamW with a learning rate of `1e-6`. - Batch Size: 32 for training, 8 for evaluation. - Epochs: 2. - Data Augmentation: - Random rotation (±90 degrees). - Random sharpness adjustment. - Random resizing and cropping. - Loss Function: Cross-Entropy Loss. - Evaluation Metrics: Accuracy, F1 Score, and Confusion Matrix. Dataset The model is fine-tuned on the dataset, which contains: - Real Images: Authentic images of human faces. - Fake Images: Deepfake images generated using advanced AI techniques. Limitations The model is trained on a specific dataset and may not generalize well to other deepfake datasets or domains. - Performance may degrade on low-resolution or heavily compressed images. - The model is designed for image classification and does not detect deepfake videos directly. Misuse: This model should not be used for malicious purposes, such as creating or spreading deepfakes. Bias: The model may inherit biases from the training dataset. Care should be taken to ensure fairness and inclusivity. Transparency: Users should be informed when deepfake detection tools are used to analyze their content. Future Work - Extend the model to detect deepfake videos. - Improve generalization by training on larger and more diverse datasets. - Incorporate explainability techniques to provide insights into model predictions. ```bibtex @misc{Deep-Fake-Detector-v2-Model, author = {prithivMLmods}, title = {Deep-Fake-Detector-v2-Model}, initial = {21 Mar 2024}, secondupdated = {31 Jan 2025}, latestupdated = {02 Feb 2025} }

NaNK
license:apache-2.0
28,129
17

Castor-3D-Sketchfab-Flux-LoRA

26,220
17

SmolLM2-135M-GGUF

llama
21,639
1

Qwen3-VL-8B-Instruct-abliterated-v1

NaNK
license:apache-2.0
18,982
13

Qwen3-VL-8B-Instruct-abliterated

NaNK
license:apache-2.0
18,982
13

Dog-Breed-120

NaNK
license:apache-2.0
15,542
0

Watermark-Detection-SigLIP2

> Watermark-Detection-SigLIP2 is a vision-language encoder model fine-tuned from google/siglip2-base-patch16-224 for binary image classification. It is trained to detect whether an image contains a watermark or not, using the SiglipForImageClassification architecture. > [!note] > Watermark detection works best with crisp and high-quality images. Noisy images are not recommended for validation. > [!note] SigLIP 2: Multilingual Vision-Language Encoders with Improved Semantic Understanding, Localization, and Dense Features https://arxiv.org/pdf/2502.14786 Watermark-Detection-SigLIP2 is useful in scenarios such as: - Content Moderation – Automatically detect watermarked content on image sharing platforms. - Dataset Cleaning – Filter out watermarked images from training datasets. - Copyright Enforcement – Monitor and flag usage of watermarked media. - Digital Forensics – Support analysis of tampered or protected media assets.

license:apache-2.0
14,248
25

Canopus-LoRA-Flux-FaceRealism

7,203
78

chandra-ocr-2-GGUF

NaNK
6,905
8

deepfake-detector-model-v1

NaNK
license:apache-2.0
5,889
25

Qwen2-VL-OCR-2B-Instruct

> The Qwen2-VL-OCR-2B-Instruct model is a fine-tuned version of Qwen/Qwen2-VL-2B-Instruct, tailored for tasks that involve Optical Character Recognition (OCR), image-to-text conversion, and math problem solving with LaTeX formatting. This model integrates a conversational approach with visual and textual understanding to handle multi-modal tasks effectively. [](https://huggingface.co/prithivMLmods/Qwen2-VL-OCR-2B-Instruct/blob/main/Demo/ocrtestqwen.ipynb) SoTA understanding of images of various resolution & ratio: Qwen2-VL achieves state-of-the-art performance on visual understanding benchmarks, including MathVista, DocVQA, RealWorldQA, MTVQA, etc. Understanding videos of 20min+: Qwen2-VL can understand videos over 20 minutes for high-quality video-based question answering, dialog, content creation, etc. Agent that can operate your mobiles, robots, etc.: with the abilities of complex reasoning and decision making, Qwen2-VL can be integrated with devices like mobile phones, robots, etc., for automatic operation based on visual environment and text instructions. Multilingual Support: to serve global users, besides English and Chinese, Qwen2-VL now supports the understanding of texts in different languages inside images, including most European languages, Japanese, Korean, Arabic, Vietnamese, etc. | File Name | Size | Description | Upload Status | |---------------------------|------------|------------------------------------------------|-------------------| | `.gitattributes` | 1.52 kB | Configures LFS tracking for specific model files. | Initial commit | | `README.md` | 203 Bytes | Minimal details about the uploaded model. | Updated | | `addedtokens.json` | 408 Bytes | Additional tokens used by the model tokenizer. | Uploaded | | `chattemplate.json` | 1.05 kB | Template for chat-based model input/output. | Uploaded | | `config.json` | 1.24 kB | Model configuration metadata. | Uploaded | | `generationconfig.json` | 252 Bytes | Configuration for text generation settings. | Uploaded | | `merges.txt` | 1.82 MB | BPE merge rules for tokenization. | Uploaded | | `model.safetensors` | 4.42 GB | Serialized model weights in a secure format. | Uploaded (LFS) | | `preprocessorconfig.json`| 596 Bytes | Preprocessing configuration for input data. | Uploaded | | `vocab.json` | 2.78 MB | Vocabulary file for tokenization. | Uploaded | 1. Vision-Language Integration: - Combines image understanding with natural language processing to convert images into text. 2. Optical Character Recognition (OCR): - Extracts and processes textual information from images with high accuracy. 3. Math and LaTeX Support: - Solves math problems and outputs equations in LaTeX format. 4. Conversational Capabilities: - Designed to handle multi-turn interactions, providing context-aware responses. 5. Image-Text-to-Text Generation: - Inputs can include images, text, or a combination, and the model generates descriptive or problem-solving text. 6. Secure Weight Format: - Uses Safetensors for faster and more secure model weight loading. - Base Model: Qwen/Qwen2-VL-2B-Instruct - Model Size: - 2.21 Billion parameters - Optimized for BF16 tensor type, enabling efficient inference. - Specializations: - OCR tasks in images containing text. - Mathematical reasoning and LaTeX output for equations.

NaNK
license:apache-2.0
4,138
101

open-scene-detection

> open-scene-detection is a vision-language encoder model fine-tuned from `siglip2-base-patch16-512` for multi-class scene classification. It is trained to recognize and categorize natural and urban scenes using a curated visual dataset. The model uses the `SiglipForImageClassification` architecture. The model classifies an image into one of the following scenes: Scene Recognition – Automatically classify natural and urban scenes. Environmental Mapping – Support geographic and ecological analysis from visual data. Dataset Annotation – Efficiently label large-scale image datasets by scene. Visual Search and Organization – Enable smart scene-based filtering or retrieval. Autonomous Systems – Assist navigation and perception modules with scene understanding.

NaNK
license:apache-2.0
3,562
2

Canopus-Clothing-Flux-LoRA

3,446
29

Camel-Doc-OCR-080125-GGUF

llama.cpp
3,292
1

Digital-Chaos-Flux-LoRA

3,233
9

Flux-Long-Toon-LoRA

3,106
22

Qwen2.5-VL-Abliterated-Caption-GGUF

NaNK
license:apache-2.0
2,794
11

Qwen3-VL-4B-Thinking-abliterated-v1

NaNK
license:apache-2.0
2,690
13

Qwen3-VL-4B-Thinking-abliterated

> Qwen3-VL-4B-Thinking-abliterated is an abliterated (v1.0) variant of Qwen3-VL-4B-Thinking, designed for Abliterated Reasoning and Captioning. This model generates detailed captions and reasoning outputs across a wide range of visual and multimodal contexts, including complex, sensitive, or nuanced content, and supports diverse aspect ratios and resolutions. Abliterated / Uncensored Captioning: Fine-tuned to bypass standard content filters while preserving factual, descriptive, and reasoning-rich outputs. High-Fidelity Descriptions: Produces comprehensive captions and reasoning for general, artistic, technical, abstract, or low-context images. Robust Across Aspect Ratios: Supports wide, tall, square, and irregular image dimensions with consistent accuracy. Variational Detail Control: Generates outputs ranging from high-level summaries to fine-grained, intricate descriptions and reasoning. Foundation on Qwen3-VL-4B-Thinking Architecture: Leverages Qwen3-VL-4B-Thinking’s multimodal reasoning and instruction-following capabilities. Multilingual Output Capability: Primarily English, with adaptability for multilingual prompts via prompt engineering. Generating detailed, uncensored captions and reasoning for general-purpose or artistic datasets. Research in content moderation, red-teaming, and generative safety evaluation. Enabling descriptive captioning and reasoning for visual datasets typically excluded from mainstream models. Creative applications such as storytelling, art generation, or multimodal reasoning tasks. Captioning and reasoning for non-standard aspect ratios and stylized visual content. May produce explicit, sensitive, or offensive descriptions depending on image content and prompts. Not recommended for production systems requiring strict content moderation. Output style, tone, and reasoning can vary depending on input phrasing. Accuracy may vary for unfamiliar, synthetic, or highly abstract visual content.

NaNK
license:apache-2.0
2,690
13

Street-Bokeh-Flux-LoRA

2,677
4

Qwen3-VL-4B-Instruct-abliterated-v1

NaNK
license:apache-2.0
2,631
15

Qwen3-VL-4B-Instruct-abliterated

> Qwen3-VL-4B-Instruct-abliterated is an abliterated (v1.0) variant of Qwen3-VL-4B-Instruct, tailored for Abliterated Reasoning and Captioning. This model is designed to generate detailed and descriptive captions, as well as reasoning outputs, across a wide range of visual and multimodal contexts—including complex, sensitive, or nuanced content—while supporting diverse aspect ratios and resolutions. Abliterated / Uncensored Captioning: Fine-tuned to bypass conventional content filters while preserving factual, descriptive, and reasoning-rich outputs. High-Fidelity Descriptions: Generates comprehensive captions and reasoning for general, artistic, technical, abstract, or low-context images. Robust Across Aspect Ratios: Supports wide, tall, square, and irregular image dimensions with consistent accuracy. Variational Detail Control: Produces outputs ranging from high-level summaries to fine-grained, intricate descriptions and reasoning. Foundation on Qwen3-VL-4B Architecture: Leverages Qwen3-VL-4B’s multimodal reasoning and instruction-following capabilities. Multilingual Output Capability: Primarily English, with adaptability for multilingual prompts via prompt engineering. Generating detailed, uncensored captions and reasoning for general-purpose or artistic datasets. Research in content moderation, red-teaming, and generative safety evaluation. Enabling descriptive captioning and reasoning for visual datasets typically excluded from mainstream models. Creative applications such as storytelling, art generation, or multimodal reasoning tasks. Captioning and reasoning for non-standard aspect ratios and stylized visual content. May produce explicit, sensitive, or offensive descriptions depending on image content and prompts. Not recommended for production systems requiring strict content moderation. Output style, tone, and reasoning can vary depending on input phrasing. Accuracy may vary for unfamiliar, synthetic, or highly abstract visual content.

NaNK
license:apache-2.0
2,631
15

Dark-Thing-Flux-LoRA

2,615
11

Flux-Toonic-2.5D-LoRA

2,409
8

Qwen3-VL-8B-Thinking-abliterated-v1

NaNK
license:apache-2.0
2,289
3

Qwen3-4B-2507-abliterated-GGUF

> The Huihui-Qwen3-4B-Instruct-2507-abliterated model is an uncensored, proof-of-concept version of the Qwen3-4B-Instruct-2507 large language model, created using a novel abliteration method designed to remove refusal responses without using TransformerLens. This approach offers a faster and more effective way to bypass the model's standard refusal behaviors, resulting in a less filtered and more raw output experience, though it lacks rigorous safety filtering and may generate sensitive or controversial content. The model is quantized (4-bit) for efficient use, can be used directly in Hugging Face’s transformers library, and is intended primarily for research or experimental use rather than production due to the reduced content restrictions and associated risks. Users are advised to carefully monitor outputs and ensure ethical and legal compliance when deploying this model. | Model Variant | Link | |--------------|------| | Qwen3-4B-Thinking-2507-abliterated-GGUF | Hugging Face | | Qwen3-4B-Instruct-2507-abliterated-GGUF | Hugging Face | | File Name | Size | Quant Type | |-----------|------|------------| | Qwen3-4B-Thinking-2507-abliterated.BF16.gguf | 8.05 GB | BF16 | | Qwen3-4B-Thinking-2507-abliterated.F16.gguf | 8.05 GB | F16 | | Qwen3-4B-Thinking-2507-abliterated.F32.gguf | 16.1 GB | F32 | | Qwen3-4B-Thinking-2507-abliterated.Q2K.gguf | 1.67 GB | Q2K | | Qwen3-4B-Thinking-2507-abliterated.Q3KL.gguf | 2.24 GB | Q3KL | | Qwen3-4B-Thinking-2507-abliterated.Q3KM.gguf | 2.08 GB | Q3KM | | Qwen3-4B-Thinking-2507-abliterated.Q3KS.gguf | 1.89 GB | Q3KS | | Qwen3-4B-Thinking-2507-abliterated.Q4KM.gguf | 2.5 GB | Q4KM | | Qwen3-4B-Thinking-2507-abliterated.Q4KS.gguf | 2.38 GB | Q4KS | | Qwen3-4B-Thinking-2507-abliterated.Q5KM.gguf | 2.89 GB | Q5KM | | Qwen3-4B-Thinking-2507-abliterated.Q5KS.gguf | 2.82 GB | Q5KS | | Qwen3-4B-Thinking-2507-abliterated.Q6K.gguf | 3.31 GB | Q6K | | Qwen3-4B-Thinking-2507-abliterated.Q80.gguf | 4.28 GB | Q80 | | File Name | Size | Quant Type | |-----------|------|------------| | Qwen3-4B-Instruct-2507-abliterated.BF16.gguf | 8.05 GB | BF16 | | Qwen3-4B-Instruct-2507-abliterated.F16.gguf | 8.05 GB | F16 | | Qwen3-4B-Instruct-2507-abliterated.F32.gguf | 16.1 GB | F32 | | Qwen3-4B-Instruct-2507-abliterated.Q2K.gguf | 1.67 GB | Q2K | | Qwen3-4B-Instruct-2507-abliterated.Q3KL.gguf | 2.24 GB | Q3KL | | Qwen3-4B-Instruct-2507-abliterated.Q3KM.gguf | 2.08 GB | Q3KM | | Qwen3-4B-Instruct-2507-abliterated.Q3KS.gguf | 1.89 GB | Q3KS | | Qwen3-4B-Instruct-2507-abliterated.Q4KM.gguf | 2.5 GB | Q4KM | | Qwen3-4B-Instruct-2507-abliterated.Q4KS.gguf | 2.38 GB | Q4KS | | Qwen3-4B-Instruct-2507-abliterated.Q5KM.gguf | 2.89 GB | Q5KM | | Qwen3-4B-Instruct-2507-abliterated.Q5KS.gguf | 2.82 GB | Q5KS | | Qwen3-4B-Instruct-2507-abliterated.Q6K.gguf | 3.31 GB | Q6K | | Qwen3-4B-Instruct-2507-abliterated.Q80.gguf | 4.28 GB | Q80 | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
license:apache-2.0
2,288
5

Realistic-Gender-Classification

license:apache-2.0
2,274
7

QIE-2509-Object-Remover-Bbox-v3

license:apache-2.0
2,265
9

Shadow-Projection-Flux-LoRA

2,192
14

Canopus-Pixar-3D-Flux-LoRA

2,101
29

Human-vs-NonHuman-Detection

NaNK
license:apache-2.0
2,047
0

3D-Render-Flux-LoRA

2,046
12

Megalodon-OCR-Sync-0713-AIO-GGUF

NaNK
llama.cpp
1,954
1

Gender-Classifier-Mini

NaNK
license:apache-2.0
1,713
2

Nanonets-OCR2-3B-AIO-GGUF

NaNK
llama.cpp
1,530
1

Qwen-Image-Edit-2511-Unblur-Upscale

license:apache-2.0
1,516
43

Llama-SmolTalk-3.2-1B-Instruct

NaNK
llama
1,496
3

Canopus-LoRA-Flux-UltraRealism-2.0

1,493
126

Age-Classification-SigLIP2

> Age-Classification-SigLIP2 is an image classification vision-language encoder model fine-tuned from google/siglip2-base-patch16-224 for a single-label classification task. It is designed to predict the age group of a person from an image using the SiglipForImageClassification architecture. The model categorizes images into five age groups: - Class 0: "Child 0-12" - Class 1: "Teenager 13-20" - Class 2: "Adult 21-44" - Class 3: "Middle Age 45-64" - Class 4: "Aged 65+" The Age-Classification-SigLIP2 model is designed to classify images into five age categories. Potential use cases include: - Demographic Analysis: Helping businesses and researchers analyze age distribution. - Health & Fitness Applications: Assisting in age-based health recommendations. - Security & Access Control: Implementing age verification in digital systems. - Retail & Marketing: Enhancing personalized customer experiences. - Forensics & Surveillance: Aiding in age estimation for security purposes.

NaNK
license:apache-2.0
1,469
5

Qwen3.5-abliterated-MAX-AIO-GGUF

NaNK
llama.cpp
1,452
1

Logo-Design-Flux-LoRA

1,447
40

DeepSeek-OCR-Latest-BF16.I64

> DeepSeek-OCR-Latest-BF16.I64 is an optimized and updated version of the original DeepSeek-OCR. It is an open-source vision-language OCR model designed to extract text from images and scanned documents—including both digital and handwritten content—and can output results as plain text or Markdown. This model leverages a powerful multimodal backbone (3B VLM) to improve reading comprehension and layout understanding for both typed and cursive handwriting. It also excels at preserving document structures such as headings, tables, and lists in its outputs. The BF16 variant has been updated and tested with the following environment: This version allows flexible configuration of attention implementations—such as `flashattention` or `sdpa`—for performance optimization or standardization. Users can also opt out of specific attention implementations if desired. | Resource Type | Description | Link | |----------------|--------------|------| | Original Model Card | Official DeepSeek-OCR release by deepseek-ai | deepseek-ai/DeepSeek-OCR | | Test Model (StrangerZone HF) | Community test deployment (experimental) | strangervisionhf/deepseek-ocr-latest-transformers | | Standard Model Card | Optimized version supporting Transformers v4.57.1 (BF16 precision) | DeepSeek-OCR-Latest-BF16.I64 | | Research Paper | DeepSeek-OCR: Contexts Optical Compression | arXiv:2510.18234 | | Demo Space | Interactive demo hosted on Hugging Face Spaces | DeepSeek-OCR Experimental Demo |

license:mit
1,388
4

Gliese-OCR-7B-Post2.0-final-GGUF

NaNK
llama.cpp
1,377
1

Retro-Pixel-Flux-LoRA

- Hosted Here🧨: https://huggingface.co/spaces/prithivMLmods/FLUX-LoRA-DLC The model is still in the training phase. This is not the final version and may contain artifacts and perform poorly in some cases. | Parameter | Value | Parameter | Value | |---------------------------|--------|---------------------------|--------| | LR Scheduler | constant | Noise Offset | 0.03 | | Optimizer | AdamW | Multires Noise Discount | 0.1 | | Network Dim | 64 | Multires Noise Iterations | 10 | | Network Alpha | 32 | Repeat & Steps | 24 & 2340| | Epoch | 15 | Save Every N Epochs | 1 | You should use `Retro Pixel` to trigger the image generation. Weights for this model are available in Safetensors format.

1,371
75

Nanonets-OCR-s-AIO-GGUF

llama.cpp
1,321
3

Security-Llama3.2-3B-GGUF

> security-llama3.2-3b is a dense, decoder-only Transformer model with approximately 3 billion parameters. It is optimized for generating text, particularly in response to prompts in a chat format, with a context length of up to 4,000 tokens. The model is specialized toward cybersecurity content, drawing on a mixture of publicly available blogs, papers, reference datasets (e.g. from the PEASEC cybersecurity repository), synthetic “textbook-style” data, and academic Q&A sources to enhance performance in security-themed tasks. For usage, the model accepts chat-style inputs (e.g. alternating “user” / “assistant” messages) and can be deployed via the Hugging Face transformers library (e.g. via pipeline("text-generation", model="viettelsecurity-ai/security-llama3.2-3b")). The model weights are stored in safetensors format, configured with fp16 (half precision), and no inference provider currently hosts it by default. | File Name | Quant Type | File Size | | - | - | - | | security-llama3.2-3b.BF16.gguf | BF16 | 6.43 GB | | security-llama3.2-3b.F16.gguf | F16 | 6.43 GB | | security-llama3.2-3b.F32.gguf | F32 | 12.9 GB | | security-llama3.2-3b.Q2K.gguf | Q2K | 1.36 GB | | security-llama3.2-3b.Q3KL.gguf | Q3KL | 1.82 GB | | security-llama3.2-3b.Q3KM.gguf | Q3KM | 1.69 GB | | security-llama3.2-3b.Q3KS.gguf | Q3KS | 1.54 GB | | security-llama3.2-3b.Q40.gguf | Q40 | 1.92 GB | | security-llama3.2-3b.Q41.gguf | Q41 | 2.09 GB | | security-llama3.2-3b.Q4K.gguf | Q4K | 2.02 GB | | security-llama3.2-3b.Q4KM.gguf | Q4KM | 2.02 GB | | security-llama3.2-3b.Q4KS.gguf | Q4KS | 1.93 GB | | security-llama3.2-3b.Q50.gguf | Q50 | 2.27 GB | | security-llama3.2-3b.Q51.gguf | Q51 | 2.45 GB | | security-llama3.2-3b.Q5K.gguf | Q5K | 2.32 GB | | security-llama3.2-3b.Q5KM.gguf | Q5KM | 2.32 GB | | security-llama3.2-3b.Q5KS.gguf | Q5KS | 2.27 GB | | security-llama3.2-3b.Q6K.gguf | Q6K | 2.64 GB | | security-llama3.2-3b.Q80.gguf | Q80 | 3.42 GB | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
llama
1,310
3

SAGE-MM-Qwen3-VL-4B-SFT-GGUF

NaNK
llama.cpp
1,261
1

EBook-Creative-Cover-Flux-LoRA

1,195
22

Qwen-Image-Edit-2511-Polaroid-Photo

license:apache-2.0
1,127
4

Camel-Doc-OCR-080125

license:apache-2.0
1,125
8

Trash-Net

NaNK
license:apache-2.0
1,076
2

Gliese-Qwen3.5-9B-Abliterated-Caption

NaNK
license:apache-2.0
1,054
3

Gliese-OCR-7B-Post2.0-final

> The Gliese-OCR-7B-Post2.0-final model is a refined and optimized version of Gliese-OCR-7B-Post1.0, built upon the Qwen2.5-VL architecture. It represents the final iteration in the Gliese-OCR series, offering enhanced efficiency, precision, and visualization capabilities for document OCR, visual analysis, and information extraction. > > Fine-tuned with extended document visualization data and OCR-focused objectives, this model delivers superior accuracy across a wide range of document types, including scanned PDFs, handwritten pages, structured forms, and analytical reports. Optimized Document Visualization and OCR Pipeline: Significantly improved recognition of text, layout, and embedded visuals for structured document understanding. Context-Aware Multimodal Linking: Enhanced understanding of document context with stronger alignment between text, images, and layout components. Refined Document Retrieval: Improved retrieval accuracy from complex layouts and multi-page documents. High-Fidelity Content Extraction: Precise extraction of structured, semi-structured, and unstructured information with advanced text normalization. Analytical Recognition: Superior reasoning over charts, graphs, tables, and mathematical equations. Improved Visual Reasoning and Layout Awareness: Trained on document visualization datasets for advanced spatial and semantic comprehension. State-of-the-Art Performance Across Resolutions: Achieves top results on benchmarks such as DocVQA, InfographicVQA, MathVista, and RealWorldQA. Extended Multimodal Duration Support: Handles long document sequences and extended videos (20+ minutes). Final Release Stability: Consolidates all prior improvements for stable and reliable performance. Document visualization and OCR extraction tasks. Context-aware document retrieval and multimodal linking. Extraction and LaTeX formatting of equations and structured content. Analytical document interpretation (charts, tables, graphs, and figures). Multilingual OCR for enterprise, academic, and research use cases. Summarization, question answering, and cross-modal reasoning over long documents. Intelligent robotic or mobile automation guided by visual document input. Reduced accuracy on heavily degraded or occluded documents. High computational requirements for large-scale or real-time applications. Limited optimization for low-resource or edge devices. Occasional misalignment in text layout or minor hallucinations in outputs. Performance may vary depending on visual token configuration and context length settings.

NaNK
license:apache-2.0
1,048
6

Qwen2.5-VL-7B-Abliterated-Caption-it

> The Qwen2.5-VL-7B-Abliterated-Caption-it model is a fine-tuned version of Qwen2.5-VL-7B-Instruct, tailored for Abliterated Captioning / Uncensored Image Captioning. This variant is designed to generate highly detailed and descriptive captions across a broad range of visual categories, including images with complex, sensitive, or nuanced content—across varying aspect ratios and resolutions. Abliterated / Uncensored Captioning: Fine-tuned to bypass common content filters while preserving factual and descriptive richness across diverse visual categories. High-Fidelity Descriptions: Generates comprehensive captions for general, artistic, technical, abstract, and low-context images. Robust Across Aspect Ratios: Capable of accurately captioning images with wide, tall, square, and irregular dimensions. Variational Detail Control: Produces outputs with both high-level summaries and fine-grained descriptions as needed. Foundation on Qwen2.5-VL Architecture: Leverages the strengths of the Qwen2.5-VL-7B multimodal model for visual reasoning, comprehension, and instruction-following. Multilingual Output Capability: Can support multilingual descriptions (English as default), adaptable via prompt engineering. This model was fine-tuned using the following datasets: prithivMLmods/blip3o-caption-mini-arrow prithivMLmods/Caption3o-Opt-v2 Private/unlisted datasets curated for uncensored and domain-specific image captioning tasks. The training objective focused on enhancing performance in unconstrained, descriptive image captioning—especially for edge cases commonly filtered out in standard captioning benchmarks. > [!note] Instruction Query: Provide a detailed caption for the image Generating detailed and unfiltered image captions for general-purpose or artistic datasets. Content moderation research, red-teaming, and generative safety evaluations. Enabling descriptive captioning for visual datasets typically excluded from mainstream models. Use in creative applications (e.g., storytelling, art generation) that benefit from rich descriptive captions. Captioning for non-standard aspect ratios and stylized visual content. May produce explicit, sensitive, or offensive descriptions depending on image content and prompts. Not suitable for deployment in production systems requiring content filtering or moderation. Can exhibit variability in caption tone or style depending on input prompt phrasing. Accuracy for unfamiliar or synthetic visual styles may vary.

NaNK
license:apache-2.0
1,010
40

Flux-Product-Ad-Backdrop

995
34

Fashion-Hut-Modeling-LoRA

970
45

GA-Guard-AIO-GGUF

license:apache-2.0
953
1

Food-101-93M

> Food-101-93M is a fine-tuned image classification model built on top of google/siglip2-base-patch16-224 using the SiglipForImageClassification architecture. It is trained to classify food images into one of 101 popular dishes, derived from the Food-101 dataset. The model categorizes images into 101 food classes such as `sushi`, `hamburger`, `waffles`, `padthai`, and more. - Recipe Recommendation Engines: Automatically tagging food images to suggest recipes. - Food Logging & Calorie Tracking Apps: Categorizing meals based on photos. - Smart Kitchens: Assisting food recognition in smart appliances. - Restaurant Menu Digitization: Auto-classifying dishes for visual menus or ordering systems. - Dataset Labeling: Enabling automatic annotation of food datasets for training other ML models.

NaNK
license:apache-2.0
930
11

Ton618-Epic-Realism-Flux-LoRA

927
38

Aura-9999

866
7

Bone-Fracture-Detection

NaNK
license:apache-2.0
860
3

Qwen3-VL-8B-Thinking-Unredacted-MAX-GGUF

NaNK
llama.cpp
851
2

Alphabet-Sign-Language-Detection

NaNK
license:apache-2.0
835
6

IndoorOutdoorNet

NaNK
license:apache-2.0
833
3

Meta-Llama-3.2-1B-GGUF-QX

NaNK
llama
817
2

chandra-OCR-GGUF

llama.cpp
806
1

Flux-GArt-LoRA

768
9

Flux-Dev-Real-Anime-LoRA

751
23

Qwen3-VL-8B-Instruct-abliterated-v2.0-GGUF

NaNK
llama.cpp
748
1

Fire-Detection-Siglip2

NaNK
license:apache-2.0
734
4

Canopus-LoRA-Flux-Anime

NaNK
729
24

Flux.1-Dev-Movie-Boards-LoRA

718
11

Abstract-Cartoon-Flux-LoRA

717
8

Qwen3-VL-8B-Instruct-Unredacted-MAX-GGUF

NaNK
llama.cpp
704
2

Deepfake-Detect-Siglip2

NaNK
license:apache-2.0
676
2

Mockup-Texture-Flux-LoRA

license:apache-2.0
655
11

Gliese-OCR-7B-Post1.0

> The Gliese-OCR-7B-Post1.0 model is a fine-tuned version of Camel-Doc-OCR-062825, optimized for Document Retrieval, Content Extraction, and Analysis Recognition. Built on top of the Qwen2.5-VL architecture, this model enhances document comprehension capabilities with focused training on the Opendoc2-Analysis-Recognition dataset for superior document analysis and information extraction tasks. > [!note] This model shows significant improvements in LaTeX rendering and Markdown rendering for OCR tasks-reportlab/GlieseOCR7BPost10(4bit)reportlab.ipynb). Context-Aware Multimodal Extraction and Linking for Documents: Advanced capability for understanding document context and establishing connections between multimodal elements within documents. Enhanced Document Retrieval: Designed to efficiently locate and extract relevant information from complex document structures and layouts. Superior Content Extraction: Optimized for precise extraction of structured and unstructured content from diverse document formats. Analysis Recognition: Specialized in recognizing and interpreting analytical content, charts, tables, and visual data representations. State-of-the-Art Performance Across Resolutions: Achieves competitive results on OCR and visual QA benchmarks such as DocVQA, MathVista, RealWorldQA, and MTVQA. Video Understanding up to 20+ minutes: Supports detailed comprehension of long-duration videos for content summarization, Q\&A, and multi-modal reasoning. Visually-Grounded Device Interaction: Enables mobile/robotic device operation via visual inputs and text-based instructions using contextual understanding and decision-making logic. Context-aware multimodal extraction and linking for complex document structures. High-fidelity document retrieval and content extraction from various document formats. Analysis recognition of charts, graphs, tables, and visual data representations. Document-based question answering for educational and enterprise applications. Extraction and LaTeX formatting of mathematical expressions from printed or handwritten content. Retrieval and summarization from long documents, slides, and multi-modal inputs. Multilingual document analysis and structured content extraction for global use cases. Robotic or mobile automation with vision-guided contextual interaction. May show degraded performance on extremely low-quality or occluded images. Not optimized for real-time applications on low-resource or edge devices due to computational demands. Variable accuracy on uncommon or low-resource languages/scripts. Long video processing may require substantial memory and is not optimized for streaming applications. Visual token settings affect performance; suboptimal configurations can impact results. In rare cases, outputs may contain hallucinated or contextually misaligned information.

NaNK
license:apache-2.0
633
12

Qwen-Image-Studio-Realism

| Parameter | Value | Parameter | Value | |---------------------------|--------|---------------------------|--------| | LR Scheduler | constant | Noise Offset | 0.03 | | Optimizer | AdamW | Multires Noise Discount | 0.1 | | Network Dim | 64 | Multires Noise Iterations | 10 | | Network Alpha | 32 | Repeat & Steps | 20 & 2790 | | Epoch | 20 | Save Every N Epochs | 1 | | Source | Link | |--------------|-------------------------------------| | Playground | playground.com | | ArtStation | artstation.com | | 4K Wallpapers| 4kwallpapers.com | | Dimensions | Aspect Ratio | Recommendation | |-----------------|------------------|---------------------------| | 1472 x 1140 | 4:3 (approx.) | Best | | 1024 x 1024 | 1:1 | Default | You should use `Studio Realism` to trigger the image generation.

license:apache-2.0
624
20

Facial-Emotion-Detection-SigLIP2

NaNK
license:apache-2.0
619
6

Red-Undersea-Flux-LoRA

597
2

Flux.1-Dev-Realtime-Toon-Mix

578
21

SAGE-MM-Qwen3-VL-4B-SFT_RL-GGUF

NaNK
llama.cpp
572
1

OpenCoder-1.5B-Instruct-GGUF

NaNK
llama-cpp
565
2

Herculis-CUA-GUI-Actioner-4B-GGUF

NaNK
llama.cpp
565
1

Deneb-Qwen3-Radiation-0.6B-GGUF

NaNK
license:apache-2.0
563
0

Flux.1-Dev-Sketch-Card-LoRA

561
13

Weather-Image-Classification

NaNK
license:apache-2.0
550
0

Qwen2-VL-OCR-2B-Instruct-GGUF

NaNK
license:apache-2.0
540
3

palmyra-mini-thinking-AIO-GGUF

license:apache-2.0
536
3

Ton618-Only-Stickers-Flux-LoRA

534
39

II-Search-4B-GGUF

> II-Search-4B is a 4-billion-parameter language model fine-tuned from Qwen3-4B specifically for advanced information seeking and web-integrated reasoning tasks, demonstrating strong capabilities in multi-hop information retrieval, fact verification, and comprehensive report generation; it excels on factual QA benchmarks compared to peers, features sophisticated tool-use for search and web visits, supports distributed inference with vLLM or SGLang (including a 131,072-token context window with custom RoPE scaling), and is suitable for factual question answering, research assistance, and educational applications, with Apple Silicon support via MLX, open integration examples, and full resources available on its Hugging Face repository. | File Name | Size | Quant Type | |-----------|------|------------| | II-Search-4B-GGUF.BF16.gguf | 8.05 GB | BF16 | | II-Search-4B-GGUF.F16.gguf | 8.05 GB | F16 | | II-Search-4B-GGUF.F32.gguf | 16.1 GB | F32 | | II-Search-4B-GGUF.Q2K.gguf | 1.67 GB | Q2K | | II-Search-4B-GGUF.Q3KL.gguf | 2.24 GB | Q3KL | | II-Search-4B-GGUF.Q3KM.gguf | 2.08 GB | Q3KM | | II-Search-4B-GGUF.Q3KS.gguf | 1.89 GB | Q3KS | | II-Search-4B-GGUF.Q4KM.gguf | 2.5 GB | Q4KM | | II-Search-4B-GGUF.Q4KS.gguf | 2.38 GB | Q4KS | | II-Search-4B-GGUF.Q5KM.gguf | 2.89 GB | Q5KM | | II-Search-4B-GGUF.Q5KS.gguf | 2.82 GB | Q5KS | | II-Search-4B-GGUF.Q6K.gguf | 3.31 GB | Q6K | | II-Search-4B-GGUF.Q80.gguf | 4.28 GB | Q80 | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
license:apache-2.0
529
5

Photo Restore I2i

Photo-Restore-i2i is an adapter for black-forest-lab's FLUX.1-Kontext-dev, designed to restore old photos into mid-colorized, detailed images. The model was trained on 50 image pairs (25 start images, 25 end images). Synthetic result nodes were generated using NanoBanana from Google and SeedDream 4 (dataset for result sets), and labeled with DeepCaption-VLA-7B. The adapter is triggered with the following prompt: > [!note] [photo content], restore and enhance the image by repairing any damage, scratches, or fading. Colorize the photo naturally while preserving authentic textures and details, maintaining a realistic and historically accurate look. | Setting | Value | | ------------------------ | ------------------------ | | Module Type | Adapter | | Base Model | FLUX.1 Kontext Dev - fp8 | | Trigger Words | [photo content], restore and enhance the image by repairing any damage, scratches, or fading. Colorize the photo naturally while preserving authentic textures and details, maintaining a realistic and historically accurate look. | | Image Processing Repeats | 50 | | Epochs | 28 | | Save Every N Epochs | 1 | Labeling: DeepCaption-VLA-7B(natural language & English) Total Images Used for Training : 50 Image Pairs (25 Start, 25 End) Synthetic Result Node generated by NanoBanana from Google (Image Result Sets Dataset) | Setting | Value | | --------------------------- | --------- | | Seed | - | | Clip Skip | - | | Text Encoder LR | 0.00001 | | UNet LR | 0.00005 | | LR Scheduler | constant | | Optimizer | AdamW8bit | | Network Dimension | 64 | | Network Alpha | 32 | | Gradient Accumulation Steps | - | | Setting | Value | | --------------- | ----- | | Shuffle Caption | - | | Keep N Tokens | - | | Setting | Value | | ------------------------- | ----- | | Noise Offset | 0.03 | | Multires Noise Discount | 0.1 | | Multires Noise Iterations | 10 | | Conv Dimension | - | | Conv Alpha | - | | Batch Size | - | | Steps | 4100 | | Sampler | euler | You should use `[photo content]` to trigger the image generation. You should use `restore and enhance the image by repairing any damage` to trigger the image generation. You should use `scratches` to trigger the image generation. You should use `or fading. Colorize the photo naturally while preserving authentic textures and details` to trigger the image generation. You should use `maintaining a realistic and historically accurate look.` to trigger the image generation.

527
17

Delorme_1-OCR-7B-Post1.0-GGUF

NaNK
license:apache-2.0
510
1

SD3.5-Large-Photorealistic-LoRA

499
61

Qwen3-VisionCaption-2B-GGUF

NaNK
llama.cpp
496
3

Speech-Emotion-Classification

> Speech-Emotion-Classification is a fine-tuned version of `facebook/wav2vec2-base-960h` for multi-class audio classification, specifically trained to detect emotions in speech. This model utilizes the `Wav2Vec2ForSequenceClassification` architecture to accurately classify speaker emotions from audio signals. > \[!note] > Wav2Vec2: Self-Supervised Learning for Speech Recognition > https://arxiv.org/pdf/2006.11477 Speech Emotion Analytics – Analyze speaker emotions in call centers, interviews, or therapeutic sessions. Conversational AI Personalization – Adjust voice assistant responses based on detected emotion. Mental Health Monitoring – Support emotion recognition in voice-based wellness or teletherapy apps. Voice Dataset Curation – Tag or filter speech datasets by emotion for research or model training. Media Annotation – Automatically annotate podcasts, audiobooks, or videos with speaker emotion metadata.

NaNK
license:apache-2.0
495
2

Qwen3-4B-Valiant-Polaris-f32-GGUF

> ZeroXClem/Qwen-4B-Valiant-Polaris is a thoughtfully merged 4B-parameter language model built upon Qwen3-4B, combining the structured reasoning of Polaris, the creative and expressive capabilities of Dot-Goat and RP-V3, and the scientific depth of ShiningValiant3, resulting in a lightweight yet powerful architecture designed for advanced reasoning, rich roleplay, scientific and analytical tasks, and seamless agentic workflows; with robust support for long contexts, multilingual reasoning, and tool integration, it is ideal for conversational agents, tutoring, problem solving, creative writing, and autonomous agent applications. | File name | Size | Quant type | |-----------|------|------------| | Qwen3-4B-Valiant-Polaris.BF16.gguf | 8.05 GB | BF16 | | Qwen3-4B-Valiant-Polaris.F16.gguf | 8.05 GB | F16 | | Qwen3-4B-Valiant-Polaris.F32.gguf | 16.1 GB | F32 | | Qwen3-4B-Valiant-Polaris.Q2K.gguf | 1.67 GB | Q2K | | Qwen3-4B-Valiant-Polaris.Q3KL.gguf | 2.24 GB | Q3KL | | Qwen3-4B-Valiant-Polaris.Q3KM.gguf | 2.08 GB | Q3KM | | Qwen3-4B-Valiant-Polaris.Q3KS.gguf | 1.89 GB | Q3KS | | Qwen3-4B-Valiant-Polaris.Q4KM.gguf | 2.5 GB | Q4KM | | Qwen3-4B-Valiant-Polaris.Q4KS.gguf | 2.38 GB | Q4KS | | Qwen3-4B-Valiant-Polaris.Q5KM.gguf | 2.89 GB | Q5KM | | Qwen3-4B-Valiant-Polaris.Q5KS.gguf | 2.82 GB | Q5KS | | Qwen3-4B-Valiant-Polaris.Q6K.gguf | 3.31 GB | Q6K | | Qwen3-4B-Valiant-Polaris.Q80.gguf | 4.28 GB | Q80 | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
license:apache-2.0
486
3

Golden-Dust-Flux-LoRA

484
5

Open-R1-Mini-Experimental-GGUF

license:apache-2.0
483
5

jupyter-agent-qwen3-4b-AIO-GGUF

> Jupyter-agent-qwen3-4b-instruct and jupyter-agent-qwen3-4b-thinking are specialized 4B-parameter models built on Qwen3-4B for agentic reasoning and data science tasks in Jupyter notebooks, supporting both general instruction following and step-by-step, notebook-native logical analysis. The instruct variant excels at delivering fast, efficient responses without generating detailed reasoning traces, while the thinking variant provides comprehensive intermediate computations and analysis, including tool calling and dataset-grounded reasoning on realistic Kaggle workflows, making both models state-of-the-art for code execution, data exploration, and practical problem solving in Python and multi-turn notebook environments. | Model Name | Download Link | |------------------------------------------|-----------------------------------------------------------------------------------------------| | jupyter-agent-qwen3-4b-instruct-GGUF | Link | | jupyter-agent-qwen3-4b-thinking-GGUF | Link | | File Name | Quant Type | File Size | | - | - | - | | jupyter-agent-qwen3-4b-instruct.BF16.gguf | BF16 | 8.05 GB | | jupyter-agent-qwen3-4b-instruct.F16.gguf | F16 | 8.05 GB | | jupyter-agent-qwen3-4b-instruct.F32.gguf | F32 | 16.1 GB | | jupyter-agent-qwen3-4b-instruct.Q2K.gguf | Q2K | 1.67 GB | | jupyter-agent-qwen3-4b-instruct.Q3KL.gguf | Q3KL | 2.24 GB | | jupyter-agent-qwen3-4b-instruct.Q3KM.gguf | Q3KM | 2.08 GB | | jupyter-agent-qwen3-4b-instruct.Q3KS.gguf | Q3KS | 1.89 GB | | jupyter-agent-qwen3-4b-instruct.Q4KM.gguf | Q4KM | 2.5 GB | | jupyter-agent-qwen3-4b-instruct.Q4KS.gguf | Q4KS | 2.38 GB | | jupyter-agent-qwen3-4b-instruct.Q5KM.gguf | Q5KM | 2.89 GB | | jupyter-agent-qwen3-4b-instruct.Q5KS.gguf | Q5KS | 2.82 GB | | jupyter-agent-qwen3-4b-instruct.Q6K.gguf | Q6K | 3.31 GB | | jupyter-agent-qwen3-4b-instruct.Q80.gguf | Q80 | 4.28 GB | | File Name | Quant Type | File Size | | - | - | - | | jupyter-agent-qwen3-4b-thinking.BF16.gguf | BF16 | 8.05 GB | | jupyter-agent-qwen3-4b-thinking.F16.gguf | F16 | 8.05 GB | | jupyter-agent-qwen3-4b-thinking.F32.gguf | F32 | 16.1 GB | | jupyter-agent-qwen3-4b-thinking.Q2K.gguf | Q2K | 1.67 GB | | jupyter-agent-qwen3-4b-thinking.Q3KL.gguf | Q3KL | 2.24 GB | | jupyter-agent-qwen3-4b-thinking.Q3KM.gguf | Q3KM | 2.08 GB | | jupyter-agent-qwen3-4b-thinking.Q3KS.gguf | Q3KS | 1.89 GB | | jupyter-agent-qwen3-4b-thinking.Q4KM.gguf | Q4KM | 2.5 GB | | jupyter-agent-qwen3-4b-thinking.Q4KS.gguf | Q4KS | 2.38 GB | | jupyter-agent-qwen3-4b-thinking.Q5KM.gguf | Q5KM | 2.89 GB | | jupyter-agent-qwen3-4b-thinking.Q5KS.gguf | Q5KS | 2.82 GB | | jupyter-agent-qwen3-4b-thinking.Q6K.gguf | Q6K | 3.31 GB | | jupyter-agent-qwen3-4b-thinking.Q80.gguf | Q80 | 4.28 GB | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
license:apache-2.0
475
4

Llama-3.2-1B-Instruct-f32-GGUF

NaNK
llama
471
1

Minimal-Futuristic-Flux-LoRA

470
15

Jan-v1-AIO-GGUF

> Jan-v1-4B is a 4-billion-parameter language model built on the Qwen3-4B-thinking architecture, meticulously fine-tuned for agentic reasoning, problem-solving, and tool utilization with support for web search tasks and large context lengths up to 256,000 tokens. Achieving 91.1% accuracy on the SimpleQA benchmark, Jan-v1-4B excels at factual question answering and conversation while running efficiently on local hardware for enhanced privacy and offline use, making it a strong choice for advanced Q&A, reasoning, and integration with the Jan desktop application or compatible inference engines. Jan-v1-edge is a lightweight agentic model built for fast, reliable on-device execution. As the second release in the Jan Family, it is distilled from the larger Jan-v1 model, preserving strong reasoning and problem-solving ability in a smaller footprint suitable for resource-constrained environments. Jan-v1-edge was developed through a two-phase post-training process. The first phase, Supervised Fine-Tuning (SFT), transferred core capabilities from the Jan-v1 teacher model to the smaller student. The second phase, Reinforcement Learning with Verifiable Rewards (RLVR) —the same method used in Jan-v1 and Lucy—further optimized reasoning efficiency, tool use, and correctness. This staged approach delivers reliable results on complex, interactive workloads. | Model Name | Hugging Face Link | |---------------|-------------------| | Jan-v1-2509-GGUF | 🔗 Link | | Jan-v1-edge-GGUF | 🔗 Link | | Jan-v1-4B-GGUF | 🔗 Link | | File Name | Quant Type | File Size | | - | - | - | | Jan-v1-2509.BF16.gguf | BF16 | 8.05 GB | | Jan-v1-2509.F16.gguf | F16 | 8.05 GB | | Jan-v1-2509.F32.gguf | F32 | 16.1 GB | | Jan-v1-2509.Q2K.gguf | Q2K | 1.67 GB | | Jan-v1-2509.Q3KL.gguf | Q3KL | 2.24 GB | | Jan-v1-2509.Q3KM.gguf | Q3KM | 2.08 GB | | Jan-v1-2509.Q3KS.gguf | Q3KS | 1.89 GB | | Jan-v1-2509.Q4KM.gguf | Q4KM | 2.5 GB | | Jan-v1-2509.Q4KS.gguf | Q4KS | 2.38 GB | | Jan-v1-2509.Q5KM.gguf | Q5KM | 2.89 GB | | Jan-v1-2509.Q5KS.gguf | Q5KS | 2.82 GB | | Jan-v1-2509.Q6K.gguf | Q6K | 3.31 GB | | Jan-v1-2509.Q80.gguf | Q80 | 4.28 GB | | File Name | Quant Type | File Size | | - | - | - | | Jan-v1-edge.BF16.gguf | BF16 | 3.45 GB | | Jan-v1-edge.F16.gguf | F16 | 3.45 GB | | Jan-v1-edge.F32.gguf | F32 | 6.89 GB | | Jan-v1-edge.Q2K.gguf | Q2K | 778 MB | | Jan-v1-edge.Q3KL.gguf | Q3KL | 1 GB | | Jan-v1-edge.Q3KM.gguf | Q3KM | 940 MB | | Jan-v1-edge.Q3KS.gguf | Q3KS | 867 MB | | Jan-v1-edge.Q40.gguf | Q40 | 1.05 GB | | Jan-v1-edge.Q41.gguf | Q41 | 1.14 GB | | Jan-v1-edge.Q4K.gguf | Q4K | 1.11 GB | | Jan-v1-edge.Q4KM.gguf | Q4KM | 1.11 GB | | Jan-v1-edge.Q4KS.gguf | Q4KS | 1.06 GB | | Jan-v1-edge.Q50.gguf | Q50 | 1.23 GB | | Jan-v1-edge.Q51.gguf | Q51 | 1.32 GB | | Jan-v1-edge.Q5K.gguf | Q5K | 1.26 GB | | Jan-v1-edge.Q5KM.gguf | Q5KM | 1.26 GB | | Jan-v1-edge.Q5KS.gguf | Q5KS | 1.23 GB | | Jan-v1-edge.Q6K.gguf | Q6K | 1.42 GB | | Jan-v1-edge.Q80.gguf | Q80 | 1.83 GB | | File Name | Quant Type | File Size | | - | - | - | | Jan-v1-4B.BF16.gguf | BF16 | 8.05 GB | | Jan-v1-4B.F16.gguf | F16 | 8.05 GB | | Jan-v1-4B.F32.gguf | F32 | 16.1 GB | | Jan-v1-4B.Q2K.gguf | Q2K | 1.67 GB | | Jan-v1-4B.Q3KL.gguf | Q3KL | 2.24 GB | | Jan-v1-4B.Q3KM.gguf | Q3KM | 2.08 GB | | Jan-v1-4B.Q3KS.gguf | Q3KS | 1.89 GB | | Jan-v1-4B.Q4KM.gguf | Q4KM | 2.5 GB | | Jan-v1-4B.Q4KS.gguf | Q4KS | 2.38 GB | | Jan-v1-4B.Q5KM.gguf | Q5KM | 2.89 GB | | Jan-v1-4B.Q5KS.gguf | Q5KS | 2.82 GB | | Jan-v1-4B.Q6K.gguf | Q6K | 3.31 GB | | Jan-v1-4B.Q80.gguf | Q80 | 4.28 GB | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
license:apache-2.0
462
1

Super-Pencil-Flux-LoRA

460
16

Qwen3 Medical GRPO GGUF

> Qwen3MedicalGRPO is a specialized medical language model fine-tuned from the Qwen3 base using Supervised Fine-Tuning (SFT) and enhanced with Group Relative Policy Optimization (GRPO) to deliver advanced performance in clinical case analysis, differential diagnosis, and medical reasoning tasks. The model is designed to provide both detailed, step-by-step reasoning (chain-of-thought) and clear, structured final answers, enabling greater transparency and reliability for healthcare professionals and research applications. By separating its internal analysis from synthesized conclusions, Qwen3MedicalGRPO allows users to trace the logic behind clinical recommendations, optimizing accuracy and trustworthiness in complex medical scenarios. | File Name | Quant Type | File Size | | - | - | - | | Qwen3-Medical-GRPO.BF16.gguf | BF16 | 8.05 GB | | Qwen3-Medical-GRPO.F16.gguf | F16 | 8.05 GB | | Qwen3-Medical-GRPO.F32.gguf | F32 | 16.1 GB | | Qwen3-Medical-GRPO.Q2K.gguf | Q2K | 1.67 GB | | Qwen3-Medical-GRPO.Q3KL.gguf | Q3KL | 2.24 GB | | Qwen3-Medical-GRPO.Q3KM.gguf | Q3KM | 2.08 GB | | Qwen3-Medical-GRPO.Q3KS.gguf | Q3KS | 1.89 GB | | Qwen3-Medical-GRPO.Q4KM.gguf | Q4KM | 2.5 GB | | Qwen3-Medical-GRPO.Q4KS.gguf | Q4KS | 2.38 GB | | Qwen3-Medical-GRPO.Q5KM.gguf | Q5KM | 2.89 GB | | Qwen3-Medical-GRPO.Q5KS.gguf | Q5KS | 2.82 GB | | Qwen3-Medical-GRPO.Q6K.gguf | Q6K | 3.31 GB | | Qwen3-Medical-GRPO.Q80.gguf | Q80 | 4.28 GB | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

license:apache-2.0
458
5

siglip2-x256-explicit-content

NaNK
license:apache-2.0
454
13

Castor-3D-Portrait-Flux-LoRA

451
20

visionOCR-3B-061125

NaNK
license:apache-2.0
449
3

AI-vs-Deepfake-vs-Real-v2.0

NaNK
license:apache-2.0
446
3

Flux.1-Dev-Poster-HQ-LoRA

431
14

Fathom-4B-AIO-GGUF

NaNK
license:mit
428
0

Augmented-Waste-Classifier-SigLIP2

NaNK
license:apache-2.0
414
2

AI-vs-Deepfake-vs-Real

license:apache-2.0
410
9

Polaris-VGA-2B-Post1.0

NaNK
llama.cpp
410
1

Flux.1-Dev-Indo-Realism-LoRA

406
25

Callisto-OCR3-2B-Instruct

NaNK
license:apache-2.0
398
6

MedScholar-1.5B-f32-GGUF

NaNK
license:apache-2.0
383
3

Aya-Expanse-8B-GGUF

NaNK
llama-cpp
382
4

Qwen3-Reranker-0.6B-seq-cls-GGUF

> Qwen3-Reranker-0.6B-seq-cls is a sequence classification adaptation of the Qwen3-Reranker-0.6B model, designed for advanced text reranking and classification tasks across over 100 languages, including code and multilingual retrieval scenarios. Built on the Qwen3 series, this 0.6B parameter model offers a 32k context window and supports instruction-aware customization for specific tasks, typically improving performance by 1% to 5% when using tailored English instructions. It inherits the robust reasoning, long-text understanding, and versatility of its parent models, excelling in text retrieval, code retrieval, clustering, classification, and bitext mining, and ranks among the top models in benchmarks such as the MTEB multilingual leaderboard. | File name | Size | Quant Type | |-----------|------|------------| | Qwen3-4B-abliterated.F32.gguf | 16.1 GB | F32 | | Qwen3-4B-abliterated.BF16.gguf | 8.05 GB | BF16 | | Qwen3-4B-abliterated.F16.gguf | 8.05 GB | F16 | | Qwen3-4B-abliterated.Q80.gguf | 4.28 GB | Q80 | | Qwen3-4B-abliterated.Q6K.gguf | 3.31 GB | Q6K | | Qwen3-4B-abliterated.Q5KM.gguf | 2.89 GB | Q5KM | | Qwen3-4B-abliterated.Q5KS.gguf | 2.82 GB | Q5KS | | Qwen3-4B-abliterated.Q4KM.gguf | 2.5 GB | Q4KM | | Qwen3-4B-abliterated.Q4KS.gguf | 2.38 GB | Q4KS | | Qwen3-4B-abliterated.Q3KL.gguf | 2.24 GB | Q3KL | | Qwen3-4B-abliterated.Q3KM.gguf | 2.08 GB | Q3KM | | Qwen3-4B-abliterated.Q3KS.gguf | 1.89 GB | Q3KS | | Qwen3-4B-abliterated.Q2K.gguf | 1.67 GB | Q2K | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
license:apache-2.0
379
2

Mnist-Digits-SigLIP2

NaNK
license:apache-2.0
371
4

Marco-o1-GGUF

NaNK
Llama-Cpp
362
6

Coloring-Book-Flux-LoRA

358
66

Qwen2.5-Coder-1.5B-GGUF

NaNK
Llama-cpp
358
4

Llama-3.2-8B-GGUF-200K

NaNK
llama
354
9

VyvoTTS-v0-Qwen3-0.6B-GGUF

> VyvoTTS-v0-Qwen3-0.6B is an English Text-to-Speech (TTS) model built on the Qwen3-0.6B architecture and trained using a 10,000-hour dataset to produce natural-sounding speech. With approximately 810 million parameters and licensed under MIT, the model offers flexible usage as a pretrained base for further development, especially recommended to be enhanced by leveraging the Emilia dataset and fine-tuning for single-speaker scenarios. Users can integrate VyvoTTS with the unsloth and SNAC frameworks for speech generation, and the model supports sequence lengths up to 8,192 tokens. Although it currently exhibits a high Word Error Rate (WER), its open-source nature and compatibility with popular Python libraries make it an accessible starting point for advanced speech synthesis projects. | File Name | Size | Quant Type | |-----------|------|------------| | VyvoTTS-v0-Qwen3-0.6B.BF16.gguf | 1.26 GB | BF16 | | VyvoTTS-v0-Qwen3-0.6B.F16.gguf | 1.26 GB | F16 | | VyvoTTS-v0-Qwen3-0.6B.F32.gguf | 2.51 GB | F32 | | VyvoTTS-v0-Qwen3-0.6B.Q2K.gguf | 321 MB | Q2K | | VyvoTTS-v0-Qwen3-0.6B.Q3KL.gguf | 393 MB | Q3KL | | VyvoTTS-v0-Qwen3-0.6B.Q3KM.gguf | 372 MB | Q3KM | | VyvoTTS-v0-Qwen3-0.6B.Q3KS.gguf | 348 MB | Q3KS | | VyvoTTS-v0-Qwen3-0.6B.Q40.gguf | 406 MB | Q40 | | VyvoTTS-v0-Qwen3-0.6B.Q41.gguf | 434 MB | Q41 | | VyvoTTS-v0-Qwen3-0.6B.Q4K.gguf | 421 MB | Q4K | | VyvoTTS-v0-Qwen3-0.6B.Q4KM.gguf | 421 MB | Q4KM | | VyvoTTS-v0-Qwen3-0.6B.Q4KS.gguf | 408 MB | Q4KS | | VyvoTTS-v0-Qwen3-0.6B.Q50.gguf | 461 MB | Q50 | | VyvoTTS-v0-Qwen3-0.6B.Q51.gguf | 489 MB | Q51 | | VyvoTTS-v0-Qwen3-0.6B.Q5K.gguf | 469 MB | Q5K | | VyvoTTS-v0-Qwen3-0.6B.Q5KM.gguf | 469 MB | Q5KM | | VyvoTTS-v0-Qwen3-0.6B.Q5KS.gguf | 461 MB | Q5KS | | VyvoTTS-v0-Qwen3-0.6B.Q6K.gguf | 520 MB | Q6K | | VyvoTTS-v0-Qwen3-0.6B.Q80.gguf | 671 MB | Q80 | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
license:mit
352
4

Qwen3-1.7B-ShiningValiant3-f32-GGUF

NaNK
license:apache-2.0
345
2

facial-age-detection

NaNK
license:apache-2.0
342
5

Kontext Watermark Remover

The Kontext-Watermark-Remover is an adapter for black-forest-lab's FLUX.1-Kontext-dev, designed to precisely remove watermarks and textual content from images while maintaining the original image quality and context. The model was trained on 150 image pairs (75 start images and 75 end images) to ensure accurate and artifact-free watermark removal. > [!note] [photo content], remove any watermark text or logos from the image while preserving the background, texture, lighting, and overall realism. Ensure the edited areas blend seamlessly with surrounding details, leaving no visible traces of watermark removal. | Setting | Value | | ------------------------ | ------------------------ | | Module Type | Adapter | | Base Model | FLUX.1 Kontext Dev - fp8 | | Trigger Words | [photo content], remove any watermark text or logos from the image while preserving the background, texture, lighting, and overall realism. Ensure the edited areas blend seamlessly with surrounding details, leaving no visible traces of watermark removal. | | Image Processing Repeats | 50 | | Epochs | 25 | | Save Every N Epochs | 1 | Labeling: florence-community/Florence-2-large-ft (natural language & English) Total Images Used for Training : 150 Image Pairs (75 Start, 75 End) | Setting | Value | | --------------------------- | --------- | | Seed | - | | Clip Skip | - | | Text Encoder LR | 0.00001 | | UNet LR | 0.00005 | | LR Scheduler | constant | | Optimizer | AdamW8bit | | Network Dimension | 64 | | Network Alpha | 32 | | Gradient Accumulation Steps | - | | Setting | Value | | --------------- | ----- | | Shuffle Caption | - | | Keep N Tokens | - | | Setting | Value | | ------------------------- | ----- | | Noise Offset | 0.03 | | Multires Noise Discount | 0.1 | | Multires Noise Iterations | 10 | | Conv Dimension | - | | Conv Alpha | - | | Batch Size | - | | Steps | 2900 & 400(warm up) | | Sampler | euler | You should use `[photo content]` to trigger the image generation. You should use `remove any watermark text or logos from the image while preserving the background` to trigger the image generation. You should use `texture` to trigger the image generation. You should use `lighting` to trigger the image generation. You should use `and overall realism. Ensure the edited areas blend seamlessly with surrounding details` to trigger the image generation. You should use `leaving no visible traces of watermark removal.` to trigger the image generation.

license:apache-2.0
341
11

Pyxidis-Manim-CodeGen-1.7B-GGUF

NaNK
license:apache-2.0
337
3

Octans-Qwen3-UI-Code-4B-GGUF

> Octans-Qwen3-UI-Code-4B is an optimized successor of Muscae-Qwen3-UI-Code-4B, fine-tuned for enhanced UI reasoning precision, layout structuring, and frontend code synthesis. Built upon Qwen3-4B and refined through Abliterated Reasoning Optimization, it delivers balanced, structured, and production-grade UI code outputs for experimental and research use. Ideal for frontend developers, UI engineers, and design system researchers exploring next-generation code synthesis. | File Name | Quant Type | File Size | | - | - | - | | Octans-Qwen3-UI-Code-4B.BF16.gguf | BF16 | 8.05 GB | | Octans-Qwen3-UI-Code-4B.F16.gguf | F16 | 8.05 GB | | Octans-Qwen3-UI-Code-4B.F32.gguf | F32 | 16.1 GB | | Octans-Qwen3-UI-Code-4B.Q2K.gguf | Q2K | 1.67 GB | | Octans-Qwen3-UI-Code-4B.Q3KL.gguf | Q3KL | 2.24 GB | | Octans-Qwen3-UI-Code-4B.Q3KM.gguf | Q3KM | 2.08 GB | | Octans-Qwen3-UI-Code-4B.Q3KS.gguf | Q3KS | 1.89 GB | | Octans-Qwen3-UI-Code-4B.Q4KM.gguf | Q4KM | 2.5 GB | | Octans-Qwen3-UI-Code-4B.Q4KS.gguf | Q4KS | 2.38 GB | | Octans-Qwen3-UI-Code-4B.Q5KM.gguf | Q5KM | 2.89 GB | | Octans-Qwen3-UI-Code-4B.Q5KS.gguf | Q5KS | 2.82 GB | | Octans-Qwen3-UI-Code-4B.Q6K.gguf | Q6K | 3.31 GB | | Octans-Qwen3-UI-Code-4B.Q80.gguf | Q80 | 4.28 GB | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
license:apache-2.0
335
0

Polaris-VGA-0.8B-Post1.0

NaNK
llama.cpp
333
1

Flux-Realism-FineDetailed

323
24

Deepfake-Detection-Exp-02-21

Deepfake-Detection-Exp-02-21 is a minimalist, high-quality dataset trained on a ViT-based model for image classification, distinguishing between deepfake and real images. The model is based on Google's `google/vit-base-patch16-224-in21k`. Limitations 1. Generalization Issues – The model may not perform well on deepfake images generated by unseen or novel deepfake techniques. 2. Dataset Bias – The training data might not cover all variations of real and fake images, leading to biased predictions. 3. Resolution Constraints – Since the model is based on `vit-base-patch16-224-in21k`, it is optimized for 224x224 image resolution, which may limit its effectiveness on high-resolution images. 4. Adversarial Vulnerabilities – The model may be susceptible to adversarial attacks designed to fool vision transformers. 5. False Positives & False Negatives – The model may occasionally misclassify real images as deepfake and vice versa, requiring human validation in critical applications. Intended Use 1. Deepfake Detection – Designed for identifying deepfake images in media, social platforms, and forensic analysis. 2. Research & Development – Useful for researchers studying deepfake detection and improving ViT-based classification models. 3. Content Moderation – Can be integrated into platforms to detect and flag manipulated images. 4. Security & Forensics – Assists in cybersecurity applications where verifying the authenticity of images is crucial. 5. Educational Purposes – Can be used in training AI practitioners and students in the field of computer vision and deepfake detection.

license:apache-2.0
322
3

Polaris-VGA-27B-Post1.0e

NaNK
llama.cpp
322
1

PhotoCleanser-i2i

PhotoCleanser-i2i is an adapter for black-forest-lab's FLUX.1-Kontext-dev. It is an experimental LoRA designed for removing specified object(s) while preserving the remaining content in the image. The model was trained on 36 image pairs (18 start images, 18 end images). Synthetic result nodes were generated using NanoBanana from Google and labeled with DeepCaption-VLA-7B. The adapter is triggered with the following prompt: > [!note] [photo content], remove the specified object(s) from the image while preserving the background and remaining elements, maintaining realism and original details. > [photo content], remove all humans from the image while preserving the background and remaining elements, maintaining realism and original details. > [photo content], remove the ball from the image while preserving the background and remaining elements, maintaining realism and original details. > [photo content], remove the cat from the image while preserving the background and remaining elements, maintaining realism and original details. Sample Inference Comparing the Base Model with the Adapter > Note: In over 25 inferences across various scale settings, the base model struggles to properly reconstruct the basketball net after the ball is removed. | FLUX.1-Kontext-dev | PhotoCleanser-i2i | |-------------------|-----------------| | | | > No desired effect was achieved with the base model after testing many settings, but PhotoCleanser-i2i performed its best to remove the human character from the image. | FLUX.1-Kontext-dev | PhotoCleanser-i2i | |-------------------|-----------------| | | | > No desired effect was achieved with the base model after testing many settings, but PhotoCleanser-i2i performed its best to remove the human/animal character from the image. | FLUX.1-Kontext-dev | PhotoCleanser-i2i | |-------------------|-----------------| | | | | Setting | Value | | ------------------------ | ------------------------ | | Module Type | Adapter | | Base Model | FLUX.1 Kontext Dev - fp8 | | Trigger Words | [photo content], remove the specified object(s) from the image while preserving the background and remaining elements, maintaining realism and original details. | | Image Processing Repeats | 50 | | Epochs | 32 | | Save Every N Epochs | 1 | Labeling: DeepCaption-VLA-7B(natural language & English) Total Images Used for Training : 36 Image Pairs (18 Start, 18 End) Synthetic Result Node generated by NanoBanana from Google (Image Result Sets Dataset) | Setting | Value | | --------------------------- | --------- | | Seed | - | | Clip Skip | - | | Text Encoder LR | 0.00001 | | UNet LR | 0.00005 | | LR Scheduler | constant | | Optimizer | AdamW8bit | | Network Dimension | 64 | | Network Alpha | 32 | | Gradient Accumulation Steps | - | | Setting | Value | | --------------- | ----- | | Shuffle Caption | - | | Keep N Tokens | - | | Setting | Value | | ------------------------- | ----- | | Noise Offset | 0.03 | | Multires Noise Discount | 0.1 | | Multires Noise Iterations | 10 | | Conv Dimension | - | | Conv Alpha | - | | Batch Size | - | | Steps | 4370 (Low(700)) | | Sampler | euler | You should use `[photo content]` to trigger the image generation. You should use `remove the specified object(s) from the image while preserving the background and remaining elements` to trigger the image generation. You should use `maintaining realism and original details.` to trigger the image generation.

321
7

Polaris-VGA-9B-Post1.0e

NaNK
llama.cpp
312
1

Gliese-4B-OSS-0410-GGUF

> Gliese-4B-OSS-0410 is a reasoning-focused model fine-tuned on Qwen-4B for enhanced reasoning and polished token probability distributions, delivering balanced multilingual generation across mathematics and general-purpose reasoning tasks. The model is fine-tuned on curated GPT-OSS synthetic dataset entries, improving its ability to handle structured reasoning, probabilistic inference, and multilingual tasks with precision. | File Name | Quant Type | File Size | | - | - | - | | Gliese-4B-OSS-0410.BF16.gguf | BF16 | 8.05 GB | | Gliese-4B-OSS-0410.F16.gguf | F16 | 8.05 GB | | Gliese-4B-OSS-0410.F32.gguf | F32 | 16.1 GB | | Gliese-4B-OSS-0410.Q2K.gguf | Q2K | 1.67 GB | | Gliese-4B-OSS-0410.Q3KL.gguf | Q3KL | 2.24 GB | | Gliese-4B-OSS-0410.Q3KM.gguf | Q3KM | 2.08 GB | | Gliese-4B-OSS-0410.Q3KS.gguf | Q3KS | 1.89 GB | | Gliese-4B-OSS-0410.Q4KM.gguf | Q4KM | 2.5 GB | | Gliese-4B-OSS-0410.Q4KS.gguf | Q4KS | 2.38 GB | | Gliese-4B-OSS-0410.Q5KM.gguf | Q5KM | 2.89 GB | | Gliese-4B-OSS-0410.Q5KS.gguf | Q5KS | 2.82 GB | | Gliese-4B-OSS-0410.Q6K.gguf | Q6K | 3.31 GB | | Gliese-4B-OSS-0410.Q80.gguf | Q80 | 4.28 GB | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
license:apache-2.0
312
0

Electric-Blue-Flux-LoRA

306
4

Llama-3.2-3B-Instruct-f32-GGUF

NaNK
llama
298
1

MegaScience-Qwen-GGUF

> MegaScience-Qwen models are a series of large language models based on the Qwen3 and Qwen2.5 architectures, meticulously fine-tuned on the MegaScience dataset to advance scientific reasoning in AI. This dataset blends over 1.25 million high-quality, university-level scientific questions and answers sourced from open textbooks and diverse scientific benchmarks, covering seven scientific disciplines. The MegaScience-Qwen lineup includes variants from smaller Qwen2.5-1.5B up to Qwen3-30B, with models such as Qwen3-4B-MegaScience, Qwen3-8B-MegaScience, and Qwen3-14B-MegaScience, each showing pronounced gains over their official instruction-tuned counterparts—especially as model scale increases. These models demonstrate state-of-the-art or leading performance on scientific reasoning, general knowledge, and mathematical benchmarks, achieving not only higher accuracy but also more concise and efficient responses. The MegaScience project also provides a rigorous evaluation system, an open-source curation pipeline, and all model checkpoints, empowering further research and application in scientific AI reasoning and education. | Model Name | GGUF Repository Link | |--------------------------------|----------------------------------------------------------------------------------------| | Qwen3-8B-MegaScience-GGUF | Hugging Face ↗| | Qwen3-4B-MegaScience-GGUF | Hugging Face ↗ | | Qwen3-1.7B-MegaScience-GGUF | Hugging Face ↗ | | Qwen2.5-3B-MegaScience-GGUF | Hugging Face ↗ | | Qwen2.5-1.5B-MegaScience-GGUF | Hugging Face ↗ | | Qwen2.5-7B-MegaScience-GGUF | Hugging Face ↗ | | File Name | Quant Type | File Size | | - | - | - | | Qwen3-8B-MegaScience.BF16.gguf | BF16 | 16.4 GB | | Qwen3-8B-MegaScience.F16.gguf | F16 | 16.4 GB | | Qwen3-8B-MegaScience.Q80.gguf | Q80 | 8.71 GB | | File Name | Size | Quant Type | |-----------|------|------------| | Qwen3-4B-MegaScience.BF16.gguf | 8.05 GB | BF16 | | Qwen3-4B-MegaScience.F16.gguf | 8.05 GB | F16 | | Qwen3-4B-MegaScience.F32.gguf | 16.1 GB | F32 | | Qwen3-4B-MegaScience.Q2K.gguf | 1.67 GB | Q2K | | Qwen3-4B-MegaScience.Q3KL.gguf | 2.24 GB | Q3KL | | Qwen3-4B-MegaScience.Q3KM.gguf | 2.08 GB | Q3KM | | Qwen3-4B-MegaScience.Q3KS.gguf | 1.89 GB | Q3KS | | Qwen3-4B-MegaScience.Q4KM.gguf | 2.5 GB | Q4KM | | Qwen3-4B-MegaScience.Q4KS.gguf | 2.38 GB | Q4KS | | Qwen3-4B-MegaScience.Q5KM.gguf | 2.89 GB | Q5KM | | Qwen3-4B-MegaScience.Q5KS.gguf | 2.82 GB | Q5KS | | Qwen3-4B-MegaScience.Q6K.gguf | 3.31 GB | Q6K | | Qwen3-4B-MegaScience.Q80.gguf | 4.28 GB | Q80 | | File Name | Size | Quant Type | |-----------|------|------------| | Qwen3-1.7B-MegaScience.BF16.gguf | 3.45 GB | BF16 | | Qwen3-1.7B-MegaScience.F16.gguf | 3.45 GB | F16 | | Qwen3-1.7B-MegaScience.F32.gguf | 6.89 GB | F32 | | Qwen3-1.7B-MegaScience.Q2K.gguf | 778 MB | Q2K | | Qwen3-1.7B-MegaScience.Q3KL.gguf | 1 GB | Q3KL | | Qwen3-1.7B-MegaScience.Q3KM.gguf | 940 MB | Q3KM | | Qwen3-1.7B-MegaScience.Q3KS.gguf | 867 MB | Q3KS | | Qwen3-1.7B-MegaScience.Q4KM.gguf | 1.11 GB | Q4KM | | Qwen3-1.7B-MegaScience.Q4KS.gguf | 1.06 GB | Q4KS | | Qwen3-1.7B-MegaScience.Q5KM.gguf | 1.26 GB | Q5KM | | Qwen3-1.7B-MegaScience.Q5KS.gguf | 1.23 GB | Q5KS | | Qwen3-1.7B-MegaScience.Q6K.gguf | 1.42 GB | Q6K | | Qwen3-1.7B-MegaScience.Q80.gguf | 1.83 GB | Q80 | | File Name | Size | Quant Type | |-----------|------|------------| | Qwen2.5-3B-MegaScience.BF16.gguf | 6.18 GB | BF16 | | Qwen2.5-3B-MegaScience.F16.gguf | 6.18 GB | F16 | | Qwen2.5-3B-MegaScience.F32.gguf | 12.3 GB | F32 | | Qwen2.5-3B-MegaScience.Q2K.gguf | 1.27 GB | Q2K | | Qwen2.5-3B-MegaScience.Q3KL.gguf | 1.71 GB | Q3KL | | Qwen2.5-3B-MegaScience.Q3KM.gguf | 1.59 GB | Q3KM | | Qwen2.5-3B-MegaScience.Q3KS.gguf | 1.45 GB | Q3KS | | Qwen2.5-3B-MegaScience.Q4KM.gguf | 1.93 GB | Q4KM | | Qwen2.5-3B-MegaScience.Q4KS.gguf | 1.83 GB | Q4KS | | Qwen2.5-3B-MegaScience.Q5KM.gguf | 2.22 GB | Q5KM | | Qwen2.5-3B-MegaScience.Q5KS.gguf | 2.17 GB | Q5KS | | Qwen2.5-3B-MegaScience.Q6K.gguf | 2.54 GB | Q6K | | Qwen2.5-3B-MegaScience.Q80.gguf | 3.29 GB | Q80 | | File Name | Size | Quant Type | |-----------|------|------------| | Qwen2.5-1.5B-MegaScience.BF16.gguf | 3.09 GB | BF16 | | Qwen2.5-1.5B-MegaScience.F16.gguf | 3.09 GB | F16 | | Qwen2.5-1.5B-MegaScience.F32.gguf | 6.18 GB | F32 | | Qwen2.5-1.5B-MegaScience.Q2K.gguf | 676 MB | Q2K | | Qwen2.5-1.5B-MegaScience.Q3KL.gguf | 880 MB | Q3KL | | Qwen2.5-1.5B-MegaScience.Q3KM.gguf | 824 MB | Q3KM | | Qwen2.5-1.5B-MegaScience.Q3KS.gguf | 761 MB | Q3KS | | Qwen2.5-1.5B-MegaScience.Q4KM.gguf | 986 MB | Q4KM | | Qwen2.5-1.5B-MegaScience.Q4KS.gguf | 940 MB | Q4KS | | Qwen2.5-1.5B-MegaScience.Q5KM.gguf | 1.13 GB | Q5KM | | Qwen2.5-1.5B-MegaScience.Q5KS.gguf | 1.1 GB | Q5KS | | Qwen2.5-1.5B-MegaScience.Q6K.gguf | 1.27 GB | Q6K | | Qwen2.5-1.5B-MegaScience.Q80.gguf | 1.65 GB | Q80 | | File Name | Quant Type | File Size | | - | - | - | | Qwen2.5-7B-MegaScience.BF16.gguf | BF16 | 15.2 GB | | Qwen2.5-7B-MegaScience.F16.gguf | F16 | 15.2 GB | | Qwen2.5-7B-MegaScience.F32.gguf | F32 | 30.5 GB | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
license:apache-2.0
296
1

Digital-Yellow-Flux-LoRA

295
4

Qwen3-4B-abliterated-f32-GGUFs

> Qwen3-4B-abliterated is an experimental, uncensored version of the Qwen/Qwen3-4B language model that explores how refusals and latent fine-tuning work in large language models using a novel "abliteration" technique, which subtracts a computed refusal direction from hidden module states (such as oproj) to minimize refusals without degrading output quality. The process involves comparing residual streams between harmful and harmless prompts, orthogonalizing hidden states with weight factors distributed across layers, and iterative or accumulated orthogonalization methods for efficiency. | File name | Size | Quant Type | |-----------|------|------------| | Qwen3-4B-abliterated.F32.gguf | 16.1 GB | F32 | | Qwen3-4B-abliterated.BF16.gguf | 8.05 GB | BF16 | | Qwen3-4B-abliterated.F16.gguf | 8.05 GB | F16 | | Qwen3-4B-abliterated.Q80.gguf | 4.28 GB | Q80 | | Qwen3-4B-abliterated.Q6K.gguf | 3.31 GB | Q6K | | Qwen3-4B-abliterated.Q5KM.gguf | 2.89 GB | Q5KM | | Qwen3-4B-abliterated.Q5KS.gguf | 2.82 GB | Q5KS | | Qwen3-4B-abliterated.Q4KM.gguf | 2.5 GB | Q4KM | | Qwen3-4B-abliterated.Q4KS.gguf | 2.38 GB | Q4KS | | Qwen3-4B-abliterated.Q3KL.gguf | 2.24 GB | Q3KL | | Qwen3-4B-abliterated.Q3KM.gguf | 2.08 GB | Q3KM | | Qwen3-4B-abliterated.Q3KS.gguf | 1.89 GB | Q3KS | | Qwen3-4B-abliterated.Q2K.gguf | 1.67 GB | Q2K | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
license:apache-2.0
294
2

Orange-Chroma-Flux-LoRA

287
5

open-age-detection

> `open-age-detection` is a vision-language encoder model fine-tuned from `google/siglip2-base-patch16-512` for multi-class image classification. It is trained to classify the estimated age group of a person from an image. The model uses the `SiglipForImageClassification` architecture. > \[!note] > SigLIP 2: Multilingual Vision-Language Encoders with Improved Semantic Understanding, Localization, and Dense Features > https://arxiv.org/pdf/2502.14786 Demographic Analysis – Estimate age groups for statistical or analytical applications. Smart Personalization – Age-based content or product recommendation. Access Control – Assist systems requiring age verification. Social Research – Study age-related trends in image datasets. Surveillance and Security – Profile age ranges in monitored environments.

NaNK
license:apache-2.0
287
4

Regulus-Qwen3-R1-Llama-Distill-GGUF

> Regulus-Qwen3-R1-Llama-Distill-1.7B is a distilled reasoning model fine-tuned on Qwen/Qwen3-1.7B using Magpie-Align/Magpie-Reasoning-V2-250K-CoT-DeepSeek-R1-Llama-70B. The training leverages distilled traces from DeepSeek-R1-Llama-70B, transferring advanced reasoning patterns into a lightweight 1.7B parameter model. It is specialized for chain-of-thought reasoning across code, math, and science, optimized for efficiency and mid-resource deployment. | File Name | Quant Type | File Size | | - | - | - | | Regulus-Qwen3-R1-Llama-Distill-1.7B.BF16.gguf | BF16 | 3.45 GB | | Regulus-Qwen3-R1-Llama-Distill-1.7B.F16.gguf | F16 | 3.45 GB | | Regulus-Qwen3-R1-Llama-Distill-1.7B.F32.gguf | F32 | 6.89 GB | | Regulus-Qwen3-R1-Llama-Distill-1.7B.Q2K.gguf | Q2K | 778 MB | | Regulus-Qwen3-R1-Llama-Distill-1.7B.Q3KL.gguf | Q3KL | 1 GB | | Regulus-Qwen3-R1-Llama-Distill-1.7B.Q3KM.gguf | Q3KM | 940 MB | | Regulus-Qwen3-R1-Llama-Distill-1.7B.Q3KS.gguf | Q3KS | 867 MB | | Regulus-Qwen3-R1-Llama-Distill-1.7B.Q40.gguf | Q40 | 1.05 GB | | Regulus-Qwen3-R1-Llama-Distill-1.7B.Q41.gguf | Q41 | 1.14 GB | | Regulus-Qwen3-R1-Llama-Distill-1.7B.Q4K.gguf | Q4K | 1.11 GB | | Regulus-Qwen3-R1-Llama-Distill-1.7B.Q4KM.gguf | Q4KM | 1.11 GB | | Regulus-Qwen3-R1-Llama-Distill-1.7B.Q4KS.gguf | Q4KS | 1.06 GB | | Regulus-Qwen3-R1-Llama-Distill-1.7B.Q50.gguf | Q50 | 1.23 GB | | Regulus-Qwen3-R1-Llama-Distill-1.7B.Q51.gguf | Q51 | 1.32 GB | | Regulus-Qwen3-R1-Llama-Distill-1.7B.Q5K.gguf | Q5K | 1.26 GB | | Regulus-Qwen3-R1-Llama-Distill-1.7B.Q5KM.gguf | Q5KM | 1.26 GB | | Regulus-Qwen3-R1-Llama-Distill-1.7B.Q5KS.gguf | Q5KS | 1.23 GB | | Regulus-Qwen3-R1-Llama-Distill-1.7B.Q6K.gguf | Q6K | 1.42 GB | | Regulus-Qwen3-R1-Llama-Distill-1.7B.Q80.gguf | Q80 | 1.83 GB | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
dataset:Magpie-Align/Magpie-Reasoning-V2-250K-CoT-Deepseek-R1-Llama-70B
284
2

Llama-Song-Stream-3B-Instruct-GGUF

NaNK
llama
283
6

Castor-Red-Dead-Redemption-2-Flux-LoRA

license:apache-2.0
280
7

Polaris-VGA-4B-Post1.0e

NaNK
llama.cpp
279
1

Castor-Dramatic-Neon-Flux-LoRA

license:apache-2.0
277
4

Triangulum-10B-GGUF

NaNK
llama
277
2

Kimina-Prover-AI-MO-GGUF

NaNK
license:apache-2.0
276
1

Logics-Qwen3-Math-4B

> Logics-Qwen3-Math-4B is a reasoning-focused model fine-tuned on Qwen3-4B-Thinking-2507 for mathematical reasoning and logical coding, trained on OpenMathReasoning, OpenCodeReasoning, and Helios-R-6M datasets. It excels in structured mathematical problem solving, algorithmic logic, and probabilistic reasoning, making it ideal for educators, researchers, and developers focused on computational logic and math. 1. Mathematical & Logical Reasoning Fine-tuned for high-precision math reasoning, algorithmic problem-solving, and logical coding tasks. 2. Event-Driven & Probabilistic Modeling Performs probability-based simulations, structured decision-making, and multi-step logical reasoning with strong accuracy. 3. Multilingual Problem Solving Supports math and logic tasks across multiple languages, suitable for global research and education workflows. 4. Hybrid Symbolic-Algorithmic Thinking Combines structured logic, symbolic computation, and probabilistic inference to handle uncertainty-driven problems efficiently. 5. Structured Output Mastery Generates outputs in LaTeX, Markdown, JSON, CSV, and YAML, enabling smooth integration into technical and research workflows. 6. Optimized 4B Parameter Footprint Deployable on mid-range GPUs, offline clusters, and edge devices, maintaining high reasoning quality while being resource-efficient. High-precision mathematical reasoning and problem-solving Algorithmic logic, structured coding tasks, and probability analysis Educational and research-focused workflows Deployment on mid-resource environments with efficient reasoning Structured data and technical content generation Focused on math and logic—less suited for creative writing or casual conversation Very complex multi-hop reasoning may challenge the 4B parameter capacity Prioritizes structured reasoning over conversational tone Outputs may be inconsistent for extremely long or cross-domain multi-document contexts

NaNK
license:apache-2.0
274
2

Flux-Lego-Ref-LoRA

270
15

Glowing-Body-Flux-LoRA

269
10

Hermes-3-Llama-3.2-3B-f32-GGUF

NaNK
llama
269
1

Qwen3-4B-SafeRL-GGUF

> Qwen3-4B-SafeRL is a safety-aligned version of the Qwen3-4B model, trained using Reinforcement Learning (RL) with a reward signal from Qwen3Guard-Gen to boost robustness against harmful or adversarial prompts. This safety alignment process optimizes the model with a hybrid reward function that simultaneously focuses on three objectives: maximizing safety (penalizing unsafe content as detected by Qwen3Guard-Gen-4B), maximizing helpfulness (rewarding genuinely helpful responses based on the WorldPM-Helpsteer2 model), and minimizing unnecessary refusals (penalizing unnecessary refusals according to Qwen3Guard-Gen-4B). | File Name | Quant Type | File Size | | - | - | - | | Qwen3-4B-SafeRL.BF16.gguf | BF16 | 8.05 GB | | Qwen3-4B-SafeRL.F16.gguf | F16 | 8.05 GB | | Qwen3-4B-SafeRL.F32.gguf | F32 | 16.1 GB | | Qwen3-4B-SafeRL.Q2K.gguf | Q2K | 1.67 GB | | Qwen3-4B-SafeRL.Q3KL.gguf | Q3KL | 2.24 GB | | Qwen3-4B-SafeRL.Q3KM.gguf | Q3KM | 2.08 GB | | Qwen3-4B-SafeRL.Q3KS.gguf | Q3KS | 1.89 GB | | Qwen3-4B-SafeRL.Q4KM.gguf | Q4KM | 2.5 GB | | Qwen3-4B-SafeRL.Q4KS.gguf | Q4KS | 2.38 GB | | Qwen3-4B-SafeRL.Q5KM.gguf | Q5KM | 2.89 GB | | Qwen3-4B-SafeRL.Q5KS.gguf | Q5KS | 2.82 GB | | Qwen3-4B-SafeRL.Q6K.gguf | Q6K | 3.31 GB | | Qwen3-4B-SafeRL.Q80.gguf | Q80 | 4.28 GB | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
license:apache-2.0
269
1

Flux-Polaroid-Plus

264
8

Kepler-Qwen3-4B-Super-Thinking-GGUF

> Kepler-Qwen3-4B-Super-Thinking is a reasoning-focused model fine-tuned on Qwen for Abliterated Reasoning and polished token probabilities, enhancing balanced multilingual generation across mathematics and general-purpose reasoning. It specializes in event-driven logic, structured analysis, and precise probabilistic modeling—making it an ideal tool for researchers, educators, and developers working with uncertainty and structured reasoning. | File Name | Quant Type | File Size | | - | - | - | | Kepler-Qwen3-4B-Super-Thinking.BF16.gguf | BF16 | 8.05 GB | | Kepler-Qwen3-4B-Super-Thinking.F16.gguf | F16 | 8.05 GB | | Kepler-Qwen3-4B-Super-Thinking.F32.gguf | F32 | 16.1 GB | | Kepler-Qwen3-4B-Super-Thinking.Q2K.gguf | Q2K | 1.67 GB | | Kepler-Qwen3-4B-Super-Thinking.Q3KL.gguf | Q3KL | 2.24 GB | | Kepler-Qwen3-4B-Super-Thinking.Q3KM.gguf | Q3KM | 2.08 GB | | Kepler-Qwen3-4B-Super-Thinking.Q3KS.gguf | Q3KS | 1.89 GB | | Kepler-Qwen3-4B-Super-Thinking.Q4KM.gguf | Q4KM | 2.5 GB | | Kepler-Qwen3-4B-Super-Thinking.Q4KS.gguf | Q4KS | 2.38 GB | | Kepler-Qwen3-4B-Super-Thinking.Q5KM.gguf | Q5KM | 2.89 GB | | Kepler-Qwen3-4B-Super-Thinking.Q5KS.gguf | Q5KS | 2.82 GB | | Kepler-Qwen3-4B-Super-Thinking.Q6K.gguf | Q6K | 3.31 GB | | Kepler-Qwen3-4B-Super-Thinking.Q80.gguf | Q80 | 4.28 GB | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
license:apache-2.0
263
0

chandra-GGUF

llama.cpp
260
1

Dorado-WebSurf_Tool-ext-GGUF

> Dorado-WebSurfTool-ext is a function-calling and agentic reasoning model fine-tuned from Qwen3-4B, designed for web search orchestration, tool-augmented reasoning, and dynamic problem-solving. It excels at agentic decision-making, tool selection, and structured execution flow, making it ideal for retrieval-augmented generation (RAG), function calling, and tool-based query resolution. | File Name | Quant Type | File Size | | - | - | - | | Dorado-WebSurfTool-ext.BF16.gguf | BF16 | 8.05 GB | | Dorado-WebSurfTool-ext.F16.gguf | F16 | 8.05 GB | | Dorado-WebSurfTool-ext.F32.gguf | F32 | 16.1 GB | | Dorado-WebSurfTool-ext.Q2K.gguf | Q2K | 1.67 GB | | Dorado-WebSurfTool-ext.Q3KL.gguf | Q3KL | 2.24 GB | | Dorado-WebSurfTool-ext.Q3KM.gguf | Q3KM | 2.08 GB | | Dorado-WebSurfTool-ext.Q3KS.gguf | Q3KS | 1.89 GB | | Dorado-WebSurfTool-ext.Q4KM.gguf | Q4KM | 2.5 GB | | Dorado-WebSurfTool-ext.Q4KS.gguf | Q4KS | 2.38 GB | | Dorado-WebSurfTool-ext.Q5KM.gguf | Q5KM | 2.89 GB | | Dorado-WebSurfTool-ext.Q5KS.gguf | Q5KS | 2.82 GB | | Dorado-WebSurfTool-ext.Q6K.gguf | Q6K | 3.31 GB | | Dorado-WebSurfTool-ext.Q80.gguf | Q80 | 4.28 GB | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

license:apache-2.0
259
1

UIGEN-X-4B-0729-f32-GGUF

NaNK
license:apache-2.0
256
1

Ton618-Tarot-Cards-Flux-LoRA

253
34

Ophiuchi-Qwen3-14B-Instruct

NaNK
license:apache-2.0
253
8

Kepler-186f-Qwen3-Instruct-4B-GGUF

> Kepler-186f-Qwen3-Instruct-4B is a reasoning-focused model fine-tuned on Qwen for Abliterated Reasoning and polished token probabilities, enhancing balanced multilingual generation across mathematics and general-purpose reasoning. It specializes in event-driven logic, structured analysis, and precise probabilistic modeling—making it an ideal tool for researchers, educators, and developers working with uncertainty and structured reasoning. | File Name | Quant Type | File Size | | - | - | - | | Kepler-186f-Qwen3-Instruct-4B.BF16.gguf | BF16 | 8.05 GB | | Kepler-186f-Qwen3-Instruct-4B.F16.gguf | F16 | 8.05 GB | | Kepler-186f-Qwen3-Instruct-4B.F32.gguf | F32 | 16.1 GB | | Kepler-186f-Qwen3-Instruct-4B.Q2K.gguf | Q2K | 1.67 GB | | Kepler-186f-Qwen3-Instruct-4B.Q3KL.gguf | Q3KL | 2.24 GB | | Kepler-186f-Qwen3-Instruct-4B.Q3KM.gguf | Q3KM | 2.08 GB | | Kepler-186f-Qwen3-Instruct-4B.Q3KS.gguf | Q3KS | 1.89 GB | | Kepler-186f-Qwen3-Instruct-4B.Q4KM.gguf | Q4KM | 2.5 GB | | Kepler-186f-Qwen3-Instruct-4B.Q4KS.gguf | Q4KS | 2.38 GB | | Kepler-186f-Qwen3-Instruct-4B.Q5KM.gguf | Q5KM | 2.89 GB | | Kepler-186f-Qwen3-Instruct-4B.Q5KS.gguf | Q5KS | 2.82 GB | | Kepler-186f-Qwen3-Instruct-4B.Q6K.gguf | Q6K | 3.31 GB | | Kepler-186f-Qwen3-Instruct-4B.Q80.gguf | Q80 | 4.28 GB | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
license:apache-2.0
252
3

CAnime-LoRA

251
14

Flux-Fine-Detail-LoRA

246
8

Qwen3-4B-Instruct-2507-GGUF

NaNK
llama.cpp
246
1

Flux-Dalle-Mix-LoRA

245
23

Qwen3-0.6B-GGUF

NaNK
license:apache-2.0
243
1

Llama-3.2-1B-GGUF

NaNK
llama
239
2

Qwen3-VL-30B-A3B-Instruct-abliterated-v1

NaNK
license:apache-2.0
238
3

KAIROS-MM-Qwen2.5-VL-7B-RL-AIO-GGUF

NaNK
llama.cpp
231
1

Flux.1-Dev-LoRA-HDR-Realism

228
22

Clouds-Illusion-Flux-LoRA

227
7

Computron-Bots-1.7B-R1-GGUF

NaNK
license:apache-2.0
225
0

Purple-Grid-Flux-LoRA

224
4

Qwen3 VL 30B A3B Instruct Abliterated

> Qwen3-VL-30B-A3B-Instruct-abliterated is an abliterated (v1.0) variant of Qwen3-VL-30B-A3B-Instruct, designed for Abliterated Reasoning and Captioning. > This model leverages the Qwen3-VL-MoE (Mixture of Experts) architecture to deliver deeply descriptive, context-rich, and reasoning-oriented multimodal outputs. It handles complex, sensitive, and nuanced visual content while maintaining balanced interpretive coherence and multilingual adaptability. Abliterated / Uncensored Captioning and Reasoning Fine-tuned to bypass standard content filters while preserving factual accuracy, descriptive depth, and logical reasoning. High-Fidelity Reasoning and Visual Understanding Generates detailed captions and structured reasoning for diverse visual categories—artistic, technical, abstract, or low-context. Mixture of Experts (MoE) Efficiency Built on Qwen3-VL-MoE, dynamically routing computation through specialized experts for enhanced precision and scalability. Aspect-Ratio Robustness Performs consistently across wide, tall, square, panoramic, and irregular visual formats. Variational Detail Control Supports both concise summaries and highly detailed reasoning narratives, depending on prompt configuration. Multilingual Output Capability Defaults to English but adaptable for multilingual use through prompt engineering. Generating detailed, uncensored captions and reasoning for complex or creative visual datasets. Research in multimodal reasoning, safety evaluation, and content moderation studies. Enabling descriptive captioning and analytical reasoning for datasets excluded from mainstream models. Creative applications such as narrative generation, artistic interpretation, and visual storytelling. Advanced reasoning over diverse visual structures and aspect ratios. May produce explicit, sensitive, or offensive content depending on input and prompt. Not recommended for deployment in production systems that require strict moderation or filtering. Style, tone, and reasoning detail can vary based on prompt phrasing. May show variable performance on synthetic, abstract, or highly stylized visual inputs.

NaNK
license:apache-2.0
222
3

Fractured-Line-Flare

219
3

Mintaka-Qwen3-1.6B-V3.1-GGUF

NaNK
license:apache-2.0
219
1

MemOperator-4B-f32-GGUF

NaNK
license:apache-2.0
217
1

Deepfake-vs-Real-8000

> Deepfake-vs-Real-8000 is an image classification vision-language encoder model fine-tuned from google/siglip2-base-patch16-224 for a single-label classification task. It is designed to detect whether an image is a deepfake or a real one using the SiglipForImageClassification architecture. The model categorizes images into two classes: - Class 0: "Deepfake" - Class 1: "Real one" The Deepfake-vs-Real-8000 model is designed to detect deepfake images from real ones. Potential use cases include: - Deepfake Detection: Assisting cybersecurity experts and forensic teams in detecting synthetic media. - Media Verification: Helping journalists and fact-checkers verify the authenticity of images. - AI Ethics & Research: Contributing to studies on AI-generated content detection. - Social Media Moderation: Enhancing tools to prevent misinformation and digital deception.

NaNK
license:apache-2.0
217
0

Novaeus-Promptist-7B-Instruct-GGUF

NaNK
Ollama
211
4

RL-Quantum-4B-GGUF

> QUASAR is a 4B-parameter language model fine-tuned from Qwen3-4B-Instruct-2507 through supervised learning followed by agentic reinforcement learning with tool-augmented feedback. Specially designed for generating OpenQASM 3.0 quantum circuits for tasks like QAOA and VQE, the model optimizes for both syntactic validity and semantic fidelity, using external quantum simulation for reward calculation across hierarchical criteria (syntax, distribution alignment, expectation value, and optimization progress). QUASAR is best suited for natural language to quantum circuit generation and quantum optimization algorithm design in research or integration scenarios, though users are advised to validate its outputs using external quantum simulators to address limitations in problem generalization. Training used a dataset with QASM 3.0 circuits and quantum optimization problems, employing SFT and RL (with GRPO and hierarchical reward). In evaluations, the model substantially outperforms comparable baselines—achieving leading results in syntactic correctness, distributional alignment, expectation-value matching, and high-quality circuit yield in both Pass@1 and Pass@10 metrics. | File Name | Quant Type | File Size | | - | - | - | | rlquantum4b.BF16.gguf | BF16 | 8.05 GB | | rlquantum4b.F16.gguf | F16 | 8.05 GB | | rlquantum4b.F32.gguf | F32 | 16.1 GB | | rlquantum4b.Q2K.gguf | Q2K | 1.67 GB | | rlquantum4b.Q3KL.gguf | Q3KL | 2.24 GB | | rlquantum4b.Q3KM.gguf | Q3KM | 2.08 GB | | rlquantum4b.Q3KS.gguf | Q3KS | 1.89 GB | | rlquantum4b.Q4KM.gguf | Q4KM | 2.5 GB | | rlquantum4b.Q4KS.gguf | Q4KS | 2.38 GB | | rlquantum4b.Q5KM.gguf | Q5KM | 2.89 GB | | rlquantum4b.Q5KS.gguf | Q5KS | 2.82 GB | | rlquantum4b.Q6K.gguf | Q6K | 3.31 GB | | rlquantum4b.Q80.gguf | Q80 | 4.28 GB | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
license:apache-2.0
210
0

docscopeOCR-7B-050425-exp-GGUF

NaNK
license:apache-2.0
209
3

Camel-Doc-OCR-062825

NaNK
license:apache-2.0
207
12

Qwen2.5-VL-3B-Abliterated-Caption-it

NaNK
204
4

Camel-Doc-OCR-062825-mmp-GGUF

NaNK
license:apache-2.0
203
0

Castor-Gta6-Theme-Flux-LoRA

201
10

Fashion-Mnist-SigLIP2

NaNK
license:apache-2.0
199
4

AI Vs Deepfake Vs Real 9999

NaNK
license:apache-2.0
199
1

Qwen3-Code-Reasoning-4B-f32-GGUF

NaNK
license:apache-2.0
198
0

Flux.1-Dev-Frosted-Container-LoRA

196
11

SmolLM2-Rethink-135M-GGUF

llama
196
0

rStar-Coder-Qwen3-0.6B-GGUF

NaNK
license:apache-2.0
196
0

Seamless-Pattern-Design-Flux-LoRA

195
13

LFM2.5-350M-F32-GGUF

llama.cpp
194
1

Flood-Image-Detection

> Flood-Image-Detection is a vision-language encoder model fine-tuned from `google/siglip2-base-patch16-512` for binary image classification. It is trained to detect whether an image contains a flooded scene or non-flooded environment. The model uses the `SiglipForImageClassification` architecture. > [!note] SigLIP 2: Multilingual Vision-Language Encoders with Improved Semantic Understanding, Localization, and Dense Features : https://arxiv.org/pdf/2502.14786 Disaster Monitoring – Rapid detection of flood-affected areas from imagery. Environmental Analysis – Track flooding patterns across regions using image datasets. Crisis Response – Assist emergency services in identifying critical zones. Surveillance and Safety – Monitor infrastructure or locations for flood exposure. Smart Alert Systems – Integrate with IoT or camera feeds for automated flood alerts.

NaNK
license:apache-2.0
190
0

GRAM-LLaMA3.2-3B-RewardModel-GGUF

> GRAM-LLaMA3.2-3B-RewardModel is a generative reward model fine-tuned from the Llama-3.2-3B-Instruct base model released by NiuTrans. It is designed to improve reward generalization for large language models (LLMs) by leveraging a novel training approach that first pre-trains on large unlabeled datasets and then fine-tunes using supervised labeled data. The training uses label smoothing and optimizes a regularized ranking loss, bridging generative and discriminative reward modeling techniques. This enables the model to be applied flexibly across a variety of tasks without the usual need for extensive fine-tuning on task-specific datasets. > GRAM-LLaMA3.2-3B-RewardModel is evaluated on the JudgeBench benchmark, which covers domains such as Chat, Code, Math, and Safety. It achieves a competitive average score of 69.9 across these categories, demonstrating strong capability for use as an open-source plug-and-play reward model that can align LLMs effectively without retraining reward models from scratch. The repository includes usage examples that let users directly apply this reward model for assessing and ranking the quality of AI-generated responses in an impartial manner. | Model File name | Size | QuantType | |---|---|---| | GRAM-LLaMA3.2-3B-RewardModel.BF16.gguf | 6.43 GB | BF16 | | GRAM-LLaMA3.2-3B-RewardModel.F16.gguf | 6.43 GB | F16 | | GRAM-LLaMA3.2-3B-RewardModel.F32.gguf | 12.9 GB | F32 | | GRAM-LLaMA3.2-3B-RewardModel.Q2K.gguf | 1.36 GB | Q2K | | GRAM-LLaMA3.2-3B-RewardModel.Q3KL.gguf | 1.82 GB | Q3KL | | GRAM-LLaMA3.2-3B-RewardModel.Q3KM.gguf | 1.69 GB | Q3KM | | GRAM-LLaMA3.2-3B-RewardModel.Q3KS.gguf | 1.54 GB | Q3KS | | GRAM-LLaMA3.2-3B-RewardModel.Q4KM.gguf | 2.02 GB | Q4KM | | GRAM-LLaMA3.2-3B-RewardModel.Q4KS.gguf | 1.93 GB | Q4KS | | GRAM-LLaMA3.2-3B-RewardModel.Q5KM.gguf | 2.32 GB | Q5KM | | GRAM-LLaMA3.2-3B-RewardModel.Q5KS.gguf | 2.27 GB | Q5KS | | GRAM-LLaMA3.2-3B-RewardModel.Q6K.gguf | 2.64 GB | Q6K | | GRAM-LLaMA3.2-3B-RewardModel.Q80.gguf | 3.42 GB | Q80 | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
llama
189
0

Smoothie-Qwen3-AIO-GGUF

> Smoothie-Qwen3 models are enhancements of Qwen3 language models, applying post-processing techniques to smooth token distributions and promote balanced, multilingual output, especially across varied Unicode ranges. These models are particularly effective for applications needing reduced language bias and improved representation consistency, maintaining the strong reasoning, coding, and dialogue abilities of Qwen3 while producing more stable and diverse generations. | File Name | Quant Type | File Size | | - | - | - | | Smoothie-Qwen3-0.6B.BF16.gguf | BF16 | 1.2 GB | | Smoothie-Qwen3-0.6B.F16.gguf | F16 | 1.2 GB | | Smoothie-Qwen3-0.6B.F32.gguf | F32 | 2.39 GB | | Smoothie-Qwen3-0.6B.Q2K.gguf | Q2K | 296 MB | | Smoothie-Qwen3-0.6B.Q3KL.gguf | Q3KL | 368 MB | | Smoothie-Qwen3-0.6B.Q3KM.gguf | Q3KM | 347 MB | | Smoothie-Qwen3-0.6B.Q3KS.gguf | Q3KS | 323 MB | | Smoothie-Qwen3-0.6B.Q40.gguf | Q40 | 382 MB | | Smoothie-Qwen3-0.6B.Q41.gguf | Q41 | 409 MB | | Smoothie-Qwen3-0.6B.Q4K.gguf | Q4K | 397 MB | | Smoothie-Qwen3-0.6B.Q4KM.gguf | Q4KM | 397 MB | | Smoothie-Qwen3-0.6B.Q4KS.gguf | Q4KS | 383 MB | | Smoothie-Qwen3-0.6B.Q50.gguf | Q50 | 437 MB | | Smoothie-Qwen3-0.6B.Q51.gguf | Q51 | 464 MB | | Smoothie-Qwen3-0.6B.Q5K.gguf | Q5K | 444 MB | | Smoothie-Qwen3-0.6B.Q5KM.gguf | Q5KM | 444 MB | | Smoothie-Qwen3-0.6B.Q5KS.gguf | Q5KS | 437 MB | | Smoothie-Qwen3-0.6B.Q6K.gguf | Q6K | 495 MB | | Smoothie-Qwen3-0.6B.Q80.gguf | Q80 | 639 MB | | File Name | Quant Type | File Size | | - | - | - | | Smoothie-Qwen3-1.7B.BF16.gguf | BF16 | 3.45 GB | | Smoothie-Qwen3-1.7B.F16.gguf | F16 | 3.45 GB | | Smoothie-Qwen3-1.7B.F32.gguf | F32 | 6.89 GB | | Smoothie-Qwen3-1.7B.Q2K.gguf | Q2K | 778 MB | | Smoothie-Qwen3-1.7B.Q3KL.gguf | Q3KL | 1 GB | | Smoothie-Qwen3-1.7B.Q3KM.gguf | Q3KM | 940 MB | | Smoothie-Qwen3-1.7B.Q3KS.gguf | Q3KS | 867 MB | | Smoothie-Qwen3-1.7B.Q40.gguf | Q40 | 1.05 GB | | Smoothie-Qwen3-1.7B.Q41.gguf | Q41 | 1.14 GB | | Smoothie-Qwen3-1.7B.Q4K.gguf | Q4K | 1.11 GB | | Smoothie-Qwen3-1.7B.Q4KM.gguf | Q4KM | 1.11 GB | | Smoothie-Qwen3-1.7B.Q4KS.gguf | Q4KS | 1.06 GB | | Smoothie-Qwen3-1.7B.Q50.gguf | Q50 | 1.23 GB | | Smoothie-Qwen3-1.7B.Q51.gguf | Q51 | 1.32 GB | | Smoothie-Qwen3-1.7B.Q5K.gguf | Q5K | 1.26 GB | | Smoothie-Qwen3-1.7B.Q5KM.gguf | Q5KM | 1.26 GB | | Smoothie-Qwen3-1.7B.Q5KS.gguf | Q5KS | 1.23 GB | | Smoothie-Qwen3-1.7B.Q6K.gguf | Q6K | 1.42 GB | | Smoothie-Qwen3-1.7B.Q80.gguf | Q80 | 1.83 GB | | File Name | Quant Type | File Size | | - | - | - | | Smoothie-Qwen3-4B.BF16.gguf | BF16 | 8.05 GB | | Smoothie-Qwen3-4B.F16.gguf | F16 | 8.05 GB | | Smoothie-Qwen3-4B.F32.gguf | F32 | 16.1 GB | | Smoothie-Qwen3-4B.Q2K.gguf | Q2K | 1.67 GB | | Smoothie-Qwen3-4B.Q3KL.gguf | Q3KL | 2.24 GB | | Smoothie-Qwen3-4B.Q3KM.gguf | Q3KM | 2.08 GB | | Smoothie-Qwen3-4B.Q3KS.gguf | Q3KS | 1.89 GB | | Smoothie-Qwen3-4B.Q4KM.gguf | Q4KM | 2.5 GB | | Smoothie-Qwen3-4B.Q4KS.gguf | Q4KS | 2.38 GB | | Smoothie-Qwen3-4B.Q5KM.gguf | Q5KM | 2.89 GB | | Smoothie-Qwen3-4B.Q5KS.gguf | Q5KS | 2.82 GB | | Smoothie-Qwen3-4B.Q6K.gguf | Q6K | 3.31 GB | | Smoothie-Qwen3-4B.Q80.gguf | Q80 | 4.28 GB | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
license:apache-2.0
188
0

Polaroid-Warm-i2i

186
2

Carinae-Qwen3-Radiation-4B-GGUF

> Carinae-Qwen3-Radiation-4B is a reasoning-focused model fine-tuned on Qwen for Abliterated Reasoning and polished token probabilities, enhancing balanced multilingual generation across mathematics and general-purpose reasoning. > It specializes in event-driven logic, structured analysis, and precise probabilistic modeling—making it an ideal tool for researchers, educators, and developers working with uncertainty and structured reasoning. | File Name | Quant Type | File Size | | - | - | - | | Carinae-Qwen3-Radiation-4B.BF16.gguf | BF16 | 8.05 GB | | Carinae-Qwen3-Radiation-4B.F16.gguf | F16 | 8.05 GB | | Carinae-Qwen3-Radiation-4B.F32.gguf | F32 | 16.1 GB | | Carinae-Qwen3-Radiation-4B.Q2K.gguf | Q2K | 1.67 GB | | Carinae-Qwen3-Radiation-4B.Q3KL.gguf | Q3KL | 2.24 GB | | Carinae-Qwen3-Radiation-4B.Q3KM.gguf | Q3KM | 2.08 GB | | Carinae-Qwen3-Radiation-4B.Q3KS.gguf | Q3KS | 1.89 GB | | Carinae-Qwen3-Radiation-4B.Q4KM.gguf | Q4KM | 2.5 GB | | Carinae-Qwen3-Radiation-4B.Q4KS.gguf | Q4KS | 2.38 GB | | Carinae-Qwen3-Radiation-4B.Q5KM.gguf | Q5KM | 2.89 GB | | Carinae-Qwen3-Radiation-4B.Q5KS.gguf | Q5KS | 2.82 GB | | Carinae-Qwen3-Radiation-4B.Q6K.gguf | Q6K | 3.31 GB | | Carinae-Qwen3-Radiation-4B.Q80.gguf | Q80 | 4.28 GB | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
license:apache-2.0
183
0

Intense-Red-Flux-LoRA

182
6

Face-Mask-Detection

> Face-Mask-Detection is a binary image classification model based on `google/siglip2-base-patch16-224`, trained to detect whether a person is wearing a face mask or not. This model can be used in public health monitoring, access control systems, and workplace compliance enforcement. The model distinguishes between the following face mask statuses: COVID-19 Compliance Monitoring Security and Access Control Automated Surveillance Systems Health Safety Enforcement in Public Spaces

NaNK
license:apache-2.0
182
0

Purple-Dreamy-Flux-LoRA

181
17

Qwen-Image-Anime-LoRA

| Parameter | Value | Parameter | Value | |---------------------------|--------|---------------------------|--------| | LR Scheduler | constant | Noise Offset | 0.03 | | Optimizer | AdamW | Multires Noise Discount | 0.1 | | Network Dim | 64 | Multires Noise Iterations | 10 | | Network Alpha | 32 | Repeat & Steps | 25 & 3000 | | Epoch | 20 | Save Every N Epochs | 1 | | Source | Link | |--------------|-------------------------------------| | Playground | playground.com | | ArtStation | artstation.com | | 4K Wallpapers| 4kwallpapers.com | | Dimensions | Aspect Ratio | Recommendation | |-----------------|------------------|---------------------------| | 1664 x 928 | 16:9 (approx.) | Best | | 1024 x 1024 | 1:1 | Default | You should use `Qwen Anime` to trigger the image generation.

license:apache-2.0
179
8

SmolLM2-135M-F32-GGUF

llama
178
1

Gliese-Query_Tool-0.6B-GGUF

> Gliese-QueryTool-0.6B is a function-calling and query-oriented reasoning model fine-tuned from Qwen3-0.6B using Salesforce/xlam-function-calling-60k, designed for tool orchestration, structured query resolution, and operation chaining across diverse tasks. It excels in dynamic function execution, structured reasoning pipelines, and multi-tool decision workflows, making it a powerful lightweight solution for developers, tooling platforms, and automation systems. Model Files | File Name | Quant Type | File Size | | - | - | - | | Gliese-QueryTool-0.6B.BF16.gguf | BF16 | 1.2 GB | | Gliese-QueryTool-0.6B.F16.gguf | F16 | 1.2 GB | | Gliese-QueryTool-0.6B.F32.gguf | F32 | 2.39 GB | | Gliese-QueryTool-0.6B.Q2K.gguf | Q2K | 296 MB | | Gliese-QueryTool-0.6B.Q3KL.gguf | Q3KL | 368 MB | | Gliese-QueryTool-0.6B.Q3KM.gguf | Q3KM | 347 MB | | Gliese-QueryTool-0.6B.Q3KS.gguf | Q3KS | 323 MB | | Gliese-QueryTool-0.6B.Q40.gguf | Q40 | 382 MB | | Gliese-QueryTool-0.6B.Q41.gguf | Q41 | 409 MB | | Gliese-QueryTool-0.6B.Q4K.gguf | Q4K | 397 MB | | Gliese-QueryTool-0.6B.Q4KM.gguf | Q4KM | 397 MB | | Gliese-QueryTool-0.6B.Q4KS.gguf | Q4KS | 383 MB | | Gliese-QueryTool-0.6B.Q50.gguf | Q50 | 437 MB | | Gliese-QueryTool-0.6B.Q51.gguf | Q51 | 464 MB | | Gliese-QueryTool-0.6B.Q5K.gguf | Q5K | 444 MB | | Gliese-QueryTool-0.6B.Q5KM.gguf | Q5KM | 444 MB | | Gliese-QueryTool-0.6B.Q5KS.gguf | Q5KS | 437 MB | | Gliese-QueryTool-0.6B.Q6K.gguf | Q6K | 495 MB | | Gliese-QueryTool-0.6B.Q80.gguf | Q80 | 639 MB | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
license:apache-2.0
178
0

Qwen3-VL-2B-Instruct-abliterated-v1

NaNK
license:apache-2.0
174
1

Polaris-4B-Preview-F32-GGUF

NaNK
license:apache-2.0
174
1

Qwen3-VL-2B-Instruct-abliterated

> Qwen3-VL-2B-Instruct-abliterated is an abliterated (v1.0) variant of Qwen3-VL-2B-Instruct, designed for Abliterated Reasoning and Captioning. > This model is optimized to generate detailed, descriptive captions and reasoning outputs across a wide range of visual and multimodal contexts—including complex, sensitive, or nuanced content—while supporting diverse aspect ratios and resolutions. Abliterated / Uncensored Captioning – Fine-tuned to bypass conventional content filters while preserving factual, descriptive, and reasoning-rich outputs. High-Fidelity Descriptions – Generates comprehensive captions and reasoning for general, artistic, technical, abstract, or low-context images. Robust Across Aspect Ratios – Performs consistently across wide, tall, square, and irregular image dimensions. Variational Detail Control – Capable of producing outputs ranging from concise summaries to fine-grained, intricate descriptions and reasoning. Foundation on Qwen3-VL-2B Architecture – Built upon Qwen3-VL-2B-Instruct’s strong multimodal reasoning and instruction-following capabilities. Multilingual Output Capability – Primarily optimized for English, with adaptability for multilingual prompts through prompt engineering. Generating detailed, uncensored captions and reasoning for general-purpose or artistic datasets. Research in content moderation, red-teaming, and generative safety evaluation. Enabling descriptive captioning and reasoning for visual datasets typically excluded from mainstream models. Creative applications such as storytelling, art generation, or multimodal reasoning tasks. Captioning and reasoning for non-standard aspect ratios and stylized visual content. May produce explicit, sensitive, or offensive descriptions depending on the image content and prompts. Not recommended for production systems requiring strict content moderation. Output style, tone, and reasoning may vary based on input phrasing. Accuracy can fluctuate for unfamiliar, synthetic, or highly abstract visual content.

NaNK
license:apache-2.0
174
1

Neumind-Math-7B-Instruct

NaNK
Ollama
173
2

FastThink-0.5B-Tiny-GGUF

NaNK
license:apache-2.0
171
0

Qwen3-VL-30B-A3B-Thinking-abliterated-v1

NaNK
license:apache-2.0
170
4

Qwen3-VL-30B-A3B-Thinking-abliterated

NaNK
license:apache-2.0
168
4

GRAM-Qwen3-4B-RewardModel-GGUF

> GRAM-Qwen3-4B-RewardModel is a generative reward model developed to address reward generalization for Large Language Models (LLMs), released by NiuTrans. Unlike traditional models that depend heavily on task-specific labeled data, this model leverages both labeled and unlabeled data—a novel approach that allows it to generalize better across various tasks. It introduces a generative reward model framework that pre-trains on large amounts of unlabeled data and is subsequently fine-tuned with supervised data. The methodology also employs label smoothing and a regularized ranking loss to further boost performance, effectively bridging the gap between generative and discriminative reward modeling techniques. > This model is built on the Qwen3-4B base and can be directly used or adapted for aligning LLMs without the need to train a reward model from scratch on extensive datasets. In evaluations on the JudgeBench benchmark—covering Chat, Code, Math, and Safety tasks—GRAM-Qwen3-4B-RewardModel achieves a competitive average score of 65.9, making it suitable for use as an open-source, plug-and-play reward model for a variety of LLM alignment scenarios. The repository provides usage instructions and demonstration code to facilitate immediate adoption for research and development purposes | Model File name | Size | QuantType | |---|---|---| | GRAM-Qwen3-4B-RewardModel.BF16.gguf | 8.05 GB | BF16 | | GRAM-Qwen3-4B-RewardModel.F16.gguf | 8.05 GB | F16 | | GRAM-Qwen3-4B-RewardModel.F32.gguf | 16.1 GB | F32 | | GRAM-Qwen3-4B-RewardModel.Q2K.gguf | 1.67 GB | Q2K | | GRAM-Qwen3-4B-RewardModel.Q3KL.gguf | 2.24 GB | Q3KL | | GRAM-Qwen3-4B-RewardModel.Q3KM.gguf | 2.08 GB | Q3KM | | GRAM-Qwen3-4B-RewardModel.Q3KS.gguf | 1.89 GB | Q3KS | | GRAM-Qwen3-4B-RewardModel.Q4KM.gguf | 2.5 GB | Q4KM | | GRAM-Qwen3-4B-RewardModel.Q4KS.gguf | 2.38 GB | Q4KS | | GRAM-Qwen3-4B-RewardModel.Q5KM.gguf | 2.89 GB | Q5KM | | GRAM-Qwen3-4B-RewardModel.Q5KS.gguf | 2.82 GB | Q5KS | | GRAM-Qwen3-4B-RewardModel.Q6K.gguf | 3.31 GB | Q6K | | GRAM-Qwen3-4B-RewardModel.Q80.gguf | 4.28 GB | Q80 | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
license:apache-2.0
168
0

Mirage-Photo-Classifier

> Mirage-Photo-Classifier is an image classification vision-language encoder model fine-tuned from google/siglip2-base-patch16-224 for a binary image authenticity classification task. It is designed to determine whether an image is real or AI-generated (fake) using the SiglipForImageClassification architecture. The Mirage-Photo-Classifier model is designed to detect whether an image is genuine (photograph) or synthetically generated. Use cases include: - AI Image Detection: Identifying AI-generated images in social media, news, or datasets. - Digital Forensics: Helping professionals detect image authenticity in investigations. - Platform Moderation: Assisting content platforms in labeling generated content. - Dataset Validation: Cleaning and verifying training data for other AI models.

NaNK
license:apache-2.0
166
2

Codepy-Deepthink-3B-GGUF

NaNK
llama
165
7

Qwen-Image-2512-Pixel-Art-LoRA

license:apache-2.0
165
5

Llama-3.1-8B-Open-SFT-GGUF

NaNK
llama
165
1

coreOCR-7B-050325-preview

NaNK
license:apache-2.0
163
12

QVikhr-3-4B-it-F32-GGUF

NaNK
license:apache-2.0
163
0

GRAM-Qwen3-1.7B-RewardModel-GGUF

> GRAM-Qwen3-1.7B-RewardModel is a generative reward model developed by NiuTrans that follows a two-step training approach: it first pre-trains on a large amount of unlabeled data and then fine-tunes with supervised labeled data. This methodology, which incorporates label smoothing and a regularized ranking loss, enables effective reward generalization for large language models (LLMs). The model is built on the Qwen3-1.7B base, a compact language model with 1.7 billion parameters, 28 layers, and attention heads designed to handle long-context inputs (up to 32,768 tokens) and support both detailed reasoning and fast responses. GRAM-Qwen3-1.7B-RewardModel is intended for flexible application across diverse tasks, providing an open-source, plug-and-play reward model for aligning LLM outputs without requiring extensive task-specific retraining. It excels in evaluating and ranking the quality of AI-generated responses, operating effectively as a judge model in AI alignment scenarios. | Model File name | Size | QuantType | |---|---|---| | GRAM-Qwen3-1.7B-RewardModel.BF16.gguf | 3.45 GB | BF16 | | GRAM-Qwen3-1.7B-RewardModel.F16.gguf | 3.45 GB | F16 | | GRAM-Qwen3-1.7B-RewardModel.F32.gguf | 6.89 GB | F32 | | GRAM-Qwen3-1.7B-RewardModel.Q2K.gguf | 778 MB | Q2K | | GRAM-Qwen3-1.7B-RewardModel.Q3KL.gguf | 1 GB | Q3KL | | GRAM-Qwen3-1.7B-RewardModel.Q3KM.gguf | 940 MB | Q3KM | | GRAM-Qwen3-1.7B-RewardModel.Q3KS.gguf | 867 MB | Q3KS | | GRAM-Qwen3-1.7B-RewardModel.Q4KM.gguf | 1.11 GB | Q4KM | | GRAM-Qwen3-1.7B-RewardModel.Q4KS.gguf | 1.06 GB | Q4KS | | GRAM-Qwen3-1.7B-RewardModel.Q5KM.gguf | 1.26 GB | Q5KM | | GRAM-Qwen3-1.7B-RewardModel.Q5KS.gguf | 1.23 GB | Q5KS | | GRAM-Qwen3-1.7B-RewardModel.Q6K.gguf | 1.42 GB | Q6K | | GRAM-Qwen3-1.7B-RewardModel.Q80.gguf | 1.83 GB | Q80 | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
license:apache-2.0
163
0

Llama-3.2-3B-Promptist-Mini-GGUF

NaNK
llama
160
7

docscopeOCR-7B-050425-exp

NaNK
license:apache-2.0
158
7

Monochrome-Pencil

158
4

Qwen3-VL-2B-Thinking-abliterated-v1

NaNK
license:apache-2.0
158
2

Qwen3-VL-2B-Thinking-abliterated

> Qwen3-VL-2B-Thinking-abliterated is an abliterated (v1.0) variant of Qwen3-VL-2B-Thinking, designed for Abliterated Reasoning and Captioning. > This model is optimized to generate detailed, descriptive captions and reasoning outputs across a wide range of visual and multimodal contexts—including complex, sensitive, or nuanced content—while supporting diverse aspect ratios and resolutions. Abliterated / Uncensored Captioning – Fine-tuned to bypass conventional content filters while preserving factual, descriptive, and reasoning-rich outputs. High-Fidelity Descriptions – Generates comprehensive captions and reasoning for general, artistic, technical, abstract, or low-context images. Robust Across Aspect Ratios – Performs consistently across wide, tall, square, and irregular image dimensions. Variational Detail Control – Capable of producing outputs ranging from concise summaries to fine-grained, intricate descriptions and reasoning. Foundation on Qwen3-VL-2B-Thinking Architecture – Built upon Qwen3-VL-2B-Thinking’s strong multimodal reasoning and instruction-following capabilities. Multilingual Output Capability – Primarily optimized for English, with adaptability for multilingual prompts through prompt engineering. Generating detailed, uncensored captions and reasoning for general-purpose or artistic datasets. Research in content moderation, red-teaming, and generative safety evaluation. Enabling descriptive captioning and reasoning for visual datasets typically excluded from mainstream models. Creative applications such as storytelling, art generation, or multimodal reasoning tasks. Captioning and reasoning for non-standard aspect ratios and stylized visual content. May produce explicit, sensitive, or offensive descriptions depending on the image content and prompts. Not recommended for production systems requiring strict content moderation. Output style, tone, and reasoning may vary based on input phrasing. Accuracy can fluctuate for unfamiliar, synthetic, or highly abstract visual content.

NaNK
license:apache-2.0
158
2

Smoothie-Qwen3-0.6B-F32-GGUF

NaNK
license:apache-2.0
155
0

Qwen3-VL-8B-Thinking-abliterated

> Qwen3-VL-8B-Thinking-abliterated is an abliterated (v1.0) variant of Qwen3-VL-8B-Thinking, designed for Abliterated Reasoning and Captioning. > This model produces detailed captions and reasoning outputs across a wide range of visual and multimodal contexts, including complex, sensitive, or nuanced content. It supports diverse aspect ratios, resolutions, and prompt conditions while maintaining reasoning integrity and descriptive precision. Abliterated / Uncensored Captioning Fine-tuned to bypass conventional content filters while preserving factual, descriptive, and reasoning-rich outputs. High-Fidelity Reasoning and Descriptions Generates comprehensive captions and reasoning for general, artistic, technical, abstract, and low-context images. Robust Across Aspect Ratios Performs consistently on wide, tall, square, panoramic, and irregular image dimensions. Variational Detail Control Produces outputs ranging from concise summaries to fine-grained, high-context reasoning and descriptions. Foundation on Qwen3-VL-8B-Thinking Architecture Built upon the Qwen3-VL-8B-Thinking model’s advanced multimodal reasoning and instruction-following capabilities. Multilingual Output Capability Defaults to English but can be adapted to other languages via prompt engineering. Generating detailed, uncensored captions and reasoning for general-purpose, artistic, or research-oriented datasets. Research in content moderation, red-teaming, and generative safety analysis. Enabling descriptive captioning and reasoning for datasets typically excluded from mainstream models. Creative applications such as visual storytelling, art description, and multimodal reasoning exploration. Captioning and reasoning for images with non-standard or stylized visual structures. May generate explicit, sensitive, or offensive content depending on prompts and image input. Not suitable for production systems that require strict content moderation. Output style, tone, and reasoning depth may vary based on input phrasing. Accuracy may fluctuate for abstract, synthetic, or highly stylized visuals.

NaNK
license:apache-2.0
154
3

SD3.5-Large-Turbo-HyperRealistic-LoRA

153
22

GiD-Land-Cover-Classification

> GiD-Land-Cover-Classification is a multi-class image classification model based on `google/siglip2-base-patch16-224`, trained to detect land cover types in geographical or environmental imagery. This model can be used for urban planning, agriculture monitoring, and environmental analysis. The model distinguishes between the following land cover types: Urban Development Planning Agricultural Monitoring Land Use and Land Cover (LULC) Mapping Disaster Management and Flood Risk Analysis

NaNK
license:apache-2.0
153
0

Dots.OCR-Latest-BF16

> Dots.OCR-Latest-BF16 is an optimized and updated vision-language OCR model variant of the original Dots.OCR. This open-source model is designed to extract text from images and scanned documents, including handwritten and printed content. It can output results as plain text or Markdown, preserving document layout elements such as headings, tables, and lists. This model uses a powerful multimodal backbone (3B VLM) to enhance reading comprehension and layout understanding, handling cursive handwriting and complex document structures effectively. The BF16 variant has been tested and updated to work smoothly with the latest `transformers` version without compatibility issues, ensuring optimized performance. | Resource Type | Description | Link | |----------------|--------------|------| | Original Model Card | Official release of Dots.OCR by rednote-hilab | rednote-hilab/dots.ocr | | Test Model (StrangerZone HF) | Community test deployment (experimental) | strangervisionhf/dots.ocr-base-fix | | Standard Model Card | Optimized version supporting Transformers v4.57.1 (BF16 precision) | prithivMLmods/Dots.OCR-Latest-BF16 | | Demo Space | Interactive demo hosted on Hugging Face Spaces | Multimodal-OCR3 Demo |

license:mit
152
1

Acrux-500M-o1-Journey-GGUF

Ollama
151
3

Llama-3.2-3B-Instruct-GGUF

NaNK
llama
150
2

qwen3-4b-code-reasoning-f32-GGUF

NaNK
license:apache-2.0
150
1

Capella-Qwen3-DS-V3.1-4B-GGUF

NaNK
license:apache-2.0
150
0

Qwen3-4B-Thinking-2507-GGUF

NaNK
llama.cpp
149
1

Lucy-f32-GGUF

license:apache-2.0
148
1

Poseidon-Reasoning-1.7B-GGUF

NaNK
license:apache-2.0
147
0

Light-IF-4B-f32-GGUF

> Light-IF-4B is a 4-billion-parameter instruction-following language model derived from Qwen3-4B-Base, designed to overcome "lazy reasoning" in complex tasks by incorporating previewing and self-checking during inference; it is fine-tuned using entropy-preserving supervised learning (Entropy-SFT) and token-wise entropy-adaptive reinforcement learning (TEA-RL) on a carefully filtered dataset, producing strong results across instruction-following and reasoning benchmarks (such as SuperClue, IFEval, and IFBench), where it matches or outperforms even larger or closed-source models, and supports advanced features such as extended context (32k-131k tokens with YaRN), efficient deployment (via Hugging Face Transformers, sglang, or vllm), and open integration for research in robust generalizable reasoning, with further details, evaluation code, and licensing on its official Hugging Face repository and paper. | File Name | Size | Quant Type | |-----------|------|------------| | Qwen3-4B-MegaScience.BF16.gguf | 8.05 GB | BF16 | | Qwen3-4B-MegaScience.F16.gguf | 8.05 GB | F16 | | Qwen3-4B-MegaScience.F32.gguf | 16.1 GB | F32 | | Qwen3-4B-MegaScience.Q3KL.gguf | 2.24 GB | Q3KL | | Qwen3-4B-MegaScience.Q3KS.gguf | 1.89 GB | Q3KS | | Qwen3-4B-MegaScience.Q4KM.gguf | 2.5 GB | Q4KM | | Qwen3-4B-MegaScience.Q4KS.gguf | 2.38 GB | Q4KS | | Qwen3-4B-MegaScience.Q5KM.gguf | 2.89 GB | Q5KM | | Qwen3-4B-MegaScience.Q5KS.gguf | 2.82 GB | Q5KS | | Qwen3-4B-MegaScience.Q6K.gguf | 3.31 GB | Q6K | | Qwen3-4B-MegaScience.Q80.gguf | 4.28 GB | Q80 | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
license:apache-2.0
146
0

Qwen3-VL-8B-Abliterated-Caption-it

> The Qwen3-VL-8B-Abliterated-Caption-it model is a fine-tuned version of Qwen3-VL-8B-Instruct, tailored for Abliterated Captioning / Uncensored Image Captioning. This variant is designed to generate highly detailed and descriptive captions across a broad range of visual categories, including images with complex, sensitive, or nuanced content—across varying aspect ratios and resolutions. Abliterated / Uncensored Captioning: Fine-tuned to bypass common content filters while preserving factual and descriptive richness across diverse visual categories. High-Fidelity Descriptions: Generates comprehensive captions for general, artistic, technical, abstract, and low-context images. Robust Across Aspect Ratios: Capable of accurately captioning images with wide, tall, square, and irregular dimensions. Variational Detail Control: Produces outputs with both high-level summaries and fine-grained descriptions as needed. Foundation on Qwen3-VL Architecture: Leverages the strengths of the Qwen3-VL-8B multimodal model for visual reasoning, comprehension, and instruction-following. Multilingual Output Capability: Supports multilingual descriptions (English as default), adaptable via prompt engineering. > [!note] > Instruction Query: Provide a detailed caption for the image Generating detailed and unfiltered image captions for general-purpose or artistic datasets. Content moderation research, red-teaming, and generative safety evaluations. Enabling descriptive captioning for visual datasets typically excluded from mainstream models. Creative applications (e.g., storytelling, art generation) that benefit from rich descriptive captions. Captioning for non-standard aspect ratios and stylized visual content. May produce explicit, sensitive, or offensive descriptions depending on image content and prompts. Not suitable for deployment in production systems requiring content filtering or moderation. Can exhibit variability in caption tone or style depending on input prompt phrasing. Accuracy for unfamiliar or synthetic visual styles may vary.

NaNK
license:apache-2.0
144
6

Hand-Gesture-19

NaNK
license:apache-2.0
144
2

Qwen3-VL-4B-Instruct-Unredacted-MAX

NaNK
license:apache-2.0
142
7

Yellow-Laser-Flux-LoRA

140
3

Canopus-LoRA-Flux-Typography-ASCII

139
7

II-Search-CIR-4B-GGUF

> II-Search-CIR-4B is a 4-billion-parameter language model built on Qwen3-4B and enhanced with Code-Integrated Reasoning (CIR), enabling it not only to call external tools (such as web search and web visit) through code blocks during inference, but also to programmatically process, filter, and reason over results within those code blocks; optimized through supervised fine-tuning and reinforcement learning on challenging reasoning datasets, the model achieves state-of-the-art or leading results on major factual QA and information-seeking benchmarks (like OpenAI/SimpleQA, Google/Frames, and Seal0), and it can be efficiently deployed using vLLM or SGLang with up to 128k-token contexts (with YaRN RoPE scaling), supporting advanced research, educational, and web-integrated applications, with datasets, code samples, and evaluation results provided in the official Hugging Face repository. | File Name | Size | Quant Type | |-----------|------|------------| | II-Search-4B-GGUF.BF16.gguf | 8.05 GB | BF16 | | II-Search-4B-GGUF.F16.gguf | 8.05 GB | F16 | | II-Search-4B-GGUF.F32.gguf | 16.1 GB | F32 | | II-Search-4B-GGUF.Q2K.gguf | 1.67 GB | Q2K | | II-Search-4B-GGUF.Q3KL.gguf | 2.24 GB | Q3KL | | II-Search-4B-GGUF.Q3KM.gguf | 2.08 GB | Q3KM | | II-Search-4B-GGUF.Q3KS.gguf | 1.89 GB | Q3KS | | II-Search-4B-GGUF.Q4KM.gguf | 2.5 GB | Q4KM | | II-Search-4B-GGUF.Q4KS.gguf | 2.38 GB | Q4KS | | II-Search-4B-GGUF.Q5KM.gguf | 2.89 GB | Q5KM | | II-Search-4B-GGUF.Q5KS.gguf | 2.82 GB | Q5KS | | II-Search-4B-GGUF.Q6K.gguf | 3.31 GB | Q6K | | II-Search-4B-GGUF.Q80.gguf | 4.28 GB | Q80 | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
license:apache-2.0
139
2

Muscae-Qwen3-UI-Code-4B-GGUF

NaNK
license:apache-2.0
139
0

Qwen2-VL-2B-Abliterated-Caption-it

> The Qwen2-VL-2B-Abliterated-Caption-it model is a fine-tuned version of Qwen2-VL-2B-Instruct, tailored for Abliterated Captioning / Uncensored Image Captioning. This variant is designed to generate highly detailed and descriptive captions across a broad range of visual categories, including images with complex, sensitive, or nuanced content—across varying aspect ratios and resolutions. Abliterated / Uncensored Captioning: Fine-tuned to bypass common content filters while preserving factual and descriptive richness across diverse visual categories. High-Fidelity Descriptions: Generates comprehensive captions for general, artistic, technical, abstract, and low-context images. Robust Across Aspect Ratios: Capable of accurately captioning images with wide, tall, square, and irregular dimensions. Variational Detail Control: Produces outputs with both high-level summaries and fine-grained descriptions as needed. Foundation on Qwen2-VL Architecture: Leverages the strengths of the Qwen2-VL-2B multimodal model for visual reasoning, comprehension, and instruction-following. Multilingual Output Capability: Can support multilingual descriptions (English as default), adaptable via prompt engineering. This model was fine-tuned using the following datasets: prithivMLmods/blip3o-caption-mini-arrow prithivMLmods/Caption3o-Opt-v2 prithivMLmods/Caption3o-LongCap-v4 Private/unlisted datasets curated for uncensored and domain-specific image captioning tasks. The training objective focused on enhancing performance in unconstrained, descriptive image captioning—especially for edge cases commonly filtered out in standard captioning benchmarks. > \[!note] > General Query: Caption the image precisely. | Demo | |------| | [](https://huggingface.co/prithivMLmods/Qwen2-VL-2B-Abliterated-Caption-it/blob/main/Qwen2-VL-2B-Abliterated-Caption-it/Qwen2VL2BAbliteratedCaptionit.ipynb) | Generating detailed and unfiltered image captions for general-purpose or artistic datasets. Content moderation research, red-teaming, and generative safety evaluations. Enabling descriptive captioning for visual datasets typically excluded from mainstream models. Use in creative applications (e.g., storytelling, art generation) that benefit from rich descriptive captions. Captioning for non-standard aspect ratios and stylized visual content. May produce explicit, sensitive, or offensive descriptions depending on image content and prompts. Not suitable for deployment in production systems requiring content filtering or moderation. Can exhibit variability in caption tone or style depending on input prompt phrasing. Accuracy for unfamiliar or synthetic visual styles may vary.

NaNK
license:apache-2.0
137
4

LZO-1-Preview

135
2

Qwen2.5-Coder-7B-Instruct-GGUF

NaNK
Llama-cpp
131
2

Knitted-Character-Flux-LoRA

license:apache-2.0
130
11

QIE-2511-Zoom-Master

license:apache-2.0
130
4

WikiArt Style

> WikiArt-Style is a vision model fine-tuned from google/siglip2-base-patch16-224 using the SiglipForImageClassification architecture. It classifies art images into one of 137 painting style categories. {'0': 'Abstract Art', '1': 'Abstract Expressionism', '2': 'Academicism', '3': 'Action painting', '4': 'American Realism', '5': 'Analytical Cubism', '6': 'Analytical\xa0Realism', '7': 'Art Brut', '8': 'Art Deco', '9': 'Art Informel', '10': 'Art Nouveau (Modern)', '11': 'Automatic Painting', '12': 'Baroque', '13': 'Biedermeier', '14': 'Byzantine', '15': 'Cartographic Art', '16': 'Classicism', '17': 'Cloisonnism', '18': 'Color Field Painting', '19': 'Conceptual Art', '20': 'Concretism', '21': 'Constructivism', '22': 'Contemporary Realism', '23': 'Costumbrismo', '24': 'Cubism', '25': 'Cubo-Expressionism', '26': 'Cubo-Futurism', '27': 'Dada', '28': 'Divisionism', '29': 'Early Renaissance', '30': 'Environmental (Land) Art', '31': 'Existential Art', '32': 'Expressionism', '33': 'Fantastic Realism', '34': 'Fauvism', '35': 'Feminist Art', '36': 'Figurative Expressionism', '37': 'Futurism', '38': 'Gongbi', '39': 'Gothic', '40': 'Hard Edge Painting', '41': 'High Renaissance', '42': 'Hyper-Realism', '43': 'Ilkhanid', '44': 'Impressionism', '45': 'Indian Space painting', '46': 'Ink and wash painting', '47': 'International Gothic', '48': 'Intimism', '49': 'Japonism', '50': 'Joseon Dynasty', '51': 'Kinetic Art', '52': 'Kitsch', '53': 'Lettrism', '54': 'Light and Space', '55': 'Luminism', '56': 'Lyrical Abstraction', '57': 'Magic Realism', '58': 'Mail Art', '59': 'Mannerism (Late Renaissance)', '60': 'Mechanistic Cubism', '61': 'Metaphysical art', '62': 'Minimalism', '63': 'Miserablism', '64': 'Modernismo', '65': 'Mosan art', '66': 'Muralism', '67': 'Nanga (Bunjinga)', '68': 'Nas-Taliq', '69': 'Native Art', '70': 'Naturalism', '71': 'Naïve Art (Primitivism)', '72': 'Neo-Byzantine', '73': 'Neo-Concretism', '74': 'Neo-Dada', '75': 'Neo-Expressionism', '76': 'Neo-Figurative Art', '77': 'Neo-Rococo', '78': 'Neo-Romanticism', '79': 'Neo-baroque', '80': 'Neoclassicism', '81': 'Neoplasticism', '82': 'New Casualism', '83': 'New European Painting', '84': 'New Realism', '85': 'Nihonga', '86': 'None', '87': 'Northern Renaissance', '88': 'Nouveau Réalisme', '89': 'Op Art', '90': 'Orientalism', '91': 'Orphism', '92': 'Ottoman Period', '93': 'Outsider art', '94': 'Perceptism ', '95': 'Photorealism', '96': 'Pointillism', '97': 'Pop Art', '98': 'Post-Impressionism', '99': 'Post-Minimalism', '100': 'Post-Painterly Abstraction', '101': 'Poster Art Realism', '102': 'Precisionism', '103': 'Primitivism', '104': 'Proto Renaissance', '105': 'Purism', '106': 'Rayonism', '107': 'Realism', '108': 'Regionalism', '109': 'Renaissance', '110': 'Rococo', '111': 'Romanesque', '112': 'Romanticism', '113': 'Safavid Period', '114': 'Shin-hanga', '115': 'Social Realism', '116': 'Socialist Realism', '117': 'Spatialism', '118': 'Spectralism', '119': 'Street art', '120': 'Suprematism', '121': 'Surrealism', '122': 'Symbolism', '123': 'Synchromism', '124': 'Synthetic Cubism', '125': 'Synthetism', '126': 'Sōsaku hanga', '127': 'Tachisme', '128': 'Tenebrism', '129': 'Timurid Period', '130': 'Tonalism', '131': 'Transautomatism', '132': 'Tubism', '133': 'Ukiyo-e', '134': 'Verism', '135': 'Yamato-e', '136': 'Zen'} The model predicts one of the following painting style categories: 1. Style Classification in Machine Learning Models - Used as labels for training and evaluating models that classify artworks based on their artistic styles. - Ideal for deep learning applications involving convolutional neural networks (CNNs) or transformer-based vision models. 2. Style Transfer Applications - Acts as a style reference for neural style transfer algorithms (e.g., applying "Baroque" or "Cubism" to photos). - Can guide users to select a target style from a curated list. 3. Dataset Annotation - Used to annotate images in large datasets of paintings with consistent style names. - Ensures compatibility with datasets like WikiArt, Kaggle’s Painter by Numbers, or custom curation. 4. Educational and Exploratory Interfaces - Powers interfaces or apps for exploring art history, with filterable and searchable styles. - Great for building art recommender systems or virtual museums. 5. Generative Art Prompting - Assists in text-to-image prompting for generative models (e.g., Stable Diffusion, DALL·E) to specify desired styles. - Example: "Generate a portrait in the style of Neo-Expressionism." 6. Metadata Categorization in Art Databases - Useful for tagging and organizing artworks by style in digital archives or NFT marketplaces.

NaNK
license:apache-2.0
130
2

SmolLM2-1.7B-F32-GGUF

NaNK
llama
130
1

Llama-Sentient-3.2-3B-Instruct-GGUF

NaNK
llama
129
8

Qwen3-Bifrost-SOL-4B-GUFF

NaNK
license:apache-2.0
129
0

Human-Action-Recognition

> Human-Action-Recognition is an image classification vision-language encoder model fine-tuned from google/siglip2-base-patch16-224 for multi-class human action recognition. It uses the SiglipForImageClassification architecture to predict human activities from still images. The model categorizes images into 15 action classes: - 0: calling - 1: clapping - 2: cycling - 3: dancing - 4: drinking - 5: eating - 6: fighting - 7: hugging - 8: laughing - 9: listeningtomusic - 10: running - 11: sitting - 12: sleeping - 13: texting - 14: usinglaptop The Human-Action-Recognition model is designed to detect and classify human actions from images. Example applications: - Surveillance & Monitoring: Recognizing suspicious or specific activities in public spaces. - Sports Analytics: Identifying player activities or movements. - Social Media Insights: Understanding trends in user-posted visuals. - Healthcare: Monitoring elderly or patients for activity patterns. - Robotics & Automation: Enabling context-aware AI systems with visual understanding.

NaNK
license:apache-2.0
128
3

Eta-Aurigae-0.6B-Echelon1-GGUF

NaNK
license:apache-2.0
128
0

Qwen3-VL-32B-Thinking-abliterated-v1

NaNK
license:apache-2.0
126
4

Canum-med-Qwen3-Reasoning-GGUF

> Canum-med-Qwen3-Reasoning is an experimental medical reasoning and advisory model fine-tuned on Qwen/Qwen3-1.7B using the MTEB/rawmedrxiv dataset. It is designed to support clinical reasoning, biomedical understanding, and structured advisory outputs, making it a useful tool for researchers, educators, and medical professionals in experimental workflows. | File Name | Quant Type | File Size | | - | - | - | | Canum-med-Qwen3-Reasoning.BF16.gguf | BF16 | 3.45 GB | | Canum-med-Qwen3-Reasoning.F16.gguf | F16 | 3.45 GB | | Canum-med-Qwen3-Reasoning.F32.gguf | F32 | 6.89 GB | | Canum-med-Qwen3-Reasoning.Q2K.gguf | Q2K | 778 MB | | Canum-med-Qwen3-Reasoning.Q3KL.gguf | Q3KL | 1 GB | | Canum-med-Qwen3-Reasoning.Q3KM.gguf | Q3KM | 940 MB | | Canum-med-Qwen3-Reasoning.Q3KS.gguf | Q3KS | 867 MB | | Canum-med-Qwen3-Reasoning.Q40.gguf | Q40 | 1.05 GB | | Canum-med-Qwen3-Reasoning.Q41.gguf | Q41 | 1.14 GB | | Canum-med-Qwen3-Reasoning.Q4K.gguf | Q4K | 1.11 GB | | Canum-med-Qwen3-Reasoning.Q4KM.gguf | Q4KM | 1.11 GB | | Canum-med-Qwen3-Reasoning.Q4KS.gguf | Q4KS | 1.06 GB | | Canum-med-Qwen3-Reasoning.Q50.gguf | Q50 | 1.23 GB | | Canum-med-Qwen3-Reasoning.Q51.gguf | Q51 | 1.32 GB | | Canum-med-Qwen3-Reasoning.Q5K.gguf | Q5K | 1.26 GB | | Canum-med-Qwen3-Reasoning.Q5KM.gguf | Q5KM | 1.26 GB | | Canum-med-Qwen3-Reasoning.Q5KS.gguf | Q5KS | 1.23 GB | | Canum-med-Qwen3-Reasoning.Q6K.gguf | Q6K | 1.42 GB | | Canum-med-Qwen3-Reasoning.Q80.gguf | Q80 | 1.83 GB | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

license:apache-2.0
126
0

Qwen3-VL-32B-Thinking-abliterated

> Qwen3-VL-32B-Thinking-abliterated is an abliterated (v1.0) variant of Qwen3-VL-32B-Thinking, designed for Abliterated Reasoning and Captioning. > This model is optimized to generate detailed, descriptive captions and reasoning outputs across a wide range of visual and multimodal contexts—including complex, sensitive, or nuanced content—while supporting diverse aspect ratios and resolutions. Abliterated / Uncensored Captioning – Fine-tuned to bypass conventional content filters while preserving factual, descriptive, and reasoning-rich outputs. High-Fidelity Descriptions – Generates comprehensive captions and reasoning for general, artistic, technical, abstract, or low-context images. Robust Across Aspect Ratios – Performs consistently across wide, tall, square, and irregular image dimensions. Variational Detail Control – Capable of producing outputs ranging from concise summaries to fine-grained, intricate descriptions and reasoning. Foundation on Qwen3-VL-32B Architecture – Built upon Qwen3-VL-32B-Thinking’s advanced multimodal reasoning and instruction-following capabilities. Multilingual Output Capability – Primarily optimized for English, with adaptability for multilingual prompts through prompt engineering. Generating detailed, uncensored captions and reasoning for general-purpose or artistic datasets. Research in content moderation, red-teaming, and generative safety evaluation. Enabling descriptive captioning and reasoning for visual datasets typically excluded from mainstream models. Creative applications such as storytelling, art generation, or multimodal reasoning tasks. Captioning and reasoning for non-standard aspect ratios and stylized visual content. May produce explicit, sensitive, or offensive descriptions depending on the image content and prompts. Not recommended for production systems requiring strict content moderation. Output style, tone, and reasoning may vary based on input phrasing. Accuracy can fluctuate for unfamiliar, synthetic, or highly abstract visual content.

NaNK
license:apache-2.0
124
4

SmolLM2-360M-GGUF

llama
124
1

Common-Voice-Gender-Detection-ONNX

This is an ONNX version of prithivMLmods/Common-Voice-Gender-Detection. It was automatically converted and uploaded using this space. > Common-Voice-Gender-Detection is a fine-tuned version of `facebook/wav2vec2-base-960h` for binary audio classification, specifically trained to detect speaker gender as female or male. This model leverages the `Wav2Vec2ForSequenceClassification` architecture for efficient and accurate voice-based gender classification. > [!note] Wav2Vec2: Self-Supervised Learning for Speech Recognition : https://arxiv.org/pdf/2006.11477 Speech Analytics – Assist in analyzing speaker demographics in call centers or customer service recordings. Conversational AI Personalization – Adjust tone or dialogue based on gender detection for more personalized voice assistants. Voice Dataset Curation – Automatically tag or filter voice datasets by speaker gender for better dataset management. Research Applications – Enable linguistic and acoustic research involving gender-specific speech patterns. Multimedia Content Tagging – Automate metadata generation for gender identification in podcasts, interviews, or video content.

NaNK
license:apache-2.0
123
1

UIGEN-T3-4B-Preview-MAX-GGUF

NaNK
license:apache-2.0
121
0

Gym-Workout-Classifier-SigLIP2

NaNK
license:apache-2.0
119
5

LatexMind-2B-Codec-GGUF

NaNK
license:apache-2.0
118
8

Galactic-Qwen-14B-Exp2

Galactic-Qwen-14B-Exp2 is based on the Qwen 2.5 14B modality architecture, designed to enhance the reasoning capabilities of 14B-parameter models. This model is optimized for general-purpose reasoning and answering, excelling in contextual understanding, logical deduction, and multi-step problem-solving. It has been fine-tuned using a long chain-of-thought reasoning model and specialized datasets to improve comprehension, structured responses, and conversational intelligence. Key Improvements 1. Enhanced General Knowledge: The model provides broad knowledge across various domains, improving capabilities in answering questions accurately and generating coherent responses. 2. Improved Instruction Following: Significant advancements in understanding and following complex instructions, generating structured responses, and maintaining coherence over extended interactions. 3. Versatile Adaptability: More resilient to diverse prompts, enhancing its ability to handle a wide range of topics and conversation styles, including open-ended and structured inquiries. 4. Long-Context Support: Supports up to 128K tokens for input context and can generate up to 8K tokens in a single output, making it ideal for detailed responses. 5. Multilingual Proficiency: Supports over 29 languages, including English, Chinese, French, Spanish, Portuguese, German, Italian, Russian, Japanese, Korean, Vietnamese, Thai, Arabic, and more. Here is a code snippet with `applychattemplate` to show you how to load the tokenizer and model and generate content: Intended Use 1. General-Purpose Reasoning: Designed for broad applicability, assisting with logical reasoning, answering diverse questions, and solving general knowledge problems. 2. Educational and Informational Assistance: Suitable for providing explanations, summaries, and research-based responses for students, educators, and general users. 3. Conversational AI and Chatbots: Ideal for building intelligent conversational agents that require contextual understanding and dynamic response generation. 4. Multilingual Applications: Supports global communication, translations, and multilingual content generation. 5. Structured Data Processing: Capable of analyzing and generating structured outputs, such as tables and JSON, useful for data science and automation. 6. Long-Form Content Generation: Can generate extended responses, including articles, reports, and guides, maintaining coherence over large text outputs. Limitations 1. Hardware Requirements: Requires high-memory GPUs or TPUs due to its large parameter size and long-context support. 2. Potential Bias in Responses: While designed to be neutral, outputs may still reflect biases present in training data. 3. Inconsistent Outputs in Creative Tasks: May produce variable results in storytelling and highly subjective topics. 4. Limited Real-World Awareness: Does not have access to real-time events beyond its training cutoff. 5. Error Propagation in Extended Outputs: Minor errors in early responses may affect overall coherence in long-form outputs. 6. Prompt Sensitivity: The effectiveness of responses may depend on how well the input prompt is structured. Open LLM Leaderboard Evaluation Results Detailed results can be found here! Summarized results can be found here! | Metric |Value (%)| |-------------------|--------:| |Average | 43.56| |IFEval (0-Shot) | 66.20| |BBH (3-Shot) | 59.92| |MATH Lvl 5 (4-Shot)| 34.74| |GPQA (0-shot) | 19.91| |MuSR (0-shot) | 28.49| |MMLU-PRO (5-shot) | 52.12|

NaNK
license:apache-2.0
116
10

Spiral-Qwen3-4B-F32-GGUF

NaNK
license:apache-2.0
116
3

MemReader-4B-f32-GGUF

NaNK
license:apache-2.0
116
0

Flux.1-Dev-Hand-Sticky-LoRA

113
10

DeepCaption-VLA-V2.0-7B

NaNK
license:apache-2.0
113
7

Canopus-Snoopy-Charlie-Brown-Flux-LoRA

113
6

Qwen3.5-9B-abliterated-v2-MAX

NaNK
llama.cpp
113
2

OpenR1-Distill-7B-F32-GGUF

NaNK
license:apache-2.0
113
1

Outfit Cut Specified

Outfit-Cut is an adapter for black-forest-lab's FLUX.1-Kontext-dev, designed to extract outfits from images based on precisely specified subjects. The model was trained on 200 image pairs (100 start images and 100 end images). The adapter can be triggered with the following prompt: > [!note] [photo content], extract only the specified clothing item [full outfit, top wear, bottom wear, t-shirt, jacket, dress, etc.] from the image and place it over a clean, plain background. Present the result in a product photography style — well-lit, crisp, and professional — while preserving the garment’s original textures, colors, shapes, and fine details. > Prompt: [photo content], extract only the specified clothing item [t-shirt] from the image and place it over a clean, plain background. Present the result in a product photography style — well-lit, crisp, and professional — while preserving the garment’s original textures, colors, shapes, and fine details. > Prompt: [photo content], extract only the specified clothing item [top wear] from the image and place it over a clean, plain background. Present the result in a product photography style — well-lit, crisp, and professional — while preserving the garment’s original textures, colors, shapes, and fine details. > Prompt: [photo content], extract only the specified clothing item [t-shirt] from the image and place it over a clean, plain background. Present the result in a product photography style — well-lit, crisp, and professional — while preserving the garment’s original textures, colors, shapes, and fine details. > Prompt: [photo content], extract only the specified clothing item [t-shirt] from the image and place it over a clean, plain background. Present the result in a product photography style — well-lit, crisp, and professional — while preserving the garment’s original textures, colors, shapes, and fine details. > Prompt: [photo content], extract only the specified clothing item [t-shirt] from the image and place it over a clean, plain background. Present the result in a product photography style — well-lit, crisp, and professional — while preserving the garment’s original textures, colors, shapes, and fine details. > Prompt: [photo content], extract only the specified clothing item [t-shirt] from the image and place it over a clean, plain background. Present the result in a product photography style — well-lit, crisp, and professional — while preserving the garment’s original textures, colors, shapes, and fine details. > [!note] Note: This adapter works well for extracting top wear (t-shirts, shirts, jackets, hoodies). The model may perform sub-optimally in more challenging cases, such as full clothing extraction, poorly lit images, and other difficult scenarios. | Setting | Value | | ------------------------ | ------------------------ | | Module Type | Adapter | | Base Model | FLUX.1 Kontext Dev - fp8 | | Trigger Words | [photo content], extract only the specified clothing item [top wear, bottom wear] from the image and place it over a clean, plain background. Present the result in a product photography style — well-lit, crisp, and professional — while preserving the garment’s original textures, colors, shapes, and fine details. | | Image Processing Repeats | 50 | | Epochs | 25 | | Save Every N Epochs | 1 | Labeling: DeepCaption-VLA-7B(natural language & English) Total Images Used for Training : 200 Image Pairs (100 Start, 100 End) | Setting | Value | | --------------------------- | --------- | | Seed | - | | Clip Skip | - | | Text Encoder LR | 0.00001 | | UNet LR | 0.00005 | | LR Scheduler | constant | | Optimizer | AdamW8bit | | Network Dimension | 64 | | Network Alpha | 32 | | Gradient Accumulation Steps | - | | Setting | Value | | --------------- | ----- | | Shuffle Caption | - | | Keep N Tokens | - | | Setting | Value | | ------------------------- | ----- | | Noise Offset | 0.03 | | Multires Noise Discount | 0.1 | | Multires Noise Iterations | 10 | | Conv Dimension | - | | Conv Alpha | - | | Batch Size | - | | Steps | 2700 | | Sampler | euler | You should use `[photo content]` to trigger the image generation. You should use `extract only the specified clothing item [full outfit` to trigger the image generation. You should use `top wear` to trigger the image generation. You should use `bottom wear` to trigger the image generation. You should use `t-shirt` to trigger the image generation. You should use `jacket` to trigger the image generation. You should use `dress` to trigger the image generation. You should use `etc.] from the image and place it over a clean` to trigger the image generation. You should use `plain background. Present the result in a product photography style — well-lit` to trigger the image generation. You should use `crisp` to trigger the image generation. You should use `and professional — while preserving the garment’s original textures` to trigger the image generation. You should use `colors` to trigger the image generation. You should use `shapes` to trigger the image generation. You should use `and fine details.` to trigger the image generation.

license:apache-2.0
112
12

llama-3.2-3b-it-grpo-250404-GGUF

> ReZero-v0.1-llama-3.2-3b-it-grpo-250404 is a research project focused on enhancing the search abilities of small language models by training them to develop robust search strategies rather than memorizing static data. The model, built on a Llama-3.2-3B backbone, interacts with multiple synthetic search engines that each have unique retrieval mechanisms, enabling it to refine queries iteratively and persist in finding exact answers using reinforcement learning. The repository provides setup instructions, including environment configuration and dependency installation, as well as scripts to train the model or regenerate synthetic training data. Demonstrations can be run through a Gradio interface, and the release includes comprehensive experiment logs on reward strategies and search quality. The model and associated resources are open-source and accessible to the research community, with further details on experiments and references provided in the documentation. | File name | Size | Quant Type | |-----------|------|------------| | llama-3.2-3b-it-grpo-250404.F32.gguf | 12.9 GB | F32 | | llama-3.2-3b-it-grpo-250404.BF16.gguf | 6.43 GB | BF16 | | llama-3.2-3b-it-grpo-250404.F16.gguf | 6.43 GB | F16 | | llama-3.2-3b-it-grpo-250404.Q80.gguf | 3.42 GB | Q80 | | llama-3.2-3b-it-grpo-250404.Q6K.gguf | 2.64 GB | Q6K | | llama-3.2-3b-it-grpo-250404.Q5KM.gguf | 2.32 GB | Q5KM | | llama-3.2-3b-it-grpo-250404.Q5KS.gguf | 2.27 GB | Q5KS | | llama-3.2-3b-it-grpo-250404.Q4KM.gguf | 2.02 GB | Q4KM | | llama-3.2-3b-it-grpo-250404.Q4KS.gguf | 1.93 GB | Q4KS | | llama-3.2-3b-it-grpo-250404.Q3KL.gguf | 1.82 GB | Q3KL | | llama-3.2-3b-it-grpo-250404.Q3KM.gguf | 1.69 GB | Q3KM | | llama-3.2-3b-it-grpo-250404.Q3KS.gguf | 1.54 GB | Q3KS | | llama-3.2-3b-it-grpo-250404.Q2K.gguf | 1.36 GB | Q2K | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
llama
112
0

Imgscope-OCR-2B-0527

> The Imgscope-OCR-2B-0527 model is a fine-tuned version of Qwen2-VL-2B-Instruct, specifically optimized for messy handwriting recognition, document OCR, realistic handwritten OCR, and math problem solving with LaTeX formatting. This model is trained on custom datasets for document and handwriting OCR tasks and integrates a conversational approach with strong visual and textual understanding for multi-modal applications. > [!note] Colab Demo : https://huggingface.co/prithivMLmods/Imgscope-OCR-2B-0527/blob/main/Imgscope%20OCR%202B%200527%20Demo/Imgscope-OCR-2B-0527.ipynb > [!note] Video Understanding Demo : https://huggingface.co/prithivMLmods/Imgscope-OCR-2B-0527/blob/main/Imgscope-OCR-2B-05270-Video-Understanding/Imgscope-OCR-2B-0527-Video-Understanding.ipynb SoTA Understanding of Images of Various Resolution & Ratio Imgscope-OCR-2B-0527 achieves state-of-the-art performance on visual understanding benchmarks such as MathVista, DocVQA, RealWorldQA, and MTVQA. Enhanced Handwriting OCR Specifically optimized for recognizing and interpreting realistic and messy handwriting with high accuracy. Ideal for digitizing handwritten documents and notes. Document OCR Fine-Tuning Fine-tuned with curated and realistic document OCR datasets, enabling accurate extraction of text from various structured and unstructured layouts. Understanding Videos of 20+ Minutes Capable of processing long videos for video-based question answering, transcription, and content generation. Device Control Agent Supports decision-making and control capabilities for integration with mobile devices, robots, and automation systems using visual-textual commands. Multilingual OCR Support In addition to English and Chinese, the model supports OCR in multiple languages including European languages, Japanese, Korean, Arabic, and Vietnamese. Fine-tuned for complex and hard-to-read handwritten inputs using real-world handwriting datasets. Accurately extracts text from structured documents, including scanned pages, forms, and academic papers. Combines vision-language capabilities for tasks like captioning, answering image-based queries, and understanding image+text prompts. Converts mathematical expressions and problem-solving steps into LaTeX format. Supports dialogue-based reasoning, retaining context for follow-up questions. Accepts inputs from videos, images, or combined media with text, and generates relevant output accordingly. Handwritten and printed document digitization OCR pipelines for educational institutions and businesses Academic and scientific content parsing, especially math-heavy documents Assistive tools for visually impaired users Robotic and mobile automation agents interpreting screen or camera data Multilingual OCR processing for document translation or archiving

NaNK
license:apache-2.0
111
2

Forest-Fire-Detection

> `Forest-Fire-Detection` is a vision-language encoder model fine-tuned from `google/siglip2-base-patch16-512` for multi-class image classification. It is trained to detect whether an image contains fire, smoke, or a normal (non-fire) scene. The model uses the `SiglipForImageClassification` architecture. > [!note] SigLIP 2: Multilingual Vision-Language Encoders with Improved Semantic Understanding, Localization, and Dense Features : https://arxiv.org/pdf/2502.14786 Wildfire Monitoring – Rapid identification of forest fire and smoke zones. Environmental Protection – Surveillance of forest areas for early fire warning. Disaster Management – Support in emergency response and evacuation decisions. Smart Surveillance – Integrate with drones or camera feeds for automated fire detection. Research and Analysis – Analyze visual datasets for fire-prone region identification.

NaNK
license:apache-2.0
110
0

Qwen-Image-HeadshotX

> [!note] Qwen-Image-HeadshotX is a super-realistic headshot adapter for Qwen-Image, an image generation model by Qwen. It is an advanced LoRA adaptation of the Qwen-Image model and an upgraded version of Qwen-Image-Studio-Realism, offering more precise portrait rendering with a strong focus on realism. The model was trained on diverse face types from across the world, labeled with florence2-en and caption-optimized using DeepCaption-VLA-7B. Total Images Used for Training: 55 RAW images \[11(types) × 5 different face types: Asian, Hispanic, Caucasian, Latina, Middle Eastern, etc.]. | Parameter | Value | Parameter | Value | |---------------------------|--------|---------------------------|--------| | LR Scheduler | constant | Noise Offset | 0.03 | | Optimizer | AdamW | Multires Noise Discount | 0.1 | | Network Dim | 64 | Multires Noise Iterations | 10 | | Network Alpha | 32 | Repeat & Steps | 25 & 4000 | | Epoch | 30 | Save Every N Epochs | 1 | Labeling: florence2-en(natural language & English) + 🔥 Optimized with Long-Caption VLA Multimodal : https://huggingface.co/prithivMLmods/DeepCaption-VLA-7B Total Images Used for Training: 55 RAW [11(types)×5 different face types (Asian, Hispanic, Caucasian, Latina, Middle Eastern, etc.)] | Source | Link | |--------------|-------------------------------------| | Playground | playground.com | | ArtStation | artstation.com | | 4K Wallpapers| 4kwallpapers.com | | Dimensions | Aspect Ratio | Recommendation | |-----------------|------------------|---------------------------| | 1472 x 1140 | 4:3 (approx.) | Best | | 1024 x 1024 | 1:1 | Default | - Recommended Inference Steps: 45-50 (approx. ~ `100 Seconds Inference`) You should use `face headshot` to trigger the image generation.

license:apache-2.0
109
40

Octantis-QwenR1-1.5B

NaNK
license:apache-2.0
108
0

Gliese-OCR-7B-Post1.0-GGUF

NaNK
llama.cpp
106
1

Qwen2.5-Coder-7B-GGUF

NaNK
Llama-cpp
105
4

Canopus-Realism-LoRA

NaNK
104
10

Llama-Chat-Summary-3.2-3B-GGUF

NaNK
llama
102
4

Chinda-Qwen3-4B-F32-GGUF

NaNK
license:apache-2.0
100
2

Qwen3-4B-PlumEsper-GGUF

NaNK
license:apache-2.0
100
0

cudaLLM-8B-GGUF

NaNK
license:apache-2.0
100
0

granite-docling-258M-f32-GGUF

llama.cpp
99
1

Llama-3.2-3B-GGUF

NaNK
llama
98
2

Demeter-LongCoT-Qwen3-1.7B-GGUF

> Demeter-LongCoT-Qwen3-1.7B is a reasoning-focused model fine-tuned on Qwen/Qwen3-1.7B using the Demeter-LongCoT-400K dataset. It is designed for math and code chain-of-thought reasoning, blending symbolic precision, scientific logic, and structured output fluency—making it an effective tool for developers, educators, and researchers seeking reliable step-by-step reasoning. | File Name | Quant Type | File Size | | - | - | - | | Demeter-LongCoT-Qwen-1.7B.BF16.gguf | BF16 | 3.45 GB | | Demeter-LongCoT-Qwen-1.7B.F16.gguf | F16 | 3.45 GB | | Demeter-LongCoT-Qwen-1.7B.F32.gguf | F32 | 6.89 GB | | Demeter-LongCoT-Qwen-1.7B.Q2K.gguf | Q2K | 778 MB | | Demeter-LongCoT-Qwen-1.7B.Q3KL.gguf | Q3KL | 1 GB | | Demeter-LongCoT-Qwen-1.7B.Q3KM.gguf | Q3KM | 940 MB | | Demeter-LongCoT-Qwen-1.7B.Q3KS.gguf | Q3KS | 867 MB | | Demeter-LongCoT-Qwen-1.7B.Q40.gguf | Q40 | 1.05 GB | | Demeter-LongCoT-Qwen-1.7B.Q41.gguf | Q41 | 1.14 GB | | Demeter-LongCoT-Qwen-1.7B.Q4K.gguf | Q4K | 1.11 GB | | Demeter-LongCoT-Qwen-1.7B.Q4KM.gguf | Q4KM | 1.11 GB | | Demeter-LongCoT-Qwen-1.7B.Q4KS.gguf | Q4KS | 1.06 GB | | Demeter-LongCoT-Qwen-1.7B.Q50.gguf | Q50 | 1.23 GB | | Demeter-LongCoT-Qwen-1.7B.Q51.gguf | Q51 | 1.32 GB | | Demeter-LongCoT-Qwen-1.7B.Q5K.gguf | Q5K | 1.26 GB | | Demeter-LongCoT-Qwen-1.7B.Q5KM.gguf | Q5KM | 1.26 GB | | Demeter-LongCoT-Qwen-1.7B.Q5KS.gguf | Q5KS | 1.23 GB | | Demeter-LongCoT-Qwen-1.7B.Q6K.gguf | Q6K | 1.42 GB | | Demeter-LongCoT-Qwen-1.7B.Q80.gguf | Q80 | 1.83 GB | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
license:apache-2.0
98
2

Leporis-Qwen3-Radiation-1.7B-GGUF

> Leporis-Qwen3-Radiation-1.7B is a reasoning-focused model fine-tuned on Qwen for Abliterated Reasoning and polished token probabilities, enhancing balanced multilingual generation across mathematics and general-purpose reasoning. It specializes in event-driven logic, structured analysis, and precise probabilistic modeling—making it an ideal tool for researchers, educators, and developers working with uncertainty and structured reasoning. | File Name | Quant Type | File Size | | - | - | - | | Leporis-Qwen3-Radiation-1.7B.BF16.gguf | BF16 | 3.45 GB | | Leporis-Qwen3-Radiation-1.7B.F16.gguf | F16 | 3.45 GB | | Leporis-Qwen3-Radiation-1.7B.F32.gguf | F32 | 6.89 GB | | Leporis-Qwen3-Radiation-1.7B.Q2K.gguf | Q2K | 778 MB | | Leporis-Qwen3-Radiation-1.7B.Q3KL.gguf | Q3KL | 1 GB | | Leporis-Qwen3-Radiation-1.7B.Q3KM.gguf | Q3KM | 940 MB | | Leporis-Qwen3-Radiation-1.7B.Q3KS.gguf | Q3KS | 867 MB | | Leporis-Qwen3-Radiation-1.7B.Q40.gguf | Q40 | 1.05 GB | | Leporis-Qwen3-Radiation-1.7B.Q41.gguf | Q41 | 1.14 GB | | Leporis-Qwen3-Radiation-1.7B.Q4K.gguf | Q4K | 1.11 GB | | Leporis-Qwen3-Radiation-1.7B.Q4KM.gguf | Q4KM | 1.11 GB | | Leporis-Qwen3-Radiation-1.7B.Q4KS.gguf | Q4KS | 1.06 GB | | Leporis-Qwen3-Radiation-1.7B.Q50.gguf | Q50 | 1.23 GB | | Leporis-Qwen3-Radiation-1.7B.Q51.gguf | Q51 | 1.32 GB | | Leporis-Qwen3-Radiation-1.7B.Q5K.gguf | Q5K | 1.26 GB | | Leporis-Qwen3-Radiation-1.7B.Q5KM.gguf | Q5KM | 1.26 GB | | Leporis-Qwen3-Radiation-1.7B.Q5KS.gguf | Q5KS | 1.23 GB | | Leporis-Qwen3-Radiation-1.7B.Q6K.gguf | Q6K | 1.42 GB | | Leporis-Qwen3-Radiation-1.7B.Q80.gguf | Q80 | 1.83 GB | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
license:apache-2.0
98
2

Qwen3-VL-4B-Instruct-Unredacted-MAX-FP8

NaNK
license:apache-2.0
98
1

Canopus-Pencil-Art-LoRA

NaNK
license:apache-2.0
97
5

Procyon-1.5B-Theorem-GGUF

NaNK
license:apache-2.0
97
0

Flux.1-Dev-Quote-LoRA

96
11

Spam-Bert-Uncased

95
4

UIGEN-FX-4B-Preview-GGUF

NaNK
license:apache-2.0
95
1

Teen-Outfit

94
14

Qwen2.5-Coder-3B-GGUF

NaNK
Llama-cpp
94
3

Kimina-Prover-Distill-1.7B-F32-GGUF

NaNK
license:mit
94
1

Explora-0.6B-GGUF

NaNK
license:apache-2.0
94
0

Rice-Leaf-Disease

NaNK
license:apache-2.0
93
0

Qwen3-Reranker-4B-F32-GGUF

NaNK
license:apache-2.0
93
0

DynaGuard-AIO-GGUF

> DynaGuard is a dynamic guardrail model framework that enables user-defined content moderation policies for generative AI, providing real-time, configurable controls on both model inputs and outputs. It employs ultralightweight, efficient models optimized for on-device or cloud deployments, allowing organizations to protect against risks like data leakage, PII exposure, toxicity, jailbreaking, and prompt injection with sub-50ms latency and minimal compute overhead. Users can create custom guardrails in natural language, tailoring moderation behaviors, targeted content categories, and downstream actions, making DynaGuard ideal for enterprise AI security, compliance, and safe application development across diverse hardware platforms. | File Name | Quant Type | File Size | | - | - | - | | DynaGuard-1.7B.BF16.gguf | BF16 | 3.45 GB | | DynaGuard-1.7B.F16.gguf | F16 | 3.45 GB | | DynaGuard-1.7B.F32.gguf | F32 | 6.89 GB | | DynaGuard-1.7B.Q2K.gguf | Q2K | 778 MB | | DynaGuard-1.7B.Q3KL.gguf | Q3KL | 1 GB | | DynaGuard-1.7B.Q3KM.gguf | Q3KM | 940 MB | | DynaGuard-1.7B.Q3KS.gguf | Q3KS | 867 MB | | DynaGuard-1.7B.Q40.gguf | Q40 | 1.05 GB | | DynaGuard-1.7B.Q41.gguf | Q41 | 1.14 GB | | DynaGuard-1.7B.Q4K.gguf | Q4K | 1.11 GB | | DynaGuard-1.7B.Q4KM.gguf | Q4KM | 1.11 GB | | DynaGuard-1.7B.Q4KS.gguf | Q4KS | 1.06 GB | | DynaGuard-1.7B.Q50.gguf | Q50 | 1.23 GB | | DynaGuard-1.7B.Q51.gguf | Q51 | 1.32 GB | | DynaGuard-1.7B.Q5K.gguf | Q5K | 1.26 GB | | DynaGuard-1.7B.Q5KM.gguf | Q5KM | 1.26 GB | | DynaGuard-1.7B.Q5KS.gguf | Q5KS | 1.23 GB | | DynaGuard-1.7B.Q6K.gguf | Q6K | 1.42 GB | | DynaGuard-1.7B.Q80.gguf | Q80 | 1.83 GB | | File Name | Quant Type | File Size | | - | - | - | | DynaGuard-4B.BF16.gguf | BF16 | 8.05 GB | | DynaGuard-4B.F16.gguf | F16 | 8.05 GB | | DynaGuard-4B.F32.gguf | F32 | 16.1 GB | | DynaGuard-4B.Q2K.gguf | Q2K | 1.67 GB | | DynaGuard-4B.Q3KL.gguf | Q3KL | 2.24 GB | | DynaGuard-4B.Q3KM.gguf | Q3KM | 2.08 GB | | DynaGuard-4B.Q3KS.gguf | Q3KS | 1.89 GB | | DynaGuard-4B.Q4KM.gguf | Q4KM | 2.5 GB | | DynaGuard-4B.Q4KS.gguf | Q4KS | 2.38 GB | | DynaGuard-4B.Q5KM.gguf | Q5KM | 2.89 GB | | DynaGuard-4B.Q5KS.gguf | Q5KS | 2.82 GB | | DynaGuard-4B.Q6K.gguf | Q6K | 3.31 GB | | DynaGuard-4B.Q80.gguf | Q80 | 4.28 GB | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
license:apache-2.0
93
0

QwQ-LCoT-3B-Instruct-GGUF

NaNK
92
7

Deepfake-Detection-Exp-02-22

Deepfake-Detection-Exp-02-22 is a minimalist, high-quality dataset trained on a ViT-based model for image classification, distinguishing between deepfake and real images. The model is based on Google's `google/vit-base-patch32-224-in21k`. Limitations 1. Generalization Issues – The model may not perform well on deepfake images generated by unseen or novel deepfake techniques. 2. Dataset Bias – The training data might not cover all variations of real and fake images, leading to biased predictions. 3. Resolution Constraints – Since the model is based on `vit-base-patch32-224-in21k`, it is optimized for 224x224 image resolution, which may limit its effectiveness on high-resolution images. 4. Adversarial Vulnerabilities – The model may be susceptible to adversarial attacks designed to fool vision transformers. 5. False Positives & False Negatives – The model may occasionally misclassify real images as deepfake and vice versa, requiring human validation in critical applications. Intended Use 1. Deepfake Detection – Designed for identifying deepfake images in media, social platforms, and forensic analysis. 2. Research & Development – Useful for researchers studying deepfake detection and improving ViT-based classification models. 3. Content Moderation – Can be integrated into platforms to detect and flag manipulated images. 4. Security & Forensics – Assists in cybersecurity applications where verifying the authenticity of images is crucial. 5. Educational Purposes – Can be used in training AI practitioners and students in the field of computer vision and deepfake detection.

license:apache-2.0
92
4

Herculis-CUA-GUI-Actioner-4B

NaNK
92
1

Open-Xi-Math-Preview-GGUF

license:apache-2.0
92
0

Arch-Router-1.5B-GGUF

NaNK
license:apache-2.0
92
0

Llama-Doctor-3.2-3B-Instruct-GGUF

NaNK
llama
91
7

visionOCR-3B-061125-GGUF

NaNK
license:apache-2.0
91
2

SmolLM-1.7B-Instruct-GGUF

NaNK
llama
90
1

Deepthink-Reasoning-7B-GGUF

NaNK
ollama
89
5

OpenMath-8B-GGUF

NaNK
llama
89
1

Radiology-Infer-Mini

NaNK
license:apache-2.0
88
13

Neumind-Math-7B-Instruct-GGUF

NaNK
Ollama
88
5

SmolLM2-Rethink-360M-GGUF

llama
88
0

Canopus-Cute-Kawaii-Flux-LoRA

87
20

Cerium-Qwen3-R1-Dev-GGUF

license:apache-2.0
87
0

Llama-SmolTalk-3.2-1B-Instruct-GGUF

NaNK
llama
86
4

SmolLM2-360M-F32-GGUF

llama
86
1

Pyxidis-Manim-CodeGen-1.7B

NaNK
license:apache-2.0
85
3

Lucy-128k-GGUF

license:apache-2.0
84
0

Nous-V1-4B-GGUF

NaNK
license:apache-2.0
84
0

Evac-Opus-14B-Exp

Evac-Opus-14B-Exp [abliterated] is an advanced language model based on the Qwen 2.5 14B modality architecture, designed to enhance reasoning, explanation, and conversational capabilities. This model is optimized for general-purpose tasks, excelling in contextual understanding, logical deduction, and multi-step problem-solving. It has been fine-tuned using a long chain-of-thought reasoning model and specialized datasets to improve comprehension, structured responses, and conversational intelligence. Key Improvements 1. Enhanced General Knowledge: The model provides broad knowledge across various domains, improving capabilities in answering questions accurately and generating coherent responses. 2. Improved Instruction Following: Significant advancements in understanding and following complex instructions, generating structured responses, and maintaining coherence over extended interactions. 3. Versatile Adaptability: More resilient to diverse prompts, enhancing its ability to handle a wide range of topics and conversation styles, including open-ended and structured inquiries. 4. Long-Context Support: Supports up to 128K tokens for input context and can generate up to 8K tokens in a single output, making it ideal for detailed responses. 5. Multilingual Proficiency: Supports over 29 languages, including English, Chinese, French, Spanish, Portuguese, German, Italian, Russian, Japanese, Korean, Vietnamese, Thai, Arabic, and more. Here is a code snippet with `applychattemplate` to show you how to load the tokenizer and model and generate content: Intended Use 1. General-Purpose Reasoning: Designed for broad applicability, assisting with logical reasoning, answering diverse questions, and solving general knowledge problems. 2. Educational and Informational Assistance: Suitable for providing explanations, summaries, and research-based responses for students, educators, and general users. 3. Conversational AI and Chatbots: Ideal for building intelligent conversational agents that require contextual understanding and dynamic response generation. 4. Multilingual Applications: Supports global communication, translations, and multilingual content generation. 5. Structured Data Processing: Capable of analyzing and generating structured outputs, such as tables and JSON, useful for data science and automation. 6. Long-Form Content Generation: Can generate extended responses, including articles, reports, and guides, maintaining coherence over large text outputs. Limitations 1. Hardware Requirements: Requires high-memory GPUs or TPUs due to its large parameter size and long-context support. 2. Potential Bias in Responses: While designed to be neutral, outputs may still reflect biases present in training data. 3. Inconsistent Outputs in Creative Tasks: May produce variable results in storytelling and highly subjective topics. 4. Limited Real-World Awareness: Does not have access to real-time events beyond its training cutoff. 5. Error Propagation in Extended Outputs: Minor errors in early responses may affect overall coherence in long-form outputs. 6. Prompt Sensitivity: The effectiveness of responses may depend on how well the input prompt is structured. Open LLM Leaderboard Evaluation Results Detailed results can be found here! Summarized results can be found here! | Metric |Value (%)| |-------------------|--------:| |Average | 39.32| |IFEval (0-Shot) | 59.16| |BBH (3-Shot) | 49.58| |MATH Lvl 5 (4-Shot)| 42.15| |GPQA (0-shot) | 18.46| |MuSR (0-shot) | 18.63| |MMLU-PRO (5-shot) | 47.96|

NaNK
license:apache-2.0
83
4

UIGEN-T3-4B-Preview-GGUF

NaNK
license:apache-2.0
83
0

Flux-Chill-Guy-Zone

82
14

Bold-Shadows-Flux-LoRA

82
5

Math-IIO-7B-Instruct-GGUF

NaNK
81
3

Triangulum-1B-GGUF

NaNK
llama
80
1

Llama-3.1-5B-Instruct

Llama-3.1 is a collection of multilingual large language models (LLMs) that includes pretrained and instruction-tuned generative models in various sizes. The Llama-3.1-5B-Instruct model is part of the series optimized for multilingual dialogue use cases, offering powerful conversational abilities and outperforming many open-source and closed chat models on key industry benchmarks. - Size: 5B parameters - Model Architecture: Llama-3.1 is an auto-regressive language model using an optimized transformer architecture. - Training: The model is fine-tuned using Supervised Fine-Tuning (SFT) and Reinforcement Learning with Human Feedback (RLHF) to align with human preferences, ensuring helpfulness, safety, and natural conversations. The Llama-3.1-5B-Instruct model is optimized for multilingual text generation and excels in a variety of dialog-based use cases. It is designed to handle a wide array of tasks, including question answering, translation, and instruction following. - Ensure you have PyTorch installed with support for `bfloat16`: Below is an example of how to use the Llama-3.1-5B-Instruct model for conversational inference: - Model Type: Instruction-Tuned Large Language Model (LLM) - Training: Trained using supervised fine-tuning and reinforcement learning with human feedback. - Supported Tasks: Dialogue generation, question answering, translation, and other text-based tasks. The Llama-3.1-5B-Instruct model outperforms many existing models on several benchmarks, making it a reliable choice for conversational AI tasks in multilingual environments. - This model is optimized for safety and helpfulness, ensuring a positive user experience. - The torchdtype is set to `bfloat16` to optimize memory usage and performance. ---

NaNK
llama
80
1

Smoothie-Qwen3-4B-F32-GGUF

NaNK
license:apache-2.0
80
0

Deepfake-Detection-Exp-02-21-ONNX

NaNK
license:apache-2.0
79
1

Qwen3-4B-MegaScience-GGUF

> Qwen3-4B-MegaScience is a large language model fine-tuned on the MegaScience dataset, specifically designed for advanced scientific reasoning, and built on top of Qwen3-4B-Base; it leverages a meticulously curated set of 1.25 million high-quality scientific questions and answers sourced from university-level textbooks and various open datasets, covering seven scientific disciplines and evaluated across 15 benchmarks, demonstrating superior reasoning ability and training efficiency compared to existing open-source science models; the model supports seamless integration via the Hugging Face transformers library, operates efficiently with bfloat16 precision, and comes with an open-source dataset, evaluation pipeline, and reproducibility code, facilitating research and applications in scientific AI reasoning, with full resources, paper, and code available via the MegaScience official website and GitHub repository. | File Name | Size | Quant Type | |-----------|------|------------| | Qwen3-4B-MegaScience.BF16.gguf | 8.05 GB | BF16 | | Qwen3-4B-MegaScience.F16.gguf | 8.05 GB | F16 | | Qwen3-4B-MegaScience.F32.gguf | 16.1 GB | F32 | | Qwen3-4B-MegaScience.Q3KL.gguf | 2.24 GB | Q3KL | | Qwen3-4B-MegaScience.Q3KS.gguf | 1.89 GB | Q3KS | | Qwen3-4B-MegaScience.Q4KM.gguf | 2.5 GB | Q4KM | | Qwen3-4B-MegaScience.Q4KS.gguf | 2.38 GB | Q4KS | | Qwen3-4B-MegaScience.Q5KM.gguf | 2.89 GB | Q5KM | | Qwen3-4B-MegaScience.Q5KS.gguf | 2.82 GB | Q5KS | | Qwen3-4B-MegaScience.Q6K.gguf | 3.31 GB | Q6K | | Qwen3-4B-MegaScience.Q80.gguf | 4.28 GB | Q80 | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
license:apache-2.0
79
1

Qwen3-4B-Esper3-F32-GGUF

NaNK
license:apache-2.0
79
0

Blitzar-Coder-4B-F.1-GGUF

NaNK
license:apache-2.0
78
7

Qwen-UMLS-7B-Instruct-GGUF

NaNK
Llama-cpp
77
2

FastThink-0.5B-Tiny

FastThink-0.5B-Tiny is a reasoning-focused model based on Qwen2.5. We have released a range of base language models and instruction-tuned language models, spanning from 0.5 billion to 72 billion parameters. Qwen2.5 introduces the following improvements over Qwen2: - Significantly enhanced knowledge and greatly improved capabilities in coding and mathematics, thanks to specialized expert models in these domains. - Major improvements in instruction following, generating long texts (over 8K tokens), understanding structured data (e.g., tables), and generating structured outputs, especially JSON. It is more resilient to diverse system prompts, enhancing role-play implementation and condition-setting for chatbots. - Long-context support for up to 128K tokens and the ability to generate outputs up to 8K tokens. - Multilingual support for over 29 languages, including Chinese, English, French, Spanish, Portuguese, German, Italian, Russian, Japanese, Korean, Vietnamese, Thai, Arabic, and more. Architecture: Transformers with RoPE, SwiGLU, RMSNorm, Attention QKV bias, and tied word embeddings. Here provides a code snippet with `applychattemplate` to show you how to load the tokenizer and model and how to generate contents. This script is designed to load, process, and combine multiple datasets into a single, standardized format suitable for training conversational AI models. The script uses the `datasets` library to load and manipulate the datasets, and the `chattemplates` library to standardize the conversation format. Intended Use 1. Reasoning Tasks: FastThink-0.5B-Tiny is optimized for reasoning-focused applications, such as logical problem-solving, decision-making, and analytical workflows. 2. Instruction Following: Ideal for scenarios where precise adherence to instructions is required, including generating structured outputs like JSON or tables. 3. Multilingual Support: Suitable for use in multilingual environments, supporting over 29 languages, making it versatile for global applications. 4. Coding and Mathematics: Highly effective in tasks involving coding, debugging, or solving mathematical problems, leveraging expert domain knowledge. 5. Role-play Scenarios: Can simulate conversational agents or personas for role-playing, enhancing chatbot and virtual assistant implementations. 6. Long-form Content Creation: Designed to generate and manage long-form text (up to 8K tokens) while maintaining context, making it ideal for tasks like report writing or storytelling. 7. Understanding and Processing Structured Data: Efficient at interpreting and working with structured data, such as tables or hierarchical formats. 8. Low-Resource Applications: With a smaller parameter size (0.5B), it is well-suited for applications with limited computational resources or edge deployment. Limitations 1. Limited Model Size: As a 0.5B-parameter model, its reasoning and comprehension capabilities are less advanced compared to larger models, particularly for highly complex tasks. 2. Contextual Limitations: Although it supports a context length of up to 128K tokens, its ability to effectively utilize such a long context may vary, particularly in tasks requiring intricate cross-referencing of earlier inputs. 3. Accuracy in Domain-Specific Tasks: While capable in coding and mathematics, it may struggle with highly specialized or esoteric domain knowledge compared to models fine-tuned specifically for those areas. 4. Ambiguity Handling: May misinterpret vague or poorly structured prompts, leading to less accurate or unintended results. 5. Long-Context Tradeoffs: Generating or processing very long outputs (e.g., close to the 8K token limit) could result in decreased coherence or relevance toward the end. 6. Multilingual Performance: Although it supports 29 languages, its proficiency and fluency may vary across languages, with some underrepresented languages possibly seeing reduced performance. 7. Resource-Intensive for Long Contexts: Using its long-context capabilities (128K tokens) can be computationally demanding, requiring significant memory and processing power. 8. Dependence on Fine-Tuning: For highly specialized tasks or domains, additional fine-tuning may be necessary to achieve optimal performance.

NaNK
license:apache-2.0
76
7

Bootes-Qwen3_Coder-Reasoning-Q4_K_M-GGUF

llama-cpp
76
1

DeepCaption-VLA-7B

> The DeepCaption-VLA-7B model is a fine-tuned version of Qwen2.5-VL-7B-Instruct, tailored for Image Captioning and Vision Language Attribution. This variant is designed to generate precise, highly descriptive captions with a focus on defining visual properties, object attributes, and scene details across a wide spectrum of images and aspect ratios. [](https://github.com/PRITHIVSAKTHIUR/Multimodal-Outpost-Notebooks/blob/main/DeepCaption-VLA-7B%5B4bit%20-%20notebook%20demo%5D/DeepCaption-VLA-7B.ipynb) 1. Vision Language Attribution (VLA): Specially fine-tuned to attribute and define visual properties of objects, scenes, and environments. 2. Detailed Object Definitions: Generates captions with rich attribute descriptions, making outputs more precise than generic captioners. 3. High-Fidelity Descriptions: Handles general, artistic, technical, abstract, and low-context images with descriptive depth. 4. Robust Across Aspect Ratios: Accurately captions images regardless of format—wide, tall, square, or irregular. 5. Variational Detail Control: Supports both concise summaries and fine-grained attributions depending on prompt structure. 6. Foundation on Qwen2.5-VL Architecture: Leverages Qwen2.5-VL-7B’s multimodal reasoning for visual comprehension and instruction-following. 7. Multilingual Capability: Default in English, but adaptable for multilingual captioning through prompt engineering. This model was fine-tuned with a curated mix of datasets focused on caption richness and object-attribute alignment: prithivMLmods/blip3o-caption-mini-arrow prithivMLmods/Caption3o-Opt-v3 prithivMLmods/Caption3o-Opt-v2 Multimodal-Fatima/Caltech101\not\background\test\facebook\opt\2.7b\Attributes\Caption\ns\5647 Private/unlisted datasets for domain-specific image captioning tasks. The training objective emphasized Vision Language Attribution: defining image properties, attributes, and objects with clarity, while preserving descriptive fluency. > [!note] General Query: Caption the image precisely. [](https://colab.research.google.com/#fileId=https%3A//huggingface.co/prithivMLmods/DeepCaption-VLA-7B/blob/main/DeepCaption-VLA-7B%5B4bit%20-%20notebook%20demo%5D/DeepCaption-VLA-7B.ipynb) Generating attribute-rich image captions for research, dataset creation, and AI training. Vision-language attribution for object detection, scene understanding, and dataset annotation. Supporting creative, artistic, and technical applications requiring detailed descriptions. Captioning across varied aspect ratios, unusual visual styles, and non-standard datasets. May over-attribute or infer properties not explicitly visible in ambiguous images. Outputs can vary in tone depending on prompt phrasing. Not intended for filtered captioning tasks (explicit or sensitive content may appear). Accuracy may degrade on synthetic or highly abstract visual domains.

NaNK
license:apache-2.0
75
22

Castor-Character-Polygon-Flux-LoRA

74
14

Fashion-Product-Season

NaNK
license:apache-2.0
74
1

Castor-Flux-Concept-Gta6-Character-Design

73
6

Qwen2.5-Coder-3B-Instruct-GGUF

NaNK
Llama
72
5

AI-vs-Deepfake-vs-Real-Siglip2

NaNK
license:apache-2.0
72
2

Luth-Instruct-GGUF

NaNK
license:apache-2.0
72
1

Omega-Qwen2.5-Coder-3B-GGUF

NaNK
license:apache-2.0
72
0

Qwen2.5-Coder-1.5B-Instruct-GGUF

NaNK
Llama-cpp
71
3

Flux.1-Dev-Stamp-Art-LoRA

70
9

Delorme_1-OCR-7B-Post1.0

NaNK
license:apache-2.0
69
3

SD3.5-Large-Anime-LoRA

68
12

Qwen3-VL-32B-Instruct-abliterated-v1

NaNK
license:apache-2.0
68
4

Minc-Materials-23

NaNK
license:apache-2.0
68
0

Lynx-TinySync-0.6B

NaNK
license:apache-2.0
68
0

Traffic-Density-Classification

NaNK
license:apache-2.0
67
2

Qwen3-4B-ShiningValiant3-GGUF

NaNK
license:apache-2.0
67
0

zerank-1-GGUF

NaNK
license:apache-2.0
67
0

Osmosis-Apply-1.7B-GGUF

NaNK
license:apache-2.0
66
2

Montuno-Omega-Anime-LoRA

NaNK
65
3

Basically-Human-4B-f32-GGUF

> Basically-Human-4B is a 4 billion parameter language model based on the Qwen3 4B architecture, fine-tuned specifically for immersive and emotionally resonant roleplaying and character interaction. It excels in maintaining in-character consistency, crafting believable dialogue, and driving dynamic storytelling, making it ideal for text-based roleplay, NPC simulation, and interactive fiction applications. The model uses the ChatML instruction format to structure multi-turn conversations with clear role delineation. Basically-Human-4B was fine-tuned on a diverse set of instruction and roleplaying datasets, including various curated and cleaned instruction data from multiple sources. It offers GGUF quantized versions for ease of deployment while delivering compact yet capable performance in roleplay scenarios. | File Name | Quant Type | File Size | | - | - | - | | Basically-Human-4B.BF16.gguf | BF16 | 8.05 GB | | Basically-Human-4B.F16.gguf | F16 | 8.05 GB | | Basically-Human-4B.F32.gguf | F32 | 16.1 GB | | Basically-Human-4B.Q2K.gguf | Q2K | 1.67 GB | | Basically-Human-4B.Q3KL.gguf | Q3KL | 2.24 GB | | Basically-Human-4B.Q3KM.gguf | Q3KM | 2.08 GB | | Basically-Human-4B.Q3KS.gguf | Q3KS | 1.89 GB | | Basically-Human-4B.Q4KM.gguf | Q4KM | 2.5 GB | | Basically-Human-4B.Q4KS.gguf | Q4KS | 2.38 GB | | Basically-Human-4B.Q5KM.gguf | Q5KM | 2.89 GB | | Basically-Human-4B.Q5KS.gguf | Q5KS | 2.82 GB | | Basically-Human-4B.Q6K.gguf | Q6K | 3.31 GB | | Basically-Human-4B.Q80.gguf | Q80 | 4.28 GB | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
license:apache-2.0
65
2

Qwen-Image-Edit-2511-Object-Remover

license:apache-2.0
64
9

Deepsync-240-GGUF

NaNK
llama
64
2

Qwen3-1.7B-GGUF

NaNK
license:apache-2.0
63
1

Flux.1-Dev-Ctoon-LoRA

62
18

Qwen3-VL-32B-Instruct-abliterated

> Qwen3-VL-32B-Instruct-abliterated is an abliterated (v1.0) variant of Qwen3-VL-32B-Instruct, designed for Abliterated Reasoning and Captioning. > This model is optimized to generate detailed, descriptive captions and reasoning outputs across a wide range of visual and multimodal contexts—including complex, sensitive, or nuanced content—while supporting diverse aspect ratios and resolutions. Abliterated / Uncensored Captioning – Fine-tuned to bypass conventional content filters while preserving factual, descriptive, and reasoning-rich outputs. High-Fidelity Descriptions – Generates comprehensive captions and reasoning for general, artistic, technical, abstract, or low-context images. Robust Across Aspect Ratios – Performs consistently across wide, tall, square, and irregular image dimensions. Variational Detail Control – Capable of producing outputs ranging from concise summaries to fine-grained, intricate descriptions and reasoning. Foundation on Qwen3-VL-32B Architecture – Built upon Qwen3-VL-32B-Instruct’s advanced multimodal reasoning and instruction-following capabilities. Multilingual Output Capability – Primarily optimized for English, with adaptability for multilingual prompts through prompt engineering. Generating detailed, uncensored captions and reasoning for general-purpose or artistic datasets. Research in content moderation, red-teaming, and generative safety evaluation. Enabling descriptive captioning and reasoning for visual datasets typically excluded from mainstream models. Creative applications such as storytelling, art generation, or multimodal reasoning tasks. Captioning and reasoning for non-standard aspect ratios and stylized visual content. May produce explicit, sensitive, or offensive descriptions depending on the image content and prompts. Not recommended for production systems requiring strict content moderation. Output style, tone, and reasoning may vary based on input phrasing. Accuracy can fluctuate for unfamiliar, synthetic, or highly abstract visual content.

NaNK
license:apache-2.0
62
4

OpenScienceReasoning-Qwen-e10-GGUF

> OpenScienceReasoning-Qwen-e10 is a high-efficiency scientific reasoning model fine-tuned from Qwen3-1.7B using the nvidia/OpenScienceReasoning-2 dataset, encompassing 10,000 curated science and math entries that strengthen analytical problem-solving, chain-of-thought exploration, and code reasoning. The model excels at hybrid symbolic-AI thinking by performing structured logic, scientific derivations, multi-language coding, and generating outputs in formats such as LaTeX, Markdown, JSON, CSV, and YAML, making it ideal for research, education, and technical documentation on mid-range GPUs and edge clusters. Optimized for STEM applications, OpenScienceReasoning-Qwen-e10 delivers robust performance for tutoring, research assistance, and structured data generation while maintaining a lightweight deployment footprint. | File Name | Quant Type | File Size | | - | - | - | | OpenScienceReasoning-Qwen-e10.BF16.gguf | BF16 | 3.45 GB | | OpenScienceReasoning-Qwen-e10.F16.gguf | F16 | 3.45 GB | | OpenScienceReasoning-Qwen-e10.F32.gguf | F32 | 6.89 GB | | OpenScienceReasoning-Qwen-e10.Q2K.gguf | Q2K | 778 MB | | OpenScienceReasoning-Qwen-e10.Q3KL.gguf | Q3KL | 1 GB | | OpenScienceReasoning-Qwen-e10.Q3KM.gguf | Q3KM | 940 MB | | OpenScienceReasoning-Qwen-e10.Q3KS.gguf | Q3KS | 867 MB | | OpenScienceReasoning-Qwen-e10.Q40.gguf | Q40 | 1.05 GB | | OpenScienceReasoning-Qwen-e10.Q41.gguf | Q41 | 1.14 GB | | OpenScienceReasoning-Qwen-e10.Q4K.gguf | Q4K | 1.11 GB | | OpenScienceReasoning-Qwen-e10.Q4KM.gguf | Q4KM | 1.11 GB | | OpenScienceReasoning-Qwen-e10.Q4KS.gguf | Q4KS | 1.06 GB | | OpenScienceReasoning-Qwen-e10.Q50.gguf | Q50 | 1.23 GB | | OpenScienceReasoning-Qwen-e10.Q51.gguf | Q51 | 1.32 GB | | OpenScienceReasoning-Qwen-e10.Q5K.gguf | Q5K | 1.26 GB | | OpenScienceReasoning-Qwen-e10.Q5KM.gguf | Q5KM | 1.26 GB | | OpenScienceReasoning-Qwen-e10.Q5KS.gguf | Q5KS | 1.23 GB | | OpenScienceReasoning-Qwen-e10.Q6K.gguf | Q6K | 1.42 GB | | OpenScienceReasoning-Qwen-e10.Q80.gguf | Q80 | 1.83 GB | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
license:apache-2.0
62
0

Minitron-8B-Instruct-200K-GGUF

NaNK
Llama
61
1

Nous-1-2B-f32-GGUF

NaNK
license:apache-2.0
60
0

Telescopium-Acyclic-Qwen3-0.6B-GGUF

NaNK
license:apache-2.0
60
0

SmolLM2-CoT-360M-GGUF

llama
59
9

Bird-Species-Classifier-526

NaNK
license:apache-2.0
59
3

epsilon-ocr-d.markdown-post3.0.m

59
2

Yellow-Pop-Flux-Dev-LoRA

58
9

Llama-Deepsync-3B-GGUF

NaNK
llama
58
4

Tulu-MathLingo-8B-GGUF

NaNK
llama
58
1

Recycling-Net-11

NaNK
license:apache-2.0
58
0

Fashion-Product-baseColour

NaNK
license:apache-2.0
57
0

Past-Present-Deep-Mix-Flux-LoRA

56
10

Canopus-Pixar-Art

NaNK
license:apache-2.0
56
5

QVikhr-3-8B-Instruction-f32-GGUF

NaNK
license:apache-2.0
56
2

Polaris-1.7B-Preview-f32-GGUF

NaNK
license:apache-2.0
56
0

crm-01-4b-f32-GGUF

NaNK
license:apache-2.0
56
0

Canes-Cars-Model-LoRA

NaNK
55
4

FLUX.1-Kontext-Cinematic-Relighting

54
12

Octans Qwen3 UI Code 4B

> Octans-Qwen3-UI-Code-4B is an optimized successor of Muscae-Qwen3-UI-Code-4B, fine-tuned for enhanced UI reasoning precision, layout structuring, and frontend code synthesis. > Built upon Qwen3-4B and refined through Abliterated Reasoning Optimization, it delivers balanced, structured, and production-grade UI code outputs for experimental and research use. > Ideal for frontend developers, UI engineers, and design system researchers exploring next-generation code synthesis. > [!note] > GGUF: https://huggingface.co/prithivMLmods/Octans-Qwen3-UI-Code-4B-GGUF 1. Enhanced UI-Oriented Reasoning Upgraded reasoning calibration from Muscae with deeper token optimization for frontend logic, layout reasoning, and component cohesion. 2. Refined Web UI Component Generation Generates responsive, accessible, and semantic UI components with Tailwind, React, and HTML5, ensuring cleaner syntax and reduced redundancy. 3. Improved Layout-Aware Structure Demonstrates superior understanding of hierarchical design, stateful components, and responsive alignment, enhancing multi-device compatibility. 4. Optimized Hybrid Reasoning Engine Integrates symbolic and probabilistic logic for event-driven UI workflows, conditional rendering, and state synchronization in code outputs. 5. Structured Output Excellence Produces consistent results in HTML, React, Markdown, JSON, and YAML, suitable for UI prototyping, design systems, and auto-documentation. 6. Lightweight and Deployable Maintains a 4B parameter scale, optimized for mid-range GPUs, edge inference, or offline environments, without compromising structure or reasoning depth. Advanced web UI and component code generation Responsive frontend prototyping with Tailwind/React Research on structured reasoning in code synthesis Semantic, design-system-aligned component generation Experimental projects exploring UI intelligence modeling Research-focused model – not fine-tuned for production-critical pipelines Specialized for UI code – not suitable for general text generation or long-form reasoning May exhibit variability with cross-framework or overextended prompts Prioritizes code structure and logic clarity over aesthetic or creative expression.

NaNK
license:apache-2.0
54
6

Qwen2.5-0.5B-200K-GGUF

NaNK
Llama-cpp
54
5

Deepfake-QualityAssess2.0-85M

license:apache-2.0
54
1

Flux-Art-Nightmare-99

53
7

Mature-Content-Detection

NaNK
license:apache-2.0
53
7

OpenReasoning-Nemotron-1.5B-F32-GGUF

NaNK
license:cc-by-4.0
53
1

Alzheimer-Stage-Classifier

NaNK
license:apache-2.0
51
1

Anime-Classification-v1.0

NaNK
license:apache-2.0
51
0

mem-agent-f32-gguf

> driaforall/mem-agent is an agentic model based on Qwen3-4B-Thinking-2507, fine-tuned using GSPO (Zheng et al., 2025) to interact with an Obsidian-inspired, markdown-based memory system for advanced retrieval, updating, and clarification tasks. It is structured around agentic scaffolding that leverages dedicated tags and tool APIs for file and directory operations, supporting memory filtering and obfuscation, and evaluated on the md-memory-bench where it outperformed most open and closed models except qwen/qwen3-235b-a22b-thinking-2507, with an overall benchmark score of 0.75. The model is designed for use as an MCP server or standalone, and relies on linked markdown files to manage user and entity data, enabling seamless, flexible document-like memory manipulation for agentic or personal assistant scenarios. | File Name | Quant Type | File Size | | - | - | - | | mem-agent.BF16.gguf | BF16 | 8.05 GB | | mem-agent.F16.gguf | F16 | 8.05 GB | | mem-agent.F32.gguf | F32 | 16.1 GB | | mem-agent.Q2K.gguf | Q2K | 1.67 GB | | mem-agent.Q3KL.gguf | Q3KL | 2.24 GB | | mem-agent.Q3KM.gguf | Q3KM | 2.08 GB | | mem-agent.Q3KS.gguf | Q3KS | 1.89 GB | | mem-agent.Q4KM.gguf | Q4KM | 2.5 GB | | mem-agent.Q4KS.gguf | Q4KS | 2.38 GB | | mem-agent.Q5KM.gguf | Q5KM | 2.89 GB | | mem-agent.Q5KS.gguf | Q5KS | 2.82 GB | | mem-agent.Q6K.gguf | Q6K | 3.31 GB | | mem-agent.Q80.gguf | Q80 | 4.28 GB | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

license:apache-2.0
51
0

Lumian-VLR-7B-Thinking

> The Lumian-VLR-7B-Thinking model is a high-fidelity vision-language reasoning (experimental model) system designed for fine-grained multimodal understanding. Built on Qwen2.5-VL-7B-Instruct, this model enhances image captioning, sampled video reasoning, and document comprehension through explicit grounded reasoning. It produces structured reasoning traces aligned with visual coordinates, enabling explainable multimodal reasoning. Trained via supervised fine-tuning (SFT) on visually-grounded reasoning traces and further refined using GRPO reinforcement learning, Lumian delivers superior step-by-step chain-of-thought reasoning with strong visual grounding. > [!NOTE] Model Subfolder: Lumian-VLR-7B-Thinking(think-preview) > > Model Folder: Lumian-VLR-7B-Thinking(no-think-single-shot) Visually-Grounded Reasoning and Thinking Traces: Generates explicit reasoning traces tied to image regions and document structures for transparent and explainable outputs. Advanced Image Captioning: Produces detailed, grounded captions with reasoning steps for improved scene understanding. Sampled Video Reasoning: Handles long-duration videos with temporal reasoning for question answering and summarization. Context-Aware Document Analysis: Excels at structured and unstructured content extraction with visual grounding. Fine-Grained Visual Grounding: Accurately links reasoning steps to tables, charts, and graphical elements. Reinforcement-Learned Thinking: GRPO training incentivizes accurate, grounded reasoning with minimal hallucinations. > [!TIP] Colab Demo : https://huggingface.co/prithivMLmods/Lumian-VLR-7B-Thinking/blob/main/think-preview/Lumian-VLR-7B-Thinking-Demo-Notebook/Lumian-VLR-7B-Thinking.ipynb The model outputs reasoning and answers in a structured format: Visual reasoning with grounded, step-by-step thinking traces. Explainable image captioning and sampled video reasoning. Multimodal document retrieval, extraction, and analytical interpretation. Transparent chain-of-thought reasoning for educational, research, and enterprise use. Multilingual reasoning and structured content extraction. Robotic and mobile vision-based automation with grounded decision-making. High memory requirements for long videos and large document batches. Degraded accuracy on extremely low-resolution or obscured visuals. Suboptimal for real-time inference on edge devices. Visual token configuration strongly influences reasoning fidelity. Occasional reasoning drift or partial grounding errors. YaRN: Efficient Context Window Extension of Large Language Models Qwen2-VL: Enhancing Vision-Language Model's Perception of the World at Any Resolution Qwen-VL: A Versatile Vision-Language Model for Understanding, Localization, Text Reading, and Beyond A Comprehensive and Challenging OCR Benchmark for Evaluating Large Multimodal Models in Literacy Ground-R1: Incentivizing Grounded Visual Reasoning via Reinforcement Learning

NaNK
license:apache-2.0
50
4

Triangulum-5B-GGUF

. . / | || | | \ \\ \| |\ \ / \ / \ | | \| | | | \ / \ | | | | \/| | / \| | \/ // >| | /| || | /| Y Y \ || || ||( /|| /\ / |/ |/|/ ||| / \/ \/// \/ Triangulum 5B GGUF: Multilingual Large Language Models (LLMs) Triangulum 5B is a collection of pretrained and instruction-tuned generative models, designed for multilingual applications. These models are trained using synthetic datasets based on long chains of thought, enabling them to perform complex reasoning tasks effectively. - Foundation Model: Built upon LLaMA's autoregressive language model, leveraging an optimized transformer architecture for enhanced performance. - Instruction Tuning: Includes supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to align model outputs with human preferences for helpfulness and safety. - Multilingual Support: Designed to handle multiple languages, ensuring broad applicability across diverse linguistic contexts. 1. Synthetic Datasets: Utilizes long chain-of-thought synthetic data to enhance reasoning capabilities. 2. Supervised Fine-Tuning (SFT): Aligns the model to specific tasks through curated datasets. 3. Reinforcement Learning with Human Feedback (RLHF): Ensures the model adheres to human values and safety guidelines through iterative training processes. Starting with `transformers >= 4.43.0` onward, you can run conversational inference using the Transformers `pipeline` abstraction or by leveraging the Auto classes with the `generate()` function. Make sure to update your transformers installation via `pip install --upgrade transformers`. Key Adjustments 1. System Prompts: Each prompt defines a different role or persona for the AI to adopt. 2. User Prompts: These specify the context or task for the assistant, ranging from teaching to storytelling or career advice. 3. Looping Through Prompts: Each prompt is processed in a loop to showcase the model's versatility. You can expand the list of prompts to explore a variety of scenarios and responses. - Multilingual content generation - Question answering and dialogue systems - Text summarization and analysis - Translation and localization tasks Triangulum 10B employs a state-of-the-art autoregressive architecture inspired by LLaMA. The optimized transformer framework ensures both efficiency and scalability, making it suitable for a variety of use cases. bash ollama run triangulum-5b-f16.gguf plaintext pulling manifest pulling 8934d96d3f08... 100% ▕██████████████████████████████████████████████████████████████████████████████████████████▏ 3.8 GB pulling 8c17c2ebb0ea... 100% ▕██████████████████████████████████████████████████████████████████████████████████████████▏ 7.0 KB pulling 7c23fb36d801... 100% ▕██████████████████████████████████████████████████████████████████████████████████████████▏ 4.8 KB pulling 2e0493f67d0c... 100% ▕██████████████████████████████████████████████████████████████████████████████████████████▏ 59 B pulling fa304d675061... 100% ▕██████████████████████████████████████████████████████████████████████████████████████████▏ 91 B pulling 42ba7f8a01dd... 100% ▕██████████████████████████████████████████████████████████████████████████████████████████▏ 557 B verifying sha256 digest writing manifest removing any unused layers success >>> Send a message (/? for help) plaintext >>> What can you do for me? plaintext As a responsible AI language model, I am here to assist you with any questions or tasks you may have. Here are some examples of things I can help with: 1. Answering questions: I can provide information on a wide range of topics, from science and technology to history and culture. 2. Generating ideas: I can help you brainstorm ideas for creative projects, or provide suggestions for solving problems. 3. Writing assistance: I can help you with writing tasks such as proofreading, editing, and suggesting alternative words or phrases. 4. Translation: I can translate text from one language to another. 5. Summarizing content: I can summarize long pieces of text, such as articles or documents, into shorter, more digestible versions. 6. Creativity: I can help you generate creative ideas for stories, poems, or other forms of writing. 7. Language learning: I can assist you in learning a new language by providing grammar explanations, vocabulary lists, and practice exercises. 8. Chatting: I'm here to chat with you and provide a response to any question or topic you'd like to discuss. Please let me know if there is anything specific you would like me to help you with. plaintext /exit ``` Ollama supports running multi-modal models where you can send images and ask questions based on them. This section will be updated as more models become available. Quantized models like triangulum-5b-f16.gguf are optimized for performance on resource-constrained hardware, making it accessible for local inference. 1. Ensure your system has sufficient VRAM or CPU resources. 2. Use the `.gguf` model format for compatibility with Ollama. Running the Triangulum-5B model with Ollama provides a robust way to leverage open-source LLMs locally for diverse use cases. By following these steps, you can explore the capabilities of other open-source models in the future.

NaNK
llama
50
2

Wolf-Rayet-2B-Prime3-GGUF

NaNK
license:apache-2.0
50
0

Castor-Collage-Dim-Flux-LoRA

49
7

Green-Cartoon-Flux-LoRA

49
6

Nemo-Minitron-8B-Instruct-GGUF

NaNK
license:apache-2.0
49
1

Piaget-1.7B-f32-GGUF

NaNK
license:apache-2.0
49
0

Gacrux-R1-Qwen3-1.7B-MoD-GGUF

> Gacrux-R1-Qwen3-1.7B-MoD is a high-efficiency, multi-domain model fine-tuned on Qwen3-1.7B with traces of Mixture of Domains (MoD). It leverages the prithivMLmods/Gargantua-R1-Wee dataset, designed for rigorous mathematical problem-solving and enriched with multi-domain coverage across mathematics, coding, and science. This model blends symbolic precision, scientific logic, and structured output fluency—making it an ideal tool for developers, educators, and researchers seeking advanced reasoning under constrained compute. | File Name | Quant Type | File Size | | - | - | - | | Gacrux-R1-Qwen3-1.7B-MoD.BF16.gguf | BF16 | 3.45 GB | | Gacrux-R1-Qwen3-1.7B-MoD.F16.gguf | F16 | 3.45 GB | | Gacrux-R1-Qwen3-1.7B-MoD.F32.gguf | F32 | 6.89 GB | | Gacrux-R1-Qwen3-1.7B-MoD.Q2K.gguf | Q2K | 778 MB | | Gacrux-R1-Qwen3-1.7B-MoD.Q3KL.gguf | Q3KL | 1 GB | | Gacrux-R1-Qwen3-1.7B-MoD.Q3KM.gguf | Q3KM | 940 MB | | Gacrux-R1-Qwen3-1.7B-MoD.Q3KS.gguf | Q3KS | 867 MB | | Gacrux-R1-Qwen3-1.7B-MoD.Q4KM.gguf | Q4KM | 1.11 GB | | Gacrux-R1-Qwen3-1.7B-MoD.Q4KS.gguf | Q4KS | 1.06 GB | | Gacrux-R1-Qwen3-1.7B-MoD.Q5KM.gguf | Q5KM | 1.26 GB | | Gacrux-R1-Qwen3-1.7B-MoD.Q5KS.gguf | Q5KS | 1.23 GB | | Gacrux-R1-Qwen3-1.7B-MoD.Q6K.gguf | Q6K | 1.42 GB | | Gacrux-R1-Qwen3-1.7B-MoD.Q80.gguf | Q80 | 1.83 GB | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
license:apache-2.0
49
0

Llama-Thinker-3B-Preview2-GGUF

NaNK
base_model:prithivMLmods/Llama-Thinker-3B-Preview2
48
5

Castor-Happy-Halloween-Flux-LoRA

48
4

Ton618-Amxtoon-Flux-LoRA

47
11

Canopus-Art-Medium-LoRA

NaNK
47
4

AI-vs-Deepfake-vs-Real-ONNX

license:apache-2.0
47
1

PACS-DG-SigLIP2

NaNK
license:apache-2.0
47
0

Pastel-BG-Flux-LoRA

46
10

Llama-3.2-3B-Promptist-Mini

NaNK
llama
46
4

Flux.1-Dev-Pov-DoorEye-LoRA

45
8

Uncoloured-Polygon-Flux-LoRA

45
5

Lime-Green-Flux-LoRA

45
3

SmolLM-1.7B-GGUF

NaNK
llama
45
1

Qwen3-8B-GGUF

NaNK
license:apache-2.0
45
1

Canopus-Interior-Architecture-0.1

NaNK
43
24

siglip2-mini-explicit-content

> siglip2-mini-explicit-content is an image classification vision-language encoder model fine-tuned from `siglip2-base-patch16-512` for a single-label classification task. It is designed to classify images into categories related to explicit, sensual, or safe-for-work content using the SiglipForImageClassification architecture. > \[!Note] > This model is intended to promote positive, safe, and respectful digital environments. Misuse is strongly discouraged and may violate platform or regional guidelines. As a classification model, it does not generate unsafe content and is suitable for moderation purposes. > [!note] SigLIP 2: Multilingual Vision-Language Encoders with Improved Semantic Understanding, Localization, and Dense Features https://arxiv.org/pdf/2502.14786 > [!Important] Note: Explicit, sensual, and pornographic content may appear in the results; however, all of them are considered not safe for work. Class 0: Anime Picture Class 1: Extincing & Sensual Class 2: Hentai Class 3: Pornography Class 4: Safe for Work Guidelines for Use of siglip2-mini-explicit-content This model is designed for responsible content moderation and filtering. It is especially tuned for anime, hentai, and adult content. Use it ethically, with the following guidelines: Content Moderation in social media and forums Parental Controls for safer browsing environments Dataset Curation for removing NSFW images from training data Safe Search Filtering for engines and discovery systems Workplace Image Scanning for compliance Harassment, exposure, or targeting of individuals Use on private content without consent Illegal or unethical surveillance Sole reliance for legal or reputational decisions Deceptive manipulation of moderation results Optimized for anime and adult content detection. Not suitable for detecting violence, drugs, or hate symbols. Probabilistic outputs — always verify with human review where needed. This model's predictions are not legal classifications. This tool was created to enhance digital safety. Do not use it to harm, surveil, or exploit individuals or communities. By using this model, you commit to ethical and privacy-respecting practices.

NaNK
license:apache-2.0
43
2

Qwen3-VisionCaption-2B-Thinking-GGUF

NaNK
llama.cpp
43
1

trlm-135m-GGUF

> The Tiny Reasoning Language Model (trlm-135m) is a 135 million parameter research prototype aimed at exploring how smaller language models can acquire step-by-step reasoning abilities. Built on the SmolLM2-135M-Instruct model (a Llama 3 based decoder-only transformer), it undergoes a three-stage fine-tuning pipeline: Stage 1 for general instruction tuning without reasoning, Stage 2 for incorporating reasoning traces marked by tags, and Stage 3 for preference alignment to refine reasoning style using Direct Preference Optimization (DPO). If you are running in LM Studio, start with a context length of 1024 and adjust it based on the responses. It’s recommended to use high-precision quants for better performance. | File Name | Quant Type | File Size | | - | - | - | | trlm-135m.BF16.gguf | BF16 | 271 MB | | trlm-135m.F16.gguf | F16 | 271 MB | | trlm-135m.F32.gguf | F32 | 540 MB | | trlm-135m.Q2K.gguf | Q2K | 88.2 MB | | trlm-135m.Q3KL.gguf | Q3KL | 97.5 MB | | trlm-135m.Q3KM.gguf | Q3KM | 93.5 MB | | trlm-135m.Q3KS.gguf | Q3KS | 88.2 MB | | trlm-135m.Q40.gguf | Q40 | 91.7 MB | | trlm-135m.Q41.gguf | Q41 | 98.4 MB | | trlm-135m.Q4K.gguf | Q4K | 105 MB | | trlm-135m.Q4KM.gguf | Q4KM | 105 MB | | trlm-135m.Q4KS.gguf | Q4KS | 102 MB | | trlm-135m.Q50.gguf | Q50 | 105 MB | | trlm-135m.Q51.gguf | Q51 | 112 MB | | trlm-135m.Q5K.gguf | Q5K | 112 MB | | trlm-135m.Q5KM.gguf | Q5KM | 112 MB | | trlm-135m.Q5KS.gguf | Q5KS | 110 MB | | trlm-135m.Q6K.gguf | Q6K | 138 MB | | trlm-135m.Q80.gguf | Q80 | 145 MB | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

llama
43
1

Nous-1-4B-f32-GGUF

NaNK
license:apache-2.0
43
0

SD3.5-Turbo-Realism-2.0-LoRA

42
11

Codepy-Deepthink-3B

The Codepy 3B Deep Think Model is a fine-tuned version of the meta-llama/Llama-3.2-3B-Instruct base model, designed for text generation tasks that require deep reasoning, logical structuring, and problem-solving. This model leverages its optimized architecture to provide accurate and contextually relevant outputs for complex queries, making it ideal for applications in education, programming, and creative writing. With its robust natural language processing capabilities, Codepy 3B Deep Think excels in generating step-by-step solutions, creative content, and logical analyses. Its architecture integrates advanced understanding of both structured and unstructured data, ensuring precise text generation aligned with user inputs. | Model Content | Size | Description | Upload Status | |-----------------------------------|----------------|------------------------------------------------|-------------------| | `.gitattributes` | 1.57 kB | Git LFS configuration for large files. | Uploaded | | `README.md` | 221 Bytes | Basic repository information. | Updated | | `config.json` | 994 Bytes | Model configuration with architectural details. | Uploaded | | `generationconfig.json` | 248 Bytes | Default generation parameters. | Uploaded | | `pytorchmodel-00001-of-00002.bin`| 4.97 GB | Sharded PyTorch model weights (part 1 of 2). | Uploaded (LFS) | | `pytorchmodel-00002-of-00002.bin`| 1.46 GB | Sharded PyTorch model weights (part 2 of 2). | Uploaded (LFS) | | `pytorchmodel.bin.index.json` | 21.2 kB | Index file mapping model shards. | Uploaded | | `specialtokensmap.json` | 477 Bytes | Maps special tokens to their respective IDs. | Uploaded | | `tokenizer.json` | 17.2 MB | Full tokenizer vocabulary and merges. | Uploaded (LFS) | | `tokenizerconfig.json` | 57.5 kB | Tokenizer configuration details. | Uploaded | | Run with LM Studio | Details | |--------------------------|-----------------------------------------------------------------------------------------------| | Run with LM Studio | https://lmstudio.ai/ | | Demo on LM Studio | https://drive.google.com/file/d/1CHdfjYrwMnk9ACvS40Abfy3xNXnCubKG/view?usp=sharing | | Codepy-Deepthink-3B-GGUF | https://huggingface.co/prithivMLmods/Codepy-Deepthink-3B-GGUF | >>> Develop a Python program to generate random passwords that consist of 8 characters. Not applicable bash python passwordgenerator.py python import random def generatepassword(length): """ Generates a random alphanumeric password of the specified length. Ensures that at least one lowercase letter, one uppercase letter, one digit, and one space are included in the password. Args: length: The number of characters in the password. Returns: A string representing the generated password or None if the input is invalid. """ # Define a set of alphanumeric characters with spaces characters = 'abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ0123456789 ' # Validate the length if length len(characters): print("Invalid password length. It should be less than or equal to", len(characters)) return None # Ensure at least one character from each required group requiredcharacters = 'abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ0123456789 ' if length > 1: passwordlengthwithoutrequirements = length - 4 randomstring = ''.join(random.choice(requiredcharacters) for in range(passwordlengthwithoutrequirements)) # Fill the rest of the password with random characters remainingcharsneeded = length - len(randomstring) allpossiblechars = list(characters) if length > 1: randomcharacter = random.choice('abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ0123456789 ') else: randomcharacter = random.choice('abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ0123456789 ') password = randomstring + randomcharacter remainingcharsneeded # Shuffle the password to avoid predictable patterns passwordlist = list(password) random.shuffle(passwordlist) password = ''.join(passwordlist) Example Usage passwordlength = 8 generatedpassword = generatepassword(passwordlength) if generatedpassword is not None: print(f"Generated Password: {generatedpassword}") else: print("Failed to generate a password. Please ensure the length is valid (between 1 and", len(characters), ").") Generated Password: g7x 2PqA bash FROM Llama-3.2-1B.F16.gguf bash ollama create metallama -f ./metallama bash ollama list bash ollama run metallama plaintext >>> write a mini passage about space x Space X, the private aerospace company founded by Elon Musk, is revolutionizing the field of space exploration... ``` With these steps, you can easily run custom models using Ollama. Adjust as needed for your specific use case.

NaNK
llama
42
4

Bootes-Qwen3_Coder-Reasoning

> Bootes-Qwen3\Coder-Reasoning is a fine-tuned variant of the Qwen3-4B architecture, optimized for high-accuracy code reasoning and structured logical task completion. Trained on the CodeAlpaca\20K dataset and additional curated programming corpora, this model is designed to perform technical coding, reasoning, and instruction-following tasks with lightweight computational requirements. > [!note] GGUF : https://huggingface.co/prithivMLmods/Bootes-Qwen3Coder-Reasoning-Q4KM-GGUF 1. Code Reasoning with CodeAlpaca\20K and More Fine-tuned on CodeAlpaca\20K and supplementary high-quality datasets focused on: Multi-language programming tasks Code explanation, completion, and debugging Instruction-following with step-wise execution logic 2. Cross-Language Code Understanding Handles Python, JavaScript, C++, and more. Ideal for code generation, transformation, bug-fixing, and logic validation. 3. Structured Output Generation Delivers responses in Markdown, JSON, YAML, and structured code blocks. Optimized for IDE workflows, documentation tools, and reproducible computation notebooks. 4. Instruction-Tuned for Developer Use Cases Maintains strong fidelity to user prompts, especially multi-turn or step-by-step technical instructions across engineering and data workflows. 5. Multilingual Reasoning in Technical Domains Capable of technical comprehension and explanation in over 20 human languages, supporting global developer audiences. 6. Efficient 4B Architecture Based on Qwen3-4B for a performance-efficient inference model that scales well on mid-range GPUs and cloud deployment setups. Code generation, completion, and explanation Multi-step algorithmic reasoning Structured technical document generation (Markdown, JSON, YAML) Debugging assistance and refactoring suggestions Technical tutoring and developer assistant workflows Cross-lingual programming education and translation May underperform on non-code-related creative writing Limited context window versus larger models Sensitive to prompt phrasing for ambiguous instructions Occasionally over-justifies code when brevity is desired 1. Qwen2.5 Technical Report – https://arxiv.org/pdf/2412.15115 2. CodeAlpaca Dataset – https://github.com/sahil280114/codealpaca 3. YaRN: Context Window Extension for LLMs – https://arxiv.org/pdf/2309.00071

NaNK
license:apache-2.0
41
8

Qwen3-VisionCaption-2B

NaNK
llama.cpp
40
2

DeepSeek-R1-Llama-8B-F32-GGUF

NaNK
llama
40
2

Food-or-Not-SigLIP2

NaNK
license:apache-2.0
40
0

OpenCoder-1.5B-Base-GGUF

NaNK
llama-cpp
39
3

Yarn Photo I2i

Yarn-Photo-i2i is an adapter for black-forest-lab's FLUX.1-Kontext-dev, designed for converting images into yarn-stitched artwork while preserving the original characteristics of the subject. The model was trained on 28 image pairs (14 start images, 14 end images). Synthetic result nodes were generated using NanoBanana from Google and SeedDream 4 (dataset for result sets), and labeled with DeepCaption-VLA-7B. The adapter is triggered with the following prompt: > [!note] [photo content], transformed into a crochet plush doll, with visible yarn stitches, button eyes, and cozy handmade charm. | Setting | Value | | ------------------------ | ------------------------ | | Module Type | Adapter | | Base Model | FLUX.1 Kontext Dev - fp8 | | Trigger Words | [photo content], transformed into a crochet plush doll, with visible yarn stitches, button eyes, and cozy handmade charm. | | Image Processing Repeats | 50 | | Epochs | 22 | | Save Every N Epochs | 1 | Labeling: DeepCaption-VLA-7B(natural language & English) Total Images Used for Training : 28 Image Pairs (14 Start, 14 End) Synthetic Result Node generated by NanoBanana from Google (Image Result Sets Dataset) | Setting | Value | | --------------------------- | --------- | | Seed | - | | Clip Skip | - | | Text Encoder LR | 0.00001 | | UNet LR | 0.00005 | | LR Scheduler | constant | | Optimizer | AdamW8bit | | Network Dimension | 64 | | Network Alpha | 32 | | Gradient Accumulation Steps | - | | Setting | Value | | --------------- | ----- | | Shuffle Caption | - | | Keep N Tokens | - | | Setting | Value | | ------------------------- | ----- | | Noise Offset | 0.03 | | Multires Noise Discount | 0.1 | | Multires Noise Iterations | 10 | | Conv Dimension | - | | Conv Alpha | - | | Batch Size | - | | Steps | 2900 | | Sampler | euler | You should use `[photo content]` to trigger the image generation. You should use `transformed into a crochet plush doll` to trigger the image generation. You should use `with visible yarn stitches` to trigger the image generation. You should use `button eyes` to trigger the image generation. You should use `and cozy handmade charm.` to trigger the image generation.

39
3

SmolLM2-1.7B-Instruct-GGUF

NaNK
llama
39
2

Qwen2.5-Coder-14B-Instruct-F16-GGUF

NaNK
Llama-cpp
39
2

Qwen3.5-0.8B-Unredacted-MAX

NaNK
license:apache-2.0
39
1

Qwen3.5-2B-Unredacted-MAX

NaNK
license:apache-2.0
39
1

Research-Reasoning-Qwen-F32-GGUF

NaNK
license:apache-2.0
39
1

Omega-Qwen3-Atom-8B-GGUF

NaNK
license:apache-2.0
39
0

SD3.5-Large-Minimal-Blacked-LoRA

38
10

Pegasi-Minimalist-Image-Style

NaNK
38
6

KAIROS-MM-Qwen2.5-VL-7B-RL

NaNK
license:apache-2.0
37
2

Gliese-CUA-Tool-Call-8B

NaNK
license:apache-2.0
37
1

OpenRHO-2B-Thinker-GGUF

NaNK
license:apache-2.0
37
0

Llama-Chat-Summary-3.2-3B

NaNK
llama
36
5

Guard-Against-Unsafe-Content2-Siglip2

NaNK
license:apache-2.0
36
2

vit-mini-explicit-content

> vit-mini-explicit-content is an image classification vision-language model fine-tuned from vit-base-patch16-224-in21k for a single-label classification task. It categorizes images based on their explicitness using the ViTForImageClassification architecture. > \[!Note] > This model is designed to promote safe, respectful, and responsible online spaces. It does not generate explicit content; it only classifies images. Misuse may violate platform or regional policies and is strongly discouraged. > [!Note] An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale : https://arxiv.org/abs/2010.11929, Visual Transformers: Token-based Image Representation and Processing for Computer Vision: https://arxiv.org/pdf/2006.03677 > [!Important] Note: Explicit, sensual, and pornographic content may appear in the results; however, all of them are considered not safe for work. Class 0: Anime Picture Class 1: Enticing & Sensual Class 2: Hentai Class 3: Pornography Class 4: Safe for Work Image moderation pipelines Parental and institutional content filters Dataset cleansing before training Online safety and well-being platforms Enhancing search engine filtering Non-consensual or malicious monitoring Automated judgments without human review Misrepresentation of moderation systems Use in unlawful or unethical surveillance Harassment, exploitation, or shaming

NaNK
license:apache-2.0
35
5

SmolLM2-360M-Instruct-GGUF

llama
35
1

Draco-CoderMini-3B-GGUF

NaNK
license:apache-2.0
35
0

Canopus-Car-Flux-Dev-LoRA

34
3

Canopus-Flux-LoRA-Hoodies

34
3

Flux-C33-Design-LoRA

34
3

cogito-v1-preview-llama-3B-f32-GGUF

> Cogito v1 preview - 3B is a 3 billion parameter hybrid reasoning language model built on the Llama 3.2 architecture and developed by DeepCogito, uniquely designed to operate in both standard and deep “self-reflective” reasoning modes through Iterated Distillation and Amplification (IDA), supporting 128k context windows and over 30 languages, and optimized for coding, STEM, multilingual tasks, tool-calling, and instruction following, consistently outperforming other models of similar size on industry benchmarks while being available under an open license for commercial use. | File name | Size | Quant type | |-----------|------|------------| | cogito-v1-preview-llama-3B.F32.gguf | 12.9 GB | F32 | | cogito-v1-preview-llama-3B.BF16.gguf | 6.43 GB | BF16 | | cogito-v1-preview-llama-3B.F16.gguf | 6.43 GB | F16 | | cogito-v1-preview-llama-3B.Q80.gguf | 3.42 GB | Q80 | | cogito-v1-preview-llama-3B.Q6K.gguf | 2.64 GB | Q6K | | cogito-v1-preview-llama-3B.Q5KM.gguf | 2.32 GB | Q5KM | | cogito-v1-preview-llama-3B.Q5KS.gguf | 2.27 GB | Q5KS | | cogito-v1-preview-llama-3B.Q4KM.gguf | 2.02 GB | Q4KM | | cogito-v1-preview-llama-3B.Q4KS.gguf | 1.93 GB | Q4KS | | cogito-v1-preview-llama-3B.Q3KL.gguf | 1.82 GB | Q3KL | | cogito-v1-preview-llama-3B.Q3KM.gguf | 1.69 GB | Q3KM | | cogito-v1-preview-llama-3B.Q3KS.gguf | 1.54 GB | Q3KS | | cogito-v1-preview-llama-3B.Q2K.gguf | 1.36 GB | Q2K | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
llama
34
0

Llama-Magpie-3.2-3B-Instruct-GGUF

NaNK
llama
33
2

Muscae-Qwen3-UI-Code-4B

NaNK
license:apache-2.0
33
2

Canopus-Isometric-InteriorDesign-3D

NaNK
32
9

Pegasi-Beta-GTA-LoRA

NaNK
32
6

BetaCeti-Beta-4B-Prime1-GGUF

NaNK
license:apache-2.0
32
0

tooth-agenesis-siglip2

> tooth-agenesis-siglip2 is a vision-language encoder model fine-tuned from `google/siglip2-base-patch16-512` for multi-class image classification. It is trained to detect various dental anomalies and conditions such as Calculus, Caries, Gingivitis, Mouth Ulcer, Tooth Discoloration, and Hypodontia. The model uses the `SiglipForImageClassification` architecture. > \[!note] > SigLIP 2: Multilingual Vision-Language Encoders with Improved Semantic Understanding, Localization, and Dense Features > https://arxiv.org/pdf/2502.14786 Dental Diagnosis Support – Assists dentists and clinicians in identifying common dental conditions from images. Oral Health Monitoring – A tool for regular monitoring of dental health in clinical or remote settings. Tele-dentistry – Enables automated screening in virtual consultations and rural healthcare setups. Research and Education – Useful for academic institutions and training platforms for demonstrating AI in dental diagnostics. Early Detection – Helps identify oral health issues early to prevent progression.

NaNK
license:apache-2.0
31
0

Megalodon-OCR-Sync-0713

> The Megalodon-OCR-Sync-0713 model is a fine-tuned version of Qwen2.5-VL-3B-Instruct, optimized for Document Retrieval, Content Extraction, and Analysis Recognition. Built on top of the Qwen2.5-VL architecture, this model enhances document comprehension capabilities with focused training on 200K image pairs from a mixture of captioning datasets, including 70K from Corvus-OCR-Caption-Mix dataset, and other document modular datasets from modular combination of opensource datasets best for doc OCR captioning, image reasoning, visual analysis, working on all category of images with variational dimension. Context-Aware Multimodal Extraction and Linking for Documents: Advanced capability for understanding document context and establishing connections between multimodal elements within documents. Enhanced Document Retrieval: Designed to efficiently locate and extract relevant information from complex document structures and layouts. Superior Content Extraction: Optimized for precise extraction of structured and unstructured content from diverse document formats. Analysis Recognition: Specialized in recognizing and interpreting analytical content, charts, tables, and visual data representations. State-of-the-Art Performance Across Resolutions: Achieves competitive results on OCR and visual QA benchmarks such as DocVQA, MathVista, RealWorldQA, and MTVQA. Video Understanding up to 20+ minutes: Supports detailed comprehension of long-duration videos for content summarization, Q\&A, and multi-modal reasoning. Visually-Grounded Device Interaction: Enables mobile/robotic device operation via visual inputs and text-based instructions using contextual understanding and decision-making logic. > [!warning] Not expected to work as well in Indian languages. Context-aware multimodal extraction and linking for complex document structures. High-fidelity document retrieval and content extraction from various document formats. Analysis recognition of charts, graphs, tables, and visual data representations. Document-based question answering for educational and enterprise applications. Extraction and LaTeX formatting of mathematical expressions from printed or handwritten content. Retrieval and summarization from long documents, slides, and multi-modal inputs. Multilingual document analysis and structured content extraction for global use cases. Robotic or mobile automation with vision-guided contextual interaction. May show degraded performance on extremely low-quality or occluded images. Not optimized for real-time applications on low-resource or edge devices due to computational demands. Variable accuracy on uncommon or low-resource languages/scripts. Long video processing may require substantial memory and is not optimized for streaming applications. Visual token settings affect performance; suboptimal configurations can impact results. In rare cases, outputs may contain hallucinated or contextually misaligned information.

NaNK
30
4

proxima-ocr-d.markdown-post3.0.l

license:apache-2.0
30
1

Fire-Detection-Engine

license:apache-2.0
28
1

Magpie-Qwen-CortexDual-0.6B-GGUF

NaNK
license:apache-2.0
28
0

Canopus-Photo-Shoot-Mini-LoRA

NaNK
27
5

DREX-062225-exp

NaNK
license:apache-2.0
27
5

Face-Diffusion-v0.1

NaNK
27
1

SmolLM2-1.7B-GGUF

NaNK
llama
27
1

Qwen3-4B-Thinking-2507-DAG-GGUF

> The sequelbox/Qwen3-4B-Thinking-2507-DAG-Reasoning model is an experimental specialist reasoning AI finetuned for multi-step causal analysis and reasoning, producing structured Directed Acyclic Graphs (DAGs) in response to user input across fields like programming, science, business, law, and more; it structures its output in a readable JSON graph format with confidence measures, enabling easy visualization and further analysis, and supports prompt-driven graph generation using a custom Qwen3-4B-Thinking-2507 inference approach, making it suitable for advanced reasoning tasks on both desktop and server environments; the model is part of open-source research, benefits from a custom DAG dataset, and is recommended for those needing clear causal structure in analysis or automation outputs. | File Name | Quant Type | File Size | | - | - | - | | Qwen3-4B-Thinking-2507-DAG-Reasoning.BF16.gguf | BF16 | 8.05 GB | | Qwen3-4B-Thinking-2507-DAG-Reasoning.F16.gguf | F16 | 8.05 GB | | Qwen3-4B-Thinking-2507-DAG-Reasoning.F32.gguf | F32 | 16.1 GB | | Qwen3-4B-Thinking-2507-DAG-Reasoning.Q2K.gguf | Q2K | 1.67 GB | | Qwen3-4B-Thinking-2507-DAG-Reasoning.Q3KL.gguf | Q3KL | 2.24 GB | | Qwen3-4B-Thinking-2507-DAG-Reasoning.Q3KM.gguf | Q3KM | 2.08 GB | | Qwen3-4B-Thinking-2507-DAG-Reasoning.Q3KS.gguf | Q3KS | 1.89 GB | | Qwen3-4B-Thinking-2507-DAG-Reasoning.Q4KM.gguf | Q4KM | 2.5 GB | | Qwen3-4B-Thinking-2507-DAG-Reasoning.Q4KS.gguf | Q4KS | 2.38 GB | | Qwen3-4B-Thinking-2507-DAG-Reasoning.Q5KM.gguf | Q5KM | 2.89 GB | | Qwen3-4B-Thinking-2507-DAG-Reasoning.Q5KS.gguf | Q5KS | 2.82 GB | | Qwen3-4B-Thinking-2507-DAG-Reasoning.Q6K.gguf | Q6K | 3.31 GB | | Qwen3-4B-Thinking-2507-DAG-Reasoning.Q80.gguf | Q80 | 4.28 GB | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
license:apache-2.0
27
0

Berenices-Mini-Emoji-LoRA

NaNK
26
5

Qwen2-VL-Math-Prase-2B-Instruct

NaNK
license:apache-2.0
26
4

QwQ-LCoT-7B-Instruct-GGUF

NaNK
Llama-Cpp
26
3

GWQ-9B-Preview2-Q8_0-GGUF

NaNK
llama-cpp
26
1

Qwen3-Coder-30B-A3B-Instruct-GGUF

> Qwen3-Coder-30B-A3B-Instruct is a state-of-the-art large language model from the Qwen series, specifically optimized for advanced agentic coding, browser-based automation, and foundational programming tasks. Featuring 30.5 billion total parameters with 3.3 billion activated in a Mixture-of-Experts (MoE) architecture, it delivers strong performance and efficiency for complex code and tool-use scenarios. Its standout long-context capability natively processes up to 262,144 tokens—expandable to 1 million with Yarn—making it ideal for repository-scale code understanding and generation. > The model supports agentic coding with advanced function-call handling, and is compatible with popular local inference platforms like Ollama, LMStudio, and llama.cpp. Designed for both pretraining and post-training stages, Qwen3-Coder-30B-A3B-Instruct runs exclusively in non-thinking mode, ensuring fast, high-quality outputs for coding and automation workflows without requiring explicit configuration for thinking blocks Use with llama.cpp Install llama.cpp through brew (works on Mac and Linux) Note: You can also use this checkpoint directly through the usage steps listed in the Llama.cpp repo as well. Step 2: Move into the llama.cpp folder and build it with `LLAMACURL=1` flag along with other hardware-specific flags (for ex: LLAMACUDA=1 for Nvidia GPUs on Linux).

NaNK
llama-cpp
26
1

Panacea-MegaScience-Qwen3-1.7B-GGUF

> Panacea-MegaScience-Qwen3-1.7B is a high-efficiency, multi-domain model fine-tuned on Qwen3-1.7B using the MegaScience/MegaScience dataset. MegaScience exhibits greater effectiveness for larger and stronger models, suggesting a scaling benefit for scientific instruction tuning. This model blends symbolic precision, scientific logic, and structured output fluency—making it an ideal tool for developers, educators, and researchers seeking advanced reasoning under constrained compute. | File Name | Quant Type | File Size | | - | - | - | | Panacea-MegaScience-Qwen3-1.7B.BF16.gguf | BF16 | 3.45 GB | | Panacea-MegaScience-Qwen3-1.7B.F16.gguf | F16 | 3.45 GB | | Panacea-MegaScience-Qwen3-1.7B.F32.gguf | F32 | 6.89 GB | | Panacea-MegaScience-Qwen3-1.7B.Q2K.gguf | Q2K | 778 MB | | Panacea-MegaScience-Qwen3-1.7B.Q3KL.gguf | Q3KL | 1 GB | | Panacea-MegaScience-Qwen3-1.7B.Q3KM.gguf | Q3KM | 940 MB | | Panacea-MegaScience-Qwen3-1.7B.Q3KS.gguf | Q3KS | 867 MB | | Panacea-MegaScience-Qwen3-1.7B.Q4KM.gguf | Q4KM | 1.11 GB | | Panacea-MegaScience-Qwen3-1.7B.Q4KS.gguf | Q4KS | 1.06 GB | | Panacea-MegaScience-Qwen3-1.7B.Q5KM.gguf | Q5KM | 1.26 GB | | Panacea-MegaScience-Qwen3-1.7B.Q5KS.gguf | Q5KS | 1.23 GB | | Panacea-MegaScience-Qwen3-1.7B.Q6K.gguf | Q6K | 1.42 GB | | Panacea-MegaScience-Qwen3-1.7B.Q80.gguf | Q80 | 1.83 GB | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
license:apache-2.0
26
0

Astral-4B-Preview-GGUF

> Astral-4B-Preview is a reasoning-centric language model from the Astral series, built on Qwen3-4b-thinking-2507 and fine-tuned on the nvidia/AceReason-1.1-SFT dataset to deliver configurable, step-by-step logical reasoning for research and development use. By including a "Reasoning-level" directive in the system prompt, users can control the model’s depth of reasoning—from direct answers to ultra-detailed reasoning traces—enabling nuanced, structured responses tailored to diverse problem-solving needs. As a preview release, Astral-4B-Preview is ideal for evaluating advanced reasoning capabilities with user-guided depth control in scientific and technical tasks. | File Name | Quant Type | File Size | | - | - | - | | Astral-4B-Preview.BF16.gguf | BF16 | 8.05 GB | | Astral-4B-Preview.F16.gguf | F16 | 8.05 GB | | Astral-4B-Preview.F32.gguf | F32 | 16.1 GB | | Astral-4B-Preview.Q2K.gguf | Q2K | 1.67 GB | | Astral-4B-Preview.Q3KL.gguf | Q3KL | 2.24 GB | | Astral-4B-Preview.Q3KM.gguf | Q3KM | 2.08 GB | | Astral-4B-Preview.Q3KS.gguf | Q3KS | 1.89 GB | | Astral-4B-Preview.Q4KM.gguf | Q4KM | 2.5 GB | | Astral-4B-Preview.Q4KS.gguf | Q4KS | 2.38 GB | | Astral-4B-Preview.Q5KM.gguf | Q5KM | 2.89 GB | | Astral-4B-Preview.Q5KS.gguf | Q5KS | 2.82 GB | | Astral-4B-Preview.Q6K.gguf | Q6K | 3.31 GB | | Astral-4B-Preview.Q80.gguf | Q80 | 4.28 GB | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
license:apache-2.0
26
0

Kepler-452b-LoRA-Flux-Dev-3D-Bubbly

NaNK
25
6

Dorado-WebSurf_Tool-ext

> Dorado-WebSurfTool-ext is a function-calling and agentic reasoning model fine-tuned from Qwen3-4B, designed for web search orchestration, tool-augmented reasoning, and dynamic problem-solving. > It excels at agentic decision-making, tool selection, and structured execution flow, making it ideal for retrieval-augmented generation (RAG), function calling, and tool-based query resolution. > [!note] > GGUF: https://huggingface.co/prithivMLmods/Dorado-WebSurfTool-ext-GGUF 1. Agentic Reasoning & Tool-Oriented Execution Built for orchestrating function calls, selecting and sequencing tools, and solving queries through structured multi-step reasoning. 2. Web Search Query Orchestration Integrates web search planning, retrieval grounding, and fact-checking, enabling intelligent query resolution from live data sources. 3. Dynamic Tool Selection & Execution Chains Chooses from an array of available tools — including web search, APIs, mathematical solvers, and structured data processors — to solve complex tasks. 4. Hybrid Symbolic-Probabilistic Logic Combines structured reasoning with probabilistic inference, ensuring accurate outcomes even in uncertainty-driven or multi-source contexts. 5. Structured Output Generation Generates responses in JSON, YAML, Markdown, or tool call schema formats, ideal for automation pipelines and agent frameworks. 6. Optimized Lightweight Footprint Maintains strong reasoning and tool orchestration capabilities in a 4B parameter model, deployable on mid-range GPUs, edge devices, and offline clusters. Function calling, tool orchestration, and agentic reasoning Web search query resolution and retrieval-based answering Dynamic tool selection and structured problem solving Automation workflows, API integration, and decision-making agents Technical structured output generation for RAG and agent frameworks Optimized for tool-assisted reasoning — less suited for standalone creative writing May require careful prompt engineering for complex multi-tool workflows Tool orchestration performance depends on external tool availability and integration quality Context length limits may affect very large multi-document tasks

license:apache-2.0
25
2

Lacaille-MoT-4B-Supreme2-GGUF

NaNK
license:apache-2.0
25
0

Llama-3B-Mono-Ceylia

NaNK
llama
24
4

gemma-3-270m-it-GGUF

> Gemma 3 270M and Gemma 3 270M-IT are lightweight, state-of-the-art multimodal models from Google, developed with the same research and technology as the Gemini family, supporting both text and image inputs (for instruction-tuned sizes above 270M), and delivering high-quality text outputs for tasks like question answering, summarization, image understanding, and code generation over a 32K context window. These models are trained on a diverse, multilingual dataset (over 140 languages) spanning web text, code, math, and images, with rigorous safety and quality filtering, and are designed for efficient deployment in resource-constrained environments. Both pre-trained (270M) and instruction-tuned (270M-IT) variants are openly available, offering robust benchmark performance, responsible AI development practices, and a broad range of academic, creative, and practical applications while emphasizing ethical use, safety, and transparency. | Model Name | Hugging Face Repository URL | |----------------------------|---------------------------------------------------------------------------------------------| | gemma-3-270m-it-GGUF | https://huggingface.co/prithivMLmods/gemma-3-270m-it-GGUF/tree/main/gemma-3-270m-it-GGUF | | gemma-3-270m-GGUF | https://huggingface.co/prithivMLmods/gemma-3-270m-it-GGUF/tree/main/gemma-3-270m-GGUF | | File Name | Quant Type | File Size | | - | - | - | | gemma-3-270m-it-BF16.gguf | BF16 | 543 MB | | gemma-3-270m-it-F16.gguf | F16 | 543 MB | | gemma-3-270m-it-Q80.gguf | Q80 | 292 MB | | File Name | Quant Type | File Size | | - | - | - | | gemma-3-270m-F16.gguf | F16 | 543 MB | | gemma-3-270m-Q80.gguf | Q80 | 292 MB | | gemma-3-270m-it-BF16.gguf | BF16 | 543 MB | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

24
1

Qwen-Image-Fragmented-Portraiture

license:apache-2.0
24
1

Llama3.1-8B-Grpo-Reasoning

NaNK
llama
24
0

Canopus-Textile-Pattern-adp-LoRA

NaNK
license:apache-2.0
23
11

Simple-Doodle-SD3.5-Turbo

23
5

Bpe-vocab-n-OCR

NaNK
license:apache-2.0
23
4

Fashion-Product-Gender

NaNK
license:apache-2.0
23
0

Multilabel-Portrait-SigLIP2

NaNK
license:apache-2.0
23
0

Anonymizer-4B-f32-GGUF

> Anonymizer-4B is the most powerful model in the Enchanted anonymizer series, built on the Qwen3-4B base and trained with a combination of supervised fine-tuning and GRPO using GPT-4.1 as a judge to achieve highly accurate, semantically consistent anonymization of personally identifiable information (PII). It provides precise PII replacement by generating semantically similar alternatives that maintain context, scoring 9.55/10 in anonymization quality, and is primarily intended for enterprise or research settings requiring top-quality privacy protection, though it requires higher-end hardware for real-time use. | File Name | Quant Type | File Size | | - | - | - | | Anonymizer-4B.BF16.gguf | BF16 | 8.05 GB | | Anonymizer-4B.F16.gguf | F16 | 8.05 GB | | Anonymizer-4B.F32.gguf | F32 | 16.1 GB | | Anonymizer-4B.Q2K.gguf | Q2K | 1.67 GB | | Anonymizer-4B.Q3KL.gguf | Q3KL | 2.24 GB | | Anonymizer-4B.Q3KM.gguf | Q3KM | 2.08 GB | | Anonymizer-4B.Q3KS.gguf | Q3KS | 1.89 GB | | Anonymizer-4B.Q4KM.gguf | Q4KM | 2.5 GB | | Anonymizer-4B.Q4KS.gguf | Q4KS | 2.38 GB | | Anonymizer-4B.Q5KM.gguf | Q5KM | 2.89 GB | | Anonymizer-4B.Q5KS.gguf | Q5KS | 2.82 GB | | Anonymizer-4B.Q6K.gguf | Q6K | 3.31 GB | | Anonymizer-4B.Q80.gguf | Q80 | 4.28 GB | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
license:apache-2.0
23
0

Ton618-Space-Wallpaper-LoRA

22
5

Qwen-UMLS-7B-Instruct

NaNK
22
4

Llama-Deepsync-1B-GGUF

NaNK
llama
22
3

Berenices-Alpha-DW-LoRA

NaNK
22
2

Lh41-1042-Magellanic-7B-0711

NaNK
license:apache-2.0
21
3

Spatial-VU

NaNK
license:apache-2.0
21
2

Flerovium-Llama-3B-GGUF

NaNK
llama
21
0

Qwen3-8B-CK-Pro-f32-GGUF

> The CognitiveKernel/Qwen3-8B-CK-Pro model is a fine-tuned variant of the Qwen3-8B base language model, trained using self-collected trajectories from queries as detailed in the Cognitive Kernel-Pro research. It is designed as a deep research agent and foundation model, achieving strong performance with Pass@1/3 scores of 32.7%/38.2% on the full GAIA dev set and 40.3%/49.3% on the text-only subset. This model builds upon the strengths of Qwen3-8B, which supports advanced reasoning, instruction-following, and multilingual capabilities, specifically optimized for research agent tasks through the Cognitive Kernel-Pro framework. It is not currently deployed by any inference provider on Hugging Face. The model leverages the underlying Qwen3-8B base and its finetuned versions to deliver enhanced agent capabilities for complex question-answering and information synthesis scenarios. `ollama run hf.co/prithivMLmods/Qwen3-8B-CK-Pro-f32-GGUF:Q2K` | File Name | Quant Type | File Size | | - | - | - | | Qwen3-8B-CK-Pro.BF16.gguf | BF16 | 16.4 GB | | Qwen3-8B-CK-Pro.F16.gguf | F16 | 16.4 GB | | Qwen3-8B-CK-Pro.F32.gguf | F32 | 32.8 GB | | Qwen3-8B-CK-Pro.Q2K.gguf | Q2K | 3.28 GB | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
license:apache-2.0
21
0

Canopus-Clothing-Adp-LoRA

NaNK
20
9

Guard-Against-Unsafe-Content-Siglip2

NaNK
license:apache-2.0
20
6

JSONify-Flux

NaNK
license:apache-2.0
20
3

Qwen3-VL-8B-Instruct-abliterated-v2

NaNK
license:apache-2.0
20
2

Llama-3.2-3B-Math-Oct

Llama-3.2-3B-Math-Oct is a math role-play model designed to solve mathematical problems and enhance the reasoning capabilities of 3B-parameter models. These models have proven highly effective in context understanding, reasoning, and mathematical problem-solving, based on the Llama 3.2 is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions use supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to align with human preferences for helpfulness and safety. Starting with `transformers >= 4.43.0` onward, you can run conversational inference using the Transformers `pipeline` abstraction or by leveraging the Auto classes with the `generate()` function. Make sure to update your transformers installation via `pip install --upgrade transformers`. Intended Use 1. Mathematical Problem Solving: Llama-3.2-3B-Math-Oct is designed for solving a wide range of mathematical problems, including arithmetic, algebra, calculus, and probability. 2. Reasoning Enhancement: It enriches logical reasoning capabilities, helping users understand and solve complex mathematical concepts. 3. Context Understanding: The model is highly effective in interpreting problem statements, mathematical scenarios, and context-heavy equations. 4. Educational Support: It serves as a learning tool for students, educators, and enthusiasts, providing step-by-step explanations for mathematical solutions. 5. Scenario Simulation: The model can role-play specific mathematical scenarios, such as tutoring, creating math problems, or acting as a math assistant. Limitations 1. Accuracy Constraints: While effective in many cases, the model may occasionally provide incorrect solutions, particularly for highly complex or unconventional problems. 2. Parameter Limitation: Being a 3B-parameter model, it might lack the precision and capacity of larger models for intricate problem-solving. 3. Lack of Domain-Specific Expertise: The model may struggle with problems requiring niche mathematical knowledge or specialized fields like advanced topology or quantum mechanics. 4. Dependency on Input Clarity: Ambiguous or poorly worded problem statements might lead to incorrect interpretations and solutions. 5. Inability to Learn Dynamically: The model cannot improve its understanding or reasoning dynamically without retraining. 6. Non-Mathematical Queries: While optimized for mathematics, the model may underperform in general-purpose tasks compared to models designed for broader use cases. 7. Computational Resources: Deploying the model may require significant computational resources for real-time usage.

NaNK
llama
20
1

WEBGEN-4B-Preview-f32-GGUF

> WEBGEN-4B-Preview is a 4B parameter model purpose-built for generating modern, responsive web pages with clean semantic HTML, CSS, and Tailwind, optimized for single-file sites and component blocks. With its compact size for local and fast iteration, the model consistently produces production-quality layouts—favoring structured markup, balanced spacing, and contemporary design patterns—making it ideal for quickly prototyping or deploying landing pages, marketing sites, and web components directly from a natural language prompt.WEBGEN-4B-Preview is a compact, 4B parameter web generation model designed to turn natural prompts into clean, production-ready HTML, CSS, and Tailwind markup, optimized for fast, local runs and consistent, modern layouts suitable for single-file sites and reusable components. This web-only generator emphasizes semantic structure, responsive spacing, and opinionated design, making it ideal for quick prototyping and web development without dependencies on external JavaScript libraries. | File Name | Quant Type | File Size | | - | - | - | | WEBGEN-4B-Preview.BF16.gguf | BF16 | 8.05 GB | | WEBGEN-4B-Preview.F16.gguf | F16 | 8.05 GB | | WEBGEN-4B-Preview.F32.gguf | F32 | 16.1 GB | | WEBGEN-4B-Preview.Q2K.gguf | Q2K | 1.67 GB | | WEBGEN-4B-Preview.Q3KL.gguf | Q3KL | 2.24 GB | | WEBGEN-4B-Preview.Q3KM.gguf | Q3KM | 2.08 GB | | WEBGEN-4B-Preview.Q3KS.gguf | Q3KS | 1.89 GB | | WEBGEN-4B-Preview.Q4KM.gguf | Q4KM | 2.5 GB | | WEBGEN-4B-Preview.Q4KS.gguf | Q4KS | 2.38 GB | | WEBGEN-4B-Preview.Q5KM.gguf | Q5KM | 2.89 GB | | WEBGEN-4B-Preview.Q5KS.gguf | Q5KS | 2.82 GB | | WEBGEN-4B-Preview.Q6K.gguf | Q6K | 3.31 GB | | WEBGEN-4B-Preview.Q80.gguf | Q80 | 4.28 GB | (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

NaNK
license:apache-2.0
20
1

GCIRS-Reasoning-1.5B-R1-GGUF

NaNK
license:apache-2.0
20
0

Phi-4-Empathetic

Phi-4 Empathetic [ Responsible Reasoning & Emotional Thought Generation ] `[Phi-4 Empathetic finetuned]` from Microsoft's Phi-4 is an advanced open model built upon a blend of high-quality synthetic datasets, data from filtered public domain websites, and carefully selected academic resources. It excels at responsible human-like reasoning, empathetic dialogue, and emotional thought generation. The model is designed to engage in nuanced, thoughtful conversations, with outputs that can include special characters and emojis for expressive communication. 🌟 Phi-4 Empathetic employs a sophisticated safety post-training approach, leveraging both open-source and proprietary datasets. Safety alignment is achieved using a combination of SFT (Supervised Fine-Tuning) and DPO (Direct Preference Optimization), targeting responsible interaction and emotional awareness in diverse contexts. Phi-4 Empathetic is fine-tuned on a carefully curated dataset tailored for empathetic and responsible reasoning tasks. The dataset incorporates the Chain of Thought (CoT) methodology, emphasizing logical reasoning, emotional nuance, and step-by-step thought processes. Additionally, it includes data optimized for generating responses that resonate with human emotions, making it ideal for: - Emotional Support Applications 🤗 - Responsible Conversations 💬 - Thoughtful Problem-Solving 🧠 You can ensure correct formatting for empathetic dialogue by using `tokenizer.applychattemplate` as follows: The Phi-4 Empathetic model is optimized for applications that require thoughtful and emotionally aware interactions. Below are some suggested use cases: 1. Emotional Support & Counseling 💖 - Providing thoughtful responses to users seeking emotional encouragement or advice. - Generating empathetic messages for mental health and well-being applications. 2. Responsible Dialogue Generation 🗣️ - Engaging in nuanced conversations with a focus on fairness, safety, and ethical considerations. - Ensuring that interactions remain respectful and aligned with safety guidelines. 3. Creative Writing Assistance ✍️ - Helping users craft emotionally engaging content, including stories, poems, and personal messages. - Assisting in generating content enriched with special characters and emojis for expressive communication. 4. Educational Tools 🎓 - Offering step-by-step explanations with an empathetic tone for better understanding. - Generating thoughtful Q&A responses for various subjects. 5. Customer Support 🤝 - Automating empathetic responses to customer queries. - Handling emotionally sensitive customer service interactions with care. 6. Social Media Engagement 📱 - Generating creative, engaging, and emotionally resonant posts for social media platforms. - Providing personalized message suggestions enriched with emojis and special characters. While Phi-4 Empathetic is highly capable, it has certain limitations users should be aware of: 1. Bias and Fairness: Despite extensive safety alignment, biases may still emerge in the model’s responses. Users should exercise discretion, particularly in sensitive contexts. 2. Emotional Nuance: The model may occasionally misinterpret the emotional tone of a prompt, leading to less relevant or inappropriate responses. 3. Real-Time Knowledge: The model's knowledge is based on the data it was trained on and does not include real-time or post-training updates. It may not reflect recent events or changes in knowledge. 4. Safety and Harmlessness: Although the model is aligned with safety standards, there may still be cases where outputs require human oversight to ensure appropriateness. 5. Resource Requirements: Running the model efficiently may require significant computational resources, especially in large-scale or real-time applications. 6. Ethical Considerations: The model must be used responsibly, avoiding any malicious applications such as generating harmful content or spreading misinformation. 7. Domain-Specific Limitations: While it performs well in general-purpose tasks, it may need further fine-tuning for highly specialized domains, such as legal, medical, or financial applications. 1. Emojis & Special Characters 🎉💡 The model can generate responses with emojis and special characters for expressive communication, making it ideal for social media and personal messaging applications. 2. Human-Like Reasoning 🧠 Fine-tuned for responsible reasoning and empathetic dialogue, it excels at generating thoughtful and human-like responses. 3. Advanced Safety Alignment 🔒 The model employs iterative SFT and DPO techniques to ensure that its outputs are helpful, harmless, and aligned with ethical standards.

NaNK
llama
19
12