kingabzpro

67 models • 1 total models in database
Sort by:

wav2vec2-large-xls-r-300m-Urdu

NaNK
license:apache-2.0
603,242
13

wav2vec2-large-xls-r-1b-Swedish

NaNK
license:apache-2.0
1,144
1

whisper-large-v3-turbo-urdu

NaNK
license:apache-2.0
507
1

wav2vec2-large-xls-r-1b-Indonesian

NaNK
license:apache-2.0
217
1

llama-3.2-3b-it-Ecommerce-ChatBot

NaNK
llama
169
8

wav2vec2-large-xlsr-53-punjabi

NaNK
license:apache-2.0
92
4

whisper-tiny-urdu

license:apache-2.0
76
0

Llama-3.2-3b-it-customer-support

NaNK
llama
68
0

wav2vec2-large-xls-r-300m-Swedish

NaNK
license:apache-2.0
47
1

Llama-3.1-8B-MATH

NaNK
llama
37
0

whisper-large-v3-urdu

NaNK
license:apache-2.0
36
0

wav2vec2-urdu

NaNK
license:apache-2.0
32
4

llama-3.2-3b-it-Ecommerce-ChatBot-Q4_K_M-GGUF

NaNK
llama3.2
32
0

wav2vec2-large-xls-r-1b-Irish

NaNK
license:apache-2.0
31
1

llama-3-8b-chat-doctor

NaNK
llama
31
0

wav2vec2-large-xlsr-53-wolof

license:apache-2.0
27
3

Llama-3-8B-ORPO

NaNK
llama
25
0

functiongemma-hermes-3k-ft

license:apache-2.0
22
1

Llama-3.1-8B-Instruct-Mental-Health-Classification

NaNK
llama
19
10

wav2vec2-large-xlsr-300-arabic

NaNK
license:apache-2.0
19
2

wav2vec2-60-urdu

license:apache-2.0
16
1

dnq-SpaceInvadersNoFrameskip-V4

16
1

Gemma-2-9b-it-chat-doctor-Q4_K_M-GGUF

NaNK
llama-cpp
16
0

a2c-AntBulletEnv-v0

15
0

a2c-HalfCheetahBulletEnv-v0

14
0

wav2vec2-60-Urdu-V8

NaNK
license:apache-2.0
13
1

Llama-3.2-3b-it-mental-health

NaNK
llama
12
0

whisper-large-v3-urdu-ct2

NaNK
license:apache-2.0
12
0

dnq-BreakoutNoFrameskip-v4

NaNK
10
1

poca-SoccerTwos-v1

10
0

wav2vec2-large-xls-r-300m-Indonesian

license:apache-2.0
9
0

Moonman-Lunar-Landing-v2

9
0

llama-3-8b-chat-doctor-Q4_K_M-GGUF

NaNK
llama-cpp
9
0

Full-Force-MountainCar-v0

8
1

DialoGPT-small-Rick-Bot

7
5

MLAgents-Pyramids

7
0

MLAgents-Worm

6
0

a2c-PandaReachDense-v2

6
0

mistral_7b_guanaco

NaNK
dataset:mlabonne/guanaco-llama2-1k
5
3

Helsinki-NLP-opus-yor-mul-en

license:apache-2.0
5
2

Gemma-3-4B-Python-Reasoning

- Developed by: kingabzpro - License: apache-2.0 - Finetuned from model : unsloth/gemma-3-4b-it-unsloth-bnb-4bit This gemma3 model was trained 2x faster with Unsloth and Huggingface's TRL library.

NaNK
license:apache-2.0
5
1

wav2vec2-large-xls-r-300m-Tatar

NaNK
license:apache-2.0
4
1

whisper-small-hi-cv

NaNK
license:apache-2.0
4
1

zephyr-7b-beta-Agent-Instruct

NaNK
license:apache-2.0
4
1

whisper-base-urdu-full

NaNK
license:apache-2.0
4
1

MLAgents-PushBlock

4
0

gemma-7b-it-v2-role-play

NaNK
license:apache-2.0
4
0

sdxl-lora-abid

NaNK
3
4

wav2vec2-large-xls-r-300m-hi

NaNK
license:apache-2.0
3
0

mistral_7b-instruct-guanaco

NaNK
3
0

phi-2-role-play

NaNK
license:apache-2.0
3
0

DeepSeek-R1-Medical-COT

NaNK
llama
2
8

Gemma-2-9b-it-chat-doctor

NaNK
license:apache-2.0
2
1

Phi-3.5-mini-instruct-Ecommerce-Text-Classification

license:apache-2.0
1
1

llama-2-7b-chat-guanaco

NaNK
llama
1
0

Medgemma Brain Cancer

`medgemma-brain-cancer` is a fine-tuned version of google/medgemma-4b-it, trained specifically for brain tumor diagnosis and classification from MRI scans. This model leverages vision-language learning for enhanced medical imaging interpretation. Base Model: google/medgemma-4b-it Dataset: orvile/brain-cancer-mri-dataset Fine-tuning Approach: Supervised fine-tuning (SFT) using Transformers Reinforcement Learning (TRL) Task: Brain tumor classification from MRI images Pipeline Tag: `image-text-to-text` Accuracy Improvement: Base model accuracy: 33% Fine-tuned model accuracy: 89% Explore the training pipeline, evaluation results, and experiments in the notebook: This model is intended for research and educational purposes related to medical imaging, specifically brain tumor classification. It is not a certified diagnostic tool and should not be used in clinical decision-making without further validation.

NaNK
license:apache-2.0
0
13

Qwen-3-32B-Medical-Reasoning

NaNK
license:apache-2.0
0
12

Magistral-Small-Medical-QA

NaNK
license:apache-2.0
0
4

DeepSeek-R1-0528-Qwen3-8B-Medical-Reasoning

NaNK
license:apache-2.0
0
3

CELEB-GANs

license:apache-2.0
0
2

Llama-4-Scout-17B-16E-Instruct-Medical-ChatBot

Fine-tuning Llama 4 (Scout 17B 16E) in 4-bit Quantization for Medical Reasoning This project fine-tunes the `meta-llama/Llama-4-Scout-17B-16E-Instruct` model using a medical reasoning dataset (`FreedomIntelligence/medical-o1-reasoning-SFT`) with 4-bit quantization for memory-efficient training. Make sure your Hugging Face token is stored in an environment variable: The notebook will automatically log you in using this token. 1. Load the Model and Tokenizer The script downloads the Llama 4 Scout model and applies 4-bit quantization with `BitsAndBytesConfig` for efficient memory usage. 2. Prepare the Dataset - The notebook uses `FreedomIntelligence/medical-o1-reasoning-SFT` (first 500 samples). - It formats each example into an instruction-following prompt with step-by-step chain-of-thought reasoning. 3. Fine-tuning - Fine-tuning is set up with PEFT (LoRA / Adapter Tuning style) to modify a small subset of model parameters. - TRL (Transformer Reinforcement Learning) is used to fine-tune efficiently. 4. Push Fine-tuned Model - After training, the fine-tuned model and tokenizer are pushed back to your Hugging Face account. Here is the training notebook: Finetuningllama4.ipynb) - Base Model: `meta-llama/Llama-4-Scout-17B-16E-Instruct` - Quantization: 4-bit (NF4) - Training: PEFT + TRL - Dataset: 500 examples from medical reasoning dataset - GPU Required: Make sure you have access to 3X H200s. Get it from RunPod for an hours. Training took only 7 minutes. - Environment: The notebook expects an environment where NVIDIA CUDA drivers are available (`nvidia-smi` check is included). - Memory Efficiency: 4-bit loading greatly reduces memory footprint.

NaNK
base_model:meta-llama/Llama-4-Scout-17B-16E-Instruct
0
2

deepseek-ocr-2-urdu-ocr-1m-lora

license:apache-2.0
0
1

qwen3vl-open-schematics-lora

NaNK
license:apache-2.0
0
1

q-FrozenLake-v1-4x4-noSlippery

0
1

q-Taxi-v3

0
1

Phi-4-Reasoning-Plus-FinQA-COT

This project fine-tunes the `microsoft/Phi-4-reasoning-plus` model using a medical reasoning dataset (`TheFinAI/Fino1ReasoningPathFinQA`). Make sure your Hugging Face token is stored in an environment variable: The notebook will automatically log you in using this token. 1. Load the Model and Tokenizer The script downloads the full Phi-4-reasoning-plus model. 2. Prepare the Dataset - The notebook uses `TheFinAI/Fino1ReasoningPathFinQA` (first 1000 samples). - It formats each example into an instruction-following prompt with step-by-step chain-of-thought reasoning. 3. Fine-tuning - Fine-tuning is set up with PEFT (LoRA / Adapter Tuning style) to modify a small subset of model parameters. - TRL (Transformer Reinforcement Learning) is used to fine-tune efficiently. 4. Push Fine-tuned Model - After training, the fine-tuned model and tokenizer are pushed back to your Hugging Face account. >> Here is the training notebook: FinetuningPhi-4-Reasoning-Plus - Base Model: `microsoft/Phi-4-reasoning-plus` - Training: PEFT + TRL - Dataset: 1000 examples FinQA reasoning dataset - GPU Required: Make sure you have access to 1X A100s. Get it from RunPod for an hours. Training took only 7 minutes. - Environment: The notebook expects an environment where NVIDIA CUDA drivers are available (`nvidia-smi` check is included).

license:apache-2.0
0
1

gpt-oss-20b-dermatology-qa

NaNK
license:apache-2.0
0
1