MMLU Pro Leaderboard
Massive Multitask Language Understanding - Professional Edition
Why This Matters
Tests breadth of knowledge across 57 professional domains - critical for general-purpose AI applications
Good Scores
70%+ is strong, 80%+ is excellent, 85%+ is state-of-the-art
Use Cases
- •Research assistants
- •Educational tools
- •Professional advisory systems
- •General Q&A chatbots
Peak Score
70.03
Average
25.05
Models Tested
53,438
Median Score
26.31
Efficiency Leaders
Best performance per billion parameters - The smart choices
ChatWaifu_v1.4
100.0M params • Score: 27.50
Efficiency
274.99
ChatWaifu_v1.4
100.0M params • Score: 27.50
Efficiency
274.99
ChatWaifu_v1.4
100.0M params • Score: 27.50
Efficiency
274.99
ChatWaifu_v1.4
100.0M params • Score: 27.50
Efficiency
274.99
ChatWaifu_v1.4
100.0M params • Score: 27.50
Efficiency
274.99
ChatWaifu_v1.4
100.0M params • Score: 27.50
Efficiency
274.99
ChatWaifu_v1.4
100.0M params • Score: 27.50
Efficiency
274.99
ChatWaifu_v1.4
100.0M params • Score: 27.50
Efficiency
274.99
ChatWaifu_v1.4
100.0M params • Score: 27.50
Efficiency
274.99
ChatWaifu_v1.4
100.0M params • Score: 27.50
Efficiency
274.99
Performance by Model Size
How different size classes perform on this benchmark
xlarge
Avg Score: 45.45
🏆 Open Source Champions
Top permissively licensed models
📈 Most Downloaded Models
Popularity meets performance
📄 License Analysis
Performance by license type
🔧 Framework Analysis
Performance by framework
About MMLU Pro
MMLU Pro is an enhanced version of the MMLU benchmark, designed to test language models across 57 professional subjects including STEM, humanities, social sciences, and more. It provides a comprehensive evaluation of a model's knowledge and reasoning capabilities.
Test These Models Yourself
Run benchmarks on your own data with these platforms
Together.ai
Instant API access to this model
Production-ready inference API. Start free, scale to millions.
Try Free APIModal
Run this model on serverless GPU
Deploy in seconds with $30 free credits. Pay only for what you use.
Get $30 Free CreditsRunPod
Rent GPU starting at $0.34/hour
Deploy on cloud GPU or serverless. 70% cheaper than AWS.
Start from $0.34/hrDisclosure: We may earn a commission from these partners. This helps keep LLMYourWay free.
Complete Leaderboard
Top 50 models ranked by MMLU Pro performance
Calme 3.2 Instruct 78b
70.03
Score
Calme 3.2 Instruct 78b
70.03
Score
Calme 3.2 Instruct 78b
70.03
Score
Calme 3.2 Instruct 78b
70.03
Score
Calme 3.2 Instruct 78b
70.03
Score
Calme 3.2 Instruct 78b
70.03
Score
Calme 3.2 Instruct 78b
70.03
Score
Calme 3.2 Instruct 78b
70.03
Score
Calme 3.2 Instruct 78b
70.03
Score
Calme 3.2 Instruct 78b
70.03
Score
Calme 3.2 Instruct 78b
70.03
Score
Calme 3.2 Instruct 78b
70.03
Score
Calme 3.2 Instruct 78b
70.03
Score
Calme 3.2 Instruct 78b
70.03
Score
Calme 3.2 Instruct 78b
70.03
Score
Calme 3.2 Instruct 78b
70.03
Score
Calme 3.2 Instruct 78b
70.03
Score
Calme 3.2 Instruct 78b
70.03
Score
Calme 3.2 Instruct 78b
70.03
Score
Calme 3.2 Instruct 78b
70.03
Score
Calme 3.2 Instruct 78b
70.03
Score
Calme 3.2 Instruct 78b
70.03
Score
Calme 3.2 Instruct 78b
70.03
Score
Calme 3.2 Instruct 78b
70.03
Score
calme-3.1-instruct-78b
68.72
Score
calme-3.1-instruct-78b
68.72
Score
calme-3.1-instruct-78b
68.72
Score
calme-3.1-instruct-78b
68.72
Score
calme-3.1-instruct-78b
68.72
Score
calme-3.1-instruct-78b
68.72
Score
calme-3.1-instruct-78b
68.72
Score
calme-3.1-instruct-78b
68.72
Score
calme-3.1-instruct-78b
68.72
Score
calme-3.1-instruct-78b
68.72
Score
calme-3.1-instruct-78b
68.72
Score
calme-3.1-instruct-78b
68.72
Score
calme-3.1-instruct-78b
68.72
Score
calme-3.1-instruct-78b
68.72
Score
calme-3.1-instruct-78b
68.72
Score
calme-3.1-instruct-78b
68.72
Score
calme-3.1-instruct-78b
68.72
Score
calme-3.1-instruct-78b
68.72
Score
calme-3.1-instruct-78b
68.72
Score
calme-3.1-instruct-78b
68.72
Score
calme-3.1-instruct-78b
68.72
Score
calme-3.1-instruct-78b
68.72
Score
calme-3.1-instruct-78b
68.72
Score
calme-3.1-instruct-78b
68.72
Score
CalmeRys-78B-Orpo-v0.1
66.80
Score
CalmeRys-78B-Orpo-v0.1
66.80
Score