Models
Model Comparison
14 open-source models across Groq, Together, and Fireworks — text LLMs and speech-to-text. Pellet routes each request to the best model for cost, speed, and quality.
Whisper Large v3
whisper-large-v3
1.5B
95%
$0.111
—
~500ms Med
Whisper Large v3 Turbo
whisper-large-v3-turbo
809M
90%
$0.040
—
~300ms Fast
DeepSeek R1
deepseek-ai/DeepSeek-R1
685B
89%
$0.55
$2.19
~2.5s Slow
Llama 3.3 70B
llama-3.3-70b-versatile
70B
88%
$0.59
$0.79
~400ms Fast
Llama 3.3 70B Turbo
meta-llama/Llama-3.3-70B-Instruct-Turbo
70B
88%
$0.59
$0.79
~1.5s Slow
DeepSeek V3.1
deepseek-ai/DeepSeek-V3.1
685B
86%
$0.50
$1.50
~2.0s Slow
Distil Whisper v3 EN
distil-whisper-large-v3-en
756M
85%
$0.020
—
~200ms Fast
Mistral Small 24B
mistralai/Mistral-Small-24B-Instruct-2501
24B
80%
$0.10
$0.30
~1.2s Slow
Qwen 3.5 9B
Qwen/Qwen3.5-9B
9B
78%
$0.06
$0.10
~900ms Med
Mixtral 8x7B
mistralai/Mixtral-8x7B-Instruct-v0.1
47B
72%
$0.24
$0.24
~1.0s Med
Qwen 2.5 7B
Qwen/Qwen2.5-7B-Instruct-Turbo
7B
72%
$0.05
$0.08
~800ms Med
Llama 3.1 8B
llama-3.1-8b-instant
8B
71%
$0.05
$0.08
~200ms Fast
Llama 3 8B Lite
meta-llama/Meta-Llama-3-8B-Instruct-Lite
8B
63%
$0.05
$0.08
~800ms Med
Gemma 3n E4B
google/gemma-3n-E4B-it
4B
61%
$0.03
$0.06
~800ms Med
Performance by Task
Routing Confidence Scores
How confidently Pellet routes each task type to each model. Higher = better fit.
| Model | Classification | Code Gen | Content Gen | Extraction | Formatting | Moderation | Q&A | Reasoning | Sentiment | Speech | Structured | Summary | Translation |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Gemma 3n E4B 4B | 65 | 40 | 60 | 62 | 58 | 60 | 65 | 83 | 62 | 0 | 58 | 60 | 58 |
Llama 3.1 8B 8B | 68 | 65 | 70 | 68 | 79 | 68 | 72 | 75 | 70 | 0 | 79 | 72 | 68 |
Qwen 2.5 7B 7B | 74 | 85 | 72 | 70 | 68 | 70 | 74 | 78 | 70 | 0 | 68 | 70 | 68 |
Qwen 3.5 9B 9B | 76 | 82 | 78 | 76 | 74 | 76 | 80 | 88 | 76 | 0 | 74 | 78 | 76 |
Llama 3 8B Lite 8B | 65 | 60 | 65 | 62 | 62 | 62 | 65 | 68 | 62 | 0 | 62 | 65 | 62 |
Mixtral 8x7B 47B | 72 | 70 | 74 | 72 | 70 | 72 | 74 | 76 | 72 | 0 | 70 | 74 | 72 |
Mistral Small 24B 24B | 81 | 85 | 80 | 79 | 78 | 79 | 82 | 85 | 79 | 0 | 78 | 80 | 79 |
Llama 3.3 70B 70B | 87 | 88 | 86 | 87 | 92 | 87 | 88 | 90 | 87 | 0 | 92 | 86 | 85 |
Llama 3.3 70B Turbo 70B | 87 | 88 | 86 | 87 | 92 | 87 | 88 | 90 | 87 | 0 | 92 | 86 | 85 |
DeepSeek V3.1 685B | 89 | 87 | 84 | 89 | 82 | 89 | 89 | 88 | 85 | 0 | 82 | 84 | 82 |
DeepSeek R1 685B | 91 | 89 | 86 | 91 | 85 | 91 | 91 | 95 | 88 | 0 | 85 | 86 | 91 |
Whisper Large v3 Turbo 809M | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 90 | 0 | 0 | 0 |
Whisper Large v3 1.5B | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 95 | 0 | 0 | 0 |
Distil Whisper v3 EN 756M | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 85 | 0 | 0 | 0 |