name
stringlengths
8
58
hf_id
stringlengths
12
49
author
stringlengths
3
21
providers
listlengths
1
16
median_input_cost
float64
0
75
median_output_cost
float64
0
150
low_input_cost
float64
0
75
low_output_cost
float64
0
150
high_input_cost
float64
0
75
high_output_cost
float64
0
150
is_open_weights
bool
2 classes
Microsoft: Phi 4 Multimodal Instruct
microsoft/Phi-4-multimodal-instruct
microsoft
[ { "name": "DeepInfra", "context": 131000, "max_output": 131000, "input": 0.07, "output": 0.14, "latency": 0.47, "throughput": 70 } ]
0.07
0.14
0.07
0.14
0.07
0.14
true
Qwen: QwQ 32B
Qwen/QwQ-32B
Qwen
[ { "name": "DeepInfra", "context": 131000, "max_output": 131000, "input": 0.12, "output": 0.18, "latency": 0.15, "throughput": 39.52 }, { "name": "Nebius AI Studio", "context": 131000, "max_output": 131000, "input": 0.15, "output": 0.45, "latency": 0.55, "throughput": 22.54 }, { "name": "NovitaAI", "context": 33000, "max_output": 33000, "input": 0.18, "output": 0.2, "latency": 1.22, "throughput": 32.7 }, { "name": "Groq", "context": 131000, "max_output": 131000, "input": 0.29, "output": 0.39, "latency": 0.24, "throughput": 426.7 }, { "name": "Hyperbolic", "context": 131000, "max_output": 131000, "input": 0.4, "output": 0.4, "latency": 0.34, "throughput": 35.14 }, { "name": "Parasail", "context": 131000, "max_output": 131000, "input": 0.5, "output": 0.5, "latency": 0.77, "throughput": 60.56 }, { "name": "Nebius AI Studio (Fast)", "context": 131000, "max_output": 131000, "input": 0.5, "output": 1.5, "latency": 0.52, "throughput": 80.88 }, { "name": "Fireworks", "context": 131000, "max_output": 131000, "input": 0.9, "output": 0.9, "latency": 0.65, "throughput": 143.5 } ]
0.4
0.4
0.12
0.18
0.5
1.5
true
Qwen: Qwen2.5 32B Instruct
Qwen/Qwen2.5-32B-Instruct
Qwen
[ { "name": "Groq", "context": 131000, "max_output": 131000, "input": 0.79, "output": 0.79, "latency": 0.28, "throughput": 190.5 } ]
0.79
0.79
0.79
0.79
0.79
0.79
true
Perplexity: R1 1776
perplexity-ai/r1-1776
perplexity-ai
[ { "name": "Perplexity", "context": 128000, "max_output": 128000, "input": 2, "output": 8, "latency": 1.5, "throughput": 38.86 } ]
2
8
2
8
2
8
true
Llama Guard 3 8B
meta-llama/Llama-Guard-3-8B
meta-llama
[ { "name": "Groq", "context": 8000, "max_output": 8000, "input": 0.2, "output": 0.2, "latency": null, "throughput": null }, { "name": "SambaNova", "context": 16000, "max_output": 16000, "input": 0.3, "output": 0.3, "latency": 0.72, "throughput": 924.6 } ]
0.3
0.3
0.2
0.2
0.3
0.3
true
Llama 3.1 Tulu 3 405B
allenai/Llama-3.1-Tulu-3-405B
allenai
[ { "name": "SambaNova", "context": 16000, "max_output": 16000, "input": 5, "output": 10, "latency": 3.36, "throughput": 84.19 } ]
5
10
5
10
5
10
true
DeepSeek: R1 Distill Llama 8B
meta-llama/Llama-3.1-8B
meta-llama
[ { "name": "NovitaAI", "context": 32000, "max_output": 32000, "input": 0.04, "output": 0.04, "latency": 0.99, "throughput": 50.36 } ]
0.04
0.04
0.04
0.04
0.04
0.04
true
AionLabs: Aion-1.0-Mini
FuseAI/FuseO1-DeepSeekR1-QwQ-SkyT1-32B-Preview
FuseAI
[ { "name": "AionLabs", "context": 33000, "max_output": 33000, "input": 0.7, "output": 1.4, "latency": 1.01, "throughput": 207.3 } ]
0.7
1.4
0.7
1.4
0.7
1.4
true
Qwen: Qwen2.5 VL 72B Instruct
Qwen/Qwen2.5-VL-72B-Instruct
Qwen
[ { "name": "Parasail", "context": 32000, "max_output": 32000, "input": 0.7, "output": 0.7, "latency": 1.61, "throughput": 35.92 } ]
0.7
0.7
0.7
0.7
0.7
0.7
true
DeepSeek: R1 Distill Qwen 1.5B
Qwen/Qwen2.5-Math-1.5B
Qwen
[ { "name": "Together", "context": 131000, "max_output": 33000, "input": 0.18, "output": 0.18, "latency": 0.26, "throughput": 393.8 } ]
0.18
0.18
0.18
0.18
0.18
0.18
true
Mistral: Mistral Small 3
mistralai/Mistral-Small-24B-Instruct-2501
mistralai
[ { "name": "DeepInfra", "context": 33000, "max_output": 8000, "input": 0.07, "output": 0.14, "latency": 0.74, "throughput": 59.34 }, { "name": "Mistral", "context": 32000, "max_output": 32000, "input": 0.1, "output": 0.3, "latency": 0.34, "throughput": 101.8 }, { "name": "Ubicloud", "context": 33000, "max_output": 33000, "input": 0.3, "output": 0.3, "latency": 1.25, "throughput": 31.15 }, { "name": "Together", "context": 33000, "max_output": 2000, "input": 0.8, "output": 0.8, "latency": 1.19, "throughput": 48.56 }, { "name": "Fireworks", "context": 33000, "max_output": 33000, "input": 0.9, "output": 0.9, "latency": 2.16, "throughput": 27.53 } ]
0.3
0.3
0.07
0.14
0.9
0.9
true
DeepSeek: R1 Distill Qwen 32B
Qwen/Qwen2.5-32B
Qwen
[ { "name": "DeepInfra", "context": 131000, "max_output": 8000, "input": 0.12, "output": 0.18, "latency": 0.13, "throughput": 36.65 }, { "name": "Cloudflare", "context": 80000, "max_output": 80000, "input": 0.5, "output": 4.88, "latency": 0.32, "throughput": 30.32 }, { "name": "Groq", "context": 131000, "max_output": 131000, "input": 0.79, "output": 0.79, "latency": 0.13, "throughput": 137.4 }, { "name": "NovitaAI", "context": 64000, "max_output": 64000, "input": 0.3, "output": 0.3, "latency": 6.37, "throughput": 30.85 } ]
0.79
0.79
0.12
0.18
0.5
4.88
true
DeepSeek: R1 Distill Qwen 14B
deepseek-ai/DeepSeek-R1-Distill-Qwen-14B
deepseek-ai
[ { "name": "Together", "context": 131000, "max_output": 33000, "input": 1.6, "output": 1.6, "latency": 0.25, "throughput": 140.6 }, { "name": "NovitaAI", "context": 64000, "max_output": 64000, "input": 0.15, "output": 0.15, "latency": 0.94, "throughput": 38.73 } ]
1.6
1.6
0.15
0.15
1.6
1.6
true
DeepSeek: R1 Distill Llama 70B
deepseek-ai/DeepSeek-R1-Distill-Llama-70B
deepseek-ai
[ { "name": "DeepInfra", "context": 131000, "max_output": 8000, "input": 0.23, "output": 0.69, "latency": 9.61, "throughput": 59.42 }, { "name": "Nebius AI Studio", "context": 128000, "max_output": 128000, "input": 0.25, "output": 0.75, "latency": 5.37, "throughput": 72.33 }, { "name": "inference.net", "context": 16000, "max_output": 16000, "input": 0.4, "output": 0.4, "latency": 10.16, "throughput": 23.81 }, { "name": "SambaNova", "context": 16000, "max_output": 16000, "input": 0.7, "output": 1.4, "latency": 2.79, "throughput": 489.8 }, { "name": "Groq", "context": 131000, "max_output": 131000, "input": 0.75, "output": 0.99, "latency": 1.28, "throughput": 623.3 }, { "name": "Together", "context": 131000, "max_output": 16000, "input": 2, "output": 2, "latency": 3.12, "throughput": 210.9 }, { "name": "NovitaAI", "context": 32000, "max_output": 32000, "input": 0.8, "output": 0.8, "latency": 36.12, "throughput": 34.55 } ]
0.8
0.8
0.4
0.4
2
2
true
DeepSeek: R1
deepseek-ai/DeepSeek-R1
deepseek-ai
[ { "name": "Minimax", "context": 64000, "max_output": 64000, "input": 0.55, "output": 2.19, "latency": 1.74, "throughput": 19.54 }, { "name": "Nebius AI Studio", "context": 128000, "max_output": 128000, "input": 0.8, "output": 2.4, "latency": 0.44, "throughput": 13.93 }, { "name": "DeepInfra Turbo", "context": 33000, "max_output": 33000, "input": 2, "output": 6, "latency": 0.54, "throughput": 38.08 }, { "name": "inference.net", "context": 131000, "max_output": 33000, "input": 3, "output": 3, "latency": 0.74, "throughput": 35.03 }, { "name": "Parasail", "context": 128000, "max_output": 128000, "input": 3, "output": 3, "latency": 0.96, "throughput": 55.96 }, { "name": "Together", "context": 164000, "max_output": 8000, "input": 3, "output": 7, "latency": 0.65, "throughput": 70.15 }, { "name": "Friendli", "context": 164000, "max_output": 164000, "input": 3, "output": 7, "latency": 0.8, "throughput": 48.69 }, { "name": "Fireworks", "context": 164000, "max_output": 164000, "input": 3, "output": 8, "latency": 5.48, "throughput": 77.05 }, { "name": "SambaNova", "context": 8000, "max_output": 8000, "input": 5, "output": 7, "latency": 3.79, "throughput": 158.4 }, { "name": "kluster.ai", "context": 128000, "max_output": 128000, "input": 7, "output": 7, "latency": 8.55, "throughput": 31.84 }, { "name": "DeepSeek", "context": 64000, "max_output": 8000, "input": 0.55, "output": 2.19, "latency": 23.57, "throughput": 28.43 }, { "name": "DeepInfra", "context": 66000, "max_output": 8000, "input": 0.75, "output": 2.4, "latency": 2.43, "throughput": 9.75 }, { "name": "Azure", "context": 164000, "max_output": 4000, "input": 1.485, "output": 5.94, "latency": 13.92, "throughput": 28.88 }, { "name": "NovitaAI", "context": 64000, "max_output": 8000, "input": 4, "output": 4, "latency": 23.19, "throughput": 23.56 }, { "name": "Featherless", "context": 33000, "max_output": 4000, "input": 6.5, "output": 8, "latency": 43.56, "throughput": 20.13 } ]
2
6
0.55
2.19
6.5
8
true
MiniMax: MiniMax-01
MiniMaxAI/MiniMax-Text-01
MiniMaxAI
[ { "name": "Minimax", "context": 1000000, "max_output": 1000000, "input": 0.2, "output": 1.1, "latency": 1.49, "throughput": 25.41 } ]
0.2
1.1
0.2
1.1
0.2
1.1
true
Microsoft: Phi 4
microsoft/phi-4
microsoft
[ { "name": "DeepInfra", "context": 16000, "max_output": 8000, "input": 0.07, "output": 0.14, "latency": 0.5, "throughput": 33.12 }, { "name": "Nebius AI Studio", "context": 16000, "max_output": 16000, "input": 0.1, "output": 0.3, "latency": 0.14, "throughput": 105.1 } ]
0.1
0.3
0.07
0.14
0.1
0.3
true
Sao10K: Llama 3.1 70B Hanami x1
Sao10K/L3.1-70B-Hanami-x1
Sao10K
[ { "name": "Infermatic", "context": 16000, "max_output": 16000, "input": 3, "output": 3, "latency": 2.39, "throughput": 31.08 } ]
3
3
3
3
3
3
true
DeepSeek: DeepSeek V3
deepseek-ai/DeepSeek-V3
deepseek-ai
[ { "name": "inference.net", "context": 131000, "max_output": 66000, "input": 1.2, "output": 1.2, "latency": 1.22, "throughput": 7.59 }, { "name": "Together", "context": 131000, "max_output": 12000, "input": 1.25, "output": 1.25, "latency": 0.75, "throughput": 30.13 }, { "name": "DeepSeek", "context": 64000, "max_output": 8000, "input": 0.27, "output": 1.1, "latency": 0.42, "throughput": 12.26 }, { "name": "NovitaAI", "context": 64000, "max_output": 16000, "input": 0.4, "output": 1.3, "latency": 1.43, "throughput": 31.15 }, { "name": "DeepInfra", "context": 66000, "max_output": 8000, "input": 0.49, "output": 0.89, "latency": 0.89, "throughput": 7.78 }, { "name": "Nebius AI Studio", "context": 131000, "max_output": 131000, "input": 0.5, "output": 1.5, "latency": 0.33, "throughput": 20.68 }, { "name": "Fireworks", "context": 131000, "max_output": 131000, "input": 0.9, "output": 0.9, "latency": 1.25, "throughput": 38.84 } ]
0.9
0.9
0.27
1.1
1.25
1.25
true
Sao10K: Llama 3.3 Euryale 70B
Sao10K/L3.3-70B-Euryale-v2.3
Sao10K
[ { "name": "DeepInfra", "context": 131000, "max_output": 8000, "input": 0.7, "output": 0.8, "latency": 0.2, "throughput": 39.25 }, { "name": "Infermatic", "context": 16000, "max_output": 16000, "input": 1.5, "output": 1.5, "latency": 0.82, "throughput": 41.21 } ]
1.5
1.5
0.7
0.8
1.5
1.5
true
EVA Llama 3.33 70B
EVA-UNIT-01/EVA-LLaMA-3.33-70B-v0.1
EVA-UNIT-01
[ { "name": "Featherless", "context": 16000, "max_output": 4000, "input": 4, "output": 6, "latency": 2.32, "throughput": 10.75 } ]
4
6
4
6
4
6
true
Meta: Llama 3.3 70B Instruct
meta-llama/Llama-3.3-70B-Instruct
meta-llama
[ { "name": "Lambda", "context": 131000, "max_output": 131000, "input": 0.12, "output": 0.3, "latency": 0.81, "throughput": 11.46 }, { "name": "DeepInfra", "context": 131000, "max_output": 8000, "input": 0.12, "output": 0.3, "latency": 0.25, "throughput": 33.18 }, { "name": "Nebius AI Studio", "context": 131000, "max_output": 131000, "input": 0.13, "output": 0.4, "latency": 0.37, "throughput": 21.01 }, { "name": "inference.net", "context": 16000, "max_output": 16000, "input": 0.4, "output": 0.4, "latency": 1.04, "throughput": 11 }, { "name": "Hyperbolic", "context": 131000, "max_output": 131000, "input": 0.4, "output": 0.4, "latency": 1.17, "throughput": 57.18 }, { "name": "Groq", "context": 33000, "max_output": 33000, "input": 0.59, "output": 0.79, "latency": 0.26, "throughput": 301.3 }, { "name": "Friendli", "context": 131000, "max_output": 131000, "input": 0.6, "output": 0.6, "latency": 0.37, "throughput": 111 }, { "name": "SambaNova", "context": 128000, "max_output": 128000, "input": 0.6, "output": 1.2, "latency": 1.83, "throughput": 132.2 }, { "name": "kluster.ai", "context": 131000, "max_output": 131000, "input": 0.7, "output": 0.7, "latency": 0.86, "throughput": 17.33 }, { "name": "Parasail", "context": 131000, "max_output": 131000, "input": 0.7, "output": 0.7, "latency": 0.69, "throughput": 51.47 }, { "name": "Cloudflare", "context": 131000, "max_output": 131000, "input": 0.75, "output": 0.75, "latency": 0.85, "throughput": 29.91 }, { "name": "Lepton", "context": 131000, "max_output": 131000, "input": 0.8, "output": 0.8, "latency": 0.33, "throughput": 40.34 }, { "name": "Together", "context": 131000, "max_output": 2000, "input": 0.88, "output": 0.88, "latency": 0.47, "throughput": 135.3 }, { "name": "Avian.io", "context": 131000, "max_output": 131000, "input": 0.9, "output": 0.9, "latency": 0.24, "throughput": 122.4 }, { "name": "Fireworks", "context": 131000, "max_output": 131000, "input": 0.9, "output": 0.9, "latency": 0.34, "throughput": 125.4 }, { "name": "NovitaAI", "context": 131000, "max_output": 131000, "input": 0.39, "output": 0.39, "latency": 1.17, "throughput": 26.84 } ]
0.7
0.7
0.12
0.3
0.9
0.9
true
Qwen: QwQ 32B Preview
Qwen/QwQ-32B-Preview
Qwen
[ { "name": "Hyperbolic", "context": 33000, "max_output": 33000, "input": 0.2, "output": 0.2, "latency": 1.21, "throughput": 65.09 }, { "name": "Fireworks", "context": 33000, "max_output": 33000, "input": 0.9, "output": 0.9, "latency": 0.29, "throughput": 67.07 }, { "name": "Together", "context": 33000, "max_output": 33000, "input": 1.2, "output": 1.2, "latency": 0.33, "throughput": 64.6 }, { "name": "SambaNova", "context": 16000, "max_output": 16000, "input": 1.5, "output": 3, "latency": 0.52, "throughput": 240.7 }, { "name": "NovitaAI", "context": 33000, "max_output": 33000, "input": 0.18, "output": 0.6, "latency": null, "throughput": null } ]
0.9
0.9
0.2
0.2
1.5
3
true
EVA Qwen2.5 72B
EVA-UNIT-01/EVA-Qwen2.5-72B-v0.1
EVA-UNIT-01
[ { "name": "Parasail", "context": 32000, "max_output": 32000, "input": 0.7, "output": 0.7, "latency": 1.31, "throughput": 30.03 }, { "name": "Featherless", "context": 16000, "max_output": 4000, "input": 4, "output": 6, "latency": 3.35, "throughput": 10.69 } ]
4
6
0.7
0.7
4
6
true
Infermatic: Mistral Nemo Inferor 12B
Infermatic/MN-12B-Inferor-v0.0
Infermatic
[ { "name": "Featherless", "context": 16000, "max_output": 4000, "input": 0.8, "output": 1.2, "latency": 2.74, "throughput": 15.11 } ]
0.8
1.2
0.8
1.2
0.8
1.2
true
Qwen2.5 Coder 32B Instruct
Qwen/Qwen2.5-Coder-32B-Instruct
Qwen
[ { "name": "Lambda", "context": 33000, "max_output": 3000, "input": 0.07, "output": 0.16, "latency": 0.47, "throughput": 68.36 }, { "name": "DeepInfra", "context": 33000, "max_output": 8000, "input": 0.07, "output": 0.16, "latency": 0.25, "throughput": 29.97 }, { "name": "Hyperbolic", "context": 128000, "max_output": 8000, "input": 0.2, "output": 0.2, "latency": 1.03, "throughput": 53.2 }, { "name": "Parasail", "context": 128000, "max_output": 128000, "input": 0.5, "output": 0.5, "latency": 0.59, "throughput": 51.69 }, { "name": "Groq", "context": 131000, "max_output": 131000, "input": 0.79, "output": 0.79, "latency": 0.39, "throughput": 389.8 }, { "name": "Together", "context": 16000, "max_output": 2000, "input": 0.8, "output": 0.8, "latency": 0.48, "throughput": 82.14 }, { "name": "Fireworks", "context": 131000, "max_output": 4000, "input": 0.9, "output": 0.9, "latency": 0.37, "throughput": 63.7 }, { "name": "Mancer", "context": 33000, "max_output": 2000, "input": 1.5, "output": 2.813, "latency": 0.84, "throughput": 19.13 }, { "name": "SambaNova", "context": 16000, "max_output": 16000, "input": 1.5, "output": 3, "latency": 0.79, "throughput": 390.9 }, { "name": "Mancer (private)", "context": 33000, "max_output": 2000, "input": 2, "output": 3.75, "latency": null, "throughput": null }, { "name": "Featherless", "context": 16000, "max_output": 4000, "input": 2.6, "output": 3.4, "latency": 4.61, "throughput": 12.84 } ]
0.8
0.8
0.07
0.16
2.6
3.4
true
SorcererLM 8x22B
rAIfle/SorcererLM-8x22b-bf16
rAIfle
[ { "name": "Infermatic", "context": 16000, "max_output": 16000, "input": 4.5, "output": 4.5, "latency": 0.8, "throughput": 7.14 } ]
4.5
4.5
4.5
4.5
4.5
4.5
true
EVA Qwen2.5 32B
EVA-UNIT-01/EVA-Qwen2.5-32B-v0.2
EVA-UNIT-01
[ { "name": "Featherless", "context": 16000, "max_output": 4000, "input": 2.6, "output": 3.4, "latency": 2.25, "throughput": 8.53 } ]
2.6
3.4
2.6
3.4
2.6
3.4
true
Unslopnemo 12B
TheDrummer/UnslopNemo-12B-v4.1
TheDrummer
[ { "name": "Infermatic", "context": 32000, "max_output": 32000, "input": 0.5, "output": 0.5, "latency": 0.77, "throughput": 64.44 } ]
0.5
0.5
0.5
0.5
0.5
0.5
true
NeverSleep: Lumimaid v0.2 70B
NeverSleep/Lumimaid-v0.2-70B
NeverSleep
[ { "name": "Mancer", "context": 16000, "max_output": 2000, "input": 3.375, "output": 4.5, "latency": 3.1, "throughput": 12.96 }, { "name": "Featherless", "context": 16000, "max_output": 4000, "input": 4, "output": 6, "latency": 2.75, "throughput": 11.7 }, { "name": "Mancer (private)", "context": 16000, "max_output": 2000, "input": 4.5, "output": 6, "latency": 1.45, "throughput": 11.78 } ]
4
6
3.375
4.5
4.5
6
true
Magnum v4 72B
anthracite-org/magnum-v4-72b
anthracite-org
[ { "name": "Mancer", "context": 16000, "max_output": 1000, "input": 1.875, "output": 2.25, "latency": 0.94, "throughput": 13.65 }, { "name": "Mancer (private)", "context": 16000, "max_output": 1000, "input": 2.5, "output": 3, "latency": 1.09, "throughput": 12.53 }, { "name": "Infermatic", "context": 33000, "max_output": 33000, "input": 3, "output": 3, "latency": 0.33, "throughput": 18.03 }, { "name": "Featherless", "context": 16000, "max_output": 4000, "input": 4, "output": 6, "latency": 2.57, "throughput": 10.33 } ]
3
3
1.875
2.25
4
6
true
Qwen2.5 7B Instruct
Qwen/Qwen2.5-7B-Instruct
Qwen
[ { "name": "DeepInfra", "context": 33000, "max_output": 8000, "input": 0.025, "output": 0.05, "latency": 0.19, "throughput": 53.19 }, { "name": "Together", "context": 33000, "max_output": 2000, "input": 0.3, "output": 0.3, "latency": 0.37, "throughput": 112.9 } ]
0.3
0.3
0.025
0.05
0.3
0.3
true
NVIDIA: Llama 3.1 Nemotron 70B Instruct
nvidia/Llama-3.1-Nemotron-70B-Instruct-HF
nvidia
[ { "name": "Lambda", "context": 131000, "max_output": 131000, "input": 0.12, "output": 0.3, "latency": 0.91, "throughput": 35.25 }, { "name": "DeepInfra", "context": 131000, "max_output": 8000, "input": 0.12, "output": 0.3, "latency": 0.55, "throughput": 28.07 }, { "name": "Together", "context": 33000, "max_output": 33000, "input": 0.88, "output": 0.88, "latency": 0.51, "throughput": 70.6 }, { "name": "Infermatic", "context": 32000, "max_output": 32000, "input": 1, "output": 1, "latency": 2.61, "throughput": 13.76 } ]
0.88
0.88
0.12
0.3
1
1
true
Magnum v2 72B
anthracite-org/magnum-v2-72b
anthracite-org
[ { "name": "Infermatic", "context": 33000, "max_output": 33000, "input": 3, "output": 3, "latency": 0.46, "throughput": 40.66 }, { "name": "Featherless", "context": 16000, "max_output": 4000, "input": 4, "output": 6, "latency": 3.5, "throughput": 10.75 } ]
4
6
3
3
4
6
true
Rocinante 12B
TheDrummer/Rocinante-12B-v1.1
TheDrummer
[ { "name": "Infermatic", "context": 33000, "max_output": 33000, "input": 0.25, "output": 0.5, "latency": 0.76, "throughput": 23.44 }, { "name": "Featherless", "context": 16000, "max_output": 4000, "input": 0.8, "output": 1.2, "latency": 1.5, "throughput": 13.2 } ]
0.8
1.2
0.25
0.5
0.8
1.2
true
Meta: Llama 3.2 3B Instruct
meta-llama/Llama-3.2-3B-Instruct
meta-llama
[ { "name": "Lambda", "context": 131000, "max_output": 131000, "input": 0.015, "output": 0.025, "latency": 0.44, "throughput": 237.9 }, { "name": "DeepInfra", "context": 131000, "max_output": 8000, "input": 0.015, "output": 0.025, "latency": 0.21, "throughput": 147 }, { "name": "inference.net", "context": 16000, "max_output": 16000, "input": 0.02, "output": 0.02, "latency": 1.04, "throughput": 82.6 }, { "name": "Lepton", "context": 131000, "max_output": 131000, "input": 0.03, "output": 0.03, "latency": 0.3, "throughput": 199.5 }, { "name": "Together", "context": 131000, "max_output": 16000, "input": 0.06, "output": 0.06, "latency": 0.32, "throughput": 73.29 }, { "name": "SambaNova", "context": 4000, "max_output": 2000, "input": 0.08, "output": 0.16, "latency": 0.26, "throughput": 1271 }, { "name": "Fireworks", "context": 131000, "max_output": 131000, "input": 0.1, "output": 0.1, "latency": 0.76, "throughput": 150.3 }, { "name": "Hyperbolic", "context": 131000, "max_output": 131000, "input": 0.1, "output": 0.1, "latency": 1.23, "throughput": 173.8 }, { "name": "Cloudflare", "context": 131000, "max_output": 131000, "input": 0.1, "output": 0.1, "latency": 0.91, "throughput": 172.1 }, { "name": "NovitaAI", "context": 33000, "max_output": 33000, "input": 0.03, "output": 0.05, "latency": 1.09, "throughput": 78.99 } ]
0.06
0.06
0.015
0.025
0.08
0.16
true
Meta: Llama 3.2 1B Instruct
meta-llama/Llama-3.2-1B-Instruct
meta-llama
[ { "name": "Lepton", "context": 131000, "max_output": 131000, "input": 0.01, "output": 0.01, "latency": 0.33, "throughput": 402.7 }, { "name": "inference.net", "context": 16000, "max_output": 16000, "input": 0.01, "output": 0.01, "latency": 0.81, "throughput": 135.2 }, { "name": "DeepInfra", "context": 131000, "max_output": 8000, "input": 0.01, "output": 0.01, "latency": 0.17, "throughput": 169.1 }, { "name": "SambaNova", "context": 4000, "max_output": 2000, "input": 0.04, "output": 0.08, "latency": 0.76, "throughput": 2039 }, { "name": "Cloudflare", "context": 131000, "max_output": 131000, "input": 0.1, "output": 0.1, "latency": 0.88, "throughput": 253 }, { "name": "NovitaAI", "context": 131000, "max_output": 131000, "input": 0.02, "output": 0.02, "latency": null, "throughput": null } ]
0.02
0.02
0.01
0.01
0.1
0.1
true
Meta: Llama 3.2 90B Vision Instruct
meta-llama/Llama-3.2-90B-Vision-Instruct
meta-llama
[ { "name": "SambaNova", "context": 4000, "max_output": 2000, "input": 0.8, "output": 1.6, "latency": 0.55, "throughput": 261.7 }, { "name": "Fireworks", "context": 131000, "max_output": 131000, "input": 0.9, "output": 0.9, "latency": 1.13, "throughput": 28.98 }, { "name": "Together", "context": 131000, "max_output": 131000, "input": 1.2, "output": 1.2, "latency": 0.52, "throughput": 27.08 }, { "name": "DeepInfra", "context": 33000, "max_output": 8000, "input": 0.35, "output": 0.4, "latency": 0.13, "throughput": 20.1 } ]
1.2
1.2
0.35
0.4
0.8
1.6
true
Meta: Llama 3.2 11B Vision Instruct
meta-llama/Llama-3.2-11B-Vision-Instruct
meta-llama
[ { "name": "inference.net", "context": 16000, "max_output": 16000, "input": 0.055, "output": 0.055, "latency": 1.71, "throughput": 33.56 }, { "name": "DeepInfra", "context": 131000, "max_output": 8000, "input": 0.055, "output": 0.055, "latency": 0.14, "throughput": 55.62 }, { "name": "SambaNova", "context": 4000, "max_output": 2000, "input": 0.15, "output": 0.3, "latency": 0.57, "throughput": 504.6 }, { "name": "Together", "context": 131000, "max_output": 131000, "input": 0.18, "output": 0.18, "latency": 0.52, "throughput": 161.8 }, { "name": "Cloudflare", "context": 131000, "max_output": 131000, "input": 0.2, "output": 0.2, "latency": 0.64, "throughput": 42.9 }, { "name": "Fireworks", "context": 131000, "max_output": 131000, "input": 0.2, "output": 0.2, "latency": 4.04, "throughput": 36.58 }, { "name": "NovitaAI", "context": 33000, "max_output": 33000, "input": 0.06, "output": 0.06, "latency": null, "throughput": null } ]
0.18
0.18
0.055
0.055
0.15
0.3
true
Qwen2.5 72B Instruct
Qwen/Qwen2.5-72B-Instruct
Qwen
[ { "name": "Nebius AI Studio", "context": 128000, "max_output": 128000, "input": 0.13, "output": 0.4, "latency": 0.39, "throughput": 25 }, { "name": "DeepInfra", "context": 33000, "max_output": 8000, "input": 0.13, "output": 0.4, "latency": 0.44, "throughput": 42.07 }, { "name": "Hyperbolic", "context": 33000, "max_output": 33000, "input": 0.4, "output": 0.4, "latency": 1.16, "throughput": 53.32 }, { "name": "Fireworks", "context": 33000, "max_output": 33000, "input": 0.9, "output": 0.9, "latency": 0.34, "throughput": 41.79 }, { "name": "Together", "context": 131000, "max_output": 2000, "input": 1.2, "output": 1.2, "latency": 0.4, "throughput": 103.2 }, { "name": "SambaNova", "context": 16000, "max_output": 16000, "input": 2, "output": 4, "latency": 0.45, "throughput": 376.3 }, { "name": "NovitaAI", "context": 32000, "max_output": 4000, "input": 0.38, "output": 0.4, "latency": 0.86, "throughput": 20.33 } ]
0.4
0.4
0.13
0.4
2
4
true
Qwen: Qwen2.5-VL 72B Instruct
Qwen/Qwen2.5-VL-72B-Instruct
Qwen
[ { "name": "Hyperbolic", "context": 4000, "max_output": 4000, "input": 0.4, "output": 0.4, "latency": 3.23, "throughput": 27.88 } ]
0.4
0.4
0.4
0.4
0.4
0.4
true
NeverSleep: Lumimaid v0.2 8B
NeverSleep/Lumimaid-v0.2-8B
NeverSleep
[ { "name": "Mancer", "context": 33000, "max_output": 2000, "input": 0.1875, "output": 1.125, "latency": 0.91, "throughput": 33.73 }, { "name": "Mancer (private)", "context": 33000, "max_output": 2000, "input": 0.25, "output": 1.5, "latency": 0.67, "throughput": 33.78 }, { "name": "Featherless", "context": 16000, "max_output": 4000, "input": 0.8, "output": 1.2, "latency": 1.62, "throughput": 26.35 } ]
0.25
1.5
0.1875
1.125
0.8
1.2
true
Mistral: Pixtral 12B
mistralai/Pixtral-12B-2409
mistralai
[ { "name": "Hyperbolic", "context": 4000, "max_output": 4000, "input": 0.1, "output": 0.1, "latency": 1.67, "throughput": 69.2 }, { "name": "Mistral", "context": 4000, "max_output": 4000, "input": 0.15, "output": 0.15, "latency": 0.85, "throughput": 88.97 } ]
0.15
0.15
0.1
0.1
0.15
0.15
true
Sao10K: Llama 3.1 Euryale 70B v2.2
Sao10K/L3.1-70B-Euryale-v2.2
Sao10K
[ { "name": "DeepInfra", "context": 131000, "max_output": 8000, "input": 0.7, "output": 0.8, "latency": 0.3, "throughput": 37.01 }, { "name": "Infermatic", "context": 16000, "max_output": 16000, "input": 1.5, "output": 1.5, "latency": 1.13, "throughput": 19.14 }, { "name": "NovitaAI", "context": 16000, "max_output": 16000, "input": 1.48, "output": 1.48, "latency": 1.6, "throughput": 21.19 } ]
1.48
1.48
0.7
0.8
1.5
1.5
true
Qwen: Qwen2.5-VL 7B Instruct
Qwen/Qwen2.5-VL-7B-Instruct
Qwen
[ { "name": "Hyperbolic", "context": 4000, "max_output": 4000, "input": 0.1, "output": 0.1, "latency": 0.68, "throughput": 57.96 } ]
0.1
0.1
0.1
0.1
0.1
0.1
true
Microsoft: Phi-3.5 Mini 128K Instruct
microsoft/Phi-3.5-mini-instruct
microsoft
[ { "name": "Azure", "context": 128000, "max_output": 128000, "input": 0.1, "output": 0.1, "latency": 4.11, "throughput": 17 } ]
0.1
0.1
0.1
0.1
0.1
0.1
true
Nous: Hermes 3 70B Instruct
NousResearch/Hermes-3-Llama-3.1-70B
NousResearch
[ { "name": "Lambda", "context": 131000, "max_output": 131000, "input": 0.12, "output": 0.3, "latency": 0.75, "throughput": 32.07 }, { "name": "Hyperbolic", "context": 12000, "max_output": 12000, "input": 0.4, "output": 0.4, "latency": 0.94, "throughput": 29.52 } ]
0.4
0.4
0.12
0.3
0.4
0.4
true
Nous: Hermes 3 405B Instruct
NousResearch/Hermes-3-Llama-3.1-405B
NousResearch
[ { "name": "Lambda", "context": 131000, "max_output": 131000, "input": 0.8, "output": 0.8, "latency": 1.04, "throughput": 27.08 }, { "name": "inference.net", "context": 33000, "max_output": 33000, "input": 0.8, "output": 0.8, "latency": 0.92, "throughput": 25.74 }, { "name": "DeepInfra", "context": 131000, "max_output": 8000, "input": 0.8, "output": 0.8, "latency": 0.77, "throughput": 12.78 }, { "name": "Nebius AI Studio", "context": 131000, "max_output": 131000, "input": 1, "output": 3, "latency": 0.31, "throughput": 27.65 } ]
0.8
0.8
0.8
0.8
1
3
true
Sao10K: Llama 3 8B Lunaris
Sao10K/L3-8B-Lunaris-v1
Sao10K
[ { "name": "DeepInfra", "context": 8000, "max_output": 8000, "input": 0.03, "output": 0.06, "latency": 0.38, "throughput": 69.42 }, { "name": "NovitaAI", "context": 8000, "max_output": 8000, "input": 0.05, "output": 0.05, "latency": 0.93, "throughput": 66.61 } ]
0.05
0.05
0.03
0.06
0.05
0.05
true
Aetherwiing: Starcannon 12B
intervitens/mini-magnum-12b-v1.1
intervitens
[ { "name": "Featherless", "context": 16000, "max_output": 4000, "input": 0.8, "output": 1.2, "latency": 3.93, "throughput": 15.71 } ]
0.8
1.2
0.8
1.2
0.8
1.2
true
Meta: Llama 3.1 405B (base)
meta-llama/llama-3.1-405B
meta-llama
[ { "name": "Hyperbolic (quantized)", "context": 33000, "max_output": 33000, "input": 2, "output": 2, "latency": 0.86, "throughput": 23.83 }, { "name": "Hyperbolic", "context": 33000, "max_output": 33000, "input": 4, "output": 4, "latency": 1.32, "throughput": 16.23 } ]
4
4
2
2
4
4
true
Mistral Nemo 12B Celeste
nothingiisreal/MN-12B-Celeste-V1.9
nothingiisreal
[ { "name": "Featherless", "context": 16000, "max_output": 4000, "input": 0.8, "output": 1.2, "latency": 3.54, "throughput": 14.64 } ]
0.8
1.2
0.8
1.2
0.8
1.2
true
Meta: Llama 3.1 405B Instruct
meta-llama/Meta-Llama-3.1-405B-Instruct
meta-llama
[ { "name": "DeepInfra", "context": 33000, "max_output": 8000, "input": 0.8, "output": 0.8, "latency": 0.17, "throughput": 3.55 }, { "name": "Lambda", "context": 131000, "max_output": 131000, "input": 0.8, "output": 0.8, "latency": 0.84, "throughput": 17.46 }, { "name": "Nebius AI Studio", "context": 128000, "max_output": 128000, "input": 1, "output": 3, "latency": 0.14, "throughput": 23.97 }, { "name": "Fireworks", "context": 131000, "max_output": 131000, "input": 3, "output": 3, "latency": 3.45, "throughput": 39.73 }, { "name": "Together", "context": 131000, "max_output": 131000, "input": 3.5, "output": 3.5, "latency": 6.1, "throughput": 27.34 }, { "name": "kluster.ai", "context": 131000, "max_output": 131000, "input": 3.5, "output": 3.5, "latency": 0.83, "throughput": 17.03 }, { "name": "Hyperbolic", "context": 33000, "max_output": 33000, "input": 4, "output": 4, "latency": 4.35, "throughput": 3.8 }, { "name": "SambaNova", "context": 8000, "max_output": 4000, "input": 5, "output": 10, "latency": 1.52, "throughput": 81.33 } ]
3.5
3.5
0.8
0.8
5
10
true
Meta: Llama 3.1 8B Instruct
meta-llama/Meta-Llama-3.1-8B-Instruct
meta-llama
[ { "name": "DeepInfra", "context": 131000, "max_output": 8000, "input": 0.02, "output": 0.05, "latency": 0.17, "throughput": 116 }, { "name": "Lambda", "context": 131000, "max_output": 131000, "input": 0.025, "output": 0.04, "latency": 0.48, "throughput": 168 }, { "name": "inference.net", "context": 16000, "max_output": 16000, "input": 0.03, "output": 0.03, "latency": 0.94, "throughput": 58.13 }, { "name": "Groq", "context": 131000, "max_output": 8000, "input": 0.05, "output": 0.08, "latency": 0.35, "throughput": 642 }, { "name": "Lepton", "context": 131000, "max_output": 131000, "input": 0.07, "output": 0.07, "latency": 0.21, "throughput": 97.29 }, { "name": "Friendli", "context": 131000, "max_output": 131000, "input": 0.1, "output": 0.1, "latency": 0.43, "throughput": 223.1 }, { "name": "Hyperbolic", "context": 33000, "max_output": 33000, "input": 0.1, "output": 0.1, "latency": 1.17, "throughput": 102.9 }, { "name": "SambaNova", "context": 8000, "max_output": 4000, "input": 0.1, "output": 0.2, "latency": 0.32, "throughput": 800 }, { "name": "Cloudflare", "context": 131000, "max_output": 131000, "input": 0.15, "output": 0.15, "latency": 4.27, "throughput": 26.08 }, { "name": "kluster.ai", "context": 131000, "max_output": 131000, "input": 0.18, "output": 0.18, "latency": 0.33, "throughput": 13.91 }, { "name": "Together", "context": 131000, "max_output": 131000, "input": 0.18, "output": 0.18, "latency": 0.25, "throughput": 288 }, { "name": "Fireworks", "context": 131000, "max_output": 131000, "input": 0.2, "output": 0.2, "latency": 0.65, "throughput": 211.1 }, { "name": "Avian.io", "context": 131000, "max_output": 131000, "input": 0.2, "output": 0.2, "latency": 0.29, "throughput": 302.6 }, { "name": "NovitaAI", "context": 16000, "max_output": 8000, "input": 0.05, "output": 0.05, "latency": null, "throughput": null } ]
0.1
0.1
0.03
0.03
0.2
0.2
true
Meta: Llama 3.1 70B Instruct
meta-llama/Meta-Llama-3.1-70B-Instruct
meta-llama
[ { "name": "DeepInfra", "context": 131000, "max_output": 8000, "input": 0.12, "output": 0.3, "latency": 0.43, "throughput": 32.15 }, { "name": "Lambda", "context": 131000, "max_output": 131000, "input": 0.12, "output": 0.3, "latency": 0.61, "throughput": 32.2 }, { "name": "Nebius AI Studio", "context": 128000, "max_output": 128000, "input": 0.13, "output": 0.4, "latency": 0.15, "throughput": 32.57 }, { "name": "inference.net", "context": 16000, "max_output": 16000, "input": 0.4, "output": 0.4, "latency": 0.99, "throughput": 18.3 }, { "name": "Hyperbolic", "context": 33000, "max_output": 33000, "input": 0.4, "output": 0.4, "latency": 0.77, "throughput": 108.8 }, { "name": "Friendli", "context": 131000, "max_output": 131000, "input": 0.6, "output": 0.6, "latency": 0.19, "throughput": 112.7 }, { "name": "SambaNova", "context": 8000, "max_output": 4000, "input": 0.6, "output": 1.2, "latency": 0.54, "throughput": 267.7 }, { "name": "Cloudflare", "context": 131000, "max_output": 131000, "input": 0.75, "output": 0.75, "latency": 0.89, "throughput": 26.16 }, { "name": "Lepton", "context": 131000, "max_output": 131000, "input": 0.8, "output": 0.8, "latency": 0.08, "throughput": 39.8 }, { "name": "Together", "context": 131000, "max_output": 131000, "input": 0.88, "output": 0.88, "latency": 0.31, "throughput": 114.5 }, { "name": "Fireworks", "context": 131000, "max_output": 131000, "input": 0.9, "output": 0.9, "latency": 0.41, "throughput": 93.77 }, { "name": "NovitaAI", "context": 33000, "max_output": 33000, "input": 0.34, "output": 0.39, "latency": 2.27, "throughput": 50.46 } ]
0.6
0.6
0.12
0.3
0.9
0.9
true
Mistral: Mistral Nemo
mistralai/Mistral-Nemo-Instruct-2407
mistralai
[ { "name": "DeepInfra", "context": 131000, "max_output": 8000, "input": 0.035, "output": 0.08, "latency": 0.23, "throughput": 63.49 }, { "name": "Nebius AI Studio", "context": 128000, "max_output": 128000, "input": 0.04, "output": 0.12, "latency": 0.7, "throughput": 15.59 }, { "name": "inference.net", "context": 16000, "max_output": 16000, "input": 0.1, "output": 0.1, "latency": 0.83, "throughput": 65.67 }, { "name": "Parasail", "context": 128000, "max_output": 128000, "input": 0.11, "output": 0.11, "latency": 0.71, "throughput": 131.8 }, { "name": "Mistral", "context": 128000, "max_output": 128000, "input": 0.15, "output": 0.15, "latency": 0.34, "throughput": 125.9 }, { "name": "Lepton", "context": 128000, "max_output": 128000, "input": 0.18, "output": 0.18, "latency": 0.28, "throughput": 50.82 }, { "name": "Azure", "context": 128000, "max_output": 128000, "input": 0.3, "output": 0.3, "latency": 0.58, "throughput": 97.82 }, { "name": "NovitaAI", "context": 131000, "max_output": 131000, "input": 0.17, "output": 0.17, "latency": 0.77, "throughput": 51.15 } ]
0.15
0.15
0.035
0.08
0.3
0.3
true
Mistral: Codestral Mamba
mistralai/mamba-codestral-7B-v0.1
mistralai
[ { "name": "Mistral", "context": 256000, "max_output": 256000, "input": 0.25, "output": 0.25, "latency": 0.48, "throughput": 108.1 } ]
0.25
0.25
0.25
0.25
0.25
0.25
true
Qwen 2 7B Instruct
Qwen/Qwen2-7B-Instruct
Qwen
[ { "name": "NovitaAI", "context": 33000, "max_output": 33000, "input": 0.054, "output": 0.054, "latency": 0.95, "throughput": 47.27 } ]
0.054
0.054
0.054
0.054
0.054
0.054
true
Google: Gemma 2 27B
google/gemma-2-27b-it
google
[ { "name": "DeepInfra", "context": 8000, "max_output": 8000, "input": 0.27, "output": 0.27, "latency": 0.5, "throughput": 37.61 }, { "name": "Together", "context": 8000, "max_output": 2000, "input": 0.8, "output": 0.8, "latency": 0.45, "throughput": 71.67 } ]
0.8
0.8
0.27
0.27
0.8
0.8
true
Magnum 72B
alpindale/magnum-72b-v1
alpindale
[ { "name": "Mancer", "context": 16000, "max_output": 1000, "input": 1.875, "output": 2.25, "latency": 0.54, "throughput": 22.53 }, { "name": "Mancer (private)", "context": 16000, "max_output": 1000, "input": 2.5, "output": 3, "latency": 1.33, "throughput": 22.81 }, { "name": "Featherless", "context": 16000, "max_output": 4000, "input": 4, "output": 6, "latency": 6.09, "throughput": 11.94 } ]
2.5
3
1.875
2.25
4
6
true
Google: Gemma 2 9B
google/gemma-2-9b-it
google
[ { "name": "DeepInfra", "context": 8000, "max_output": 8000, "input": 0.03, "output": 0.06, "latency": 0.23, "throughput": 39.15 }, { "name": "Lepton", "context": 8000, "max_output": 8000, "input": 0.07, "output": 0.07, "latency": 0.35, "throughput": 102.3 }, { "name": "Groq", "context": 8000, "max_output": 8000, "input": 0.2, "output": 0.2, "latency": 0.2, "throughput": 565 }, { "name": "Together", "context": 8000, "max_output": 8000, "input": 0.3, "output": 0.3, "latency": 0.34, "throughput": 107.8 }, { "name": "NovitaAI", "context": 8000, "max_output": 8000, "input": 0.08, "output": 0.08, "latency": 10.15, "throughput": 29.6 } ]
0.08
0.08
0.03
0.06
0.3
0.3
true
Sao10k: Llama 3 Euryale 70B v2.1
Sao10K/L3-70B-Euryale-v2.1
Sao10K
[ { "name": "DeepInfra", "context": 8000, "max_output": 8000, "input": 0.7, "output": 0.8, "latency": 0.21, "throughput": 37.88 }, { "name": "NovitaAI", "context": 16000, "max_output": 16000, "input": 1.48, "output": 1.48, "latency": 3.01, "throughput": 19.66 } ]
1.48
1.48
0.7
0.8
1.48
1.48
true
Dolphin 2.9.2 Mixtral 8x22B 🐬
cognitivecomputations/dolphin-2.9.2-mixtral-8x22b
cognitivecomputations
[ { "name": "NovitaAI", "context": 16000, "max_output": 16000, "input": 0.9, "output": 0.9, "latency": 4.12, "throughput": 9.66 } ]
0.9
0.9
0.9
0.9
0.9
0.9
true
Qwen 2 72B Instruct
Qwen/Qwen2-72B-Instruct
Qwen
[ { "name": "Together", "context": 33000, "max_output": 4000, "input": 0.9, "output": 0.9, "latency": 0.37, "throughput": 66.68 } ]
0.9
0.9
0.9
0.9
0.9
0.9
true
Mistral: Mistral 7B Instruct
mistralai/Mistral-7B-Instruct-v0.3
mistralai
[ { "name": "DeepInfra", "context": 33000, "max_output": 8000, "input": 0.03, "output": 0.055, "latency": 0.29, "throughput": 82.53 }, { "name": "Parasail", "context": 33000, "max_output": 33000, "input": 0.11, "output": 0.11, "latency": 0.66, "throughput": 128.7 }, { "name": "Together", "context": 33000, "max_output": 4000, "input": 0.2, "output": 0.2, "latency": 0.34, "throughput": 163.9 }, { "name": "NovitaAI", "context": 33000, "max_output": 33000, "input": 0.059, "output": 0.059, "latency": 1.26, "throughput": 121.9 }, { "name": "Lepton", "context": 33000, "max_output": 33000, "input": 0.07, "output": 0.07, "latency": 0.55, "throughput": 105.5 } ]
0.07
0.07
0.03
0.055
0.2
0.2
true
Mistral: Mistral 7B Instruct v0.3
mistralai/Mistral-7B-Instruct-v0.3
mistralai
[ { "name": "DeepInfra", "context": 33000, "max_output": 8000, "input": 0.03, "output": 0.055, "latency": 0.18, "throughput": 102.4 }, { "name": "Lepton", "context": 33000, "max_output": 33000, "input": 0.07, "output": 0.07, "latency": 0.38, "throughput": 105.6 }, { "name": "Together", "context": 33000, "max_output": 4000, "input": 0.2, "output": 0.2, "latency": 0.3, "throughput": 173.8 }, { "name": "NovitaAI", "context": 33000, "max_output": 33000, "input": 0.059, "output": 0.059, "latency": null, "throughput": null } ]
0.07
0.07
0.03
0.055
0.2
0.2
true
NousResearch: Hermes 2 Pro - Llama-3 8B
NousResearch/Hermes-2-Pro-Llama-3-8B
NousResearch
[ { "name": "Lambda", "context": 131000, "max_output": 131000, "input": 0.025, "output": 0.04, "latency": 0.53, "throughput": 196.1 }, { "name": "NovitaAI", "context": 8000, "max_output": 8000, "input": 0.14, "output": 0.14, "latency": 1.16, "throughput": 122.3 } ]
0.14
0.14
0.025
0.04
0.14
0.14
true
Microsoft: Phi-3 Mini 128K Instruct
microsoft/Phi-3-mini-128k-instruct
microsoft
[ { "name": "Azure", "context": 128000, "max_output": 128000, "input": 0.1, "output": 0.1, "latency": 0.59, "throughput": 67.71 } ]
0.1
0.1
0.1
0.1
0.1
0.1
true
Microsoft: Phi-3 Medium 128K Instruct
microsoft/Phi-3-medium-128k-instruct
microsoft
[ { "name": "Azure", "context": 128000, "max_output": 128000, "input": 1, "output": 1, "latency": 0.59, "throughput": 52.04 } ]
1
1
1
1
1
1
true
NeverSleep: Llama 3 Lumimaid 70B
NeverSleep/Llama-3-Lumimaid-70B-v0.1
NeverSleep
[ { "name": "Mancer", "context": 8000, "max_output": 2000, "input": 3.375, "output": 4.5, "latency": 0.88, "throughput": 13.56 }, { "name": "Featherless", "context": 8000, "max_output": 4000, "input": 4, "output": 6, "latency": 2.58, "throughput": 11.46 }, { "name": "Mancer (private)", "context": 8000, "max_output": 2000, "input": 4.5, "output": 6, "latency": 1.11, "throughput": 12.8 } ]
4
6
3.375
4.5
4.5
6
true
Meta: LlamaGuard 2 8B
meta-llama/LlamaGuard-7b
meta-llama
[ { "name": "Together", "context": 8000, "max_output": 8000, "input": 0.2, "output": 0.2, "latency": 0.88, "throughput": 76.09 } ]
0.2
0.2
0.2
0.2
0.2
0.2
true
NeverSleep: Llama 3 Lumimaid 8B (extended)
NeverSleep/Llama-3-Lumimaid-8B-v0.1
NeverSleep
[ { "name": "Mancer", "context": 25000, "max_output": 2000, "input": 0.1875, "output": 1.125, "latency": 0.74, "throughput": 51.57 }, { "name": "Mancer (private)", "context": 25000, "max_output": 2000, "input": 0.25, "output": 1.5, "latency": 0.52, "throughput": 51.92 } ]
0.25
1.5
0.1875
1.125
0.25
1.5
true
NeverSleep: Llama 3 Lumimaid 8B
NeverSleep/Llama-3-Lumimaid-8B-v0.1
NeverSleep
[ { "name": "Mancer", "context": 25000, "max_output": 2000, "input": 0.1875, "output": 1.125, "latency": 0.74, "throughput": 51.57 }, { "name": "Mancer (private)", "context": 25000, "max_output": 2000, "input": 0.25, "output": 1.5, "latency": 0.52, "throughput": 51.92 }, { "name": "Featherless", "context": 8000, "max_output": 4000, "input": 0.8, "output": 1.2, "latency": 1.24, "throughput": 25.1 } ]
0.25
1.5
0.1875
1.125
0.8
1.2
true
Fimbulvetr 11B v2
Sao10K/Fimbulvetr-11B-v2
Sao10K
[ { "name": "Featherless", "context": 4000, "max_output": 4000, "input": 0.8, "output": 1.2, "latency": 0.97, "throughput": 27.78 } ]
0.8
1.2
0.8
1.2
0.8
1.2
true
Meta: Llama 3 8B Instruct
meta-llama/Meta-Llama-3-8B-Instruct
meta-llama
[ { "name": "DeepInfra", "context": 8000, "max_output": 8000, "input": 0.03, "output": 0.06, "latency": 0.16, "throughput": 113.2 }, { "name": "Groq", "context": 8000, "max_output": 8000, "input": 0.05, "output": 0.08, "latency": 0.31, "throughput": 893.2 }, { "name": "Cloudflare", "context": 8000, "max_output": 8000, "input": 0.15, "output": 0.15, "latency": 0.76, "throughput": 15.78 }, { "name": "Mancer", "context": 16000, "max_output": 2000, "input": 0.1875, "output": 1.125, "latency": 0.67, "throughput": 50.42 }, { "name": "Fireworks", "context": 8000, "max_output": 8000, "input": 0.2, "output": 0.2, "latency": 0.51, "throughput": 123.5 }, { "name": "Mancer (private)", "context": 16000, "max_output": 2000, "input": 0.25, "output": 1.5, "latency": 0.54, "throughput": 50.82 }, { "name": "NovitaAI", "context": 8000, "max_output": 8000, "input": 0.04, "output": 0.04, "latency": null, "throughput": null } ]
0.15
0.15
0.04
0.04
0.25
1.5
true
Meta: Llama 3 70B Instruct
meta-llama/Meta-Llama-3-70B-Instruct
meta-llama
[ { "name": "DeepInfra", "context": 8000, "max_output": 8000, "input": 0.23, "output": 0.4, "latency": 0.41, "throughput": 25.65 }, { "name": "Groq", "context": 8000, "max_output": 8000, "input": 0.59, "output": 0.79, "latency": 0.19, "throughput": 213.2 }, { "name": "Together", "context": 8000, "max_output": 8000, "input": 0.88, "output": 0.88, "latency": 0.44, "throughput": 19.92 }, { "name": "Fireworks", "context": 8000, "max_output": 8000, "input": 0.9, "output": 0.9, "latency": 0.61, "throughput": 150.3 }, { "name": "NovitaAI", "context": 8000, "max_output": 8000, "input": 0.51, "output": 0.74, "latency": 5.46, "throughput": 19.8 } ]
0.59
0.79
0.23
0.4
0.9
0.9
true
Mistral: Mixtral 8x22B Instruct
mistralai/Mixtral-8x22B-Instruct-v0.1
mistralai
[ { "name": "Fireworks", "context": 66000, "max_output": 66000, "input": 0.9, "output": 0.9, "latency": 0.67, "throughput": 88.81 }, { "name": "Together", "context": 66000, "max_output": 2000, "input": 1.2, "output": 1.2, "latency": 0.74, "throughput": 93.93 }, { "name": "Mistral", "context": 66000, "max_output": 66000, "input": 2, "output": 6, "latency": 0.25, "throughput": 86.28 } ]
1.2
1.2
0.9
0.9
2
6
true
WizardLM-2 8x22B
microsoft/WizardLM-2-8x22B
microsoft
[ { "name": "DeepInfra", "context": 66000, "max_output": 8000, "input": 0.5, "output": 0.5, "latency": 0.1, "throughput": 31.3 }, { "name": "Lepton", "context": 66000, "max_output": 66000, "input": 1, "output": 1, "latency": 0.36, "throughput": 28.57 }, { "name": "NovitaAI", "context": 66000, "max_output": 66000, "input": 0.62, "output": 0.62, "latency": 1.06, "throughput": 19.54 }, { "name": "Together", "context": 66000, "max_output": 66000, "input": 1.2, "output": 1.2, "latency": 0.57, "throughput": 69.67 } ]
1
1
0.5
0.5
1.2
1.2
true
WizardLM-2 7B
microsoft/WizardLM-2-7B
microsoft
[ { "name": "Lepton", "context": 32000, "max_output": 32000, "input": 0.07, "output": 0.07, "latency": 0.37, "throughput": 102.9 } ]
0.07
0.07
0.07
0.07
0.07
0.07
true
Databricks: DBRX 132B Instruct
databricks/dbrx-instruct
databricks
[ { "name": "Together", "context": 33000, "max_output": 2000, "input": 1.2, "output": 1.2, "latency": 0.33, "throughput": 83.02 } ]
1.2
1.2
1.2
1.2
1.2
1.2
true
Midnight Rose 70B
sophosympatheia/Wizard-Tulu-Dolphin-70B-v1.0
sophosympatheia
[ { "name": "NovitaAI", "context": 4000, "max_output": 4000, "input": 0.8, "output": 0.8, "latency": 2.21, "throughput": 15.59 } ]
0.8
0.8
0.8
0.8
0.8
0.8
true
Google: Gemma 7B
google/gemma-1.1-7b-it
google
[ { "name": "Cloudflare", "context": 8000, "max_output": 8000, "input": 0.15, "output": 0.15, "latency": 1.22, "throughput": 13.04 } ]
0.15
0.15
0.15
0.15
0.15
0.15
true
Nous: Hermes 2 Mixtral 8x7B DPO
NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO
NousResearch
[ { "name": "Together", "context": 33000, "max_output": 2000, "input": 0.6, "output": 0.6, "latency": 0.4, "throughput": 109 } ]
0.6
0.6
0.6
0.6
0.6
0.6
true
Dolphin 2.6 Mixtral 8x7B 🐬
cognitivecomputations/dolphin-2.6-mixtral-8x7b
cognitivecomputations
[ { "name": "Lepton", "context": 33000, "max_output": 33000, "input": 0.5, "output": 0.5, "latency": 0.11, "throughput": 84.57 } ]
0.5
0.5
0.5
0.5
0.5
0.5
true
Mistral: Mixtral 8x7B (base)
mistralai/Mixtral-8x7B-v0.1
mistralai
[ { "name": "Together", "context": 33000, "max_output": 2000, "input": 0.6, "output": 0.6, "latency": 0.46, "throughput": 118.1 } ]
0.6
0.6
0.6
0.6
0.6
0.6
true
Mistral: Mixtral 8x7B Instruct
mistralai/Mixtral-8x7B-Instruct-v0.1
mistralai
[ { "name": "DeepInfra", "context": 33000, "max_output": 8000, "input": 0.24, "output": 0.24, "latency": 0.2, "throughput": 115.1 }, { "name": "Groq", "context": 33000, "max_output": 33000, "input": 0.24, "output": 0.24, "latency": 0.37, "throughput": 678.1 }, { "name": "Fireworks", "context": 33000, "max_output": 33000, "input": 0.5, "output": 0.5, "latency": 0.75, "throughput": 191.9 }, { "name": "Lepton", "context": 33000, "max_output": 33000, "input": 0.5, "output": 0.5, "latency": 0.36, "throughput": 75.27 }, { "name": "Together", "context": 33000, "max_output": 2000, "input": 0.6, "output": 0.6, "latency": 0.49, "throughput": 90.88 } ]
0.5
0.5
0.24
0.24
0.6
0.6
true
OpenChat 3.5 7B
openchat/openchat-3.5-0106
openchat
[ { "name": "DeepInfra", "context": 8000, "max_output": 8000, "input": 0.055, "output": 0.055, "latency": 0.22, "throughput": 99.19 }, { "name": "Lepton", "context": 8000, "max_output": 8000, "input": 0.07, "output": 0.07, "latency": 0.4, "throughput": 107.9 }, { "name": "NovitaAI", "context": 4000, "max_output": 4000, "input": 0.06, "output": 0.06, "latency": 0.8, "throughput": 58.78 } ]
0.06
0.06
0.055
0.055
0.07
0.07
true
Noromaid 20B
NeverSleep/Noromaid-20b-v0.1.1
NeverSleep
[ { "name": "Mancer", "context": 8000, "max_output": 2000, "input": 1.5, "output": 2.25, "latency": 0.67, "throughput": 23.41 }, { "name": "Mancer (private)", "context": 8000, "max_output": 2000, "input": 2, "output": 3, "latency": 1.1, "throughput": 23.6 } ]
2
3
1.5
2.25
2
3
true
OpenHermes 2.5 Mistral 7B
teknium/OpenHermes-2.5-Mistral-7B
teknium
[ { "name": "NovitaAI", "context": 4000, "max_output": 4000, "input": 0.17, "output": 0.17, "latency": 0.67, "throughput": 148 } ]
0.17
0.17
0.17
0.17
0.17
0.17
true
Toppy M 7B
Undi95/Toppy-M-7B
Undi95
[ { "name": "Lepton", "context": 4000, "max_output": 4000, "input": 0.07, "output": 0.07, "latency": 0.42, "throughput": 108.2 } ]
0.07
0.07
0.07
0.07
0.07
0.07
true
Goliath 120B
alpindale/goliath-120b
alpindale
[ { "name": "Mancer", "context": 6000, "max_output": 512, "input": 9.375, "output": 9.375, "latency": 1.06, "throughput": 18.62 }, { "name": "Mancer (private)", "context": 6000, "max_output": 512, "input": 12.5, "output": 12.5, "latency": 1.53, "throughput": 15.02 } ]
12.5
12.5
9.375
9.375
12.5
12.5
true
Airoboros 70B
jondurbin/airoboros-l2-70b-2.2.1
jondurbin
[ { "name": "NovitaAI", "context": 4000, "max_output": 4000, "input": 0.5, "output": 0.5, "latency": 2.31, "throughput": 56.22 } ]
0.5
0.5
0.5
0.5
0.5
0.5
true
Xwin 70B
Xwin-LM/Xwin-LM-70B-V0.1
Xwin-LM
[ { "name": "Mancer", "context": 8000, "max_output": 512, "input": 3.75, "output": 3.75, "latency": 2.17, "throughput": 18.04 }, { "name": "Mancer (private)", "context": 8000, "max_output": 512, "input": 5, "output": 5, "latency": 2.05, "throughput": 18.1 } ]
5
5
3.75
3.75
5
5
true
Mistral: Mistral 7B Instruct v0.1
mistralai/Mistral-7B-Instruct-v0.1
mistralai
[ { "name": "Together", "context": 33000, "max_output": 2000, "input": 0.2, "output": 0.2, "latency": 0.3, "throughput": 169.1 } ]
0.2
0.2
0.2
0.2
0.2
0.2
true
Pygmalion: Mythalion 13B
PygmalionAI/mythalion-13b
PygmalionAI
[ { "name": "Featherless", "context": 4000, "max_output": 4000, "input": 0.8, "output": 1.2, "latency": 1.42, "throughput": 13.23 }, { "name": "Mancer", "context": 8000, "max_output": 512, "input": 1.125, "output": 1.125, "latency": 1.03, "throughput": 22.71 }, { "name": "Mancer (private)", "context": 8000, "max_output": 512, "input": 1.5, "output": 1.5, "latency": 1.04, "throughput": 22.3 } ]
1.125
1.125
0.8
1.2
1.5
1.5
true
Nous: Hermes 13B
NousResearch/Nous-Hermes-Llama2-13b
NousResearch
[ { "name": "Lepton", "context": 4000, "max_output": 4000, "input": 0.18, "output": 0.18, "latency": 0.31, "throughput": 83.27 }, { "name": "NovitaAI", "context": 4000, "max_output": 4000, "input": 0.17, "output": 0.17, "latency": 0.73, "throughput": 81.14 } ]
0.18
0.18
0.17
0.17
0.18
0.18
true
ReMM SLERP 13B
Undi95/ReMM-SLERP-L2-13B
Undi95
[ { "name": "Featherless", "context": 4000, "max_output": 4000, "input": 0.8, "output": 1.2, "latency": null, "throughput": null }, { "name": "Mancer", "context": 6000, "max_output": 512, "input": 1.125, "output": 1.125, "latency": 0.57, "throughput": 43.17 }, { "name": "Mancer (private)", "context": 6000, "max_output": 512, "input": 1.5, "output": 1.5, "latency": 0.56, "throughput": 43.44 } ]
1.125
1.125
0.8
1.2
1.5
1.5
true
MythoMax 13B
Gryphe/MythoMax-L2-13b
Gryphe
[ { "name": "DeepInfra", "context": 4000, "max_output": 4000, "input": 0.065, "output": 0.065, "latency": 0.2, "throughput": 106.4 }, { "name": "Parasail", "context": 4000, "max_output": 4000, "input": 0.11, "output": 0.11, "latency": 0.49, "throughput": 91.35 }, { "name": "Lepton", "context": 4000, "max_output": 4000, "input": 0.18, "output": 0.18, "latency": 0.08, "throughput": 105.5 }, { "name": "Fireworks", "context": 4000, "max_output": 4000, "input": 0.2, "output": 0.2, "latency": 1.92, "throughput": 7.63 }, { "name": "Together", "context": 4000, "max_output": 4000, "input": 0.3, "output": 0.3, "latency": 0.43, "throughput": 125.8 }, { "name": "Mancer", "context": 8000, "max_output": 512, "input": 1.125, "output": 1.125, "latency": 0.7, "throughput": 40.18 }, { "name": "Mancer (private)", "context": 8000, "max_output": 512, "input": 1.5, "output": 1.5, "latency": 0.88, "throughput": 38.69 }, { "name": "NovitaAI", "context": 4000, "max_output": 4000, "input": 0.09, "output": 0.09, "latency": 1.58, "throughput": 88.68 } ]
0.2
0.2
0.065
0.065
1.5
1.5
true
Meta: Llama 2 13B Chat
meta-llama/Llama-2-13b-chat-hf
meta-llama
[ { "name": "Together", "context": 4000, "max_output": 2000, "input": 0.22, "output": 0.22, "latency": 0.77, "throughput": 63.08 }, { "name": "Lepton", "context": 4000, "max_output": 4000, "input": 0.3, "output": 0.3, "latency": 0.5, "throughput": 98.07 } ]
0.3
0.3
0.22
0.22
0.3
0.3
true
Meta: Llama 2 70B Chat
meta-llama/Llama-2-70b-chat-hf
meta-llama
[ { "name": "Together", "context": 4000, "max_output": 4000, "input": 0.9, "output": 0.9, "latency": 0.69, "throughput": 45.46 } ]
0.9
0.9
0.9
0.9
0.9
0.9
true