name
stringlengths 8
58
| hf_id
stringlengths 12
49
⌀ | author
stringlengths 3
21
| providers
listlengths 1
16
| median_input_cost
float64 0
75
| median_output_cost
float64 0
150
| low_input_cost
float64 0
75
| low_output_cost
float64 0
150
| high_input_cost
float64 0
75
| high_output_cost
float64 0
150
| is_open_weights
bool 2
classes |
---|---|---|---|---|---|---|---|---|---|---|
Microsoft: Phi 4 Multimodal Instruct | microsoft/Phi-4-multimodal-instruct | microsoft | [
{
"name": "DeepInfra",
"context": 131000,
"max_output": 131000,
"input": 0.07,
"output": 0.14,
"latency": 0.47,
"throughput": 70
}
] | 0.07 | 0.14 | 0.07 | 0.14 | 0.07 | 0.14 | true |
Qwen: QwQ 32B | Qwen/QwQ-32B | Qwen | [
{
"name": "DeepInfra",
"context": 131000,
"max_output": 131000,
"input": 0.12,
"output": 0.18,
"latency": 0.15,
"throughput": 39.52
},
{
"name": "Nebius AI Studio",
"context": 131000,
"max_output": 131000,
"input": 0.15,
"output": 0.45,
"latency": 0.55,
"throughput": 22.54
},
{
"name": "NovitaAI",
"context": 33000,
"max_output": 33000,
"input": 0.18,
"output": 0.2,
"latency": 1.22,
"throughput": 32.7
},
{
"name": "Groq",
"context": 131000,
"max_output": 131000,
"input": 0.29,
"output": 0.39,
"latency": 0.24,
"throughput": 426.7
},
{
"name": "Hyperbolic",
"context": 131000,
"max_output": 131000,
"input": 0.4,
"output": 0.4,
"latency": 0.34,
"throughput": 35.14
},
{
"name": "Parasail",
"context": 131000,
"max_output": 131000,
"input": 0.5,
"output": 0.5,
"latency": 0.77,
"throughput": 60.56
},
{
"name": "Nebius AI Studio (Fast)",
"context": 131000,
"max_output": 131000,
"input": 0.5,
"output": 1.5,
"latency": 0.52,
"throughput": 80.88
},
{
"name": "Fireworks",
"context": 131000,
"max_output": 131000,
"input": 0.9,
"output": 0.9,
"latency": 0.65,
"throughput": 143.5
}
] | 0.4 | 0.4 | 0.12 | 0.18 | 0.5 | 1.5 | true |
Qwen: Qwen2.5 32B Instruct | Qwen/Qwen2.5-32B-Instruct | Qwen | [
{
"name": "Groq",
"context": 131000,
"max_output": 131000,
"input": 0.79,
"output": 0.79,
"latency": 0.28,
"throughput": 190.5
}
] | 0.79 | 0.79 | 0.79 | 0.79 | 0.79 | 0.79 | true |
Perplexity: R1 1776 | perplexity-ai/r1-1776 | perplexity-ai | [
{
"name": "Perplexity",
"context": 128000,
"max_output": 128000,
"input": 2,
"output": 8,
"latency": 1.5,
"throughput": 38.86
}
] | 2 | 8 | 2 | 8 | 2 | 8 | true |
Llama Guard 3 8B | meta-llama/Llama-Guard-3-8B | meta-llama | [
{
"name": "Groq",
"context": 8000,
"max_output": 8000,
"input": 0.2,
"output": 0.2,
"latency": null,
"throughput": null
},
{
"name": "SambaNova",
"context": 16000,
"max_output": 16000,
"input": 0.3,
"output": 0.3,
"latency": 0.72,
"throughput": 924.6
}
] | 0.3 | 0.3 | 0.2 | 0.2 | 0.3 | 0.3 | true |
Llama 3.1 Tulu 3 405B | allenai/Llama-3.1-Tulu-3-405B | allenai | [
{
"name": "SambaNova",
"context": 16000,
"max_output": 16000,
"input": 5,
"output": 10,
"latency": 3.36,
"throughput": 84.19
}
] | 5 | 10 | 5 | 10 | 5 | 10 | true |
DeepSeek: R1 Distill Llama 8B | meta-llama/Llama-3.1-8B | meta-llama | [
{
"name": "NovitaAI",
"context": 32000,
"max_output": 32000,
"input": 0.04,
"output": 0.04,
"latency": 0.99,
"throughput": 50.36
}
] | 0.04 | 0.04 | 0.04 | 0.04 | 0.04 | 0.04 | true |
AionLabs: Aion-1.0-Mini | FuseAI/FuseO1-DeepSeekR1-QwQ-SkyT1-32B-Preview | FuseAI | [
{
"name": "AionLabs",
"context": 33000,
"max_output": 33000,
"input": 0.7,
"output": 1.4,
"latency": 1.01,
"throughput": 207.3
}
] | 0.7 | 1.4 | 0.7 | 1.4 | 0.7 | 1.4 | true |
Qwen: Qwen2.5 VL 72B Instruct | Qwen/Qwen2.5-VL-72B-Instruct | Qwen | [
{
"name": "Parasail",
"context": 32000,
"max_output": 32000,
"input": 0.7,
"output": 0.7,
"latency": 1.61,
"throughput": 35.92
}
] | 0.7 | 0.7 | 0.7 | 0.7 | 0.7 | 0.7 | true |
DeepSeek: R1 Distill Qwen 1.5B | Qwen/Qwen2.5-Math-1.5B | Qwen | [
{
"name": "Together",
"context": 131000,
"max_output": 33000,
"input": 0.18,
"output": 0.18,
"latency": 0.26,
"throughput": 393.8
}
] | 0.18 | 0.18 | 0.18 | 0.18 | 0.18 | 0.18 | true |
Mistral: Mistral Small 3 | mistralai/Mistral-Small-24B-Instruct-2501 | mistralai | [
{
"name": "DeepInfra",
"context": 33000,
"max_output": 8000,
"input": 0.07,
"output": 0.14,
"latency": 0.74,
"throughput": 59.34
},
{
"name": "Mistral",
"context": 32000,
"max_output": 32000,
"input": 0.1,
"output": 0.3,
"latency": 0.34,
"throughput": 101.8
},
{
"name": "Ubicloud",
"context": 33000,
"max_output": 33000,
"input": 0.3,
"output": 0.3,
"latency": 1.25,
"throughput": 31.15
},
{
"name": "Together",
"context": 33000,
"max_output": 2000,
"input": 0.8,
"output": 0.8,
"latency": 1.19,
"throughput": 48.56
},
{
"name": "Fireworks",
"context": 33000,
"max_output": 33000,
"input": 0.9,
"output": 0.9,
"latency": 2.16,
"throughput": 27.53
}
] | 0.3 | 0.3 | 0.07 | 0.14 | 0.9 | 0.9 | true |
DeepSeek: R1 Distill Qwen 32B | Qwen/Qwen2.5-32B | Qwen | [
{
"name": "DeepInfra",
"context": 131000,
"max_output": 8000,
"input": 0.12,
"output": 0.18,
"latency": 0.13,
"throughput": 36.65
},
{
"name": "Cloudflare",
"context": 80000,
"max_output": 80000,
"input": 0.5,
"output": 4.88,
"latency": 0.32,
"throughput": 30.32
},
{
"name": "Groq",
"context": 131000,
"max_output": 131000,
"input": 0.79,
"output": 0.79,
"latency": 0.13,
"throughput": 137.4
},
{
"name": "NovitaAI",
"context": 64000,
"max_output": 64000,
"input": 0.3,
"output": 0.3,
"latency": 6.37,
"throughput": 30.85
}
] | 0.79 | 0.79 | 0.12 | 0.18 | 0.5 | 4.88 | true |
DeepSeek: R1 Distill Qwen 14B | deepseek-ai/DeepSeek-R1-Distill-Qwen-14B | deepseek-ai | [
{
"name": "Together",
"context": 131000,
"max_output": 33000,
"input": 1.6,
"output": 1.6,
"latency": 0.25,
"throughput": 140.6
},
{
"name": "NovitaAI",
"context": 64000,
"max_output": 64000,
"input": 0.15,
"output": 0.15,
"latency": 0.94,
"throughput": 38.73
}
] | 1.6 | 1.6 | 0.15 | 0.15 | 1.6 | 1.6 | true |
DeepSeek: R1 Distill Llama 70B | deepseek-ai/DeepSeek-R1-Distill-Llama-70B | deepseek-ai | [
{
"name": "DeepInfra",
"context": 131000,
"max_output": 8000,
"input": 0.23,
"output": 0.69,
"latency": 9.61,
"throughput": 59.42
},
{
"name": "Nebius AI Studio",
"context": 128000,
"max_output": 128000,
"input": 0.25,
"output": 0.75,
"latency": 5.37,
"throughput": 72.33
},
{
"name": "inference.net",
"context": 16000,
"max_output": 16000,
"input": 0.4,
"output": 0.4,
"latency": 10.16,
"throughput": 23.81
},
{
"name": "SambaNova",
"context": 16000,
"max_output": 16000,
"input": 0.7,
"output": 1.4,
"latency": 2.79,
"throughput": 489.8
},
{
"name": "Groq",
"context": 131000,
"max_output": 131000,
"input": 0.75,
"output": 0.99,
"latency": 1.28,
"throughput": 623.3
},
{
"name": "Together",
"context": 131000,
"max_output": 16000,
"input": 2,
"output": 2,
"latency": 3.12,
"throughput": 210.9
},
{
"name": "NovitaAI",
"context": 32000,
"max_output": 32000,
"input": 0.8,
"output": 0.8,
"latency": 36.12,
"throughput": 34.55
}
] | 0.8 | 0.8 | 0.4 | 0.4 | 2 | 2 | true |
DeepSeek: R1 | deepseek-ai/DeepSeek-R1 | deepseek-ai | [
{
"name": "Minimax",
"context": 64000,
"max_output": 64000,
"input": 0.55,
"output": 2.19,
"latency": 1.74,
"throughput": 19.54
},
{
"name": "Nebius AI Studio",
"context": 128000,
"max_output": 128000,
"input": 0.8,
"output": 2.4,
"latency": 0.44,
"throughput": 13.93
},
{
"name": "DeepInfra Turbo",
"context": 33000,
"max_output": 33000,
"input": 2,
"output": 6,
"latency": 0.54,
"throughput": 38.08
},
{
"name": "inference.net",
"context": 131000,
"max_output": 33000,
"input": 3,
"output": 3,
"latency": 0.74,
"throughput": 35.03
},
{
"name": "Parasail",
"context": 128000,
"max_output": 128000,
"input": 3,
"output": 3,
"latency": 0.96,
"throughput": 55.96
},
{
"name": "Together",
"context": 164000,
"max_output": 8000,
"input": 3,
"output": 7,
"latency": 0.65,
"throughput": 70.15
},
{
"name": "Friendli",
"context": 164000,
"max_output": 164000,
"input": 3,
"output": 7,
"latency": 0.8,
"throughput": 48.69
},
{
"name": "Fireworks",
"context": 164000,
"max_output": 164000,
"input": 3,
"output": 8,
"latency": 5.48,
"throughput": 77.05
},
{
"name": "SambaNova",
"context": 8000,
"max_output": 8000,
"input": 5,
"output": 7,
"latency": 3.79,
"throughput": 158.4
},
{
"name": "kluster.ai",
"context": 128000,
"max_output": 128000,
"input": 7,
"output": 7,
"latency": 8.55,
"throughput": 31.84
},
{
"name": "DeepSeek",
"context": 64000,
"max_output": 8000,
"input": 0.55,
"output": 2.19,
"latency": 23.57,
"throughput": 28.43
},
{
"name": "DeepInfra",
"context": 66000,
"max_output": 8000,
"input": 0.75,
"output": 2.4,
"latency": 2.43,
"throughput": 9.75
},
{
"name": "Azure",
"context": 164000,
"max_output": 4000,
"input": 1.485,
"output": 5.94,
"latency": 13.92,
"throughput": 28.88
},
{
"name": "NovitaAI",
"context": 64000,
"max_output": 8000,
"input": 4,
"output": 4,
"latency": 23.19,
"throughput": 23.56
},
{
"name": "Featherless",
"context": 33000,
"max_output": 4000,
"input": 6.5,
"output": 8,
"latency": 43.56,
"throughput": 20.13
}
] | 2 | 6 | 0.55 | 2.19 | 6.5 | 8 | true |
MiniMax: MiniMax-01 | MiniMaxAI/MiniMax-Text-01 | MiniMaxAI | [
{
"name": "Minimax",
"context": 1000000,
"max_output": 1000000,
"input": 0.2,
"output": 1.1,
"latency": 1.49,
"throughput": 25.41
}
] | 0.2 | 1.1 | 0.2 | 1.1 | 0.2 | 1.1 | true |
Microsoft: Phi 4 | microsoft/phi-4 | microsoft | [
{
"name": "DeepInfra",
"context": 16000,
"max_output": 8000,
"input": 0.07,
"output": 0.14,
"latency": 0.5,
"throughput": 33.12
},
{
"name": "Nebius AI Studio",
"context": 16000,
"max_output": 16000,
"input": 0.1,
"output": 0.3,
"latency": 0.14,
"throughput": 105.1
}
] | 0.1 | 0.3 | 0.07 | 0.14 | 0.1 | 0.3 | true |
Sao10K: Llama 3.1 70B Hanami x1 | Sao10K/L3.1-70B-Hanami-x1 | Sao10K | [
{
"name": "Infermatic",
"context": 16000,
"max_output": 16000,
"input": 3,
"output": 3,
"latency": 2.39,
"throughput": 31.08
}
] | 3 | 3 | 3 | 3 | 3 | 3 | true |
DeepSeek: DeepSeek V3 | deepseek-ai/DeepSeek-V3 | deepseek-ai | [
{
"name": "inference.net",
"context": 131000,
"max_output": 66000,
"input": 1.2,
"output": 1.2,
"latency": 1.22,
"throughput": 7.59
},
{
"name": "Together",
"context": 131000,
"max_output": 12000,
"input": 1.25,
"output": 1.25,
"latency": 0.75,
"throughput": 30.13
},
{
"name": "DeepSeek",
"context": 64000,
"max_output": 8000,
"input": 0.27,
"output": 1.1,
"latency": 0.42,
"throughput": 12.26
},
{
"name": "NovitaAI",
"context": 64000,
"max_output": 16000,
"input": 0.4,
"output": 1.3,
"latency": 1.43,
"throughput": 31.15
},
{
"name": "DeepInfra",
"context": 66000,
"max_output": 8000,
"input": 0.49,
"output": 0.89,
"latency": 0.89,
"throughput": 7.78
},
{
"name": "Nebius AI Studio",
"context": 131000,
"max_output": 131000,
"input": 0.5,
"output": 1.5,
"latency": 0.33,
"throughput": 20.68
},
{
"name": "Fireworks",
"context": 131000,
"max_output": 131000,
"input": 0.9,
"output": 0.9,
"latency": 1.25,
"throughput": 38.84
}
] | 0.9 | 0.9 | 0.27 | 1.1 | 1.25 | 1.25 | true |
Sao10K: Llama 3.3 Euryale 70B | Sao10K/L3.3-70B-Euryale-v2.3 | Sao10K | [
{
"name": "DeepInfra",
"context": 131000,
"max_output": 8000,
"input": 0.7,
"output": 0.8,
"latency": 0.2,
"throughput": 39.25
},
{
"name": "Infermatic",
"context": 16000,
"max_output": 16000,
"input": 1.5,
"output": 1.5,
"latency": 0.82,
"throughput": 41.21
}
] | 1.5 | 1.5 | 0.7 | 0.8 | 1.5 | 1.5 | true |
EVA Llama 3.33 70B | EVA-UNIT-01/EVA-LLaMA-3.33-70B-v0.1 | EVA-UNIT-01 | [
{
"name": "Featherless",
"context": 16000,
"max_output": 4000,
"input": 4,
"output": 6,
"latency": 2.32,
"throughput": 10.75
}
] | 4 | 6 | 4 | 6 | 4 | 6 | true |
Meta: Llama 3.3 70B Instruct | meta-llama/Llama-3.3-70B-Instruct | meta-llama | [
{
"name": "Lambda",
"context": 131000,
"max_output": 131000,
"input": 0.12,
"output": 0.3,
"latency": 0.81,
"throughput": 11.46
},
{
"name": "DeepInfra",
"context": 131000,
"max_output": 8000,
"input": 0.12,
"output": 0.3,
"latency": 0.25,
"throughput": 33.18
},
{
"name": "Nebius AI Studio",
"context": 131000,
"max_output": 131000,
"input": 0.13,
"output": 0.4,
"latency": 0.37,
"throughput": 21.01
},
{
"name": "inference.net",
"context": 16000,
"max_output": 16000,
"input": 0.4,
"output": 0.4,
"latency": 1.04,
"throughput": 11
},
{
"name": "Hyperbolic",
"context": 131000,
"max_output": 131000,
"input": 0.4,
"output": 0.4,
"latency": 1.17,
"throughput": 57.18
},
{
"name": "Groq",
"context": 33000,
"max_output": 33000,
"input": 0.59,
"output": 0.79,
"latency": 0.26,
"throughput": 301.3
},
{
"name": "Friendli",
"context": 131000,
"max_output": 131000,
"input": 0.6,
"output": 0.6,
"latency": 0.37,
"throughput": 111
},
{
"name": "SambaNova",
"context": 128000,
"max_output": 128000,
"input": 0.6,
"output": 1.2,
"latency": 1.83,
"throughput": 132.2
},
{
"name": "kluster.ai",
"context": 131000,
"max_output": 131000,
"input": 0.7,
"output": 0.7,
"latency": 0.86,
"throughput": 17.33
},
{
"name": "Parasail",
"context": 131000,
"max_output": 131000,
"input": 0.7,
"output": 0.7,
"latency": 0.69,
"throughput": 51.47
},
{
"name": "Cloudflare",
"context": 131000,
"max_output": 131000,
"input": 0.75,
"output": 0.75,
"latency": 0.85,
"throughput": 29.91
},
{
"name": "Lepton",
"context": 131000,
"max_output": 131000,
"input": 0.8,
"output": 0.8,
"latency": 0.33,
"throughput": 40.34
},
{
"name": "Together",
"context": 131000,
"max_output": 2000,
"input": 0.88,
"output": 0.88,
"latency": 0.47,
"throughput": 135.3
},
{
"name": "Avian.io",
"context": 131000,
"max_output": 131000,
"input": 0.9,
"output": 0.9,
"latency": 0.24,
"throughput": 122.4
},
{
"name": "Fireworks",
"context": 131000,
"max_output": 131000,
"input": 0.9,
"output": 0.9,
"latency": 0.34,
"throughput": 125.4
},
{
"name": "NovitaAI",
"context": 131000,
"max_output": 131000,
"input": 0.39,
"output": 0.39,
"latency": 1.17,
"throughput": 26.84
}
] | 0.7 | 0.7 | 0.12 | 0.3 | 0.9 | 0.9 | true |
Qwen: QwQ 32B Preview | Qwen/QwQ-32B-Preview | Qwen | [
{
"name": "Hyperbolic",
"context": 33000,
"max_output": 33000,
"input": 0.2,
"output": 0.2,
"latency": 1.21,
"throughput": 65.09
},
{
"name": "Fireworks",
"context": 33000,
"max_output": 33000,
"input": 0.9,
"output": 0.9,
"latency": 0.29,
"throughput": 67.07
},
{
"name": "Together",
"context": 33000,
"max_output": 33000,
"input": 1.2,
"output": 1.2,
"latency": 0.33,
"throughput": 64.6
},
{
"name": "SambaNova",
"context": 16000,
"max_output": 16000,
"input": 1.5,
"output": 3,
"latency": 0.52,
"throughput": 240.7
},
{
"name": "NovitaAI",
"context": 33000,
"max_output": 33000,
"input": 0.18,
"output": 0.6,
"latency": null,
"throughput": null
}
] | 0.9 | 0.9 | 0.2 | 0.2 | 1.5 | 3 | true |
EVA Qwen2.5 72B | EVA-UNIT-01/EVA-Qwen2.5-72B-v0.1 | EVA-UNIT-01 | [
{
"name": "Parasail",
"context": 32000,
"max_output": 32000,
"input": 0.7,
"output": 0.7,
"latency": 1.31,
"throughput": 30.03
},
{
"name": "Featherless",
"context": 16000,
"max_output": 4000,
"input": 4,
"output": 6,
"latency": 3.35,
"throughput": 10.69
}
] | 4 | 6 | 0.7 | 0.7 | 4 | 6 | true |
Infermatic: Mistral Nemo Inferor 12B | Infermatic/MN-12B-Inferor-v0.0 | Infermatic | [
{
"name": "Featherless",
"context": 16000,
"max_output": 4000,
"input": 0.8,
"output": 1.2,
"latency": 2.74,
"throughput": 15.11
}
] | 0.8 | 1.2 | 0.8 | 1.2 | 0.8 | 1.2 | true |
Qwen2.5 Coder 32B Instruct | Qwen/Qwen2.5-Coder-32B-Instruct | Qwen | [
{
"name": "Lambda",
"context": 33000,
"max_output": 3000,
"input": 0.07,
"output": 0.16,
"latency": 0.47,
"throughput": 68.36
},
{
"name": "DeepInfra",
"context": 33000,
"max_output": 8000,
"input": 0.07,
"output": 0.16,
"latency": 0.25,
"throughput": 29.97
},
{
"name": "Hyperbolic",
"context": 128000,
"max_output": 8000,
"input": 0.2,
"output": 0.2,
"latency": 1.03,
"throughput": 53.2
},
{
"name": "Parasail",
"context": 128000,
"max_output": 128000,
"input": 0.5,
"output": 0.5,
"latency": 0.59,
"throughput": 51.69
},
{
"name": "Groq",
"context": 131000,
"max_output": 131000,
"input": 0.79,
"output": 0.79,
"latency": 0.39,
"throughput": 389.8
},
{
"name": "Together",
"context": 16000,
"max_output": 2000,
"input": 0.8,
"output": 0.8,
"latency": 0.48,
"throughput": 82.14
},
{
"name": "Fireworks",
"context": 131000,
"max_output": 4000,
"input": 0.9,
"output": 0.9,
"latency": 0.37,
"throughput": 63.7
},
{
"name": "Mancer",
"context": 33000,
"max_output": 2000,
"input": 1.5,
"output": 2.813,
"latency": 0.84,
"throughput": 19.13
},
{
"name": "SambaNova",
"context": 16000,
"max_output": 16000,
"input": 1.5,
"output": 3,
"latency": 0.79,
"throughput": 390.9
},
{
"name": "Mancer (private)",
"context": 33000,
"max_output": 2000,
"input": 2,
"output": 3.75,
"latency": null,
"throughput": null
},
{
"name": "Featherless",
"context": 16000,
"max_output": 4000,
"input": 2.6,
"output": 3.4,
"latency": 4.61,
"throughput": 12.84
}
] | 0.8 | 0.8 | 0.07 | 0.16 | 2.6 | 3.4 | true |
SorcererLM 8x22B | rAIfle/SorcererLM-8x22b-bf16 | rAIfle | [
{
"name": "Infermatic",
"context": 16000,
"max_output": 16000,
"input": 4.5,
"output": 4.5,
"latency": 0.8,
"throughput": 7.14
}
] | 4.5 | 4.5 | 4.5 | 4.5 | 4.5 | 4.5 | true |
EVA Qwen2.5 32B | EVA-UNIT-01/EVA-Qwen2.5-32B-v0.2 | EVA-UNIT-01 | [
{
"name": "Featherless",
"context": 16000,
"max_output": 4000,
"input": 2.6,
"output": 3.4,
"latency": 2.25,
"throughput": 8.53
}
] | 2.6 | 3.4 | 2.6 | 3.4 | 2.6 | 3.4 | true |
Unslopnemo 12B | TheDrummer/UnslopNemo-12B-v4.1 | TheDrummer | [
{
"name": "Infermatic",
"context": 32000,
"max_output": 32000,
"input": 0.5,
"output": 0.5,
"latency": 0.77,
"throughput": 64.44
}
] | 0.5 | 0.5 | 0.5 | 0.5 | 0.5 | 0.5 | true |
NeverSleep: Lumimaid v0.2 70B | NeverSleep/Lumimaid-v0.2-70B | NeverSleep | [
{
"name": "Mancer",
"context": 16000,
"max_output": 2000,
"input": 3.375,
"output": 4.5,
"latency": 3.1,
"throughput": 12.96
},
{
"name": "Featherless",
"context": 16000,
"max_output": 4000,
"input": 4,
"output": 6,
"latency": 2.75,
"throughput": 11.7
},
{
"name": "Mancer (private)",
"context": 16000,
"max_output": 2000,
"input": 4.5,
"output": 6,
"latency": 1.45,
"throughput": 11.78
}
] | 4 | 6 | 3.375 | 4.5 | 4.5 | 6 | true |
Magnum v4 72B | anthracite-org/magnum-v4-72b | anthracite-org | [
{
"name": "Mancer",
"context": 16000,
"max_output": 1000,
"input": 1.875,
"output": 2.25,
"latency": 0.94,
"throughput": 13.65
},
{
"name": "Mancer (private)",
"context": 16000,
"max_output": 1000,
"input": 2.5,
"output": 3,
"latency": 1.09,
"throughput": 12.53
},
{
"name": "Infermatic",
"context": 33000,
"max_output": 33000,
"input": 3,
"output": 3,
"latency": 0.33,
"throughput": 18.03
},
{
"name": "Featherless",
"context": 16000,
"max_output": 4000,
"input": 4,
"output": 6,
"latency": 2.57,
"throughput": 10.33
}
] | 3 | 3 | 1.875 | 2.25 | 4 | 6 | true |
Qwen2.5 7B Instruct | Qwen/Qwen2.5-7B-Instruct | Qwen | [
{
"name": "DeepInfra",
"context": 33000,
"max_output": 8000,
"input": 0.025,
"output": 0.05,
"latency": 0.19,
"throughput": 53.19
},
{
"name": "Together",
"context": 33000,
"max_output": 2000,
"input": 0.3,
"output": 0.3,
"latency": 0.37,
"throughput": 112.9
}
] | 0.3 | 0.3 | 0.025 | 0.05 | 0.3 | 0.3 | true |
NVIDIA: Llama 3.1 Nemotron 70B Instruct | nvidia/Llama-3.1-Nemotron-70B-Instruct-HF | nvidia | [
{
"name": "Lambda",
"context": 131000,
"max_output": 131000,
"input": 0.12,
"output": 0.3,
"latency": 0.91,
"throughput": 35.25
},
{
"name": "DeepInfra",
"context": 131000,
"max_output": 8000,
"input": 0.12,
"output": 0.3,
"latency": 0.55,
"throughput": 28.07
},
{
"name": "Together",
"context": 33000,
"max_output": 33000,
"input": 0.88,
"output": 0.88,
"latency": 0.51,
"throughput": 70.6
},
{
"name": "Infermatic",
"context": 32000,
"max_output": 32000,
"input": 1,
"output": 1,
"latency": 2.61,
"throughput": 13.76
}
] | 0.88 | 0.88 | 0.12 | 0.3 | 1 | 1 | true |
Magnum v2 72B | anthracite-org/magnum-v2-72b | anthracite-org | [
{
"name": "Infermatic",
"context": 33000,
"max_output": 33000,
"input": 3,
"output": 3,
"latency": 0.46,
"throughput": 40.66
},
{
"name": "Featherless",
"context": 16000,
"max_output": 4000,
"input": 4,
"output": 6,
"latency": 3.5,
"throughput": 10.75
}
] | 4 | 6 | 3 | 3 | 4 | 6 | true |
Rocinante 12B | TheDrummer/Rocinante-12B-v1.1 | TheDrummer | [
{
"name": "Infermatic",
"context": 33000,
"max_output": 33000,
"input": 0.25,
"output": 0.5,
"latency": 0.76,
"throughput": 23.44
},
{
"name": "Featherless",
"context": 16000,
"max_output": 4000,
"input": 0.8,
"output": 1.2,
"latency": 1.5,
"throughput": 13.2
}
] | 0.8 | 1.2 | 0.25 | 0.5 | 0.8 | 1.2 | true |
Meta: Llama 3.2 3B Instruct | meta-llama/Llama-3.2-3B-Instruct | meta-llama | [
{
"name": "Lambda",
"context": 131000,
"max_output": 131000,
"input": 0.015,
"output": 0.025,
"latency": 0.44,
"throughput": 237.9
},
{
"name": "DeepInfra",
"context": 131000,
"max_output": 8000,
"input": 0.015,
"output": 0.025,
"latency": 0.21,
"throughput": 147
},
{
"name": "inference.net",
"context": 16000,
"max_output": 16000,
"input": 0.02,
"output": 0.02,
"latency": 1.04,
"throughput": 82.6
},
{
"name": "Lepton",
"context": 131000,
"max_output": 131000,
"input": 0.03,
"output": 0.03,
"latency": 0.3,
"throughput": 199.5
},
{
"name": "Together",
"context": 131000,
"max_output": 16000,
"input": 0.06,
"output": 0.06,
"latency": 0.32,
"throughput": 73.29
},
{
"name": "SambaNova",
"context": 4000,
"max_output": 2000,
"input": 0.08,
"output": 0.16,
"latency": 0.26,
"throughput": 1271
},
{
"name": "Fireworks",
"context": 131000,
"max_output": 131000,
"input": 0.1,
"output": 0.1,
"latency": 0.76,
"throughput": 150.3
},
{
"name": "Hyperbolic",
"context": 131000,
"max_output": 131000,
"input": 0.1,
"output": 0.1,
"latency": 1.23,
"throughput": 173.8
},
{
"name": "Cloudflare",
"context": 131000,
"max_output": 131000,
"input": 0.1,
"output": 0.1,
"latency": 0.91,
"throughput": 172.1
},
{
"name": "NovitaAI",
"context": 33000,
"max_output": 33000,
"input": 0.03,
"output": 0.05,
"latency": 1.09,
"throughput": 78.99
}
] | 0.06 | 0.06 | 0.015 | 0.025 | 0.08 | 0.16 | true |
Meta: Llama 3.2 1B Instruct | meta-llama/Llama-3.2-1B-Instruct | meta-llama | [
{
"name": "Lepton",
"context": 131000,
"max_output": 131000,
"input": 0.01,
"output": 0.01,
"latency": 0.33,
"throughput": 402.7
},
{
"name": "inference.net",
"context": 16000,
"max_output": 16000,
"input": 0.01,
"output": 0.01,
"latency": 0.81,
"throughput": 135.2
},
{
"name": "DeepInfra",
"context": 131000,
"max_output": 8000,
"input": 0.01,
"output": 0.01,
"latency": 0.17,
"throughput": 169.1
},
{
"name": "SambaNova",
"context": 4000,
"max_output": 2000,
"input": 0.04,
"output": 0.08,
"latency": 0.76,
"throughput": 2039
},
{
"name": "Cloudflare",
"context": 131000,
"max_output": 131000,
"input": 0.1,
"output": 0.1,
"latency": 0.88,
"throughput": 253
},
{
"name": "NovitaAI",
"context": 131000,
"max_output": 131000,
"input": 0.02,
"output": 0.02,
"latency": null,
"throughput": null
}
] | 0.02 | 0.02 | 0.01 | 0.01 | 0.1 | 0.1 | true |
Meta: Llama 3.2 90B Vision Instruct | meta-llama/Llama-3.2-90B-Vision-Instruct | meta-llama | [
{
"name": "SambaNova",
"context": 4000,
"max_output": 2000,
"input": 0.8,
"output": 1.6,
"latency": 0.55,
"throughput": 261.7
},
{
"name": "Fireworks",
"context": 131000,
"max_output": 131000,
"input": 0.9,
"output": 0.9,
"latency": 1.13,
"throughput": 28.98
},
{
"name": "Together",
"context": 131000,
"max_output": 131000,
"input": 1.2,
"output": 1.2,
"latency": 0.52,
"throughput": 27.08
},
{
"name": "DeepInfra",
"context": 33000,
"max_output": 8000,
"input": 0.35,
"output": 0.4,
"latency": 0.13,
"throughput": 20.1
}
] | 1.2 | 1.2 | 0.35 | 0.4 | 0.8 | 1.6 | true |
Meta: Llama 3.2 11B Vision Instruct | meta-llama/Llama-3.2-11B-Vision-Instruct | meta-llama | [
{
"name": "inference.net",
"context": 16000,
"max_output": 16000,
"input": 0.055,
"output": 0.055,
"latency": 1.71,
"throughput": 33.56
},
{
"name": "DeepInfra",
"context": 131000,
"max_output": 8000,
"input": 0.055,
"output": 0.055,
"latency": 0.14,
"throughput": 55.62
},
{
"name": "SambaNova",
"context": 4000,
"max_output": 2000,
"input": 0.15,
"output": 0.3,
"latency": 0.57,
"throughput": 504.6
},
{
"name": "Together",
"context": 131000,
"max_output": 131000,
"input": 0.18,
"output": 0.18,
"latency": 0.52,
"throughput": 161.8
},
{
"name": "Cloudflare",
"context": 131000,
"max_output": 131000,
"input": 0.2,
"output": 0.2,
"latency": 0.64,
"throughput": 42.9
},
{
"name": "Fireworks",
"context": 131000,
"max_output": 131000,
"input": 0.2,
"output": 0.2,
"latency": 4.04,
"throughput": 36.58
},
{
"name": "NovitaAI",
"context": 33000,
"max_output": 33000,
"input": 0.06,
"output": 0.06,
"latency": null,
"throughput": null
}
] | 0.18 | 0.18 | 0.055 | 0.055 | 0.15 | 0.3 | true |
Qwen2.5 72B Instruct | Qwen/Qwen2.5-72B-Instruct | Qwen | [
{
"name": "Nebius AI Studio",
"context": 128000,
"max_output": 128000,
"input": 0.13,
"output": 0.4,
"latency": 0.39,
"throughput": 25
},
{
"name": "DeepInfra",
"context": 33000,
"max_output": 8000,
"input": 0.13,
"output": 0.4,
"latency": 0.44,
"throughput": 42.07
},
{
"name": "Hyperbolic",
"context": 33000,
"max_output": 33000,
"input": 0.4,
"output": 0.4,
"latency": 1.16,
"throughput": 53.32
},
{
"name": "Fireworks",
"context": 33000,
"max_output": 33000,
"input": 0.9,
"output": 0.9,
"latency": 0.34,
"throughput": 41.79
},
{
"name": "Together",
"context": 131000,
"max_output": 2000,
"input": 1.2,
"output": 1.2,
"latency": 0.4,
"throughput": 103.2
},
{
"name": "SambaNova",
"context": 16000,
"max_output": 16000,
"input": 2,
"output": 4,
"latency": 0.45,
"throughput": 376.3
},
{
"name": "NovitaAI",
"context": 32000,
"max_output": 4000,
"input": 0.38,
"output": 0.4,
"latency": 0.86,
"throughput": 20.33
}
] | 0.4 | 0.4 | 0.13 | 0.4 | 2 | 4 | true |
Qwen: Qwen2.5-VL 72B Instruct | Qwen/Qwen2.5-VL-72B-Instruct | Qwen | [
{
"name": "Hyperbolic",
"context": 4000,
"max_output": 4000,
"input": 0.4,
"output": 0.4,
"latency": 3.23,
"throughput": 27.88
}
] | 0.4 | 0.4 | 0.4 | 0.4 | 0.4 | 0.4 | true |
NeverSleep: Lumimaid v0.2 8B | NeverSleep/Lumimaid-v0.2-8B | NeverSleep | [
{
"name": "Mancer",
"context": 33000,
"max_output": 2000,
"input": 0.1875,
"output": 1.125,
"latency": 0.91,
"throughput": 33.73
},
{
"name": "Mancer (private)",
"context": 33000,
"max_output": 2000,
"input": 0.25,
"output": 1.5,
"latency": 0.67,
"throughput": 33.78
},
{
"name": "Featherless",
"context": 16000,
"max_output": 4000,
"input": 0.8,
"output": 1.2,
"latency": 1.62,
"throughput": 26.35
}
] | 0.25 | 1.5 | 0.1875 | 1.125 | 0.8 | 1.2 | true |
Mistral: Pixtral 12B | mistralai/Pixtral-12B-2409 | mistralai | [
{
"name": "Hyperbolic",
"context": 4000,
"max_output": 4000,
"input": 0.1,
"output": 0.1,
"latency": 1.67,
"throughput": 69.2
},
{
"name": "Mistral",
"context": 4000,
"max_output": 4000,
"input": 0.15,
"output": 0.15,
"latency": 0.85,
"throughput": 88.97
}
] | 0.15 | 0.15 | 0.1 | 0.1 | 0.15 | 0.15 | true |
Sao10K: Llama 3.1 Euryale 70B v2.2 | Sao10K/L3.1-70B-Euryale-v2.2 | Sao10K | [
{
"name": "DeepInfra",
"context": 131000,
"max_output": 8000,
"input": 0.7,
"output": 0.8,
"latency": 0.3,
"throughput": 37.01
},
{
"name": "Infermatic",
"context": 16000,
"max_output": 16000,
"input": 1.5,
"output": 1.5,
"latency": 1.13,
"throughput": 19.14
},
{
"name": "NovitaAI",
"context": 16000,
"max_output": 16000,
"input": 1.48,
"output": 1.48,
"latency": 1.6,
"throughput": 21.19
}
] | 1.48 | 1.48 | 0.7 | 0.8 | 1.5 | 1.5 | true |
Qwen: Qwen2.5-VL 7B Instruct | Qwen/Qwen2.5-VL-7B-Instruct | Qwen | [
{
"name": "Hyperbolic",
"context": 4000,
"max_output": 4000,
"input": 0.1,
"output": 0.1,
"latency": 0.68,
"throughput": 57.96
}
] | 0.1 | 0.1 | 0.1 | 0.1 | 0.1 | 0.1 | true |
Microsoft: Phi-3.5 Mini 128K Instruct | microsoft/Phi-3.5-mini-instruct | microsoft | [
{
"name": "Azure",
"context": 128000,
"max_output": 128000,
"input": 0.1,
"output": 0.1,
"latency": 4.11,
"throughput": 17
}
] | 0.1 | 0.1 | 0.1 | 0.1 | 0.1 | 0.1 | true |
Nous: Hermes 3 70B Instruct | NousResearch/Hermes-3-Llama-3.1-70B | NousResearch | [
{
"name": "Lambda",
"context": 131000,
"max_output": 131000,
"input": 0.12,
"output": 0.3,
"latency": 0.75,
"throughput": 32.07
},
{
"name": "Hyperbolic",
"context": 12000,
"max_output": 12000,
"input": 0.4,
"output": 0.4,
"latency": 0.94,
"throughput": 29.52
}
] | 0.4 | 0.4 | 0.12 | 0.3 | 0.4 | 0.4 | true |
Nous: Hermes 3 405B Instruct | NousResearch/Hermes-3-Llama-3.1-405B | NousResearch | [
{
"name": "Lambda",
"context": 131000,
"max_output": 131000,
"input": 0.8,
"output": 0.8,
"latency": 1.04,
"throughput": 27.08
},
{
"name": "inference.net",
"context": 33000,
"max_output": 33000,
"input": 0.8,
"output": 0.8,
"latency": 0.92,
"throughput": 25.74
},
{
"name": "DeepInfra",
"context": 131000,
"max_output": 8000,
"input": 0.8,
"output": 0.8,
"latency": 0.77,
"throughput": 12.78
},
{
"name": "Nebius AI Studio",
"context": 131000,
"max_output": 131000,
"input": 1,
"output": 3,
"latency": 0.31,
"throughput": 27.65
}
] | 0.8 | 0.8 | 0.8 | 0.8 | 1 | 3 | true |
Sao10K: Llama 3 8B Lunaris | Sao10K/L3-8B-Lunaris-v1 | Sao10K | [
{
"name": "DeepInfra",
"context": 8000,
"max_output": 8000,
"input": 0.03,
"output": 0.06,
"latency": 0.38,
"throughput": 69.42
},
{
"name": "NovitaAI",
"context": 8000,
"max_output": 8000,
"input": 0.05,
"output": 0.05,
"latency": 0.93,
"throughput": 66.61
}
] | 0.05 | 0.05 | 0.03 | 0.06 | 0.05 | 0.05 | true |
Aetherwiing: Starcannon 12B | intervitens/mini-magnum-12b-v1.1 | intervitens | [
{
"name": "Featherless",
"context": 16000,
"max_output": 4000,
"input": 0.8,
"output": 1.2,
"latency": 3.93,
"throughput": 15.71
}
] | 0.8 | 1.2 | 0.8 | 1.2 | 0.8 | 1.2 | true |
Meta: Llama 3.1 405B (base) | meta-llama/llama-3.1-405B | meta-llama | [
{
"name": "Hyperbolic (quantized)",
"context": 33000,
"max_output": 33000,
"input": 2,
"output": 2,
"latency": 0.86,
"throughput": 23.83
},
{
"name": "Hyperbolic",
"context": 33000,
"max_output": 33000,
"input": 4,
"output": 4,
"latency": 1.32,
"throughput": 16.23
}
] | 4 | 4 | 2 | 2 | 4 | 4 | true |
Mistral Nemo 12B Celeste | nothingiisreal/MN-12B-Celeste-V1.9 | nothingiisreal | [
{
"name": "Featherless",
"context": 16000,
"max_output": 4000,
"input": 0.8,
"output": 1.2,
"latency": 3.54,
"throughput": 14.64
}
] | 0.8 | 1.2 | 0.8 | 1.2 | 0.8 | 1.2 | true |
Meta: Llama 3.1 405B Instruct | meta-llama/Meta-Llama-3.1-405B-Instruct | meta-llama | [
{
"name": "DeepInfra",
"context": 33000,
"max_output": 8000,
"input": 0.8,
"output": 0.8,
"latency": 0.17,
"throughput": 3.55
},
{
"name": "Lambda",
"context": 131000,
"max_output": 131000,
"input": 0.8,
"output": 0.8,
"latency": 0.84,
"throughput": 17.46
},
{
"name": "Nebius AI Studio",
"context": 128000,
"max_output": 128000,
"input": 1,
"output": 3,
"latency": 0.14,
"throughput": 23.97
},
{
"name": "Fireworks",
"context": 131000,
"max_output": 131000,
"input": 3,
"output": 3,
"latency": 3.45,
"throughput": 39.73
},
{
"name": "Together",
"context": 131000,
"max_output": 131000,
"input": 3.5,
"output": 3.5,
"latency": 6.1,
"throughput": 27.34
},
{
"name": "kluster.ai",
"context": 131000,
"max_output": 131000,
"input": 3.5,
"output": 3.5,
"latency": 0.83,
"throughput": 17.03
},
{
"name": "Hyperbolic",
"context": 33000,
"max_output": 33000,
"input": 4,
"output": 4,
"latency": 4.35,
"throughput": 3.8
},
{
"name": "SambaNova",
"context": 8000,
"max_output": 4000,
"input": 5,
"output": 10,
"latency": 1.52,
"throughput": 81.33
}
] | 3.5 | 3.5 | 0.8 | 0.8 | 5 | 10 | true |
Meta: Llama 3.1 8B Instruct | meta-llama/Meta-Llama-3.1-8B-Instruct | meta-llama | [
{
"name": "DeepInfra",
"context": 131000,
"max_output": 8000,
"input": 0.02,
"output": 0.05,
"latency": 0.17,
"throughput": 116
},
{
"name": "Lambda",
"context": 131000,
"max_output": 131000,
"input": 0.025,
"output": 0.04,
"latency": 0.48,
"throughput": 168
},
{
"name": "inference.net",
"context": 16000,
"max_output": 16000,
"input": 0.03,
"output": 0.03,
"latency": 0.94,
"throughput": 58.13
},
{
"name": "Groq",
"context": 131000,
"max_output": 8000,
"input": 0.05,
"output": 0.08,
"latency": 0.35,
"throughput": 642
},
{
"name": "Lepton",
"context": 131000,
"max_output": 131000,
"input": 0.07,
"output": 0.07,
"latency": 0.21,
"throughput": 97.29
},
{
"name": "Friendli",
"context": 131000,
"max_output": 131000,
"input": 0.1,
"output": 0.1,
"latency": 0.43,
"throughput": 223.1
},
{
"name": "Hyperbolic",
"context": 33000,
"max_output": 33000,
"input": 0.1,
"output": 0.1,
"latency": 1.17,
"throughput": 102.9
},
{
"name": "SambaNova",
"context": 8000,
"max_output": 4000,
"input": 0.1,
"output": 0.2,
"latency": 0.32,
"throughput": 800
},
{
"name": "Cloudflare",
"context": 131000,
"max_output": 131000,
"input": 0.15,
"output": 0.15,
"latency": 4.27,
"throughput": 26.08
},
{
"name": "kluster.ai",
"context": 131000,
"max_output": 131000,
"input": 0.18,
"output": 0.18,
"latency": 0.33,
"throughput": 13.91
},
{
"name": "Together",
"context": 131000,
"max_output": 131000,
"input": 0.18,
"output": 0.18,
"latency": 0.25,
"throughput": 288
},
{
"name": "Fireworks",
"context": 131000,
"max_output": 131000,
"input": 0.2,
"output": 0.2,
"latency": 0.65,
"throughput": 211.1
},
{
"name": "Avian.io",
"context": 131000,
"max_output": 131000,
"input": 0.2,
"output": 0.2,
"latency": 0.29,
"throughput": 302.6
},
{
"name": "NovitaAI",
"context": 16000,
"max_output": 8000,
"input": 0.05,
"output": 0.05,
"latency": null,
"throughput": null
}
] | 0.1 | 0.1 | 0.03 | 0.03 | 0.2 | 0.2 | true |
Meta: Llama 3.1 70B Instruct | meta-llama/Meta-Llama-3.1-70B-Instruct | meta-llama | [
{
"name": "DeepInfra",
"context": 131000,
"max_output": 8000,
"input": 0.12,
"output": 0.3,
"latency": 0.43,
"throughput": 32.15
},
{
"name": "Lambda",
"context": 131000,
"max_output": 131000,
"input": 0.12,
"output": 0.3,
"latency": 0.61,
"throughput": 32.2
},
{
"name": "Nebius AI Studio",
"context": 128000,
"max_output": 128000,
"input": 0.13,
"output": 0.4,
"latency": 0.15,
"throughput": 32.57
},
{
"name": "inference.net",
"context": 16000,
"max_output": 16000,
"input": 0.4,
"output": 0.4,
"latency": 0.99,
"throughput": 18.3
},
{
"name": "Hyperbolic",
"context": 33000,
"max_output": 33000,
"input": 0.4,
"output": 0.4,
"latency": 0.77,
"throughput": 108.8
},
{
"name": "Friendli",
"context": 131000,
"max_output": 131000,
"input": 0.6,
"output": 0.6,
"latency": 0.19,
"throughput": 112.7
},
{
"name": "SambaNova",
"context": 8000,
"max_output": 4000,
"input": 0.6,
"output": 1.2,
"latency": 0.54,
"throughput": 267.7
},
{
"name": "Cloudflare",
"context": 131000,
"max_output": 131000,
"input": 0.75,
"output": 0.75,
"latency": 0.89,
"throughput": 26.16
},
{
"name": "Lepton",
"context": 131000,
"max_output": 131000,
"input": 0.8,
"output": 0.8,
"latency": 0.08,
"throughput": 39.8
},
{
"name": "Together",
"context": 131000,
"max_output": 131000,
"input": 0.88,
"output": 0.88,
"latency": 0.31,
"throughput": 114.5
},
{
"name": "Fireworks",
"context": 131000,
"max_output": 131000,
"input": 0.9,
"output": 0.9,
"latency": 0.41,
"throughput": 93.77
},
{
"name": "NovitaAI",
"context": 33000,
"max_output": 33000,
"input": 0.34,
"output": 0.39,
"latency": 2.27,
"throughput": 50.46
}
] | 0.6 | 0.6 | 0.12 | 0.3 | 0.9 | 0.9 | true |
Mistral: Mistral Nemo | mistralai/Mistral-Nemo-Instruct-2407 | mistralai | [
{
"name": "DeepInfra",
"context": 131000,
"max_output": 8000,
"input": 0.035,
"output": 0.08,
"latency": 0.23,
"throughput": 63.49
},
{
"name": "Nebius AI Studio",
"context": 128000,
"max_output": 128000,
"input": 0.04,
"output": 0.12,
"latency": 0.7,
"throughput": 15.59
},
{
"name": "inference.net",
"context": 16000,
"max_output": 16000,
"input": 0.1,
"output": 0.1,
"latency": 0.83,
"throughput": 65.67
},
{
"name": "Parasail",
"context": 128000,
"max_output": 128000,
"input": 0.11,
"output": 0.11,
"latency": 0.71,
"throughput": 131.8
},
{
"name": "Mistral",
"context": 128000,
"max_output": 128000,
"input": 0.15,
"output": 0.15,
"latency": 0.34,
"throughput": 125.9
},
{
"name": "Lepton",
"context": 128000,
"max_output": 128000,
"input": 0.18,
"output": 0.18,
"latency": 0.28,
"throughput": 50.82
},
{
"name": "Azure",
"context": 128000,
"max_output": 128000,
"input": 0.3,
"output": 0.3,
"latency": 0.58,
"throughput": 97.82
},
{
"name": "NovitaAI",
"context": 131000,
"max_output": 131000,
"input": 0.17,
"output": 0.17,
"latency": 0.77,
"throughput": 51.15
}
] | 0.15 | 0.15 | 0.035 | 0.08 | 0.3 | 0.3 | true |
Mistral: Codestral Mamba | mistralai/mamba-codestral-7B-v0.1 | mistralai | [
{
"name": "Mistral",
"context": 256000,
"max_output": 256000,
"input": 0.25,
"output": 0.25,
"latency": 0.48,
"throughput": 108.1
}
] | 0.25 | 0.25 | 0.25 | 0.25 | 0.25 | 0.25 | true |
Qwen 2 7B Instruct | Qwen/Qwen2-7B-Instruct | Qwen | [
{
"name": "NovitaAI",
"context": 33000,
"max_output": 33000,
"input": 0.054,
"output": 0.054,
"latency": 0.95,
"throughput": 47.27
}
] | 0.054 | 0.054 | 0.054 | 0.054 | 0.054 | 0.054 | true |
Google: Gemma 2 27B | google/gemma-2-27b-it | google | [
{
"name": "DeepInfra",
"context": 8000,
"max_output": 8000,
"input": 0.27,
"output": 0.27,
"latency": 0.5,
"throughput": 37.61
},
{
"name": "Together",
"context": 8000,
"max_output": 2000,
"input": 0.8,
"output": 0.8,
"latency": 0.45,
"throughput": 71.67
}
] | 0.8 | 0.8 | 0.27 | 0.27 | 0.8 | 0.8 | true |
Magnum 72B | alpindale/magnum-72b-v1 | alpindale | [
{
"name": "Mancer",
"context": 16000,
"max_output": 1000,
"input": 1.875,
"output": 2.25,
"latency": 0.54,
"throughput": 22.53
},
{
"name": "Mancer (private)",
"context": 16000,
"max_output": 1000,
"input": 2.5,
"output": 3,
"latency": 1.33,
"throughput": 22.81
},
{
"name": "Featherless",
"context": 16000,
"max_output": 4000,
"input": 4,
"output": 6,
"latency": 6.09,
"throughput": 11.94
}
] | 2.5 | 3 | 1.875 | 2.25 | 4 | 6 | true |
Google: Gemma 2 9B | google/gemma-2-9b-it | google | [
{
"name": "DeepInfra",
"context": 8000,
"max_output": 8000,
"input": 0.03,
"output": 0.06,
"latency": 0.23,
"throughput": 39.15
},
{
"name": "Lepton",
"context": 8000,
"max_output": 8000,
"input": 0.07,
"output": 0.07,
"latency": 0.35,
"throughput": 102.3
},
{
"name": "Groq",
"context": 8000,
"max_output": 8000,
"input": 0.2,
"output": 0.2,
"latency": 0.2,
"throughput": 565
},
{
"name": "Together",
"context": 8000,
"max_output": 8000,
"input": 0.3,
"output": 0.3,
"latency": 0.34,
"throughput": 107.8
},
{
"name": "NovitaAI",
"context": 8000,
"max_output": 8000,
"input": 0.08,
"output": 0.08,
"latency": 10.15,
"throughput": 29.6
}
] | 0.08 | 0.08 | 0.03 | 0.06 | 0.3 | 0.3 | true |
Sao10k: Llama 3 Euryale 70B v2.1 | Sao10K/L3-70B-Euryale-v2.1 | Sao10K | [
{
"name": "DeepInfra",
"context": 8000,
"max_output": 8000,
"input": 0.7,
"output": 0.8,
"latency": 0.21,
"throughput": 37.88
},
{
"name": "NovitaAI",
"context": 16000,
"max_output": 16000,
"input": 1.48,
"output": 1.48,
"latency": 3.01,
"throughput": 19.66
}
] | 1.48 | 1.48 | 0.7 | 0.8 | 1.48 | 1.48 | true |
Dolphin 2.9.2 Mixtral 8x22B 🐬 | cognitivecomputations/dolphin-2.9.2-mixtral-8x22b | cognitivecomputations | [
{
"name": "NovitaAI",
"context": 16000,
"max_output": 16000,
"input": 0.9,
"output": 0.9,
"latency": 4.12,
"throughput": 9.66
}
] | 0.9 | 0.9 | 0.9 | 0.9 | 0.9 | 0.9 | true |
Qwen 2 72B Instruct | Qwen/Qwen2-72B-Instruct | Qwen | [
{
"name": "Together",
"context": 33000,
"max_output": 4000,
"input": 0.9,
"output": 0.9,
"latency": 0.37,
"throughput": 66.68
}
] | 0.9 | 0.9 | 0.9 | 0.9 | 0.9 | 0.9 | true |
Mistral: Mistral 7B Instruct | mistralai/Mistral-7B-Instruct-v0.3 | mistralai | [
{
"name": "DeepInfra",
"context": 33000,
"max_output": 8000,
"input": 0.03,
"output": 0.055,
"latency": 0.29,
"throughput": 82.53
},
{
"name": "Parasail",
"context": 33000,
"max_output": 33000,
"input": 0.11,
"output": 0.11,
"latency": 0.66,
"throughput": 128.7
},
{
"name": "Together",
"context": 33000,
"max_output": 4000,
"input": 0.2,
"output": 0.2,
"latency": 0.34,
"throughput": 163.9
},
{
"name": "NovitaAI",
"context": 33000,
"max_output": 33000,
"input": 0.059,
"output": 0.059,
"latency": 1.26,
"throughput": 121.9
},
{
"name": "Lepton",
"context": 33000,
"max_output": 33000,
"input": 0.07,
"output": 0.07,
"latency": 0.55,
"throughput": 105.5
}
] | 0.07 | 0.07 | 0.03 | 0.055 | 0.2 | 0.2 | true |
Mistral: Mistral 7B Instruct v0.3 | mistralai/Mistral-7B-Instruct-v0.3 | mistralai | [
{
"name": "DeepInfra",
"context": 33000,
"max_output": 8000,
"input": 0.03,
"output": 0.055,
"latency": 0.18,
"throughput": 102.4
},
{
"name": "Lepton",
"context": 33000,
"max_output": 33000,
"input": 0.07,
"output": 0.07,
"latency": 0.38,
"throughput": 105.6
},
{
"name": "Together",
"context": 33000,
"max_output": 4000,
"input": 0.2,
"output": 0.2,
"latency": 0.3,
"throughput": 173.8
},
{
"name": "NovitaAI",
"context": 33000,
"max_output": 33000,
"input": 0.059,
"output": 0.059,
"latency": null,
"throughput": null
}
] | 0.07 | 0.07 | 0.03 | 0.055 | 0.2 | 0.2 | true |
NousResearch: Hermes 2 Pro - Llama-3 8B | NousResearch/Hermes-2-Pro-Llama-3-8B | NousResearch | [
{
"name": "Lambda",
"context": 131000,
"max_output": 131000,
"input": 0.025,
"output": 0.04,
"latency": 0.53,
"throughput": 196.1
},
{
"name": "NovitaAI",
"context": 8000,
"max_output": 8000,
"input": 0.14,
"output": 0.14,
"latency": 1.16,
"throughput": 122.3
}
] | 0.14 | 0.14 | 0.025 | 0.04 | 0.14 | 0.14 | true |
Microsoft: Phi-3 Mini 128K Instruct | microsoft/Phi-3-mini-128k-instruct | microsoft | [
{
"name": "Azure",
"context": 128000,
"max_output": 128000,
"input": 0.1,
"output": 0.1,
"latency": 0.59,
"throughput": 67.71
}
] | 0.1 | 0.1 | 0.1 | 0.1 | 0.1 | 0.1 | true |
Microsoft: Phi-3 Medium 128K Instruct | microsoft/Phi-3-medium-128k-instruct | microsoft | [
{
"name": "Azure",
"context": 128000,
"max_output": 128000,
"input": 1,
"output": 1,
"latency": 0.59,
"throughput": 52.04
}
] | 1 | 1 | 1 | 1 | 1 | 1 | true |
NeverSleep: Llama 3 Lumimaid 70B | NeverSleep/Llama-3-Lumimaid-70B-v0.1 | NeverSleep | [
{
"name": "Mancer",
"context": 8000,
"max_output": 2000,
"input": 3.375,
"output": 4.5,
"latency": 0.88,
"throughput": 13.56
},
{
"name": "Featherless",
"context": 8000,
"max_output": 4000,
"input": 4,
"output": 6,
"latency": 2.58,
"throughput": 11.46
},
{
"name": "Mancer (private)",
"context": 8000,
"max_output": 2000,
"input": 4.5,
"output": 6,
"latency": 1.11,
"throughput": 12.8
}
] | 4 | 6 | 3.375 | 4.5 | 4.5 | 6 | true |
Meta: LlamaGuard 2 8B | meta-llama/LlamaGuard-7b | meta-llama | [
{
"name": "Together",
"context": 8000,
"max_output": 8000,
"input": 0.2,
"output": 0.2,
"latency": 0.88,
"throughput": 76.09
}
] | 0.2 | 0.2 | 0.2 | 0.2 | 0.2 | 0.2 | true |
NeverSleep: Llama 3 Lumimaid 8B (extended) | NeverSleep/Llama-3-Lumimaid-8B-v0.1 | NeverSleep | [
{
"name": "Mancer",
"context": 25000,
"max_output": 2000,
"input": 0.1875,
"output": 1.125,
"latency": 0.74,
"throughput": 51.57
},
{
"name": "Mancer (private)",
"context": 25000,
"max_output": 2000,
"input": 0.25,
"output": 1.5,
"latency": 0.52,
"throughput": 51.92
}
] | 0.25 | 1.5 | 0.1875 | 1.125 | 0.25 | 1.5 | true |
NeverSleep: Llama 3 Lumimaid 8B | NeverSleep/Llama-3-Lumimaid-8B-v0.1 | NeverSleep | [
{
"name": "Mancer",
"context": 25000,
"max_output": 2000,
"input": 0.1875,
"output": 1.125,
"latency": 0.74,
"throughput": 51.57
},
{
"name": "Mancer (private)",
"context": 25000,
"max_output": 2000,
"input": 0.25,
"output": 1.5,
"latency": 0.52,
"throughput": 51.92
},
{
"name": "Featherless",
"context": 8000,
"max_output": 4000,
"input": 0.8,
"output": 1.2,
"latency": 1.24,
"throughput": 25.1
}
] | 0.25 | 1.5 | 0.1875 | 1.125 | 0.8 | 1.2 | true |
Fimbulvetr 11B v2 | Sao10K/Fimbulvetr-11B-v2 | Sao10K | [
{
"name": "Featherless",
"context": 4000,
"max_output": 4000,
"input": 0.8,
"output": 1.2,
"latency": 0.97,
"throughput": 27.78
}
] | 0.8 | 1.2 | 0.8 | 1.2 | 0.8 | 1.2 | true |
Meta: Llama 3 8B Instruct | meta-llama/Meta-Llama-3-8B-Instruct | meta-llama | [
{
"name": "DeepInfra",
"context": 8000,
"max_output": 8000,
"input": 0.03,
"output": 0.06,
"latency": 0.16,
"throughput": 113.2
},
{
"name": "Groq",
"context": 8000,
"max_output": 8000,
"input": 0.05,
"output": 0.08,
"latency": 0.31,
"throughput": 893.2
},
{
"name": "Cloudflare",
"context": 8000,
"max_output": 8000,
"input": 0.15,
"output": 0.15,
"latency": 0.76,
"throughput": 15.78
},
{
"name": "Mancer",
"context": 16000,
"max_output": 2000,
"input": 0.1875,
"output": 1.125,
"latency": 0.67,
"throughput": 50.42
},
{
"name": "Fireworks",
"context": 8000,
"max_output": 8000,
"input": 0.2,
"output": 0.2,
"latency": 0.51,
"throughput": 123.5
},
{
"name": "Mancer (private)",
"context": 16000,
"max_output": 2000,
"input": 0.25,
"output": 1.5,
"latency": 0.54,
"throughput": 50.82
},
{
"name": "NovitaAI",
"context": 8000,
"max_output": 8000,
"input": 0.04,
"output": 0.04,
"latency": null,
"throughput": null
}
] | 0.15 | 0.15 | 0.04 | 0.04 | 0.25 | 1.5 | true |
Meta: Llama 3 70B Instruct | meta-llama/Meta-Llama-3-70B-Instruct | meta-llama | [
{
"name": "DeepInfra",
"context": 8000,
"max_output": 8000,
"input": 0.23,
"output": 0.4,
"latency": 0.41,
"throughput": 25.65
},
{
"name": "Groq",
"context": 8000,
"max_output": 8000,
"input": 0.59,
"output": 0.79,
"latency": 0.19,
"throughput": 213.2
},
{
"name": "Together",
"context": 8000,
"max_output": 8000,
"input": 0.88,
"output": 0.88,
"latency": 0.44,
"throughput": 19.92
},
{
"name": "Fireworks",
"context": 8000,
"max_output": 8000,
"input": 0.9,
"output": 0.9,
"latency": 0.61,
"throughput": 150.3
},
{
"name": "NovitaAI",
"context": 8000,
"max_output": 8000,
"input": 0.51,
"output": 0.74,
"latency": 5.46,
"throughput": 19.8
}
] | 0.59 | 0.79 | 0.23 | 0.4 | 0.9 | 0.9 | true |
Mistral: Mixtral 8x22B Instruct | mistralai/Mixtral-8x22B-Instruct-v0.1 | mistralai | [
{
"name": "Fireworks",
"context": 66000,
"max_output": 66000,
"input": 0.9,
"output": 0.9,
"latency": 0.67,
"throughput": 88.81
},
{
"name": "Together",
"context": 66000,
"max_output": 2000,
"input": 1.2,
"output": 1.2,
"latency": 0.74,
"throughput": 93.93
},
{
"name": "Mistral",
"context": 66000,
"max_output": 66000,
"input": 2,
"output": 6,
"latency": 0.25,
"throughput": 86.28
}
] | 1.2 | 1.2 | 0.9 | 0.9 | 2 | 6 | true |
WizardLM-2 8x22B | microsoft/WizardLM-2-8x22B | microsoft | [
{
"name": "DeepInfra",
"context": 66000,
"max_output": 8000,
"input": 0.5,
"output": 0.5,
"latency": 0.1,
"throughput": 31.3
},
{
"name": "Lepton",
"context": 66000,
"max_output": 66000,
"input": 1,
"output": 1,
"latency": 0.36,
"throughput": 28.57
},
{
"name": "NovitaAI",
"context": 66000,
"max_output": 66000,
"input": 0.62,
"output": 0.62,
"latency": 1.06,
"throughput": 19.54
},
{
"name": "Together",
"context": 66000,
"max_output": 66000,
"input": 1.2,
"output": 1.2,
"latency": 0.57,
"throughput": 69.67
}
] | 1 | 1 | 0.5 | 0.5 | 1.2 | 1.2 | true |
WizardLM-2 7B | microsoft/WizardLM-2-7B | microsoft | [
{
"name": "Lepton",
"context": 32000,
"max_output": 32000,
"input": 0.07,
"output": 0.07,
"latency": 0.37,
"throughput": 102.9
}
] | 0.07 | 0.07 | 0.07 | 0.07 | 0.07 | 0.07 | true |
Databricks: DBRX 132B Instruct | databricks/dbrx-instruct | databricks | [
{
"name": "Together",
"context": 33000,
"max_output": 2000,
"input": 1.2,
"output": 1.2,
"latency": 0.33,
"throughput": 83.02
}
] | 1.2 | 1.2 | 1.2 | 1.2 | 1.2 | 1.2 | true |
Midnight Rose 70B | sophosympatheia/Wizard-Tulu-Dolphin-70B-v1.0 | sophosympatheia | [
{
"name": "NovitaAI",
"context": 4000,
"max_output": 4000,
"input": 0.8,
"output": 0.8,
"latency": 2.21,
"throughput": 15.59
}
] | 0.8 | 0.8 | 0.8 | 0.8 | 0.8 | 0.8 | true |
Google: Gemma 7B | google/gemma-1.1-7b-it | google | [
{
"name": "Cloudflare",
"context": 8000,
"max_output": 8000,
"input": 0.15,
"output": 0.15,
"latency": 1.22,
"throughput": 13.04
}
] | 0.15 | 0.15 | 0.15 | 0.15 | 0.15 | 0.15 | true |
Nous: Hermes 2 Mixtral 8x7B DPO | NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO | NousResearch | [
{
"name": "Together",
"context": 33000,
"max_output": 2000,
"input": 0.6,
"output": 0.6,
"latency": 0.4,
"throughput": 109
}
] | 0.6 | 0.6 | 0.6 | 0.6 | 0.6 | 0.6 | true |
Dolphin 2.6 Mixtral 8x7B 🐬 | cognitivecomputations/dolphin-2.6-mixtral-8x7b | cognitivecomputations | [
{
"name": "Lepton",
"context": 33000,
"max_output": 33000,
"input": 0.5,
"output": 0.5,
"latency": 0.11,
"throughput": 84.57
}
] | 0.5 | 0.5 | 0.5 | 0.5 | 0.5 | 0.5 | true |
Mistral: Mixtral 8x7B (base) | mistralai/Mixtral-8x7B-v0.1 | mistralai | [
{
"name": "Together",
"context": 33000,
"max_output": 2000,
"input": 0.6,
"output": 0.6,
"latency": 0.46,
"throughput": 118.1
}
] | 0.6 | 0.6 | 0.6 | 0.6 | 0.6 | 0.6 | true |
Mistral: Mixtral 8x7B Instruct | mistralai/Mixtral-8x7B-Instruct-v0.1 | mistralai | [
{
"name": "DeepInfra",
"context": 33000,
"max_output": 8000,
"input": 0.24,
"output": 0.24,
"latency": 0.2,
"throughput": 115.1
},
{
"name": "Groq",
"context": 33000,
"max_output": 33000,
"input": 0.24,
"output": 0.24,
"latency": 0.37,
"throughput": 678.1
},
{
"name": "Fireworks",
"context": 33000,
"max_output": 33000,
"input": 0.5,
"output": 0.5,
"latency": 0.75,
"throughput": 191.9
},
{
"name": "Lepton",
"context": 33000,
"max_output": 33000,
"input": 0.5,
"output": 0.5,
"latency": 0.36,
"throughput": 75.27
},
{
"name": "Together",
"context": 33000,
"max_output": 2000,
"input": 0.6,
"output": 0.6,
"latency": 0.49,
"throughput": 90.88
}
] | 0.5 | 0.5 | 0.24 | 0.24 | 0.6 | 0.6 | true |
OpenChat 3.5 7B | openchat/openchat-3.5-0106 | openchat | [
{
"name": "DeepInfra",
"context": 8000,
"max_output": 8000,
"input": 0.055,
"output": 0.055,
"latency": 0.22,
"throughput": 99.19
},
{
"name": "Lepton",
"context": 8000,
"max_output": 8000,
"input": 0.07,
"output": 0.07,
"latency": 0.4,
"throughput": 107.9
},
{
"name": "NovitaAI",
"context": 4000,
"max_output": 4000,
"input": 0.06,
"output": 0.06,
"latency": 0.8,
"throughput": 58.78
}
] | 0.06 | 0.06 | 0.055 | 0.055 | 0.07 | 0.07 | true |
Noromaid 20B | NeverSleep/Noromaid-20b-v0.1.1 | NeverSleep | [
{
"name": "Mancer",
"context": 8000,
"max_output": 2000,
"input": 1.5,
"output": 2.25,
"latency": 0.67,
"throughput": 23.41
},
{
"name": "Mancer (private)",
"context": 8000,
"max_output": 2000,
"input": 2,
"output": 3,
"latency": 1.1,
"throughput": 23.6
}
] | 2 | 3 | 1.5 | 2.25 | 2 | 3 | true |
OpenHermes 2.5 Mistral 7B | teknium/OpenHermes-2.5-Mistral-7B | teknium | [
{
"name": "NovitaAI",
"context": 4000,
"max_output": 4000,
"input": 0.17,
"output": 0.17,
"latency": 0.67,
"throughput": 148
}
] | 0.17 | 0.17 | 0.17 | 0.17 | 0.17 | 0.17 | true |
Toppy M 7B | Undi95/Toppy-M-7B | Undi95 | [
{
"name": "Lepton",
"context": 4000,
"max_output": 4000,
"input": 0.07,
"output": 0.07,
"latency": 0.42,
"throughput": 108.2
}
] | 0.07 | 0.07 | 0.07 | 0.07 | 0.07 | 0.07 | true |
Goliath 120B | alpindale/goliath-120b | alpindale | [
{
"name": "Mancer",
"context": 6000,
"max_output": 512,
"input": 9.375,
"output": 9.375,
"latency": 1.06,
"throughput": 18.62
},
{
"name": "Mancer (private)",
"context": 6000,
"max_output": 512,
"input": 12.5,
"output": 12.5,
"latency": 1.53,
"throughput": 15.02
}
] | 12.5 | 12.5 | 9.375 | 9.375 | 12.5 | 12.5 | true |
Airoboros 70B | jondurbin/airoboros-l2-70b-2.2.1 | jondurbin | [
{
"name": "NovitaAI",
"context": 4000,
"max_output": 4000,
"input": 0.5,
"output": 0.5,
"latency": 2.31,
"throughput": 56.22
}
] | 0.5 | 0.5 | 0.5 | 0.5 | 0.5 | 0.5 | true |
Xwin 70B | Xwin-LM/Xwin-LM-70B-V0.1 | Xwin-LM | [
{
"name": "Mancer",
"context": 8000,
"max_output": 512,
"input": 3.75,
"output": 3.75,
"latency": 2.17,
"throughput": 18.04
},
{
"name": "Mancer (private)",
"context": 8000,
"max_output": 512,
"input": 5,
"output": 5,
"latency": 2.05,
"throughput": 18.1
}
] | 5 | 5 | 3.75 | 3.75 | 5 | 5 | true |
Mistral: Mistral 7B Instruct v0.1 | mistralai/Mistral-7B-Instruct-v0.1 | mistralai | [
{
"name": "Together",
"context": 33000,
"max_output": 2000,
"input": 0.2,
"output": 0.2,
"latency": 0.3,
"throughput": 169.1
}
] | 0.2 | 0.2 | 0.2 | 0.2 | 0.2 | 0.2 | true |
Pygmalion: Mythalion 13B | PygmalionAI/mythalion-13b | PygmalionAI | [
{
"name": "Featherless",
"context": 4000,
"max_output": 4000,
"input": 0.8,
"output": 1.2,
"latency": 1.42,
"throughput": 13.23
},
{
"name": "Mancer",
"context": 8000,
"max_output": 512,
"input": 1.125,
"output": 1.125,
"latency": 1.03,
"throughput": 22.71
},
{
"name": "Mancer (private)",
"context": 8000,
"max_output": 512,
"input": 1.5,
"output": 1.5,
"latency": 1.04,
"throughput": 22.3
}
] | 1.125 | 1.125 | 0.8 | 1.2 | 1.5 | 1.5 | true |
Nous: Hermes 13B | NousResearch/Nous-Hermes-Llama2-13b | NousResearch | [
{
"name": "Lepton",
"context": 4000,
"max_output": 4000,
"input": 0.18,
"output": 0.18,
"latency": 0.31,
"throughput": 83.27
},
{
"name": "NovitaAI",
"context": 4000,
"max_output": 4000,
"input": 0.17,
"output": 0.17,
"latency": 0.73,
"throughput": 81.14
}
] | 0.18 | 0.18 | 0.17 | 0.17 | 0.18 | 0.18 | true |
ReMM SLERP 13B | Undi95/ReMM-SLERP-L2-13B | Undi95 | [
{
"name": "Featherless",
"context": 4000,
"max_output": 4000,
"input": 0.8,
"output": 1.2,
"latency": null,
"throughput": null
},
{
"name": "Mancer",
"context": 6000,
"max_output": 512,
"input": 1.125,
"output": 1.125,
"latency": 0.57,
"throughput": 43.17
},
{
"name": "Mancer (private)",
"context": 6000,
"max_output": 512,
"input": 1.5,
"output": 1.5,
"latency": 0.56,
"throughput": 43.44
}
] | 1.125 | 1.125 | 0.8 | 1.2 | 1.5 | 1.5 | true |
MythoMax 13B | Gryphe/MythoMax-L2-13b | Gryphe | [
{
"name": "DeepInfra",
"context": 4000,
"max_output": 4000,
"input": 0.065,
"output": 0.065,
"latency": 0.2,
"throughput": 106.4
},
{
"name": "Parasail",
"context": 4000,
"max_output": 4000,
"input": 0.11,
"output": 0.11,
"latency": 0.49,
"throughput": 91.35
},
{
"name": "Lepton",
"context": 4000,
"max_output": 4000,
"input": 0.18,
"output": 0.18,
"latency": 0.08,
"throughput": 105.5
},
{
"name": "Fireworks",
"context": 4000,
"max_output": 4000,
"input": 0.2,
"output": 0.2,
"latency": 1.92,
"throughput": 7.63
},
{
"name": "Together",
"context": 4000,
"max_output": 4000,
"input": 0.3,
"output": 0.3,
"latency": 0.43,
"throughput": 125.8
},
{
"name": "Mancer",
"context": 8000,
"max_output": 512,
"input": 1.125,
"output": 1.125,
"latency": 0.7,
"throughput": 40.18
},
{
"name": "Mancer (private)",
"context": 8000,
"max_output": 512,
"input": 1.5,
"output": 1.5,
"latency": 0.88,
"throughput": 38.69
},
{
"name": "NovitaAI",
"context": 4000,
"max_output": 4000,
"input": 0.09,
"output": 0.09,
"latency": 1.58,
"throughput": 88.68
}
] | 0.2 | 0.2 | 0.065 | 0.065 | 1.5 | 1.5 | true |
Meta: Llama 2 13B Chat | meta-llama/Llama-2-13b-chat-hf | meta-llama | [
{
"name": "Together",
"context": 4000,
"max_output": 2000,
"input": 0.22,
"output": 0.22,
"latency": 0.77,
"throughput": 63.08
},
{
"name": "Lepton",
"context": 4000,
"max_output": 4000,
"input": 0.3,
"output": 0.3,
"latency": 0.5,
"throughput": 98.07
}
] | 0.3 | 0.3 | 0.22 | 0.22 | 0.3 | 0.3 | true |
Meta: Llama 2 70B Chat | meta-llama/Llama-2-70b-chat-hf | meta-llama | [
{
"name": "Together",
"context": 4000,
"max_output": 4000,
"input": 0.9,
"output": 0.9,
"latency": 0.69,
"throughput": 45.46
}
] | 0.9 | 0.9 | 0.9 | 0.9 | 0.9 | 0.9 | true |
Subsets and Splits