name
stringlengths 8
58
| hf_id
stringlengths 12
49
⌀ | author
stringlengths 3
21
| providers
listlengths 1
16
| median_input_cost
float64 0
75
| median_output_cost
float64 0
150
| low_input_cost
float64 0
75
| low_output_cost
float64 0
150
| high_input_cost
float64 0
75
| high_output_cost
float64 0
150
| is_open_weights
bool 2
classes |
---|---|---|---|---|---|---|---|---|---|---|
OpenAI: GPT-4 Turbo (older v1106) | null | OpenAI | [
{
"name": "OpenAI",
"context": 128000,
"max_output": 4000,
"input": 10,
"output": 30,
"latency": 0.84,
"throughput": 39.99
}
] | 10 | 30 | 10 | 30 | 10 | 30 | false |
Google: PaLM 2 Chat 32k | null | Google | [
{
"name": "Google Vertex",
"context": 33000,
"max_output": 8000,
"input": 1,
"output": 2,
"latency": 1.28,
"throughput": 128.7
}
] | 1 | 2 | 1 | 2 | 1 | 2 | false |
Google: PaLM 2 Code Chat 32k | null | Google | [
{
"name": "Google Vertex",
"context": 33000,
"max_output": 8000,
"input": 1,
"output": 2,
"latency": 3.21,
"throughput": 84.47
}
] | 1 | 2 | 1 | 2 | 1 | 2 | false |
OpenAI: GPT-3.5 Turbo Instruct | null | OpenAI | [
{
"name": "OpenAI",
"context": 4000,
"max_output": 4000,
"input": 1.5,
"output": 2,
"latency": 0.37,
"throughput": 116.9
}
] | 1.5 | 2 | 1.5 | 2 | 1.5 | 2 | false |
OpenAI: GPT-3.5 Turbo 16k | null | OpenAI | [
{
"name": "OpenAI",
"context": 16000,
"max_output": 4000,
"input": 3,
"output": 4,
"latency": 0.96,
"throughput": 109
},
{
"name": "Azure",
"context": 16000,
"max_output": 4000,
"input": 3,
"output": 4,
"latency": 0.21,
"throughput": 150
}
] | 3 | 4 | 3 | 4 | 3 | 4 | false |
OpenAI: GPT-4 32k | null | OpenAI | [
{
"name": "OpenAI",
"context": 33000,
"max_output": 4000,
"input": 60,
"output": 120,
"latency": 1.24,
"throughput": 30.89
},
{
"name": "Azure",
"context": 33000,
"max_output": 4000,
"input": 60,
"output": 120,
"latency": 1.79,
"throughput": 30.63
}
] | 60 | 120 | 60 | 120 | 60 | 120 | false |
OpenAI: GPT-4 32k (older v0314) | null | OpenAI | [
{
"name": "OpenAI",
"context": 33000,
"max_output": 4000,
"input": 60,
"output": 120,
"latency": 1.17,
"throughput": 31.26
}
] | 60 | 120 | 60 | 120 | 60 | 120 | false |
Mancer: Weaver (alpha) | null | Mancer | [
{
"name": "Mancer",
"context": 8000,
"max_output": 1000,
"input": 1.5,
"output": 2.25,
"latency": 0.95,
"throughput": 39.77
},
{
"name": "Mancer (private)",
"context": 8000,
"max_output": 1000,
"input": 2,
"output": 3,
"latency": 0.99,
"throughput": 40.2
}
] | 2 | 3 | 1.5 | 2.25 | 2 | 3 | false |
Anthropic: Claude v2.0 (self-moderated) | null | Anthropic | [
{
"name": "Anthropic",
"context": 100000,
"max_output": 4000,
"input": 8,
"output": 24,
"latency": 2.55,
"throughput": 29.62
}
] | 8 | 24 | 8 | 24 | 8 | 24 | false |
Anthropic: Claude v2.0 | null | Anthropic | [
{
"name": "Anthropic",
"context": 100000,
"max_output": 4000,
"input": 8,
"output": 24,
"latency": 2.55,
"throughput": 29.62
}
] | 8 | 24 | 8 | 24 | 8 | 24 | false |
Google: PaLM 2 Chat | null | Google | [
{
"name": "Google Vertex",
"context": 9000,
"max_output": 1000,
"input": 1,
"output": 2,
"latency": 1.43,
"throughput": 169.4
}
] | 1 | 2 | 1 | 2 | 1 | 2 | false |
Google: PaLM 2 Code Chat | null | Google | [
{
"name": "Google Vertex",
"context": 7000,
"max_output": 1000,
"input": 1,
"output": 2,
"latency": 1.16,
"throughput": 141.6
}
] | 1 | 2 | 1 | 2 | 1 | 2 | false |
OpenAI: GPT-3.5 Turbo | null | OpenAI | [
{
"name": "OpenAI",
"context": 16000,
"max_output": 4000,
"input": 0.5,
"output": 1.5,
"latency": 0.37,
"throughput": 108.8
}
] | 0.5 | 1.5 | 0.5 | 1.5 | 0.5 | 1.5 | false |
OpenAI: GPT-3.5 Turbo 16k | null | OpenAI | [
{
"name": "OpenAI",
"context": 16000,
"max_output": 4000,
"input": 0.5,
"output": 1.5,
"latency": 0.36,
"throughput": 116.3
}
] | 0.5 | 1.5 | 0.5 | 1.5 | 0.5 | 1.5 | false |
OpenAI: GPT-4 | null | OpenAI | [
{
"name": "OpenAI",
"context": 8000,
"max_output": 4000,
"input": 30,
"output": 60,
"latency": 0.61,
"throughput": 30.44
},
{
"name": "Azure",
"context": 8000,
"max_output": 4000,
"input": 30,
"output": 60,
"latency": 0.59,
"throughput": 29.05
}
] | 30 | 60 | 30 | 60 | 30 | 60 | false |
OpenAI: GPT-4 (older v0314) | null | OpenAI | [
{
"name": "OpenAI",
"context": 8000,
"max_output": 4000,
"input": 30,
"output": 60,
"latency": 0.82,
"throughput": 37.99
}
] | 30 | 60 | 30 | 60 | 30 | 60 | false |
xAI: Grok Beta | null | xAI | [
{
"name": "xAI",
"context": 131000,
"max_output": 131000,
"input": 5,
"output": 15,
"latency": 0.26,
"throughput": 59.3
}
] | 5 | 15 | 5 | 15 | 5 | 15 | false |
Perplexity: Llama 3.1 Sonar 70B | null | perplexity-ai | [
{
"name": "Perplexity",
"context": 131000,
"max_output": 131000,
"input": 1,
"output": 1,
"latency": 0.42,
"throughput": 49.69
}
] | 1 | 1 | 1 | 1 | 1 | 1 | false |
Subsets and Splits