name
stringlengths
8
58
hf_id
stringlengths
12
49
author
stringlengths
3
21
providers
listlengths
1
16
median_input_cost
float64
0
75
median_output_cost
float64
0
150
low_input_cost
float64
0
75
low_output_cost
float64
0
150
high_input_cost
float64
0
75
high_output_cost
float64
0
150
is_open_weights
bool
2 classes
OpenAI: GPT-4 Turbo (older v1106)
null
OpenAI
[ { "name": "OpenAI", "context": 128000, "max_output": 4000, "input": 10, "output": 30, "latency": 0.84, "throughput": 39.99 } ]
10
30
10
30
10
30
false
Google: PaLM 2 Chat 32k
null
Google
[ { "name": "Google Vertex", "context": 33000, "max_output": 8000, "input": 1, "output": 2, "latency": 1.28, "throughput": 128.7 } ]
1
2
1
2
1
2
false
Google: PaLM 2 Code Chat 32k
null
Google
[ { "name": "Google Vertex", "context": 33000, "max_output": 8000, "input": 1, "output": 2, "latency": 3.21, "throughput": 84.47 } ]
1
2
1
2
1
2
false
OpenAI: GPT-3.5 Turbo Instruct
null
OpenAI
[ { "name": "OpenAI", "context": 4000, "max_output": 4000, "input": 1.5, "output": 2, "latency": 0.37, "throughput": 116.9 } ]
1.5
2
1.5
2
1.5
2
false
OpenAI: GPT-3.5 Turbo 16k
null
OpenAI
[ { "name": "OpenAI", "context": 16000, "max_output": 4000, "input": 3, "output": 4, "latency": 0.96, "throughput": 109 }, { "name": "Azure", "context": 16000, "max_output": 4000, "input": 3, "output": 4, "latency": 0.21, "throughput": 150 } ]
3
4
3
4
3
4
false
OpenAI: GPT-4 32k
null
OpenAI
[ { "name": "OpenAI", "context": 33000, "max_output": 4000, "input": 60, "output": 120, "latency": 1.24, "throughput": 30.89 }, { "name": "Azure", "context": 33000, "max_output": 4000, "input": 60, "output": 120, "latency": 1.79, "throughput": 30.63 } ]
60
120
60
120
60
120
false
OpenAI: GPT-4 32k (older v0314)
null
OpenAI
[ { "name": "OpenAI", "context": 33000, "max_output": 4000, "input": 60, "output": 120, "latency": 1.17, "throughput": 31.26 } ]
60
120
60
120
60
120
false
Mancer: Weaver (alpha)
null
Mancer
[ { "name": "Mancer", "context": 8000, "max_output": 1000, "input": 1.5, "output": 2.25, "latency": 0.95, "throughput": 39.77 }, { "name": "Mancer (private)", "context": 8000, "max_output": 1000, "input": 2, "output": 3, "latency": 0.99, "throughput": 40.2 } ]
2
3
1.5
2.25
2
3
false
Anthropic: Claude v2.0 (self-moderated)
null
Anthropic
[ { "name": "Anthropic", "context": 100000, "max_output": 4000, "input": 8, "output": 24, "latency": 2.55, "throughput": 29.62 } ]
8
24
8
24
8
24
false
Anthropic: Claude v2.0
null
Anthropic
[ { "name": "Anthropic", "context": 100000, "max_output": 4000, "input": 8, "output": 24, "latency": 2.55, "throughput": 29.62 } ]
8
24
8
24
8
24
false
Google: PaLM 2 Chat
null
Google
[ { "name": "Google Vertex", "context": 9000, "max_output": 1000, "input": 1, "output": 2, "latency": 1.43, "throughput": 169.4 } ]
1
2
1
2
1
2
false
Google: PaLM 2 Code Chat
null
Google
[ { "name": "Google Vertex", "context": 7000, "max_output": 1000, "input": 1, "output": 2, "latency": 1.16, "throughput": 141.6 } ]
1
2
1
2
1
2
false
OpenAI: GPT-3.5 Turbo
null
OpenAI
[ { "name": "OpenAI", "context": 16000, "max_output": 4000, "input": 0.5, "output": 1.5, "latency": 0.37, "throughput": 108.8 } ]
0.5
1.5
0.5
1.5
0.5
1.5
false
OpenAI: GPT-3.5 Turbo 16k
null
OpenAI
[ { "name": "OpenAI", "context": 16000, "max_output": 4000, "input": 0.5, "output": 1.5, "latency": 0.36, "throughput": 116.3 } ]
0.5
1.5
0.5
1.5
0.5
1.5
false
OpenAI: GPT-4
null
OpenAI
[ { "name": "OpenAI", "context": 8000, "max_output": 4000, "input": 30, "output": 60, "latency": 0.61, "throughput": 30.44 }, { "name": "Azure", "context": 8000, "max_output": 4000, "input": 30, "output": 60, "latency": 0.59, "throughput": 29.05 } ]
30
60
30
60
30
60
false
OpenAI: GPT-4 (older v0314)
null
OpenAI
[ { "name": "OpenAI", "context": 8000, "max_output": 4000, "input": 30, "output": 60, "latency": 0.82, "throughput": 37.99 } ]
30
60
30
60
30
60
false
xAI: Grok Beta
null
xAI
[ { "name": "xAI", "context": 131000, "max_output": 131000, "input": 5, "output": 15, "latency": 0.26, "throughput": 59.3 } ]
5
15
5
15
5
15
false
Perplexity: Llama 3.1 Sonar 70B
null
perplexity-ai
[ { "name": "Perplexity", "context": 131000, "max_output": 131000, "input": 1, "output": 1, "latency": 0.42, "throughput": 49.69 } ]
1
1
1
1
1
1
false