Models

Warm

llama3-8b-8k

meta-llama/Meta-Llama-3-8B

6,195

5,069,672

M
Warm

llama31-8b-16k

meta-llama/Meta-Llama-3.1-8B-Instruct

3,980

4,851,980

M
Warm

qwen25-7b-lc

Qwen/Qwen2.5-7B-Instruct

688

2,594,131

Q
Warm

mistral-v02-7b-std-lc

mistralai/Mistral-7B-Instruct-v0.2

2,793

1,842,389

M
Warm

qwen25-7b-lc

Qwen/Qwen2.5-7B-Instruct-1M

328

1,759,738

Q
Warm

llama31-8b-16k

deepseek-ai/DeepSeek-R1-Distill-Llama-8B

739

1,353,996

D
Warm

llama31-8b-16k

meta-llama/Meta-Llama-3.1-8B

1,618

1,138,449

M
Warm

qwen3-8b

Qwen/Qwen3-8B

365

1,136,318

Q
Warm

llama3-8b-8k

meta-llama/Meta-Llama-3-8B-Instruct

3,989

1,097,627

M
Warm

gemma3-27b

google/gemma-3-27b-it

1,099

964,080

G
Warm

llama31-70b-16k

meta-llama/Meta-Llama-3.1-70B-Instruct

810

945,390

M
Warm

qwen3-32b

Qwen/Qwen3-32B

375

898,928

Q
Warm

mistral-24b-lc

sarvamai/sarvam-m

223

784,231

S
Warm

llama33-70b-16k

meta-llama/Llama-3.3-70B-Instruct

2,351

767,307

M
Warm

qwen3-14b

Qwen/Qwen3-14B

174

730,534

Q
Warm

mistral-v02-7b-std-lc

HuggingFaceH4/zephyr-7b-beta

1,718

594,668

H
Warm

qwen25-7b-lc

deepseek-ai/DeepSeek-R1-Distill-Qwen-7B

653

572,177

D
Warm

qwen25-32b-lc

Qwen/Qwen2.5-32B-Instruct

276

557,442

Q
Warm

qwen25-14b-lc

deepseek-ai/DeepSeek-R1-Distill-Qwen-14B

514

551,738

D
Warm

qwen25-14b-lc

Qwen/Qwen2.5-14B-Instruct

234

549,254

Q