Models

Warm

llama31-8b-16k

meta-llama/Llama-3.1-8B-Instruct

4,723

6,995,384

M
Warm

llama31-8b-16k

meta-llama/Meta-Llama-3.1-8B-Instruct

4,722

6,995,384

M
Warm

qwen3-0b6

Qwen/Qwen3-0.6B

689

6,950,849

Q
Warm

qwen25-3b

Qwen/Qwen2.5-3B-Instruct

318

6,810,955

Q
Warm

llama32-1b

meta-llama/Llama-3.2-1B-Instruct

1,102

6,205,949

M
Warm

qwen3-32b

Qwen/Qwen3-32B

544

5,850,157

Q
Warm

qwen25-7b-lc

Qwen/Qwen2.5-7B-Instruct

815

5,168,028

Q
Warm

qwen25-0b5

Gensyn/Qwen2.5-0.5B-Instruct

22

4,515,469

G
Warm

llama32-3b

context-labs/meta-llama-Llama-3.2-3B-Instruct-FP16

7

3,160,075

C
Warm

qwen25-32b-lc

deepseek-ai/DeepSeek-R1-Distill-Qwen-32B

1,449

2,938,228

D
Warm

tinyllama-1b1

TinyLlama/TinyLlama-1.1B-Chat-v1.0

1,421

2,742,062

T
Warm

qwen25-7b-lc

Qwen/Qwen2.5-7B

233

2,117,981

Q
Warm

qwen3-8b

Qwen/Qwen3-8B

655

2,085,940

Q
Warm

qwen3-4b

Qwen/Qwen3-4B-Instruct-2507

374

2,027,011

Q
Warm

llama3-8b-8k

meta-llama/Meta-Llama-3-8B

6,340

1,944,234

M
Warm

llama32-1b

meta-llama/Llama-3.2-1B

2,105

1,933,473

M
Warm

llama32-3b

meta-llama/Llama-3.2-3B-Instruct

1,752

1,917,624

M
Warm

qwen25-0b5

Qwen/Qwen2.5-0.5B-Instruct

375

1,711,759

Q
Warm

qwen25-14b-lc

Qwen/Qwen2.5-14B-Instruct

276

1,637,468

Q
Warm

qwen3-4b

Qwen/Qwen3-4B

413

1,496,056

Q