Models

Warm

qwen2-7b-lc

Qwen/Qwen2-7B-Instruct

657

8,272,855

Q
Warm

llama31-8b-16k

meta-llama/Meta-Llama-3.1-8B-Instruct

4,340

8,177,177

M
Warm

llama31-8b-16k

meta-llama/Llama-3.1-8B-Instruct

4,340

8,177,177

M
Warm

qwen3-4b

Qwen/Qwen3-4B-Base

35

6,765,514

Q
Warm

qwen25-7b-lc

Qwen/Qwen2.5-7B-Instruct

728

5,805,193

Q
Warm

qwen3-0b6

Qwen/Qwen3-0.6B

472

3,784,181

Q
Warm

qwen3-8b

Qwen/Qwen3-8B-Base

41

3,762,600

Q
Warm

qwen25-0b5

Gensyn/Qwen2.5-0.5B-Instruct

17

2,922,364

G
Warm

qwen25-7b-lc

Qwen/Qwen2.5-7B-Instruct-1M

339

2,034,862

Q
Warm

gemma3-4b

google/gemma-3-4b-it

727

1,773,158

G
Warm

qwen3-1b7

Qwen/Qwen3-1.7B

197

1,744,116

Q
Warm

llama32-3b

meta-llama/Llama-3.2-3B-Instruct

1,608

1,695,709

M
Warm

mistral-v02-7b-std-lc

mistralai/Mistral-7B-Instruct-v0.2

2,871

1,685,666

M
Warm

qwen3-8b

Qwen/Qwen3-8B

472

1,654,455

Q
Warm

qwen25-1b5

Qwen/Qwen2.5-1.5B-Instruct

477

1,516,951

Q
Warm

llama3-8b-8k

meta-llama/Meta-Llama-3-8B-Instruct

4,084

1,460,472

M
Warm

llama32-3b

context-labs/meta-llama-Llama-3.2-3B-Instruct-FP16

0

1,153,461

C
Warm

llama31-8b-16k

deepseek-ai/DeepSeek-R1-Distill-Llama-8B

776

1,098,427

D
Warm

qwen25-7b-lc

Qwen/Qwen2.5-7B

203

1,097,541

Q
Warm

qwen25-1b5

deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B

1,274

975,233

D