Models

3,525
1B32Kllama32-1b
Warm

renwei87/Llama-3.2-1B-Instruct

0
·
3
1B32Kllama32-1b
Warm

emst/Sibatikgal

0
·
3
1B32Kllama32-1b
Warm

meeksfr/Ultrachat200k-SFT-llama3.2-1B

0
·
3
1B32Kllama32-1b
Warm

Dc-4nderson/EverFlora-Llama-3.2-1B-Finetuned

0
·
3
1B32Kllama32-1b
Warm

Dc-4nderson/EverFlora-Llama-3.2-1B-Finetuned3

0
·
3
1B32Kllama32-1b
Warm

Muadil/Llama-3.2-1B-Instruct_sum_DPO_40k_1_2ep

0
·
3
1B32Kllama32-1b
Warm

Muadil/Llama-3.2-1B-Instruct_sum_DPO_1k_1_3ep

0
·
3
1B32Kllama32-1b
Warm

Plan-9/Llama3.2-docker-training

0
·
3
1B32Kllama32-1b
Warm

Muadil/Llama-3.2-1B-Instruct_sum_PPO_Skywork_10k_1_2ep_4bit

0
·
3
1B32Kllama32-1b
Warm

anish12/llama-1681_A

0
·
3
1B32Kllama32-1b
Warm

jiinking/1_layer_GQA4_llama_model

0
·
3
1B32Kllama32-1b
Warm

dmohanayogesh9/interviewer_model9

0
·
3
1B32Kllama32-1b
Warm

Muadil/Llama-3.2-1B-Instruct_sum_KTO_10k_1_1ep

0
·
3
1B32Kllama32-1b
Warm

jiinking/4_layer_GQA4_llama_model

0
·
3
1B32Kllama32-1b
Warm

Trelis/Llama-3.2-1B-Instruct-RL-gsm8k-step1

0
·
3
1B32Kllama32-1b
Warm

derickio/llama-3.2-1b-instruct-finetune_png_10k_cot_1k

0
·
3
1B32Kllama32-1b
Warm

Mattia2700/Llama-3.2-1B_AllDataSources_it.layer1_NoQuant_64_16_0.01_16CLINICALe3c-sentences_tag

0
·
3
1B32Kllama32-1b
Warm

Muadil/Llama-3.2-1B-Instruct_sum_KTO_40k_4_3ep

0
·
3
1B32Kllama32-1b
Warm

mengqizou011438/merged-llama3.2-1B-financial

0
·
3
1B32Kllama32-1b
Warm

Mattia2700/Llama-3.2-1B_ClinicalWhole_it.layer1_NoQuant_64_32_0.01_16CLINICALe3c-sentences_tag

0
·
3