tsavage68/Summary_L3_1000steps_1e7rate_SFT2
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kLicense:llama3Architecture:Transformer Cold

The tsavage68/Summary_L3_1000steps_1e7rate_SFT2 is an 8 billion parameter language model, fine-tuned from meta-llama/Meta-Llama-3-8B-Instruct. This model was trained over 1000 steps with a learning rate of 1e-07, achieving a final validation loss of 1.5908. While its specific intended uses and training dataset are not detailed, it represents a specialized iteration of the Llama 3 architecture.

Loading preview...

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p