tsavage68/Summary_L3_1000steps_1e7rate_SFT2
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kLicense:llama3Architecture:Transformer Cold
The tsavage68/Summary_L3_1000steps_1e7rate_SFT2 is an 8 billion parameter language model, fine-tuned from meta-llama/Meta-Llama-3-8B-Instruct. This model was trained over 1000 steps with a learning rate of 1e-07, achieving a final validation loss of 1.5908. While its specific intended uses and training dataset are not detailed, it represents a specialized iteration of the Llama 3 architecture.
Loading preview...
Popular Sampler Settings
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.
temperature
–
top_p
–
top_k
–
frequency_penalty
–
presence_penalty
–
repetition_penalty
–
min_p
–