tsavage68/chat_600STEPS_1e8rate_SFT
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Feb 13, 2024Architecture:Transformer Cold

The tsavage68/chat_600STEPS_1e8rate_SFT model is a 7 billion parameter language model fine-tuned from Meta's Llama-2-7b-chat-hf. It was trained for 600 steps with a learning rate of 1e-08, achieving a final validation loss of 1.6169. This model is a foundational fine-tune, with specific optimizations or primary use cases not detailed in its current documentation.

Loading preview...