tsavage68/chat_200STEPS_1e6_01beta
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Feb 14, 2024Architecture:Transformer Cold

The tsavage68/chat_200STEPS_1e6_01beta is a 7 billion parameter language model, fine-tuned from meta-llama/Llama-2-7b-chat-hf. This model was trained for 200 steps with a learning rate of 1e-06, achieving a validation loss of 0.6840. It is intended for chat-based applications, building upon the Llama 2 architecture.

Loading preview...