nlee-208/uf-mistral-it-sft-g0
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Jul 19, 2024License:llama3Architecture:Transformer Cold

The nlee-208/uf-mistral-it-sft-g0 model is an 8 billion parameter language model fine-tuned from Meta-Llama-3-8B-Instruct. It was trained on the nlee-208/uf-g0-sft dataset for one epoch with a learning rate of 5e-05 and a context length of 8192 tokens. This model is intended for general instruction-following tasks, leveraging its Llama-3 base architecture.

Loading preview...