wandb/mistral-7b-zephyr-sft
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Mar 9, 2024License:mitArchitecture:Transformer0.0K Open Weights Cold
wandb/mistral-7b-zephyr-sft is a 7.2 billion parameter GPT-like model, fine-tuned from Mistral-7B-v0.1 using the Zephyr SFT recipe. Primarily English, it is optimized for instruction-following tasks based on a mix of publicly available and synthetic datasets. This model is suitable for general-purpose conversational AI applications requiring a compact yet capable language model.
Loading preview...