UCLA-AGI/zephyr-7b-sft-full-SPIN-iter0
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:8kPublished:Jan 4, 2024License:mitArchitecture:Transformer0.0K Open Weights Cold
UCLA-AGI/zephyr-7b-sft-full-SPIN-iter0 is a 7 billion parameter GPT-like language model developed by UCLA-AGI, fine-tuned using a self-play fine-tuning (SPIN) approach. This model is based on alignment-handbook/zephyr-7b-sft-full, which in turn is derived from Mistral-7B-v0.1, and is primarily English-language focused. It leverages synthetic data from the HuggingFaceH4/ultrachat_200k dataset to enhance its capabilities through iterative self-improvement. The model demonstrates competitive performance on various benchmarks, including ARC, HellaSwag, and MMLU, making it suitable for general language understanding and generation tasks.
Loading preview...