thu-ml/STAIR-Llama-3.1-8B-SFT
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Jan 17, 2025License:apache-2.0Architecture:Transformer Open Weights Cold

thu-ml/STAIR-Llama-3.1-8B-SFT is an 8 billion parameter instruction-tuned causal language model developed by thu-ml, fine-tuned from meta-llama/Llama-3.1-8B-Instruct. It is specifically trained on the STAIR-SFT dataset, which consists of 20k prompts from UltraFeedback and PKU-SafeRLHF, to align reasoning formats and facilitate self-improvement. This model excels at generating step-level Chain-of-Thought (CoT) responses, making it suitable for tasks requiring structured reasoning and ethical response generation.

Loading preview...