allenai/open-instruct-llama2-sharegpt-dpo-7b
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Nov 12, 2023Architecture:Transformer Cold

The allenai/open-instruct-llama2-sharegpt-dpo-7b model is a 7 billion parameter language model from the Tulu series, developed by AllenAI. It is a Llama 2 variant initially fine-tuned on the ShareGPT dataset and further optimized using Direct Preference Optimization (DPO) on the UltraFeedback dataset. This model is designed to function as a helpful assistant, primarily in English, and is particularly strong in generating conversational responses due to its DPO training.

Loading preview...