princeton-nlp/Llama-3-Base-8B-SFT-SimPO

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:May 24, 2024Architecture:Transformer0.0K Warm

The princeton-nlp/Llama-3-Base-8B-SFT-SimPO is an 8 billion parameter language model developed by princeton-nlp. This model is a fine-tuned version of Llama-3-Base, utilizing Supervised Fine-Tuning (SFT) and Simplified Preference Optimization (SimPO). Its primary purpose is to serve as a foundational model for various natural language processing tasks, leveraging its base architecture and optimization methods.

Loading preview...

Model Overview

The princeton-nlp/Llama-3-Base-8B-SFT-SimPO is an 8 billion parameter language model developed by princeton-nlp. It is built upon the Llama-3-Base architecture and has undergone further optimization through Supervised Fine-Tuning (SFT) and Simplified Preference Optimization (SimPO). This model is designed to provide a robust foundation for a wide range of natural language processing applications.

Key Characteristics

  • Base Architecture: Utilizes the Llama-3-Base model as its foundation.
  • Parameter Count: Features 8 billion parameters, offering a balance between performance and computational efficiency.
  • Optimization Methods: Enhanced through Supervised Fine-Tuning (SFT) and Simplified Preference Optimization (SimPO), suggesting improvements in instruction following and alignment.

Potential Use Cases

Given its foundational nature and optimization, this model is suitable for:

  • General Text Generation: Creating coherent and contextually relevant text.
  • Instruction Following: Responding to prompts and instructions effectively due to SFT and SimPO.
  • Further Fine-tuning: Serving as a strong base model for domain-specific or task-specific fine-tuning.
  • Research and Development: Exploring the impact of SFT and SimPO on Llama-3-Base performance.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p