ContextualAI/archangel_sft-kto_llama13b

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:13BQuant:FP8Ctx Length:4kPublished:Dec 3, 2023License:apache-2.0Architecture:Transformer0.0K Open Weights Warm

The ContextualAI/archangel_sft-kto_llama13b is a 13 billion parameter language model from the Llama family, developed by Contextual AI. It is optimized using a combination of Supervised Fine-Tuning (SFT) and KTO (Kahneman-Tversky Optimization) loss functions, and aligned with SHP, Anthropic HH, and Open Assistant datasets. This model is designed for general language generation tasks, particularly excelling in conversational contexts due to its alignment process.

Loading preview...

Archangel SFT+KTO Llama 13B Overview

ContextualAI's archangel_sft-kto_llama13b is a 13 billion parameter model built on the Llama architecture. It distinguishes itself through its unique optimization strategy, combining Supervised Fine-Tuning (SFT) with Kahneman-Tversky Optimization (KTO) loss functions. This approach aims to enhance model alignment and performance.

Key Capabilities

  • Advanced Alignment: The model has been aligned using a diverse set of human preference datasets, including SHP, Anthropic HH, and Open Assistant, contributing to more human-like and helpful responses.
  • TuluV2 Prompt Format: It is designed to be prompted using the TuluV2 format, which clearly delineates user and assistant turns, facilitating structured conversations.
  • Conditional SFT Control Tokens: Models trained with conditional SFT include special tokens like <|good|> and <|bad|> in their embeddings, allowing for fine-grained control over generation by appending these to the prompt.

Good For

  • Conversational AI: Its alignment on human preference datasets makes it well-suited for dialogue systems and interactive applications.
  • Research into Alignment Techniques: The model serves as an example of applying SFT+KTO loss for alignment, offering insights for researchers in the field of LLM optimization. Further details on the methodology are available in the code repository and technical paper.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p