kaist-ai/janus-7b

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Apr 4, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Warm

Janus-7B is a 7 billion parameter language model developed by KAIST AI, built upon the Mistral-7B-v0.2 base architecture. It is fine-tuned on the Multifaceted Collection, a preference dataset with 196k unique system messages. This model excels at generating personalized responses aligned with diverse human preferences, making it adept at producing helpful and harmless outputs controlled by specific system messages.

Loading preview...

Janus-7B: System Message Generalization for Diverse Preferences

Janus-7B, developed by KAIST AI, is a 7 billion parameter language model based on Mistral-7B-v0.2. Its core innovation lies in its training on the Multifaceted Collection, a unique preference dataset comprising 196,000 distinct system messages. This supervised fine-tuning approach enables Janus-7B to generalize across a wide array of human preferences.

Key Capabilities

  • Personalized Response Generation: Excels at producing outputs tailored to specific user preferences defined via system messages.
  • Helpful and Harmless Alignment: Designed to generate responses that are generally preferred for their helpfulness and safety.
  • System Message Control: Users can precisely control the model's behavior and output style by inputting desired system messages, allowing for highly customized interactions.

Good For

  • Applications requiring adaptable AI personalities: Ideal for scenarios where the AI needs to adopt various personas or adhere to specific conversational styles.
  • Developing preference-aligned chatbots: Useful for creating agents that can generate responses consistent with diverse user expectations and guidelines.
  • Research into LLM alignment and generalization: Provides a strong base for exploring how models can be fine-tuned to understand and apply a broad spectrum of human preferences.