Overview
PJMixers-Archive/LLaMa-3-CursedStock-v2.0-8B is an 8 billion parameter merged language model built upon the Llama 3 architecture. It was created by PJMixers-Archive by combining over 40 distinct Llama 3-8B instruction-tuned models using the model_stock merge method. This extensive merge aims to consolidate a wide array of specialized capabilities into a single, versatile model.
Key Capabilities
- Roleplay & Creative Writing: Integrates models specifically fine-tuned for roleplay, story generation, and fantasy writing, including those with LORA adaptations for long stories.
- General Intelligence: Incorporates several "smart" models to enhance overall reasoning and instruction-following abilities.
- Uncensored Content Generation: Includes models designed for less restricted content generation.
- Specialized Domains: Features models for specific applications such as code generation (e.g., SQL coding) and medical text processing.
Good For
- Developers seeking a single Llama 3-8B model with a broad spectrum of capabilities, reducing the need to switch between specialized models.
- Applications requiring strong performance in creative writing, roleplay scenarios, and general conversational tasks.
- Use cases that benefit from a model trained on diverse datasets, including those with less restrictive content policies.
Limitations
- The README does not specify a particular context length for the merged model, though some constituent models mention 8K context. Users should refer to the original models for best usage guidelines.