fhai50032/RolePlayLake-7B
fhai50032/RolePlayLake-7B is a 7 billion parameter language model created by fhai50032, formed by merging SanjiWatsuki/Silicon-Maid-7B and senseable/WestLake-7B-v2. This model is specifically optimized for role-playing scenarios and chat applications, aiming for uncensored responses. It achieves an average score of 72.54 on the Open LLM Leaderboard, demonstrating strong performance across various reasoning and language understanding tasks.
Loading preview...
RolePlayLake-7B: A Merged Model for Enhanced Role-Playing
RolePlayLake-7B is a 7 billion parameter language model developed by fhai50032, created through a strategic merge of two distinct models: SanjiWatsuki/Silicon-Maid-7B and senseable/WestLake-7B-v2. The primary goal of this merge was to combine the strengths of both base models to produce a model excelling in role-playing and chat interactions, with a focus on providing uncensored responses.
Key Capabilities & Characteristics
- Optimized for Role-Playing: The model is specifically designed to enhance role-play capabilities, drawing from Silicon-Maid's charm and WestLake's role-play prowess.
- Uncensored Responses: A stated objective of the merge was to create a model that is more uncensored than its constituents, particularly WestLake.
- Strong Reasoning & Language Understanding: Evaluated on the Open LLM Leaderboard, RolePlayLake-7B achieved an average score of 72.54. Notable scores include 70.56 on AI2 Reasoning Challenge, 87.42 on HellaSwag, and 64.55 on MMLU.
- Configuration Synergy: The merge configuration supports various prompt formats, leveraging the combined strengths of the base models.
Ideal Use Cases
RolePlayLake-7B is particularly well-suited for applications requiring:
- Interactive Role-Playing: Its fine-tuning makes it effective for engaging in detailed and dynamic role-play scenarios.
- Uncensored Chat Applications: Developers seeking a model with fewer content restrictions for chat-based interactions may find this model beneficial.
- General Language Tasks: Its solid performance across various benchmarks indicates its utility for a range of general language understanding and generation tasks.