Aryanne/Open-StarLake-Swap-7B

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Mar 18, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

Aryanne/Open-StarLake-Swap-7B is a 7 billion parameter language model created by Aryanne, built upon the senseable/WestLake-7B-v2 base model. It was developed using a task_swapping merge method, combining the strengths of berkeley-nest/Starling-LM-7B-alpha, NousResearch/Nous-Hermes-2-Mistral-7B-DPO, and openchat/openchat-3.5-0106. This model is designed for conversational role-play, with a specified prompt format for generating verbose and descriptive interactions.

Loading preview...

Aryanne/Open-StarLake-Swap-7B Overview

Aryanne/Open-StarLake-Swap-7B is a 7 billion parameter language model developed by Aryanne, utilizing a unique task_swapping merge method via a custom branch of mergekit. It is built on the senseable/WestLake-7B-v2 as its foundational base model.

Key Capabilities & Merge Details

This model integrates the strengths of three distinct pre-trained language models:

  • berkeley-nest/Starling-LM-7B-alpha
  • NousResearch/Nous-Hermes-2-Mistral-7B-DPO
  • openchat/openchat-3.5-0106

The merge configuration involved specific layer ranges and weighting parameters for each contributing model, aiming to combine their respective proficiencies. The model has a context length of 4096 tokens.

Intended Use Case

Open-StarLake-Swap-7B is specifically optimized for verbose and descriptive role-play conversations. The developer provides a recommended prompt format, including start and end sequences, to facilitate engaging and detailed interactive storytelling, particularly within environments like Koboldcpp. This focus on rich, character-driven dialogue distinguishes it from more general-purpose instruction-tuned models.