PJMixers-Archive/Alpacino-SuperCOT-13B

TEXT GENERATIONConcurrency Cost:1Model Size:13BQuant:FP8Ctx Length:4kArchitecture:Transformer0.0K Cold

PJMixers-Archive/Alpacino-SuperCOT-13B is a 13 billion parameter language model created by PJMixers-Archive, formed by merging Alpacino-13B and LLaMa-SuperCOT-13B. This model combines the characteristics of its base models, offering a general-purpose language understanding and generation capability. With a 4096-token context length, it is suitable for a range of conversational and text-based tasks. Its primary differentiator lies in its merged architecture, aiming to leverage the strengths of both foundational models.

Loading preview...

Model Overview

PJMixers-Archive/Alpacino-SuperCOT-13B is a 13 billion parameter language model developed by PJMixers-Archive. This model was created through a 50%/50% merge of two existing 13B models: Alpacino-13B and LLaMa-SuperCOT-13B. The merging process was undertaken as an experimental test to explore the capabilities of combining different foundational models.

Key Characteristics

  • Merged Architecture: Combines the strengths of Alpacino-13B and LLaMa-SuperCOT-13B.
  • Parameter Count: 13 billion parameters, offering substantial language processing capabilities.
  • Context Length: Supports a context window of 4096 tokens, suitable for moderately long inputs and outputs.

Potential Use Cases

Given its merged nature, Alpacino-SuperCOT-13B is likely to be a versatile model, potentially suitable for:

  • General text generation and completion.
  • Conversational AI and chatbots.
  • Summarization and question-answering tasks.
  • Experimental applications requiring a blend of capabilities from its base models.