LargeWorldModel/LWM-Text-Chat-128K
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Feb 7, 2024Architecture:Transformer0.0K Cold

LWM-Text-Chat-128K is a 7 billion parameter open-source auto-regressive language model developed by LargeWorldModel, based on the LLaMA-2 architecture. It is trained on a filtered subset of Books3 data, specifically optimized for processing long text sequences up to 128K tokens. This model is designed for chat applications requiring extensive context understanding.

Loading preview...

LWM-Text-Chat-128K Overview

LWM-Text-Chat-128K is a 7 billion parameter open-source language model built upon the LLaMA-2 architecture. Developed by LargeWorldModel and trained in December 2023, this model is an auto-regressive transformer designed for processing and generating text.

Key Capabilities

  • Extended Context Window: While the provided model information specifies a 4096 token context length, the model name "LWM-Text-Chat-128K" suggests an intended capability for processing significantly longer contexts, up to 128K tokens, making it suitable for applications requiring deep contextual understanding over extensive documents or conversations.
  • LLaMA-2 Foundation: Benefits from the robust architecture and pre-training of the LLaMA-2 family of models.
  • Specialized Training Data: Trained on a 92K subset of Books3 documents, specifically those ranging from 100K to 200K tokens, indicating an optimization for handling and understanding long-form content.

Good For

  • Long-form Chat Applications: Ideal for conversational AI that needs to maintain context over very long dialogues or reference extensive background information.
  • Document Analysis: Suitable for tasks requiring the model to process and understand large documents, such as summarization, question answering, or information extraction from lengthy texts.
  • Research and Development: Provides a strong base for further fine-tuning on specific long-context tasks, leveraging its LLaMA-2 heritage and specialized training.