bongchoi/MoMo-70B-V1.0

TEXT GENERATIONConcurrency Cost:4Model Size:69BQuant:FP8Ctx Length:32kPublished:Nov 10, 2023License:llama2Architecture:Transformer Open Weights Cold

MoMo-70B-V1.0 by bongchoi is a 69 billion parameter language model with a 32,768 token context length. It is instruction-tuned using Orca-style and Alpaca-style datasets, focusing on general conversational and instruction-following capabilities. This model is designed for applications requiring robust understanding and generation based on diverse instruction formats.

Loading preview...

MoMo-70B-V1.0 Overview

MoMo-70B-V1.0 is a 69 billion parameter language model developed by bongchoi, featuring an extended context length of 32,768 tokens. This model has been instruction-tuned to enhance its ability to follow diverse prompts and engage in conversational tasks.

Key Capabilities

  • Instruction Following: The model is fine-tuned on a combination of Orca-style and Alpaca-style datasets, which are known for improving a model's ability to understand and execute complex instructions.
  • Extended Context: With a 32,768 token context window, MoMo-70B-V1.0 can process and generate longer, more coherent responses, making it suitable for tasks requiring extensive contextual understanding.

Training Details

The training process exclusively utilized Orca-style and Alpaca-style datasets. No other datasets, including benchmark test sets or their training counterparts, were incorporated, ensuring a focused instruction-tuning approach.

Good For

  • General-purpose instruction following and conversational AI applications.
  • Tasks requiring processing and generating long texts due to its large context window.
  • Developers seeking a model trained specifically on established instruction-tuning methodologies.