fblgit/cybertron-v4-qw7B-MGS
TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Oct 29, 2024License:qwenArchitecture:Transformer0.0K Cold

The fblgit/cybertron-v4-qw7B-MGS is a 7.6 billion parameter causal language model developed by fblgit, based on the Qwen2.5 architecture with a 131072 token context length. This model utilizes a novel 'MGS' approach to address corpora forgetfulness, distinguishing it from other LLMs. It is fine-tuned via Supervised Fine-Tuning (SFT) on the Magpie-Align/Magpie-Qwen2.5-Pro-1M-v0.1 dataset, making it suitable for general language understanding and generation tasks where robust performance in its class is desired.

Loading preview...