adamo1139/LWM-7B-1M-1000000ctx-AEZAKMI-3_1-1702
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Feb 17, 2024License:llama2Architecture:Transformer0.0K Open Weights Cold

The adamo1139/LWM-7B-1M-1000000ctx-AEZAKMI-3_1-1702 is a 7 billion parameter LargeWorldModel fine-tuned on the AEZAKMI v3.1 dataset. This model was optimized for long context capabilities, specifically trained with a maximum sequence length of 4000 tokens using QLoRA. It is designed to retain extensive contextual understanding, making it suitable for tasks requiring deep comprehension over long passages.

Loading preview...