zai-org/LongAlign-7B-64k
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Jan 29, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

LongAlign-7B-64k is a 7 billion parameter chat model developed by THUDM, based on Llama-2-7B, with an extended context window of 64k tokens. It is specifically fine-tuned for long-context instruction following, utilizing the LongAlign-10k dataset and specialized training strategies like packing with loss weighting and sorted batching. This model excels at processing and responding to queries up to 64k tokens in length, making it suitable for applications requiring extensive contextual understanding.

Loading preview...