AntNLP/TinyLlama-NoPE-HeadScale8k
TEXT GENERATIONConcurrency Cost:1Model Size:1.1BQuant:BF16Ctx Length:2kPublished:May 4, 2024License:mitArchitecture:Transformer Open Weights Cold

AntNLP/TinyLlama-NoPE-HeadScale8k is a 1.1 billion parameter causal language model developed by AntNLP. This model is notable for its use of a NoPE (No Position Encoding) architecture, allowing for length generalization without traditional positional embeddings. It is designed for tasks requiring efficient processing and understanding of sequences without explicit positional information.

Loading preview...