rachmanino/SelfExtended
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Apr 2, 2024License:mitArchitecture:Transformer Open Weights Cold

rachmanino/SelfExtended is a 7 billion parameter language model based on the Meta Llama-2-7b-chat-bf architecture. It incorporates the SelfExtend technique, implemented with Flash Attention, to significantly enhance its performance on tasks requiring longer context windows. This model is specifically optimized for applications demanding efficient processing of extended input sequences.

Loading preview...