swype/deepshard-7B-raw
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kLicense:mitArchitecture:Transformer Open Weights Cold

Deepshard-7B-raw is a 7 billion parameter foundational model from swype, provided as a raw mapping of weights to HuggingFace's format. This model serves as a base for further fine-tuning and research, offering a versatile foundation for various natural language processing tasks. Its raw nature means it is not instruction-tuned, making it suitable for developers who require direct access to foundational model weights for custom applications. It is primarily intended for pre-training or fine-tuning for specific downstream applications.

Loading preview...