4bit/Redmond-Puffin-13B
TEXT GENERATIONConcurrency Cost:1Model Size:13BQuant:FP8Ctx Length:4kLicense:mitArchitecture:Transformer0.0K Open Weights Cold

Redmond-Puffin-13B is a 13 billion parameter Llama 2-based language model developed by Nous Research, fine-tuned on 3,000 high-quality GPT-4 examples with a 4096 token context length. It excels in multi-turn conversations and long-context recall, with pretraining on 2 trillion tokens. This model is designed for applications requiring extensive conversational memory and up-to-date information recall.

Loading preview...