cookinai/titanbagel
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Feb 23, 2024License:cc-by-4.0Architecture:Transformer Open Weights Cold

cookinai/titanbagel is an 8 billion parameter language model, a finetune of Jon Durbin's Bagel DPO 7B. It has been further trained using the Hercules 3.0 dataset, offering an 8192 token context length. This model is designed for general language tasks, leveraging its DPO base and additional training for improved performance.

Loading preview...