deinon-daemon/axolotl-13b-chat-qlora-dev
TEXT GENERATIONConcurrency Cost:1Model Size:13BQuant:FP8Ctx Length:4kLicense:llama2Architecture:Transformer Open Weights Cold
The deinon-daemon/axolotl-13b-chat-qlora-dev is a 13 billion parameter instruct-tuned chat model, fine-tuned from Llama-2-13b-chat-hf. Developed by deinon-daemon, it utilizes QLORA and Flash Attention for efficient training on a 40k slice of the Open-Orca dataset. This model is a proof-of-concept demonstrating a small-is-powerful approach to chat model development, aiming for performance comparable to other Llama/Alpaca/Guanaco/Vicuna models of similar scale.
Loading preview...