YeungNLP/firefly-gemma-7b
TEXT GENERATIONConcurrency Cost:1Model Size:8.5BQuant:FP8Ctx Length:8kPublished:Feb 28, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

YeungNLP/firefly-gemma-7b is an 8.5 billion parameter instruction-tuned causal language model based on Google's Gemma-7B architecture, developed by YeungNLP. Fine-tuned using QLoRA on a single V100 GPU, this model is designed to function as a helpful and harmless AI assistant. It demonstrates strong performance on the Open LLM Leaderboard, outperforming several comparable models including gemma-7b-it and Zephyr-7B-Beta, making it suitable for general conversational AI tasks.

Loading preview...