minosu/godot_dodo_4x_60k_llama_7b
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kArchitecture:Transformer0.0K Cold
The minosu/godot_dodo_4x_60k_llama_7b is a 7 billion parameter instruction-following model, fine-tuned from the LLaMA architecture. Developed by minosu, this model specializes in adhering to instructions, particularly those related to code. It was trained on a 60,000-row instruction-following dataset, making it suitable for tasks requiring precise command execution and code-related instruction processing.
Loading preview...
Model Overview
The minosu/godot_dodo_4x_60k_llama_7b is a 7 billion parameter instruction-following model, part of the Godot-Dodo series, fine-tuned from the LLaMA architecture. This model was trained in April 2023 with a focus on robust instruction adherence.
Key Capabilities
- Instruction Following: Specifically designed and fine-tuned to accurately follow instructions provided in prompts.
- Code Instruction Processing: Evaluated using code instruction prompts, indicating a specialization in understanding and responding to code-related directives.
- Custom Training Data: Benefits from training on a unique 60,000-row instruction-following dataset, which is publicly available.
Good For
- Applications requiring a model to precisely execute given instructions.
- Tasks involving code generation, analysis, or transformation based on explicit commands.
- Developers looking for a LLaMA-based model optimized for instruction-following capabilities, particularly in technical or coding contexts.