neph1/Qwen2.5-Coder-7B-Instruct-Unity

TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Nov 17, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

neph1/Qwen2.5-Coder-7B-Instruct-Unity is a 7.6 billion parameter instruction-tuned causal language model developed by neph1, fine-tuned from Qwen2.5-Coder-7B-Instruct. Optimized for Unity3D development, it excels at answering questions and generating code related to the Unity game engine. This model is specifically trained on Unity3D Q&A datasets, making it a specialized tool for game developers using Unity.

Loading preview...

Model Overview

neph1/Qwen2.5-Coder-7B-Instruct-Unity is a 7.6 billion parameter instruction-tuned model, developed by neph1 and fine-tuned from unsloth/Qwen2.5-Coder-7B-Instruct-bnb-4bit. This model is specifically designed to assist with Unity3D development, leveraging a specialized training dataset.

Key Capabilities

  • Unity3D Code Generation: Optimized for generating code and providing answers related to the Unity game engine.
  • Specialized Knowledge: Trained on a merged dataset comprising Unity3D Q&A from sources like ibranze/codellama_unity3d_v2, Hypersniper/unity_api_2022_3, and neph1/Unity_Code_QnA.
  • Efficient Training: Fine-tuned using Unsloth and Huggingface's TRL library, enabling faster training.
  • Instruction Following: Responds well to instruction-based prompts, with preliminary testing showing good compatibility with the Mistral chat template.

Training Details

The model underwent approximately 1.5 epochs of training, with a focus on Unity-specific content. Training parameters included a per-device batch size of 2, gradient accumulation steps of 64, and a learning rate of 1e-4. Validation loss decreased steadily during training, indicating effective learning on the specialized dataset. The developer notes that further validation with general coding questions might be beneficial to prevent overfitting to the Unity domain.