Dhyey3559/gujarati-finetune-llama3b

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:1BQuant:BF16Ctx Length:32kPublished:Aug 23, 2025Architecture:Transformer0.0K Warm

The Dhyey3559/gujarati-finetune-llama3b is a 1 billion parameter experimental Gujarati language model, fine-tuned from unsloth/Llama-3.2-1B. It is designed for research and testing purposes, focusing on Gujarati language generation. This prototype model was trained on a small dataset and currently exhibits limited accuracy, with future work planned to enhance its performance and dataset.

Loading preview...

Overview

This model, developed by Dhyey3559, is an experimental Gujarati language model with 1 billion parameters, fine-tuned from the unsloth/Llama-3.2-1B architecture. It represents a prototype effort to create a language model specifically for Gujarati.

Key Characteristics

  • Base Model: Fine-tuned from unsloth/Llama-3.2-1B.
  • Language Focus: Specifically designed for the Gujarati language.
  • Training Data: Trained on a relatively small dataset of approximately 10,000 samples.
  • Current Performance: Evaluated using BLEU score, indicating that accuracy is not optimal at this prototype stage.
  • Development Status: This is a prototype intended solely for research and testing; it is not recommended for production environments due to potential inconsistencies or incorrect outputs.

Future Development

Future work for this model is planned to focus on:

  • Expanding the training dataset significantly.
  • Optimizing hyperparameters to improve model performance.
  • General enhancements to overall accuracy and output quality.

Use Cases

  • Research and Development: Ideal for researchers exploring fine-tuning techniques for low-resource languages like Gujarati.
  • Experimental Prototyping: Suitable for testing initial concepts related to Gujarati NLP applications.
  • Language Study: Can be used to analyze the behavior of fine-tuned models on specific linguistic datasets.