manurajcv/bruckeai-legal-merged

TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Mar 12, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

manurajcv/bruckeai-legal-merged is an 8 billion parameter Llama 3.1 instruction-tuned causal language model developed by manurajcv. This model was fine-tuned using Unsloth and Huggingface's TRL library, enabling faster training. It is designed for general language tasks, leveraging the capabilities of the Llama 3.1 architecture.

Loading preview...

Model Overview

manurajcv/bruckeai-legal-merged is an 8 billion parameter instruction-tuned language model based on the Llama 3.1 architecture. Developed by manurajcv, this model was fine-tuned using the Unsloth library, which facilitated a 2x faster training process, in conjunction with Huggingface's TRL library.

Key Characteristics

  • Base Model: Fine-tuned from unsloth/meta-llama-3.1-8b-instruct-bnb-4bit, inheriting the robust capabilities of the Llama 3.1 series.
  • Efficient Training: Leverages Unsloth for accelerated fine-tuning, demonstrating efficiency in model development.
  • Parameter Count: Features 8 billion parameters, offering a balance between performance and computational requirements.

Potential Use Cases

This model is suitable for a variety of general natural language processing tasks, benefiting from its Llama 3.1 foundation and instruction-tuned nature. Its efficient development process suggests it could be a good candidate for applications requiring a capable yet resource-conscious language model.