ricdomolm/lawma-8b
Lawma 8B is an 8 billion parameter instruction-tuned Llama 3 8B Instruct model developed by Ricardo Dominguez-Olmedo and his collaborators. Fine-tuned on 260 legal classification tasks from Supreme Court and Court of Appeals databases, it excels at legal multiple-choice classification. This model demonstrates significant performance improvements over general-purpose LLMs like GPT-4 and Llama 3 70B Instruct on specialized legal tasks, achieving 80.3% mean accuracy across all tasks.
Loading preview...
Lawma 8B: Specialized Legal Classification Model
Lawma 8B is an 8 billion parameter model, fine-tuned from Llama 3 8B Instruct, specifically designed for legal classification tasks. Developed by Ricardo Dominguez-Olmedo and his team, this model was trained on over 500,000 task examples, totaling 2 billion tokens, derived from 260 legal classification tasks from the Supreme Court and Songer Court of Appeals databases. Its specialization allows it to significantly outperform larger, general-purpose models on these specific legal challenges.
Key Capabilities & Performance
- Superior Legal Classification: Lawma 8B achieves a mean classification accuracy of 80.3% across 260 legal tasks, outperforming GPT-4 (62.9%) and Llama 3 70B Instruct (58.4%) by substantial margins.
- Task-Specific Optimization: The model is optimized for multiple-choice legal classification, outputting only multiple-choice letters or numbers.
- Foundation for Further Fine-tuning: While highly effective for its intended tasks, practitioners are encouraged to further fine-tune Lawma on their specific legal use cases for even greater performance gains.
Good For
- Legal Research Automation: Classifying legal documents based on specific criteria from Supreme Court and Court of Appeals databases.
- Specialized Legal AI Applications: Developing applications that require high accuracy in legal multiple-choice classification.
- Benchmarking Specialized LLMs: Demonstrating the power of specialization in achieving superior performance on niche domains compared to generalist models. More details are available in the arXiv preprint.
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.