divinetaco/L3.3-70B-Lycosa-v0.2

Warm
Public
70B
FP8
32768
4
Jan 26, 2025
License: llama3.3
Hugging Face

L3.3-70B-Lycosa-v0.2 by divinetaco is a 70 billion parameter merged language model, built using the 'sce' merge method with DeepSeek-R1-Distill-Llama-70B as its base. This model is specifically engineered to enhance intelligence, reduce positive bias, and foster creativity, making it suitable for applications requiring nuanced and imaginative responses. It integrates several Llama-3.3 based models, focusing on improved reasoning capabilities.

Overview

L3.3-70B-Lycosa-v0.2: A Merged Model for Enhanced Intelligence and Creativity

L3.3-70B-Lycosa-v0.2 is a 70 billion parameter language model developed by divinetaco, created through an 'sce' merge using mergekit. This iteration, a refinement of v0.1, specifically dropped llama-3.3-70b-instruct to further reduce positive bias and incorporated DeepSeek-R1-Distill-Llama-70B as a target model to significantly improve reasoning capabilities.

Key Characteristics

  • Enhanced Intelligence: The merge prioritizes overall model intelligence, aiming for more sophisticated and coherent outputs.
  • Reduced Positive Bias: Through strategic model selection and merging, v0.2 seeks to mitigate inherent positive biases often found in language models.
  • Increased Creativity: Designed to excel in tasks requiring imaginative and novel responses.
  • DeepSeek-R1 Influence: Utilizes deepseek-ai/DeepSeek-R1-Distill-Llama-70B as its base and a significant target model, influencing its reasoning and overall performance.

Recommended Usage

This model is particularly well-suited for use cases demanding high intelligence, creative text generation, and a more neutral, less biased output. The recommended chat template is the DeepSeek-R1-Distill-Llama-70B format, as the increased DeepSeek-R1 influence makes the Llama3 chat template less optimal for this version.