divinetaco/L3.3-70B-Lycosa-v0.2

Warm
Public
70B
FP8
32768
Jan 26, 2025
License: llama3.3
Hugging Face
Overview

L3.3-70B-Lycosa-v0.2: A Merged Model for Enhanced Intelligence and Creativity

L3.3-70B-Lycosa-v0.2 is a 70 billion parameter language model developed by divinetaco, created through an 'sce' merge using mergekit. This iteration, a refinement of v0.1, specifically dropped llama-3.3-70b-instruct to further reduce positive bias and incorporated DeepSeek-R1-Distill-Llama-70B as a target model to significantly improve reasoning capabilities.

Key Characteristics

  • Enhanced Intelligence: The merge prioritizes overall model intelligence, aiming for more sophisticated and coherent outputs.
  • Reduced Positive Bias: Through strategic model selection and merging, v0.2 seeks to mitigate inherent positive biases often found in language models.
  • Increased Creativity: Designed to excel in tasks requiring imaginative and novel responses.
  • DeepSeek-R1 Influence: Utilizes deepseek-ai/DeepSeek-R1-Distill-Llama-70B as its base and a significant target model, influencing its reasoning and overall performance.

Recommended Usage

This model is particularly well-suited for use cases demanding high intelligence, creative text generation, and a more neutral, less biased output. The recommended chat template is the DeepSeek-R1-Distill-Llama-70B format, as the increased DeepSeek-R1 influence makes the Llama3 chat template less optimal for this version.