divinetaco/L3.3-70B-Lycosa-v0.2

Hugging Face
TEXT GENERATIONConcurrency Cost:4Model Size:70BQuant:FP8Ctx Length:32kPublished:Jan 26, 2025License:llama3.3Architecture:Transformer0.0K Warm

L3.3-70B-Lycosa-v0.2 by divinetaco is a 70 billion parameter merged language model, built using the 'sce' merge method with DeepSeek-R1-Distill-Llama-70B as its base. This model is specifically engineered to enhance intelligence, reduce positive bias, and foster creativity, making it suitable for applications requiring nuanced and imaginative responses. It integrates several Llama-3.3 based models, focusing on improved reasoning capabilities.

Loading preview...

L3.3-70B-Lycosa-v0.2: A Merged Model for Enhanced Intelligence and Creativity

L3.3-70B-Lycosa-v0.2 is a 70 billion parameter language model developed by divinetaco, created through an 'sce' merge using mergekit. This iteration, a refinement of v0.1, specifically dropped llama-3.3-70b-instruct to further reduce positive bias and incorporated DeepSeek-R1-Distill-Llama-70B as a target model to significantly improve reasoning capabilities.

Key Characteristics

  • Enhanced Intelligence: The merge prioritizes overall model intelligence, aiming for more sophisticated and coherent outputs.
  • Reduced Positive Bias: Through strategic model selection and merging, v0.2 seeks to mitigate inherent positive biases often found in language models.
  • Increased Creativity: Designed to excel in tasks requiring imaginative and novel responses.
  • DeepSeek-R1 Influence: Utilizes deepseek-ai/DeepSeek-R1-Distill-Llama-70B as its base and a significant target model, influencing its reasoning and overall performance.

Recommended Usage

This model is particularly well-suited for use cases demanding high intelligence, creative text generation, and a more neutral, less biased output. The recommended chat template is the DeepSeek-R1-Distill-Llama-70B format, as the increased DeepSeek-R1 influence makes the Llama3 chat template less optimal for this version.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p