CONCISE/LLaMa_V2-13B-Instruct-Uncensored-HF

TEXT GENERATIONConcurrency Cost:1Model Size:13BQuant:FP8Ctx Length:4kArchitecture:Transformer0.0K Cold

LLaMa_V2-13B-Instruct-Uncensored-HF is a 13 billion parameter instruction-tuned LLaMa V2 model developed by CONCISE. This model is specifically designed to address model biases without sacrificing performance, leveraging datasets like wizardlm_evol_instruct_V2_196k_unfiltered_merged_split and wizard_vicuna_70k_unfiltered. Its primary strength lies in providing instruction-based responses while maintaining a neutral stance on sensitive topics.

Loading preview...

Model Overview

LLaMa_V2-13B-Instruct-Uncensored-HF is an instruction-based variant of the LLaMa V2 13 billion parameter model, developed by CONCISE. This model focuses on delivering instruction-tuned responses while actively working to mitigate inherent model biases. It achieves this balance by incorporating diverse and carefully curated datasets.

Key Characteristics

  • Instruction-Tuned: Optimized for following user instructions and generating relevant responses.
  • Bias Mitigation: Specifically trained with datasets like wizardlm_evol_instruct_V2_196k_unfiltered_merged_split and wizard_vicuna_70k_unfiltered to reduce biases.
  • Uncensored Nature: Designed to provide responses without artificial restrictions, aiming for neutrality rather than censorship.

Ideal Use Cases

  • General Instruction Following: Suitable for a wide range of tasks requiring adherence to specific instructions.
  • Research on Bias in LLMs: Can be a valuable tool for studying and understanding how models handle sensitive topics.
  • Applications Requiring Neutrality: Useful in scenarios where an unbiased and uncensored response is preferred, provided ethical guidelines are followed by the implementer.