MikCil/PREMOVE_llama3.3-70b_float16

TEXT GENERATIONConcurrency Cost:4Model Size:70BQuant:FP8Ctx Length:32kPublished:Jan 11, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

MikCil/PREMOVE_llama3.3-70b_float16 is a 70 billion parameter Llama 3.3 instruction-tuned model developed by MikCil, finetuned from unsloth/llama-3.3-70b-instruct-unsloth-bnb-4bit. This model was trained using Unsloth and Huggingface's TRL library, achieving 2x faster training speeds. It is designed for general-purpose language tasks, leveraging its large parameter count and efficient training methodology.

Loading preview...

Model Overview

MikCil/PREMOVE_llama3.3-70b_float16 is a 70 billion parameter instruction-tuned language model developed by MikCil. It is finetuned from the unsloth/llama-3.3-70b-instruct-unsloth-bnb-4bit base model, leveraging the Llama 3.3 architecture. A key aspect of this model's development is its training efficiency, having been trained 2x faster using the Unsloth library in conjunction with Huggingface's TRL library.

Key Characteristics

  • Architecture: Llama 3.3 based, instruction-tuned.
  • Parameter Count: 70 billion parameters, providing robust language understanding and generation capabilities.
  • Training Efficiency: Utilizes Unsloth for significantly faster finetuning, making it a potentially more resource-efficient option for deployment or further adaptation.
  • Context Length: Supports a context length of 32768 tokens, suitable for processing longer inputs and generating comprehensive responses.

Use Cases

This model is well-suited for a broad range of applications requiring a powerful, instruction-following language model. Its large parameter count and Llama 3.3 foundation make it effective for:

  • Complex text generation and summarization.
  • Advanced question answering and information extraction.
  • Conversational AI and chatbot development.
  • Code generation and understanding (given its Llama 3.3 base).

The Apache-2.0 license allows for flexible use in various projects.