L1nus/qwen3-4B-default-pubmed-labeled-5epoch-seq-2048

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Mar 24, 2026License:apache-2.0Architecture:Transformer Open Weights Warm

The L1nus/qwen3-4B-default-pubmed-labeled-5epoch-seq-2048 is a 4 billion parameter Qwen3-based causal language model developed by L1nus. It was fine-tuned using Unsloth and Huggingface's TRL library, specifically optimized for tasks related to PubMed-labeled data. This model is designed for applications requiring specialized knowledge from biomedical literature, leveraging its 32768 token context length for comprehensive analysis.

Loading preview...

Model Overview

This model, developed by L1nus, is a 4 billion parameter Qwen3-based causal language model. It was fine-tuned from unsloth/qwen3-4b-instruct-2507-unsloth-bnb-4bit using the Unsloth library, which facilitated a 2x faster training process, and Huggingface's TRL library. The model's training specifically targeted PubMed-labeled data over 5 epochs with a sequence length of 2048, indicating a specialization in biomedical and scientific text processing.

Key Capabilities

  • Specialized Knowledge: Fine-tuned on PubMed-labeled data, suggesting proficiency in understanding and generating content related to medical and biological research.
  • Efficient Training: Leveraged Unsloth for accelerated training, demonstrating an efficient development approach.
  • Qwen3 Architecture: Based on the Qwen3 family, providing a robust foundation for language understanding and generation.

Good For

  • Applications requiring analysis or generation of text from biomedical literature.
  • Tasks involving PubMed data, such as information extraction, summarization, or question answering in the medical domain.
  • Developers looking for a specialized Qwen3 model with efficient fine-tuning origins.