akshayballal/Qwen3-1.7B-Pubmed-16bit-GRPO

Warm
Public
2B
BF16
40960
Jan 21, 2026
License: apache-2.0
Hugging Face
Overview

Model Overview

The akshayballal/Qwen3-1.7B-Pubmed-16bit-GRPO is a 1.7 billion parameter language model based on the Qwen3 architecture. Developed by akshayballal, this model is a fine-tuned version of unsloth/qwen3-1.7b-unsloth-bnb-4bit.

Key Capabilities

  • Biomedical Text Processing: This model is specifically fine-tuned for tasks related to biomedical literature, making it adept at understanding and generating content relevant to medical research and publications.
  • Efficient Training: Leveraging Unsloth and Huggingface's TRL library, the model was trained 2x faster, indicating an optimized and efficient development process.

Good for

  • Applications requiring analysis or generation of text from biomedical databases like PubMed.
  • Research in medical natural language processing.
  • Tasks such as medical information extraction, summarization of scientific papers, or question answering in the biomedical domain.