IntelLabs/sqft-phi-3-mini-4k-30-base

Warm
Public
4B
BF16
4096
License: apache-2.0
Hugging Face
Overview

SQFT Phi-3-mini-4k-30-base Overview

This model, developed by IntelLabs, is a 4 billion parameter sparse variant of Microsoft's Phi-3-mini-4k-instruct. It incorporates the Wanda sparse method to achieve a 30% sparsity level, aiming for efficient and low-cost model adaptation. The development is part of ongoing research into hardware-aware automated machine learning and sparse foundation models.

Key Characteristics

  • Base Model: Derived from microsoft/Phi-3-mini-4k-instruct.
  • Sparsity: Achieves 30% sparsity using the Wanda method.
  • Parameter Count: 4 billion parameters.
  • Context Length: Supports a 4k (4096 token) context window.
  • Quantization: This specific base model does not include quantization.

Research Focus

This model is primarily a research artifact, detailed in papers such as "SQFT: Low-cost Model Adaptation in Low-precision Sparse Foundation Models" and "Low-Rank Adapters Meet Neural Architecture Search for LLM Compression." It serves as a foundation for exploring efficient model compression and adaptation techniques, particularly for scenarios requiring reduced computational resources.