olusegunola/phi-1.5-distill-Ablation_No_L2_Norm-merged
TEXT GENERATIONConcurrency Cost:1Model Size:1.4BQuant:BF16Ctx Length:2kPublished:Mar 22, 2026Architecture:Transformer Loading
The olusegunola/phi-1.5-distill-Ablation_No_L2_Norm-merged is a 1.4 billion parameter language model, likely based on the Microsoft Phi-1.5 architecture, with a context length of 2048 tokens. This specific version is an ablation study variant, indicating modifications related to the removal of L2 normalization during its distillation or training process. Its primary differentiator lies in exploring the impact of L2 norm absence on model performance and characteristics. Further details on its specific capabilities or intended use cases are not provided in the available information.
Loading preview...
Popular Sampler Settings
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.
temperature
–
top_p
–
top_k
–
frequency_penalty
–
presence_penalty
–
repetition_penalty
–
min_p
–