RatanRohith/NeuralPizza-7B-V0.3
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Jan 29, 2024License:apache-2.0Architecture:Transformer Open Weights Cold
NeuralPizza-7B-V0.3 by RatanRohith is a 7 billion parameter language model, fine-tuned from NeuralPizza-7B-V0.1 using Direct Preference Optimization (DPO) on the argilla/distilabel-intel-orca-dpo-pairs dataset. This model is specialized for research and experimental applications in language modeling, particularly for exploring and understanding DPO techniques. It offers a 4096-token context length and is intended for evaluating DPO's impact on language model performance.
Loading preview...