RatanRohith/NeuralPizza-7B-V0.2
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Jan 21, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

NeuralPizza-7B-V0.2 is a 7 billion parameter language model developed by RatanRohith, fine-tuned from NeuralMathChat-7B-V0.2 using Direct Preference Optimization (DPO). With a 4096-token context length, this model specializes in exploring DPO techniques for enhanced performance based on preference comparisons. It is primarily intended for research and experimental applications in language modeling.

Loading preview...