ewqr2130/7B_ppo_phiRM_2GPU_3e-7step_4000
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Jan 22, 2024License:apache-2.0Architecture:Transformer Open Weights Cold

The ewqr2130/7B_ppo_phiRM_2GPU_3e-7step_4000 is a 7 billion parameter language model developed by ewqr2130. This model is a PPO-tuned variant, building upon a Zephre 7B-SFT base. It is designed for general language generation tasks, leveraging its 4096-token context length for coherent and extended outputs.

Loading preview...