MS student, Peking University
3 papers at NeurIPS 2025
A novel alignment framework that integrates generative reward models with multi-modal RLHF.
Safe RLHF-V, the multimodal safety alignment framework.
A human preference dataset for multi-turn interleaved multimodal understanding and generatin tasks