Skip to content

Latest commit

 

History

History
59 lines (37 loc) · 3.81 KB

llm_rlhf.md

File metadata and controls

59 lines (37 loc) · 3.81 KB

LLM RLHF

Survey

RLHF

  • Evolving Alignment via Asymmetric Self-Play, arXiv, 2411.00062, arxiv, pdf, cication: -1

    Ziyu Ye, Rishabh Agarwal, Tianqi Liu, ..., Qijun Tan, Yuan Liu · (jiqizhixin)

  • Hybrid Preferences: Learning to Route Instances for Human vs. AI Feedback, arXiv, 2410.19133, arxiv, pdf, cication: -1

    Lester James V. Miranda, Yizhong Wang, Yanai Elazar, ..., Hannaneh Hajishirzi, Pradeep Dasigi

  • LongReward: Improving Long-context Large Language Models with AI Feedback, arXiv, 2410.21252, arxiv, pdf, cication: -1

    Jiajie Zhang, Zhongni Hou, Xin Lv, ..., Ling Feng, Juanzi Li · (LongReward - THUDM) Star · (huggingface)

  • Thinking LLMs: General Instruction Following With Thought Generation 𝕏

  • MIA-DPO: Multi-Image Augmented Direct Preference Optimization For Large Vision-Language Models, arXiv, 2410.17637, arxiv, pdf, cication: -1

    Ziyu Liu, Yuhang Zang, Xiaoyi Dong, ..., Dahua Lin, Jiaqi Wang

Reward Models

Projects

Misc