firstbacksecondback
6 Results
Workshop
|
Safer-Instruct: Aligning Language Models with Automated Preference Data Taiwei Shi · Kai Chen · Jieyu Zhao |
||
Workshop
|
Safer-Instruct: Aligning Language Models with Automated Preference Data Taiwei Shi · Kai Chen · Jieyu Zhao |
||
Affinity Workshop
|
Thu 7:30 |
Policy Optimization in RLHF: The Impact of Out-of-preference Data Ziniu Li · Tian Xu · Yang Yu |
|
Affinity Workshop
|
Policy Optimization in RLHF: The Impact of Out-of-preference Data Ziniu Li · Tian Xu · Yang Yu |
||
Workshop
|
Peering Through Preferences: Unraveling Feedback Acquisition for Aligning Large Language Models Hritik Bansal · John Dang · Aditya Grover |
||
Workshop
|
West-of-N: Synthetic Preference Generation for Improved Reward Modeling Alizée Pace · Jonathan Mallinson · Eric Malmi · Sebastian Krause · Aliaksei Severyn |