Skip to yearly menu bar Skip to main content


Search All 2024 Events
 

6 Results

<<   <   Page 1 of 1   >>   >
Workshop
Safer-Instruct: Aligning Language Models with Automated Preference Data
Taiwei Shi · Kai Chen · Jieyu Zhao
Workshop
Safer-Instruct: Aligning Language Models with Automated Preference Data
Taiwei Shi · Kai Chen · Jieyu Zhao
Affinity Workshop
Thu 7:30 Policy Optimization in RLHF: The Impact of Out-of-preference Data
Ziniu Li · Tian Xu · Yang Yu
Affinity Workshop
Policy Optimization in RLHF: The Impact of Out-of-preference Data
Ziniu Li · Tian Xu · Yang Yu
Workshop
Peering Through Preferences: Unraveling Feedback Acquisition for Aligning Large Language Models
Hritik Bansal · John Dang · Aditya Grover
Workshop
West-of-N: Synthetic Preference Generation for Improved Reward Modeling
Alizée Pace · Jonathan Mallinson · Eric Malmi · Sebastian Krause · Aliaksei Severyn