Workshop
|
Sat 16:00
|
Contributed Talk: Personalizing Reinforcement Learning from Human Feedback with Variational Preference Learning
|
|
Workshop
|
|
Aligning to What? Limits to RLHF Based Alignment
Logan Barnhart · Reza Akbarian Bafghi · Maziar Raissi · Stephen Becker
|
|
Workshop
|
Sat 15:00
|
Phillip Isola (MIT): Representation Learning from Human Feedback
Phillip Isola
|
|
Poster
|
Wed 16:30
|
REBORN: Reinforcement-Learned Boundary Segmentation with Iterative Training for Unsupervised ASR
Liang-Hsuan Tseng · En-Pei Hu · Cheng-Han Chiang · Yuan Tseng · Hung-yi Lee · Lin-shan Lee · Shao-Hua Sun
|
|
Workshop
|
Sat 9:00
|
Andreea Bobu: Learning a Lot from a Little: How Structure Enables Efficient and Human-Aligned Robot Learning
|
|
Workshop
|
|
Personalizing Reinforcement Learning from Human Feedback with Variational Preference Learning
Sriyash Poddar · Yanming Wan · Hamish Ivison · Abhishek Gupta · Natasha Jaques
|
|
Workshop
|
|
Computational discovery of human reinforcement learning dynamics from choice behavior
Daniel Weinhardt · Maria Eckstein · Sebastian Musslick
|
|
Poster
|
Thu 16:30
|
Personalizing Reinforcement Learning from Human Feedback with Variational Preference Learning
Sriyash Poddar · Yanming Wan · Hamish Ivison · Abhishek Gupta · Natasha Jaques
|
|
Poster
|
Wed 16:30
|
AHA: Human-Assisted Out-of-Distribution Generalization and Detection
Haoyue Bai · Jifan Zhang · Robert Nowak
|
|
Workshop
|
|
Evaluating Fine-Tuning Efficiency of Human-Inspired Learning Strategies in Medical Question Answering
Yushi Yang · Andrew M. Bean · Robert McCraith · Adam Mahdi
|
|
Poster
|
Fri 11:00
|
Online Iterative Reinforcement Learning from Human Feedback with General Preference Model
Chenlu Ye · Wei Xiong · Yuheng Zhang · Hanze Dong · Nan Jiang · Tong Zhang
|
|
Poster
|
Wed 11:00
|
When Your AIs Deceive You: Challenges of Partial Observability in Reinforcement Learning from Human Feedback
Leon Lang · Davis Foote · Stuart J Russell · Anca Dragan · Erik Jenner · Scott Emmons
|
|