Skip to yearly menu bar Skip to main content

Spotlight Poster

Continual Learning for Instruction Following from Realtime Feedback

Alane Suhr · Yoav Artzi

Great Hall & Hall B1+B2 (level 1) #330


We propose and deploy an approach to continually train an instruction-following agent from feedback provided by users during collaborative interactions. During interaction, human users instruct an agent using natural language, and provide realtime binary feedback as they observe the agent following their instructions. We design a contextual bandit learning approach, converting user feedback to immediate reward. We evaluate through thousands of human-agent interactions, demonstrating 15.4% absolute improvement in instruction execution accuracy over time. We also show our approach is robust to several design variations, and that the feedback signal is roughly equivalent to the learning signal of supervised demonstration data.

Chat is not available.