Skip to yearly menu bar Skip to main content


Search All 2022 Events
 

126 Results

<<   <   Page 2 of 11   >   >>
Workshop
Offline Policy Evaluation for Reinforcement Learning with Adaptively Collected Data
Sunil Madhow · Dan Qiao · Yu-Xiang Wang
Workshop
Confidence-Conditioned Value Functions for Offline Reinforcement Learning
Joey Hong · Aviral Kumar · Sergey Levine
Workshop
Collaborative symmetricity exploitation for offline learning of hardware design solver
HAEYEON KIM · Minsu Kim · joungho kim · Jinkyoo Park
Workshop
Offline Reinforcement Learning from Heteroskedastic Data Via Support Constraints
Anikait Singh · Aviral Kumar · Quan Vuong · Yevgen Chebotar · Sergey Levine
Workshop
Can Active Sampling Reduce Causal Confusion in Offline Reinforcement Learning?
Gunshi Gupta · Tim G. J. Rudner · Rowan McAllister · Adrien Gaidon · Yarin Gal
Workshop
Can Active Sampling Reduce Causal Confusion in Offline Reinforcement Learning?
Gunshi Gupta · Tim G. J. Rudner · Rowan McAllister · Adrien Gaidon · Yarin Gal
Workshop
Diffusion Policies as an Expressive Policy Class for Offline Reinforcement Learning
Zhendong Wang · jonathan j hunt · Mingyuan Zhou
Poster
Thu 14:00 Bellman Residual Orthogonalization for Offline Reinforcement Learning
Andrea Zanette · Martin J Wainwright
Workshop
Offline Reinforcement Learning from Heteroskedastic Data Via Support Constraints
Anikait Singh · Aviral Kumar · Quan Vuong · Yevgen Chebotar · Sergey Levine
Workshop
Can Active Sampling Reduce Causal Confusion in Offline Reinforcement Learning?
Gunshi Gupta · Tim G. J. Rudner · Rowan McAllister · Adrien Gaidon · Yarin Gal
Workshop
Keep Calm and Carry Offline: Policy refinement in offline reinforcement learning
Alex Beeson · Giovanni Montana
Workshop
Wall Street Tree Search: Risk-Aware Planning for Offline Reinforcement Learning
Dan Elbaz · Gal Novik · Oren Salzman