Timezone: »
Machine learning (ML) has been one of the premier drivers of recent advances in robotics research and has made its way into impacting several real-world robotic applications in unstructured and human-centric environments, such as transportation, healthcare, and manufacturing. At the same time, robotics has been a key motivation for numerous research problems in artificial intelligence research, from efficient algorithms to robust generalization of decision models. However, there are still considerable obstacles to fully leveraging state-of-the-art ML in real-world robotics applications. For capable robots equipped with ML models, guarantees on the robustness and additional analysis of the social implications of these models are required for their utilization in real-world robotic domains that interface with humans (e.g. autonomous vehicles, and tele-operated or assistive robots).
To support the development of robots that are safely deployable among humans, the field must consider trustworthiness as a central aspect in the development of real-world robot learning systems. Unlike many other applications of ML, the combined complexity of physical robotic platforms and learning-based perception-action loops presents unique technical challenges. These challenges include concrete technical problems such as very high performance requirements, explainability, predictability, verification, uncertainty quantification, and robust operation in dynamically distributed, open-set domains. Since robots are developed for use in human environments, in addition to these technical challenges, we must also consider the social aspects of robotics such as privacy, transparency, fairness, and algorithmic bias. Both technical and social challenges also present opportunities for robotics and ML researchers alike. Contributing to advances in the aforementioned sub-fields promises to have an important impact on real-world robot deployment in human environments, building towards robots that use human feedback, indicate when their model is uncertain, and are safe to operate autonomously in safety-critical settings such as healthcare and transportation.
This year’s robot learning workshop aims at discussing unique research challenges from the lens of trustworthy robotics. We adopt a broad definition of trustworthiness that highlights different application domains and the responsibility of the robotics and ML research communities to develop “robots for social good.” Bringing together experts with diverse backgrounds from the ML and robotics communities, the workshop will offer new perspectives on trust in the context of ML-driven robot systems.
Scope of contributions:
Specific areas of interest include but are not limited to:
* epistemic uncertainty estimation in robotics;
* explainable robot learning;
* domain adaptation and distribution shift in robot learning;
* multi-modal trustworthy sensing and sensor fusion;
* safe deployment for applications such as agriculture, space, science, and healthcare;
* privacy aware robotic perception;
* information system security in robot learning;
* learning from offline data and safe on-line learning;
* simulation-to-reality transfer for safe deployment;
* robustness and safety evaluation;
* certifiability and performance guarantees;
* robotics for social good;
* safe robot learning with humans in the loop;
* algorithmic bias in robot learning;
* ethical robotics.
Fri 7:00 a.m. - 7:15 a.m.
|
Opening Remarks
(
Workshop Introduction
)
SlidesLive Video » |
🔗 |
Fri 7:15 a.m. - 7:30 a.m.
|
DALL-E-Bot: Introducing Web-Scale Diffusion Models to Robotics
(
Contributed Talk 1
)
|
Ivan Kapelyukh · Vitalis Vosylius · Edward Johns 🔗 |
Fri 7:30 a.m. - 8:15 a.m.
|
Panel: Uncertainty-Aware Machine Learning for Robotics (Q&A 1)
(
Discussion Panel
)
SlidesLive Video » |
Georgia Chalvatzaki · Stefanie Tellex · Animesh Garg 🔗 |
Fri 8:15 a.m. - 8:30 a.m.
|
Coffee Break 1
|
🔗 |
Fri 8:30 a.m. - 9:15 a.m.
|
Panel: Scaling & Models (Q&A 2)
(
Discussion Panel
)
SlidesLive Video » |
Andy Zeng · Haoran Tang · Karol Hausman · Jackie Kay · Gabriel Barth-Maron 🔗 |
Fri 9:15 a.m. - 10:15 a.m.
|
Poster Session 1
(
Poster Session
)
|
🔗 |
Fri 10:15 a.m. - 11:00 a.m.
|
Panel: Safety and Verification for Decision-Making Systems (Q&A 3)
(
Discussion Panel
)
SlidesLive Video » |
Luca Carlone · Sarah Dean · Matthew Johnson-Roberson 🔗 |
Fri 11:00 a.m. - 3:00 p.m.
|
Long Break
|
🔗 |
Fri 3:00 p.m. - 4:00 p.m.
|
Debate: Robotics for Good
(
Discussion Panel
)
SlidesLive Video » |
Karol Hausman · Katherine Driggs-Campbell · Luca Carlone · Sarah Dean · Matthew Johnson-Roberson · Animesh Garg 🔗 |
Fri 4:00 p.m. - 4:15 p.m.
|
Robust Forecasting for Robotic Control: A Game-Theoretic Approach
(
Contributed Talk 2
)
|
Shubhankar Agarwal · David Fridovich-Keil · Sandeep Chinchali 🔗 |
Fri 4:15 p.m. - 4:30 p.m.
|
Certifiably-correct Control Policies for Safe Learning and Adaptation in Assistive Robotics
(
Contributed Talk 3
)
|
Keyvan Majd · GEOFFEY CLARK · Tanmay Khandait · Heni Ben Amor 🔗 |
Fri 4:45 p.m. - 5:00 p.m.
|
Coffee Break 2
|
🔗 |
Fri 5:00 p.m. - 6:00 p.m.
|
Poster Session 2
(
Poster Session
)
|
🔗 |
Fri 6:00 p.m. - 6:45 p.m.
|
Panel: Explainability/Predictability Robotics (Q&A 4)
(
Discussion Panel
)
SlidesLive Video » |
Katherine Driggs-Campbell · Been Kim · Leila Takayama 🔗 |
Fri 6:45 p.m. - 7:00 p.m.
|
Closing Remarks
(
Workshop Presentation
)
SlidesLive Video » |
🔗 |
-
|
Visual Backtracking Teleoperation: A Data Collection Protocol for Offline Image-Based RL
(
Poster
)
SlidesLive Video » We consider how to most efficiently leverage teleoperator time to collect data for learning robust image-based value functions and policies for sparse reward robotic tasks. To accomplish this goal, we modify the process of data collection to include more than just successful demonstrations of the desired task. Instead we develop a novel protocol that we call Visual Backtracking Teleoperation (VBT), which deliberately collects a dataset of visually similar failures, recoveries, and successes. VBT data collection is particularly useful for efficiently learning accurate value functions from small datasets of image-based observations. We demonstrate VBT on a real robot to perform continuous control from image observations for the deformable manipulation task of T-shirt grasping. We find that by adjusting the data collection process we improve the quality of both the learned value functions and policies over a variety of baseline methods for data collection. Specifically, we find that offline reinforcement learning on VBT data outperforms standard behavior cloning on successful demonstration data by 13% when both methods are given equal-sized datasets of 60 minutes of data from the real robot. |
David Brandfonbrener · Stephen Tu · Avi Singh · Stefan Welker · Chad Boodoo · Nikolai Matni · Jacob Varley 🔗 |
-
|
Conformal Semantic Keypoint Detection with Statistical Guarantees
(
Poster
)
SlidesLive Video » Detecting semantic keypoints is a critical intermediate task for object detection and pose estimation from images. Existing approaches, albeit performing well on standard benchmarks, offer no provable guarantees on the quality of the detection. In this paper, we apply the statistical machinery of inductive conformal prediction that, given a calibration dataset (e.g., 200 images) and a nonconformity function, converts a heuristic heatmap detection into a prediction set that provably covers the true keypoint location with a user-specified probability (e.g., 90%). We design three different nonconformity functions leading to circular or elliptical prediction sets that are easy to compute. On the LINEMOD Occluded dataset we demonstrate that (i) the empirical coverage rate of the prediction sets is valid; (ii) the prediction sets are tight, e.g., a ball with radius 10 pixels covers true keypoint locations on most test images; and (iii) the prediction sets are adaptive, i.e., their sizes become larger for keypoints that are difficult to detect and smaller for easy instances. |
Heng Yang · Marco Pavone 🔗 |
-
|
A Contextual Bandit Approach for Learning to Plan in Environments with Probabilistic Goal Configurations
(
Poster
)
SlidesLive Video » Object-goal navigation (Object-nav) entails searching, recognizing and navigating to a target object. Object-nav has been extensively studied by the Embodied-AI community, but most solutions are often restricted to considering static objects (e.g., television, fridge, etc.). We propose a modular framework for object-nav that is able to efficiently search indoor environments for not just static objects but also movable objects (e.g. fruits, glasses, phones, etc.) that frequently change their positions due to human interaction. Our contextual-bandit agent efficiently explores the environment by showing optimism in the face of uncertainty and learns a model of the likelihood of spotting different objects from each navigable location. The likelihoods are used as rewards in a weighted minimum latency solver to deduce a trajectory for the robot. We evaluate our algorithms in two simulated environments and a real-world setting, to demonstrate high sample efficiency and reliability. |
Sohan Rudra · Saksham Goel · Anirban Santara · Claudio Gentile · Laurent Perron · Fei Xia · Vikas Sindhwani · Carolina Parada · Gaurav Aggarwal 🔗 |
-
|
Imitating careful experts to avoid catastrophic events
(
Poster
)
SlidesLive Video » Reinforcement learning (RL) is increasingly being used to control robotic systems that interact closely with humans. This interaction raises the problem of safe RL: how to ensure that an RL-controlled robotic system never, for instance, injures a human. This problem is especially challenging in rich, realistic settings where it is not even possible to clearly write down a reward function which incorporates these outcomes. In these circumstances, perhaps the only viable approach is based on inverse reinforcement learning (IRL), which infers rewards from human demonstrations. However, IRL is massively underdetermined as many different rewards can lead to the same optimal policies; we show that this makes it difficult to distinguish catastrophic outcomes (such as injuring a human) from merely undesirable outcomes. Our key insight is that humans do display different behaviour when catastrophic outcomes are possible: they become much more careful. We incorporate carefulness signals into IRL, and find that they do indeed allow IRL to disambiguate undesirable from catastrophic outcomes, which is critical to ensuring safety in future real-world human-robot interactions. |
Jack Hanslope · Laurence Aitchison 🔗 |
-
|
Infrastructure-based End-to-End Learning and Prevention of Driver Failure
(
Poster
)
SlidesLive Video » Intelligent intersection managers can improve safety by detecting dangerous drivers or failure modes in autonomous vehicles, warning oncoming vehicles as they approach an intersection. In this work, we present FailureNet, a recurrent neural network trained end-to-end on trajectories of both nominal and reckless drivers in a scaled miniature city. FailureNet observes the poses of vehicles as they approach an intersection and detects whether a failure is present in the autonomy stack, warning cross-traffic of potentially dangerous drivers. FailureNet can accurately identify control failures, upstream perception errors, and speeding drivers, distinguishing them from nominal driving. The network is trained and deployed with autonomous vehicles in a scaled miniature city. Compared to speed or frequency-based predictors, FailureNet's recurrent neural network structure provides improved predictive power, yielding upwards of 84% accuracy when deployed on hardware. |
Noam Buckman · Shiva Sreeram · Mathias Lechner · Yutong Ban · Ramin Hasani · Sertac Karaman · Daniela Rus 🔗 |
-
|
Formal Controller Synthesis for Stochastic Dynamical Models with Epistemic Uncertainty
(
Poster
)
SlidesLive Video » Capturing both aleatoric and epistemic uncertainty in models of robotic systems is crucial to designing safe controllers. Most existing approaches for synthesizing certifiably safe controllers exclusively consider aleatoric but not epistemic uncertainty, thus requiring that model parameters and disturbances are known precisely. Our contribution to overcoming this restriction is a novel abstraction-based controller synthesis method for continuous-state models with stochastic noise, uncertain parameters, and external disturbances. By sampling techniques and robust analysis, we capture both aleatoric and epistemic uncertainty, with a user-specified confidence level, in the transition probability intervals of a so-called interval Markov decision process (iMDP). We then synthesize an optimal policy on this abstract iMDP, which translates (with the specified confidence level) to a feedback controller for the continuous model, with the same performance guarantees. Our experimental benchmarks confirm that accounting for epistemic uncertainty leads to controllers that are more robust against variations in parameter values. |
Thom Badings · Licio Romao · Alessandro Abate · Nils Jansen 🔗 |
-
|
A Benchmark for Out of Distribution Detection in Point Cloud 3D Semantic Segmentation
(
Poster
)
SlidesLive Video » Safety-critical applications like autonomous driving use Deep Neural Networks (DNNs) for object detection and segmentation. The DNNs fail to predict when theyobserve an Out-of-Distribution (OOD) input leading to catastrophic consequences.Existing OOD detection methods were extensively studied for image inputs but have not been explored much for LiDAR inputs. So in this study, we proposed two datasets for benchmarking OOD detection in 3D semantic segmentation. We used Maximum Softmax Probability and Entropy scores generated using Deep Ensembles and Flipout versions of RandLA-Net as OOD scores. We observed that Deep Ensembles out perform Flipout model in OOD detection with greater AUROC scores for both datasets. |
Lokesh Veeramacheneni · Matias Valdenegro-Toro 🔗 |
-
|
VIP: Towards Universal Visual Reward and Representation via Value-Implicit Pre-Training
(
Poster
)
SlidesLive Video » We introduce Value-Implicit Pre-training (VIP), a self-supervised pre-trained visual representation capable of generating dense and smooth reward functions for unseen robotic tasks. VIP casts representation learning from human videos as an offline goal-conditioned reinforcement learning problem and derives a self-supervised dual goal-conditioned value-function objective that does not depend on actions, enabling pre-training on unlabeled human videos. Theoretically, VIP can be understood as a novel implicit time contrastive learning that makes for temporally smooth embedding that enables the value function to be implicitly defined via the embedding distance, which can be used as the reward function for any downstream task specified through goal images. Trained on large-scale Ego4D human videos and without any fine-tuning on task-specific robot data, VIP's frozen representation can provide dense visual reward for an extensive set of simulated and real-robot tasks, enabling diverse reward-based policy learning methods, including visual trajectory optimization and online/offline RL, and significantly outperform all prior pre-trained representations. Notably, VIP can enable few-shot offline RL on a suite of real-world robot tasks with as few as 20 trajectories. |
Jason Yecheng Ma · Shagun Sodhani · Dinesh Jayaraman · Osbert Bastani · Vikash Kumar · Amy Zhang 🔗 |
-
|
Learning Certifiably Robust Controllers Using Fragile Perception
(
Poster
)
Advances in computer vision and machine learning enable robots to perceive their surroundings in powerful new ways, but these perception modules have well-known fragilities. We consider the problem of synthesizing a safe controller that is robust despite perception errors. The proposed method constructs a state estimator based on Gaussian processes with input-dependent noises. This estimator computes a high-confidence set for the actual state given a perceived state. Then, a robust neural network controller is synthesized that can provably handle the state uncertainty. Furthermore, an adaptive sampling algorithm is proposed to jointly improve the estimator and controller. Simulation experiments, including a realistic vision-based lane-keeping example in CARLA, illustrate the promise of the proposed approach in synthesizing robust controllers with deep-learning-based perception. |
Dawei Sun · Negin Musavi · Geir Dullerud · Sanjay Shakkottai · Sayan Mitra 🔗 |
-
|
PARTNR: Pick and place Ambiguity Resolving by Trustworthy iNteractive leaRning
(
Poster
)
SlidesLive Video » Several recent works show impressive results in mapping language-based human commands and image scene observations to direct robot executable policies (e.g., pick and place poses). However, these approaches do not consider the uncertainty of the trained policy and simply always execute actions suggested by the current policy as the most probable ones. This makes them vulnerable to domain shift and inefficient in the number of required demonstrations. We extend previous works and present the PARTNR algorithm that can detect ambiguities in the trained policy by analyzing multiple modalities in the pick and place poses using topological analysis. PARTNR employs an adaptive, sensitivity-based, gating function that decides if additional user demonstrations are required. User demonstrations are aggregated to the dataset and used for subsequent training. In this way, the policy can adapt promptly to domain shift and it can minimize the number of required demonstrations for a well-trained policy. The adaptive threshold enables to achieve the user-acceptable level of ambiguity to execute the policy autonomously and in turn, increase the trustworthiness of our system. We demonstrate the performance of PARTNR in a table-top pick and place task. |
Jelle Luijkx · Zlatan Ajanovic · Laura Ferranti · Jens Kober 🔗 |
-
|
Robust Forecasting for Robotic Control: A Game-Theoretic Approach
(
Poster
)
SlidesLive Video » Modern robots require accurate forecasts to make optimal decisions in the real world. For example, self-driving cars need an accurate forecast of other agents' future actions to plan safe trajectories. Current methods rely heavily on historical time series to accurately predict the future. However, relying entirely on the observed history is problematic since it could be corrupted by noise, have outliers, or not completely represent all possible outcomes. We propose a novel framework for generating robust forecasts for robotic control to solve this problem. To model real-world factors affecting future forecasts, we introduce the notion of an adversary, which perturbs observed historical time series to increase a robot's ultimate control cost. Specifically, we model this interaction as a zero-sum two-player game between a robot's forecaster and this hypothetical adversary. We show that our proposed game may be solved to a local Nash equilibrium using gradient-based optimization techniques. Furthermore, a forecaster trained with our method performs 30.14% better on out-of-distribution real-world lane change data than baselines. |
Shubhankar Agarwal · David Fridovich-Keil · Sandeep Chinchali 🔗 |
-
|
DALL-E-Bot: Introducing Web-Scale Diffusion Models to Robotics
(
Poster
)
SlidesLive Video » We introduce the first work to explore web-scale diffusion models for robotics. DALL-E-Bot enables a robot to rearrange objects in a scene, by first inferring a text description of those objects, then generating an image representing a human-like arrangement of those objects, and finally physically arranging the objects according to that image. Our implementation achieves this zero-shot using DALL-E, without any further data collection or training. Strong real-world results with human studies show that this is an exciting direction for future generations of robot learning algorithms. We propose a list of recommendations to the community for further developments in this direction. Videos: https://www.robot-learning.uk/dall-e-bot |
Ivan Kapelyukh · Vitalis Vosylius · Edward Johns 🔗 |
-
|
MAEA: Multimodal Attribution Framework for Embodied AI
(
Poster
)
SlidesLive Video » Understanding multimodal perception for embodied AI is an open question because such inputs may contain highly complementary as well as redundant information for the task. A relevant direction for multimodal policies is understanding the global trends of each modality at the fusion layer. To this end, we disentangle the attributions for visual, language, and previous action inputs across different policies trained on the ALFRED dataset. Attribution analysis can be utilized to rank and group the failure scenarios, investigate modeling and dataset biases, and critically analyze multimodal EAI policies for robustness and user trust before deployment. We present MAFEA, a framework to compute global attributions per modality of any differentiable policy. In addition, we show how attributions enable lower-level behavior analysis in EAI policies through two example case studies on language and visual attributions. |
Vidhi Jain · Jayant Sravan Tamarapalli · Sahiti Yerramilli · Yonatan Bisk 🔗 |
-
|
Safety-Guaranteed Skill Discovery for Robot Manipulation Tasks
(
Poster
)
SlidesLive Video » Recent progress in unsupervised skill discovery algorithms has shown great promise in learning an extensive collection of behaviors without extrinsic supervision. On the other hand, safety is one of the most critical factors for real-world robot applications. As skill discovery methods typically encourage exploratory and dynamic behaviors, it can often be the case that a large portion of learned skills remains too dangerous and unsafe. In this paper, we introduce the novel problem of safe skill discovery, which aims at learning, in a task-agnostic fashion, a repertoire of reusable skills that is inherently safe to be composed for solving downstream tasks. We propose \textit{Safety-Guaranteed Skill Discovery} (SGSD), an algorithm that learns a latent-conditioned skill-policy, regularized with a safety-critic modelinga user-defined safety definition. Using the pretrained safe skill repertoire, hierarchical reinforcement learning can solve downstream tasks without the need of explicit consideration of safety during training and testing. We evaluate our algorithm on a collection of force-controlled robotic manipulation tasks in simulation and show promising downstream task performance with safety guarantees.Please find \url{https://sites.google.com/view/safe-skill} for supplementary videos. |
Sunin Kim · Jaewoon Kwon · Taeyoon Lee · Younghyo Park · Julien PEREZ 🔗 |
-
|
Insights towards Sim2Real Contact-Rich Manipulation
(
Poster
)
SlidesLive Video » Recent work has shown promise towards training policies for contact-rich tasks in simulation with the hope that they can be transferred to the real world. However, to close the sim2real gap, it is important to consider the effects of partial observability that are unavoidable in the real world and particularly relevant when dealing with small parts that require precise manipulation. In this work, we perform a detailed simulation-based analysis of how pose-estimation error, object geometry variability, and controller variability affect deep reinforcement learning algorithms. We show that using asymmetric actor-critic architectures leads to more robust training under noise. |
Michael Noseworthy · Iretiayo Akinola · Yashraj Narang · Fabio Ramos · Lucas Manuelli · Ankur Handa · Dieter Fox 🔗 |
-
|
Train Offline, Test Online: A Real Robot Learning Benchmark
(
Poster
)
SlidesLive Video » Three challenges limit the progress of robot learning research: robots are expensive (few labs can participate), everyone uses different robots (findings do not generalize across labs), and we lack internet-scale robotics data. We take on these challenges via a new benchmark: Train Offline, Test Online (TOTO). TOTO provides remote users with access to shared robots for evaluating methods on common tasks and an open-source dataset of these tasks for offline training. Its manipulation task suite requires challenging generalization to unseen objects, positions, and lighting. We present initial results on TOTO comparing five pretrained visual representations and four offline policy learning baselines, remotely contributed by five institutions. The real promise of TOTO, however, lies in the future: we release the benchmark for additional submissions from any user, enabling easy, direct comparison to several methods without the need to obtain hardware or collect data. |
Gaoyue Zhou · Victoria Dean · Mohan Kumar Srirama · Aravind Rajeswaran · Jyothish Pari · Kyle Hatch · Aryan Jain · Tianhe Yu · Pieter Abbeel · Lerrel Pinto · Chelsea Finn · Abhinav Gupta
|
-
|
Learning a Meta-Controller for Dynamic Grasping
(
Poster
)
SlidesLive Video » Grasping moving objects is a challenging task that combines multiple modules such as object pose predictor, arm motion planner, etc. Each module operates under its own set of meta-parameters, for example, the prediction horizon in the pose predictor and the time budget for planning motion in the motion planner. Many previous works assign fixed values to these parameters either heuristically or through grid search; however, at different time steps within a single episode of dynamic grasping, there should be different optimal values for each parameter, depending on the current scene. In this work, we learn a meta-controller through reinforcement learning to control the prediction horizon and time budget dynamically at each time step. Our experiments show that the meta-controller significantly improves the grasping success rate and reduces grasping time, compared to baselines whose parameters are fixed or random. Our meta-controller learns to reason about the reachable workspace and through dynamically controlling the meta-parameters, it maintains the predicted pose and the planned motion within the reachable region. It also generalizes to different environment setups and can handle various target motions and obstacles. |
Yinsen Jia · Jingxi Xu · Dinesh Jayaraman · Shuran Song 🔗 |
-
|
Real World Offline Reinforcement Learning with Realistic Data Source
(
Poster
)
SlidesLive Video » Offline reinforcement learning (ORL) holds great promise for robot learning due to its ability to learn from arbitrary pre-generated experience. However, current ORL benchmarks are almost entirely in simulation and utilize contrived datasets like replay buffers of online RL agents or sub-optimal trajectories, and thus hold limited relevance for real-world robotics. In this work (Real-ORL), we posit that data collected from safe operations of closely related tasks are more practical data sources for real-world robot learning. Under these settings, we perform an extensive (6500+ trajectories collected over 800+ robot hours and 270+ human labor hour) empirical study evaluating generalization and transfer capabilities of representative ORL methods on four real-world tabletop manipulation tasks. Our study finds that ORL and imitation learning prefer different action spaces, and that ORL algorithms can generalize from leveraging offline heterogeneous data sources and outperform imitation learning. We release our dataset and implementations at URL: https://sites.google.com/view/real-orl |
Gaoyue Zhou · Liyiming Ke · Siddhartha Srinivasa · Abhinav Gupta · Aravind Rajeswaran · Vikash Kumar 🔗 |
-
|
Interactive Language: Talking to Robots in Real Time
(
Poster
)
SlidesLive Video » We present a framework for building interactive, real-time, natural language-instructable robots in the real world, and we open source related assets (dataset, environment, benchmark, and policies). Trained with behavioral cloning on a dataset of hundreds of thousands of language-annotated trajectories, a produced policy can proficiently execute an order of magnitude more commands than previous works: specifically we estimate a 93.5% success rate on a set of 87,000 unique natural language strings specifying raw end-to-end visuo-linguo-motor skills in the real world. We find that the same policy is capable of being guided by a human via real-time language to address a wide range of precise long-horizon rearrangement goals, e.g. "make a smiley face out of blocks". The dataset we release comprises nearly 600,000 language-labeled trajectories, an order of magnitude larger than prior available datasets. We hope the demonstrated results and associated assets enable further advancement of helpful, capable, natural-language-interactable robots. See videos at https://sites.google.com/view/interactive-language. |
Corey Lynch · Pete Florence · Jonathan Tompson · Ayzaan Wahid · Tianli Ding · James Betker · Robert Baruch · Travis Armstrong 🔗 |
-
|
Robotic Skill Acquistion via Instruction Augmentation with Vision-Language Models
(
Poster
)
SlidesLive Video » In recent years, much progress has been made in learning robotic manipulation policies that follow natural language instructions. Such methods typically learn from corpora of robot-language data that was either collected with specific tasks in mind or expensively re-labelled by humans with rich language descriptions in hindsight. Recently, large-scale pretrained vision-language models (VLMs) like CLIP or ViLD have been applied to robotics for learning representations and scene descriptors. Can these pretrained models serve as automatic labelers for robot data, effectively importing Internet-scale knowledge into existing datasets to make them useful even for tasks that are not reflected in their ground truth annotations? For example, if the original annotations contained simple task descriptions such as "pick up the apple", a pretrained VLM-based labeller could significantly expand the number of semantic concepts available in the data and introduce spatial concepts such as "the apple on the right side of the table" or alternative phrasings such as "the red colored fruit". To accomplish this, we introduce Data-driven Instruction Augmentation for Language-conditioned control (DIAL): we utilize semi-supervised language labels leveraging the semantic understanding of CLIP to propagate knowledge onto large datasets of unlabelled demonstration data and then train language-conditioned policies on the augmented datasets. This method enables cheaper acquisition of useful language descriptions compared to expensive human labels, allowing for more efficient label coverage of large-scale datasets. We apply DIAL to a challenging real-world robotic manipulation domain where 96.5% of the 80,000 demonstrations do not contain crowd-sourced language annotations. DIAL enables imitation learning policies to acquire new capabilities and generalize to 60 novel instructions unseen in the original dataset. |
Ted Xiao · Harris Chan · Pierre Sermanet · Ayzaan Wahid · Anthony Brohan · Karol Hausman · Sergey Levine · Jonathan Tompson 🔗 |
-
|
Certifiably-correct Control Policies for Safe Learning and Adaptation in Assistive Robotics
(
Poster
)
Guaranteeing safety in human-centric applications is critical in robot learning as the learned policies may demonstrate unsafe behaviors in formerly unseen scenarios. We present a framework to locally repair an erroneous policy network to satisfy a set of formal safety constraints using Mixed Integer Quadratic Programming (MIQP). Our MIQP formulation explicitly imposes the safety constraints to the learned policy while minimizing the original loss function. The policy network is then verified to be locally safe. We demonstrate the application of our framework to derive safe policies for a robotic lower-leg prosthesis. |
Keyvan Majd · GEOFFEY CLARK · Tanmay Khandait · · Sriram Sankaranarayanan · Georgios Fainekos · Heni Ben Amor 🔗 |
-
|
Capsa: A Unified Framework for Quantifying Risk in Deep Neural Networks
(
Poster
)
SlidesLive Video » The deployment of large-scale deep neural networks in safety-critical scenariosrequires quantifiably calibrated and reliable measures of trust. Unfortunately,existing algorithms to achieve risk-awareness are complex and adhoc. We presentcapsa, an open-source and flexible framework for unifying these methods andcreating risk-aware models. We unify state-of-the-art risk algorithms under thecapsa framework, propose a composability method for combining different riskestimators together in a single function set, and benchmark on high-dimensionalperception tasks. Code is available at: https://github.com/themis-ai/capsa |
Sadhana Lolla · Iaroslav Elistratov · Alejandro Perez · Elaheh Ahmadi · Daniela Rus · Alexander Amini 🔗 |
-
|
Language as Robot Middleware - Andy Zeng & Jacky Liang
(
Discussion Panel: Scaling + Models
)
link »
SlidesLive Video » |
Hamidreza Kasaei 🔗 |
-
|
Angela Schoellig
(
Discussion Panel: Uncertainty-Aware ML for Robotics
)
|
Hamidreza Kasaei 🔗 |
-
|
Uncertainty Aware Machine Learning for Robotics - Stefanie Tellex
(
Discussion Panel: Uncertainty-Aware ML for Robotics
)
|
Hamidreza Kasaei 🔗 |
-
|
Real Robots Learn with Structure - Georgia Chalvatzaki ( Discussion Panel: Uncertainty-Aware ML for Robotics ) link » | Hamidreza Kasaei 🔗 |
-
|
My Hopes and Dreams of Communicating with Machines and Where to Begin - Beem Kim ( Discussion Panel: Explainability/Predictability in Robotics ) link » | Hamidreza Kasaei 🔗 |
-
|
Why Robots Need Social Skills - Leila Takayama
(
Discussion Panel: Explainability/Predictability in Robotics
)
SlidesLive Video » |
Hamidreza Kasaei 🔗 |
-
|
Representing Interactions for Robot Navigation - Katherine Driggs-Campbell & Zhe Huang
(
Discussion Panel: Explainability/Predictability in Robotics
)
SlidesLive Video » |
Hamidreza Kasaei 🔗 |
-
|
Towards Certifiably Safe Nonlinear Control with Sensor and Dynamics Uncertainties - Sarah Dean & Andrew Taylor
(
Discussion Panel: Safety & Verification for Decision-Making
)
SlidesLive Video » |
Hamidreza Kasaei 🔗 |
-
|
A Vision for Certifiable Perception: from Outlier-Robust Estimation to Self-Supervised Learning - Luca Carlone & Rajat Talak
(
Discussion Panel: Safety & Verification for Decision-Making
)
SlidesLive Video » |
Hamidreza Kasaei 🔗 |
-
|
Self-driving Cars - Matthew Johnson-Roberson
(
Discussion Panel: Safety & Verification for Decision-Making
)
SlidesLive Video » |
Hamidreza Kasaei 🔗 |
-
|
Trustworthy AI Robotics for Real-world Logistics - Anusha Nagabandi
(
Discussion Panel: Scaling + Models
)
link »
SlidesLive Video » |
Hamidreza Kasaei 🔗 |
-
|
A Generalist Agent (GATO) - Scott Reed & Gabriel Barth-Maron & Jackie Kay
(
Discussion Panel: Scaling + Models
)
SlidesLive Video » |
Hamidreza Kasaei 🔗 |
-
|
Language as a Connective Tissue for Robotics - Karol Hausman & Brian Ichter
(
Discussion Panel: Safety & Verification for Decision-Making
)
link »
SlidesLive Video » |
Hamidreza Kasaei 🔗 |
-
|
Visual Backtracking Teleoperation: A Data Collection Protocol for Offline Image-Based RL
(
Oral
)
We consider how to most efficiently leverage teleoperator time to collect data for learning robust image-based value functions and policies for sparse reward robotic tasks. To accomplish this goal, we modify the process of data collection to include more than just successful demonstrations of the desired task. Instead we develop a novel protocol that we call Visual Backtracking Teleoperation (VBT), which deliberately collects a dataset of visually similar failures, recoveries, and successes. VBT data collection is particularly useful for efficiently learning accurate value functions from small datasets of image-based observations. We demonstrate VBT on a real robot to perform continuous control from image observations for the deformable manipulation task of T-shirt grasping. We find that by adjusting the data collection process we improve the quality of both the learned value functions and policies over a variety of baseline methods for data collection. Specifically, we find that offline reinforcement learning on VBT data outperforms standard behavior cloning on successful demonstration data by 13% when both methods are given equal-sized datasets of 60 minutes of data from the real robot. |
David Brandfonbrener · Stephen Tu · Avi Singh · Stefan Welker · Chad Boodoo · Nikolai Matni · Jacob Varley 🔗 |
-
|
Conformal Semantic Keypoint Detection with Statistical Guarantees
(
Oral
)
Detecting semantic keypoints is a critical intermediate task for object detection and pose estimation from images. Existing approaches, albeit performing well on standard benchmarks, offer no provable guarantees on the quality of the detection. In this paper, we apply the statistical machinery of inductive conformal prediction that, given a calibration dataset (e.g., 200 images) and a nonconformity function, converts a heuristic heatmap detection into a prediction set that provably covers the true keypoint location with a user-specified probability (e.g., 90%). We design three different nonconformity functions leading to circular or elliptical prediction sets that are easy to compute. On the LINEMOD Occluded dataset we demonstrate that (i) the empirical coverage rate of the prediction sets is valid; (ii) the prediction sets are tight, e.g., a ball with radius 10 pixels covers true keypoint locations on most test images; and (iii) the prediction sets are adaptive, i.e., their sizes become larger for keypoints that are difficult to detect and smaller for easy instances. |
Heng Yang · Marco Pavone 🔗 |
-
|
A Contextual Bandit Approach for Learning to Plan in Environments with Probabilistic Goal Configurations
(
Oral
)
Object-goal navigation (Object-nav) entails searching, recognizing and navigating to a target object. Object-nav has been extensively studied by the Embodied-AI community, but most solutions are often restricted to considering static objects (e.g., television, fridge, etc.). We propose a modular framework for object-nav that is able to efficiently search indoor environments for not just static objects but also movable objects (e.g. fruits, glasses, phones, etc.) that frequently change their positions due to human interaction. Our contextual-bandit agent efficiently explores the environment by showing optimism in the face of uncertainty and learns a model of the likelihood of spotting different objects from each navigable location. The likelihoods are used as rewards in a weighted minimum latency solver to deduce a trajectory for the robot. We evaluate our algorithms in two simulated environments and a real-world setting, to demonstrate high sample efficiency and reliability. |
Sohan Rudra · Saksham Goel · Anirban Santara · Claudio Gentile · Laurent Perron · Fei Xia · Vikas Sindhwani · Carolina Parada · Gaurav Aggarwal 🔗 |
-
|
Imitating careful experts to avoid catastrophic events
(
Oral
)
Reinforcement learning (RL) is increasingly being used to control robotic systems that interact closely with humans. This interaction raises the problem of safe RL: how to ensure that an RL-controlled robotic system never, for instance, injures a human. This problem is especially challenging in rich, realistic settings where it is not even possible to clearly write down a reward function which incorporates these outcomes. In these circumstances, perhaps the only viable approach is based on inverse reinforcement learning (IRL), which infers rewards from human demonstrations. However, IRL is massively underdetermined as many different rewards can lead to the same optimal policies; we show that this makes it difficult to distinguish catastrophic outcomes (such as injuring a human) from merely undesirable outcomes. Our key insight is that humans do display different behaviour when catastrophic outcomes are possible: they become much more careful. We incorporate carefulness signals into IRL, and find that they do indeed allow IRL to disambiguate undesirable from catastrophic outcomes, which is critical to ensuring safety in future real-world human-robot interactions. |
Jack Hanslope · Laurence Aitchison 🔗 |
-
|
Infrastructure-based End-to-End Learning and Prevention of Driver Failure
(
Oral
)
Intelligent intersection managers can improve safety by detecting dangerous drivers or failure modes in autonomous vehicles, warning oncoming vehicles as they approach an intersection. In this work, we present FailureNet, a recurrent neural network trained end-to-end on trajectories of both nominal and reckless drivers in a scaled miniature city. FailureNet observes the poses of vehicles as they approach an intersection and detects whether a failure is present in the autonomy stack, warning cross-traffic of potentially dangerous drivers. FailureNet can accurately identify control failures, upstream perception errors, and speeding drivers, distinguishing them from nominal driving. The network is trained and deployed with autonomous vehicles in a scaled miniature city. Compared to speed or frequency-based predictors, FailureNet's recurrent neural network structure provides improved predictive power, yielding upwards of 84% accuracy when deployed on hardware. |
Noam Buckman · Shiva Sreeram · Mathias Lechner · Yutong Ban · Ramin Hasani · Sertac Karaman · Daniela Rus 🔗 |
-
|
Formal Controller Synthesis for Stochastic Dynamical Models with Epistemic Uncertainty
(
Oral
)
Capturing both aleatoric and epistemic uncertainty in models of robotic systems is crucial to designing safe controllers. Most existing approaches for synthesizing certifiably safe controllers exclusively consider aleatoric but not epistemic uncertainty, thus requiring that model parameters and disturbances are known precisely. Our contribution to overcoming this restriction is a novel abstraction-based controller synthesis method for continuous-state models with stochastic noise, uncertain parameters, and external disturbances. By sampling techniques and robust analysis, we capture both aleatoric and epistemic uncertainty, with a user-specified confidence level, in the transition probability intervals of a so-called interval Markov decision process (iMDP). We then synthesize an optimal policy on this abstract iMDP, which translates (with the specified confidence level) to a feedback controller for the continuous model, with the same performance guarantees. Our experimental benchmarks confirm that accounting for epistemic uncertainty leads to controllers that are more robust against variations in parameter values. |
Thom Badings · Licio Romao · Alessandro Abate · Nils Jansen 🔗 |
-
|
A Benchmark for Out of Distribution Detection in Point Cloud 3D Semantic Segmentation
(
Oral
)
Safety-critical applications like autonomous driving use Deep Neural Networks (DNNs) for object detection and segmentation. The DNNs fail to predict when theyobserve an Out-of-Distribution (OOD) input leading to catastrophic consequences.Existing OOD detection methods were extensively studied for image inputs but have not been explored much for LiDAR inputs. So in this study, we proposed two datasets for benchmarking OOD detection in 3D semantic segmentation. We used Maximum Softmax Probability and Entropy scores generated using Deep Ensembles and Flipout versions of RandLA-Net as OOD scores. We observed that Deep Ensembles out perform Flipout model in OOD detection with greater AUROC scores for both datasets. |
Lokesh Veeramacheneni · Matias Valdenegro-Toro 🔗 |
-
|
VIP: Towards Universal Visual Reward and Representation via Value-Implicit Pre-Training
(
Oral
)
We introduce Value-Implicit Pre-training (VIP), a self-supervised pre-trained visual representation capable of generating dense and smooth reward functions for unseen robotic tasks. VIP casts representation learning from human videos as an offline goal-conditioned reinforcement learning problem and derives a self-supervised dual goal-conditioned value-function objective that does not depend on actions, enabling pre-training on unlabeled human videos. Theoretically, VIP can be understood as a novel implicit time contrastive learning that makes for temporally smooth embedding that enables the value function to be implicitly defined via the embedding distance, which can be used as the reward function for any downstream task specified through goal images. Trained on large-scale Ego4D human videos and without any fine-tuning on task-specific robot data, VIP's frozen representation can provide dense visual reward for an extensive set of simulated and real-robot tasks, enabling diverse reward-based policy learning methods, including visual trajectory optimization and online/offline RL, and significantly outperform all prior pre-trained representations. Notably, VIP can enable few-shot offline RL on a suite of real-world robot tasks with as few as 20 trajectories. |
Jason Yecheng Ma · Shagun Sodhani · Dinesh Jayaraman · Osbert Bastani · Vikash Kumar · Amy Zhang 🔗 |
-
|
Learning Certifiably Robust Controllers Using Fragile Perception
(
Oral
)
Advances in computer vision and machine learning enable robots to perceive their surroundings in powerful new ways, but these perception modules have well-known fragilities. We consider the problem of synthesizing a safe controller that is robust despite perception errors. The proposed method constructs a state estimator based on Gaussian processes with input-dependent noises. This estimator computes a high-confidence set for the actual state given a perceived state. Then, a robust neural network controller is synthesized that can provably handle the state uncertainty. Furthermore, an adaptive sampling algorithm is proposed to jointly improve the estimator and controller. Simulation experiments, including a realistic vision-based lane-keeping example in CARLA, illustrate the promise of the proposed approach in synthesizing robust controllers with deep-learning-based perception. |
Dawei Sun · Negin Musavi · Geir Dullerud · Sanjay Shakkottai · Sayan Mitra 🔗 |
-
|
PARTNR: Pick and place Ambiguity Resolving by Trustworthy iNteractive leaRning
(
Oral
)
Several recent works show impressive results in mapping language-based human commands and image scene observations to direct robot executable policies (e.g., pick and place poses). However, these approaches do not consider the uncertainty of the trained policy and simply always execute actions suggested by the current policy as the most probable ones. This makes them vulnerable to domain shift and inefficient in the number of required demonstrations. We extend previous works and present the PARTNR algorithm that can detect ambiguities in the trained policy by analyzing multiple modalities in the pick and place poses using topological analysis. PARTNR employs an adaptive, sensitivity-based, gating function that decides if additional user demonstrations are required. User demonstrations are aggregated to the dataset and used for subsequent training. In this way, the policy can adapt promptly to domain shift and it can minimize the number of required demonstrations for a well-trained policy. The adaptive threshold enables to achieve the user-acceptable level of ambiguity to execute the policy autonomously and in turn, increase the trustworthiness of our system. We demonstrate the performance of PARTNR in a table-top pick and place task. |
Jelle Luijkx · Zlatan Ajanovic · Laura Ferranti · Jens Kober 🔗 |
-
|
MAEA: Multimodal Attribution Framework for Embodied AI
(
Oral
)
Understanding multimodal perception for embodied AI is an open question because such inputs may contain highly complementary as well as redundant information for the task. A relevant direction for multimodal policies is understanding the global trends of each modality at the fusion layer. To this end, we disentangle the attributions for visual, language, and previous action inputs across different policies trained on the ALFRED dataset. Attribution analysis can be utilized to rank and group the failure scenarios, investigate modeling and dataset biases, and critically analyze multimodal EAI policies for robustness and user trust before deployment. We present MAFEA, a framework to compute global attributions per modality of any differentiable policy. In addition, we show how attributions enable lower-level behavior analysis in EAI policies through two example case studies on language and visual attributions. |
Vidhi Jain · Jayant Sravan Tamarapalli · Sahiti Yerramilli · Yonatan Bisk 🔗 |
-
|
Safety-Guaranteed Skill Discovery for Robot Manipulation Tasks
(
Oral
)
Recent progress in unsupervised skill discovery algorithms has shown great promise in learning an extensive collection of behaviors without extrinsic supervision. On the other hand, safety is one of the most critical factors for real-world robot applications. As skill discovery methods typically encourage exploratory and dynamic behaviors, it can often be the case that a large portion of learned skills remains too dangerous and unsafe. In this paper, we introduce the novel problem of safe skill discovery, which aims at learning, in a task-agnostic fashion, a repertoire of reusable skills that is inherently safe to be composed for solving downstream tasks. We propose \textit{Safety-Guaranteed Skill Discovery} (SGSD), an algorithm that learns a latent-conditioned skill-policy, regularized with a safety-critic modelinga user-defined safety definition. Using the pretrained safe skill repertoire, hierarchical reinforcement learning can solve downstream tasks without the need of explicit consideration of safety during training and testing. We evaluate our algorithm on a collection of force-controlled robotic manipulation tasks in simulation and show promising downstream task performance with safety guarantees.Please find \url{https://sites.google.com/view/safe-skill} for supplementary videos. |
Sunin Kim · Jaewoon Kwon · Taeyoon Lee · Younghyo Park · Julien PEREZ 🔗 |
-
|
Insights towards Sim2Real Contact-Rich Manipulation
(
Oral
)
Recent work has shown promise towards training policies for contact-rich tasks in simulation with the hope that they can be transferred to the real world. However, to close the sim2real gap, it is important to consider the effects of partial observability that are unavoidable in the real world and particularly relevant when dealing with small parts that require precise manipulation. In this work, we perform a detailed simulation-based analysis of how pose-estimation error, object geometry variability, and controller variability affect deep reinforcement learning algorithms. We show that using asymmetric actor-critic architectures leads to more robust training under noise. |
Michael Noseworthy · Iretiayo Akinola · Yashraj Narang · Fabio Ramos · Lucas Manuelli · Ankur Handa · Dieter Fox 🔗 |
-
|
Train Offline, Test Online: A Real Robot Learning Benchmark
(
Oral
)
Three challenges limit the progress of robot learning research: robots are expensive (few labs can participate), everyone uses different robots (findings do not generalize across labs), and we lack internet-scale robotics data. We take on these challenges via a new benchmark: Train Offline, Test Online (TOTO). TOTO provides remote users with access to shared robots for evaluating methods on common tasks and an open-source dataset of these tasks for offline training. Its manipulation task suite requires challenging generalization to unseen objects, positions, and lighting. We present initial results on TOTO comparing five pretrained visual representations and four offline policy learning baselines, remotely contributed by five institutions. The real promise of TOTO, however, lies in the future: we release the benchmark for additional submissions from any user, enabling easy, direct comparison to several methods without the need to obtain hardware or collect data. |
Gaoyue Zhou · Victoria Dean · Mohan Kumar Srirama · Aravind Rajeswaran · Jyothish Pari · Kyle Hatch · Aryan Jain · Tianhe Yu · Pieter Abbeel · Lerrel Pinto · Chelsea Finn · Abhinav Gupta
|
-
|
Learning a Meta-Controller for Dynamic Grasping
(
Oral
)
Grasping moving objects is a challenging task that combines multiple modules such as object pose predictor, arm motion planner, etc. Each module operates under its own set of meta-parameters, for example, the prediction horizon in the pose predictor and the time budget for planning motion in the motion planner. Many previous works assign fixed values to these parameters either heuristically or through grid search; however, at different time steps within a single episode of dynamic grasping, there should be different optimal values for each parameter, depending on the current scene. In this work, we learn a meta-controller through reinforcement learning to control the prediction horizon and time budget dynamically at each time step. Our experiments show that the meta-controller significantly improves the grasping success rate and reduces grasping time, compared to baselines whose parameters are fixed or random. Our meta-controller learns to reason about the reachable workspace and through dynamically controlling the meta-parameters, it maintains the predicted pose and the planned motion within the reachable region. It also generalizes to different environment setups and can handle various target motions and obstacles. |
Yinsen Jia · Jingxi Xu · Dinesh Jayaraman · Shuran Song 🔗 |
-
|
Real World Offline Reinforcement Learning with Realistic Data Source
(
Oral
)
Offline reinforcement learning (ORL) holds great promise for robot learning due to its ability to learn from arbitrary pre-generated experience. However, current ORL benchmarks are almost entirely in simulation and utilize contrived datasets like replay buffers of online RL agents or sub-optimal trajectories, and thus hold limited relevance for real-world robotics. In this work (Real-ORL), we posit that data collected from safe operations of closely related tasks are more practical data sources for real-world robot learning. Under these settings, we perform an extensive (6500+ trajectories collected over 800+ robot hours and 270+ human labor hour) empirical study evaluating generalization and transfer capabilities of representative ORL methods on four real-world tabletop manipulation tasks. Our study finds that ORL and imitation learning prefer different action spaces, and that ORL algorithms can generalize from leveraging offline heterogeneous data sources and outperform imitation learning. We release our dataset and implementations at URL: https://sites.google.com/view/real-orl |
Gaoyue Zhou · Liyiming Ke · Siddhartha Srinivasa · Abhinav Gupta · Aravind Rajeswaran · Vikash Kumar 🔗 |
-
|
Interactive Language: Talking to Robots in Real Time
(
Oral
)
We present a framework for building interactive, real-time, natural language-instructable robots in the real world, and we open source related assets (dataset, environment, benchmark, and policies). Trained with behavioral cloning on a dataset of hundreds of thousands of language-annotated trajectories, a produced policy can proficiently execute an order of magnitude more commands than previous works: specifically we estimate a 93.5% success rate on a set of 87,000 unique natural language strings specifying raw end-to-end visuo-linguo-motor skills in the real world. We find that the same policy is capable of being guided by a human via real-time language to address a wide range of precise long-horizon rearrangement goals, e.g. "make a smiley face out of blocks". The dataset we release comprises nearly 600,000 language-labeled trajectories, an order of magnitude larger than prior available datasets. We hope the demonstrated results and associated assets enable further advancement of helpful, capable, natural-language-interactable robots. See videos at https://sites.google.com/view/interactive-language. |
Corey Lynch · Pete Florence · Jonathan Tompson · Ayzaan Wahid · Tianli Ding · James Betker · Robert Baruch · Travis Armstrong 🔗 |
-
|
Robotic Skill Acquistion via Instruction Augmentation with Vision-Language Models
(
Oral
)
In recent years, much progress has been made in learning robotic manipulation policies that follow natural language instructions. Such methods typically learn from corpora of robot-language data that was either collected with specific tasks in mind or expensively re-labelled by humans with rich language descriptions in hindsight. Recently, large-scale pretrained vision-language models (VLMs) like CLIP or ViLD have been applied to robotics for learning representations and scene descriptors. Can these pretrained models serve as automatic labelers for robot data, effectively importing Internet-scale knowledge into existing datasets to make them useful even for tasks that are not reflected in their ground truth annotations? For example, if the original annotations contained simple task descriptions such as "pick up the apple", a pretrained VLM-based labeller could significantly expand the number of semantic concepts available in the data and introduce spatial concepts such as "the apple on the right side of the table" or alternative phrasings such as "the red colored fruit". To accomplish this, we introduce Data-driven Instruction Augmentation for Language-conditioned control (DIAL): we utilize semi-supervised language labels leveraging the semantic understanding of CLIP to propagate knowledge onto large datasets of unlabelled demonstration data and then train language-conditioned policies on the augmented datasets. This method enables cheaper acquisition of useful language descriptions compared to expensive human labels, allowing for more efficient label coverage of large-scale datasets. We apply DIAL to a challenging real-world robotic manipulation domain where 96.5% of the 80,000 demonstrations do not contain crowd-sourced language annotations. DIAL enables imitation learning policies to acquire new capabilities and generalize to 60 novel instructions unseen in the original dataset. |
Ted Xiao · Harris Chan · Pierre Sermanet · Ayzaan Wahid · Anthony Brohan · Karol Hausman · Sergey Levine · Jonathan Tompson 🔗 |
-
|
Capsa: A Unified Framework for Quantifying Risk in Deep Neural Networks
(
Oral
)
The deployment of large-scale deep neural networks in safety-critical scenariosrequires quantifiably calibrated and reliable measures of trust. Unfortunately,existing algorithms to achieve risk-awareness are complex and adhoc. We presentcapsa, an open-source and flexible framework for unifying these methods andcreating risk-aware models. We unify state-of-the-art risk algorithms under thecapsa framework, propose a composability method for combining different riskestimators together in a single function set, and benchmark on high-dimensionalperception tasks. Code is available at: https://github.com/themis-ai/capsa |
Sadhana Lolla · Iaroslav Elistratov · Alejandro Perez · Elaheh Ahmadi · Daniela Rus · Alexander Amini 🔗 |
-
|
Robust Forecasting for Robotic Control: A Game-Theoretic Approach
(
Oral
)
Modern robots require accurate forecasts to make optimal decisions in the real world. For example, self-driving cars need an accurate forecast of other agents' future actions to plan safe trajectories. Current methods rely heavily on historical time series to accurately predict the future. However, relying entirely on the observed history is problematic since it could be corrupted by noise, have outliers, or not completely represent all possible outcomes. We propose a novel framework for generating robust forecasts for robotic control to solve this problem. To model real-world factors affecting future forecasts, we introduce the notion of an adversary, which perturbs observed historical time series to increase a robot's ultimate control cost. Specifically, we model this interaction as a zero-sum two-player game between a robot's forecaster and this hypothetical adversary. We show that our proposed game may be solved to a local Nash equilibrium using gradient-based optimization techniques. Furthermore, a forecaster trained with our method performs 30.14% better on out-of-distribution real-world lane change data than baselines. |
Shubhankar Agarwal · David Fridovich-Keil · Sandeep Chinchali 🔗 |
-
|
DALL-E-Bot: Introducing Web-Scale Diffusion Models to Robotics
(
Oral
)
We introduce the first work to explore web-scale diffusion models for robotics. DALL-E-Bot enables a robot to rearrange objects in a scene, by first inferring a text description of those objects, then generating an image representing a human-like arrangement of those objects, and finally physically arranging the objects according to that image. Our implementation achieves this zero-shot using DALL-E, without any further data collection or training. Strong real-world results with human studies show that this is an exciting direction for future generations of robot learning algorithms. We propose a list of recommendations to the community for further developments in this direction. Videos: https://www.robot-learning.uk/dall-e-bot |
Ivan Kapelyukh · Vitalis Vosylius · Edward Johns 🔗 |
-
|
Certifiably-correct Control Policies for Safe Learning and Adaptation in Assistive Robotics
(
Oral
)
Guaranteeing safety in human-centric applications is critical in robot learning as the learned policies may demonstrate unsafe behaviors in formerly unseen scenarios. We present a framework to locally repair an erroneous policy network to satisfy a set of formal safety constraints using Mixed Integer Quadratic Programming (MIQP). Our MIQP formulation explicitly imposes the safety constraints to the learned policy while minimizing the original loss function. The policy network is then verified to be locally safe. We demonstrate the application of our framework to derive safe policies for a robotic lower-leg prosthesis. |
Keyvan Majd · GEOFFEY CLARK · Tanmay Khandait · · Sriram Sankaranarayanan · Georgios Fainekos · Heni Ben Amor 🔗 |
Author Information
Alex Bewley (Google)
Roberto Calandra (Facebook)
Anca Dragan (UC Berkeley)
Igor Gilitschenski (TRI)
Emily Hannigan (Columbia University)
Masha Itkina (Stanford University)
Hamidreza Kasaei (Dept. of AI, University of Groningen)
Hamidreza Kasaei is an Assistant Professor in the Department of Artificial Intelligence at the University of Groningen, the Netherlands. His research group focuses on Lifelong Interactive Robot Learning (IRL-Lab). These days, Hamidreza is particularly interested in enabling robots to incrementally learn from past experiences and intelligently and safely interact with non-expert human users using data-efficient open-ended machine-learning techniques.
Jens Kober (TU Delft)
Danica Kragic (KTH Royal Institute of Technology)
Nathan Lambert (Hugging Face)
Julien PEREZ (NAVER LABS Europe)
Fabio Ramos (University of Sydney, NVIDIA)
Ransalu Senanayake (Stanford University)
Jonathan Tompson (Google Brain)
Vincent Vanhoucke (Google)
Markus Wulfmeier (DeepMind)
More from the Same Authors
-
2021 : B-Pref: Benchmarking Preference-Based Reinforcement Learning »
Kimin Lee · Laura Smith · Anca Dragan · Pieter Abbeel -
2021 Spotlight: Pragmatic Image Compression for Human-in-the-Loop Decision-Making »
Sid Reddy · Anca Dragan · Sergey Levine -
2021 : Dynamic Environments with Deformable Objects »
Rika Antonova · peiyang shi · Hang Yin · Zehang Weng · Danica Kragic -
2021 : Occlusion-Aware Crowd Navigation Using People as Sensors »
Ye-Ji Mun · Masha Itkina · Katherine Driggs-Campbell -
2021 : Learning Transferable Motor Skills with Hierarchical Latent Mixture Policies »
Dushyant Rao · Fereshteh Sadeghi · Leonard Hasenclever · Markus Wulfmeier · Martina Zambelli · Giulia Vezzani · Dhruva Tirumala · Yusuf Aytar · Josh Merel · Nicolas Heess · Raia Hadsell -
2021 : Wish you were here: Hindsight Goal Selection for long-horizon dexterous manipulation »
Todor Davchev · Oleg Sushkov · Jean-Baptiste Regli · Stefan Schaal · Yusuf Aytar · Markus Wulfmeier · Jonathan Scholz -
2021 : Strength Through Diversity: Robust Behavior Learning via Mixture Policies »
Tim Seyde · Wilko Schwarting · Igor Gilitschenski · Markus Wulfmeier · Daniela Rus -
2021 : Implicit Behavioral Cloning »
Pete Florence · Corey Lynch · Andy Zeng · Oscar Ramirez · Ayzaan Wahid · Laura Downs · Adrian Wong · Igor Mordatch · Jonathan Tompson -
2021 : Improving Zero-shot Generalization in Offline Reinforcement Learning using Generalized Similarity Functions »
Bogdan Mazoure · Ilya Kostrikov · Ofir Nachum · Jonathan Tompson -
2022 : Bias Assessment of Text-to-Image Models »
Sasha Alexandra Luccioni · Clémentine Fourrier · Nathan Lambert · Unso Eun Seo Jo · Irene Solaiman · Helen Ngo · Nazneen Rajani · Giada Pistilli · Yacine Jernite · Margaret Mitchell -
2022 : Skill Acquisition by Instruction Augmentation on Offline Datasets »
Ted Xiao · Harris Chan · Pierre Sermanet · Ayzaan Wahid · Anthony Brohan · Karol Hausman · Sergey Levine · Jonathan Tompson -
2022 : PARTNR: Pick and place Ambiguity Resolving by Trustworthy iNteractive leaRning »
Jelle Luijkx · Zlatan Ajanovic · Laura Ferranti · Jens Kober -
2022 : Safety-Guaranteed Skill Discovery for Robot Manipulation Tasks »
Sunin Kim · Jaewoon Kwon · Taeyoon Lee · Younghyo Park · Julien PEREZ -
2022 : Interactive Language: Talking to Robots in Real Time »
Corey Lynch · Pete Florence · Jonathan Tompson · Ayzaan Wahid · Tianli Ding · James Betker · Robert Baruch · Travis Armstrong -
2022 : Robotic Skill Acquistion via Instruction Augmentation with Vision-Language Models »
Ted Xiao · Harris Chan · Pierre Sermanet · Ayzaan Wahid · Anthony Brohan · Karol Hausman · Sergey Levine · Jonathan Tompson -
2022 : Equivariant Representations for Non-Free Group Actions »
Luis Armando Pérez Rey · Giovanni Luca Marchetti · Danica Kragic · Dmitri Jarnikov · Mike Holenderski -
2022 : Contrastive Value Learning: Implicit Models for Simple Offline RL »
Bogdan Mazoure · Benjamin Eysenbach · Ofir Nachum · Jonathan Tompson -
2022 : Variance Reduction in Off-Policy Deep Reinforcement Learning using Spectral Normalization »
Payal Bawa · Rafael Oliveira · Fabio Ramos -
2022 : Learning Successor Feature Representations to Train Robust Policies for Multi-task Learning »
Melissa Mozifian · Dieter Fox · David Meger · Fabio Ramos · Animesh Garg -
2022 : Time-Efficient Reward Learning via Visually Assisted Cluster Ranking »
David Zhang · Micah Carroll · Andreea Bobu · Anca Dragan -
2022 : Optimal Behavior Prior: Data-Efficient Human Models for Improved Human-AI Collaboration »
Mesut Yang · Micah Carroll · Anca Dragan -
2022 : Language as Robot Middleware - Andy Zeng & Jacky Liang »
Hamidreza Kasaei -
2022 : Angela Schoellig »
Hamidreza Kasaei -
2022 : Uncertainty Aware Machine Learning for Robotics - Stefanie Tellex »
Hamidreza Kasaei -
2022 : Real Robots Learn with Structure - Georgia Chalvatzaki »
Hamidreza Kasaei -
2022 : My Hopes and Dreams of Communicating with Machines and Where to Begin - Beem Kim »
Hamidreza Kasaei -
2022 : Why Robots Need Social Skills - Leila Takayama »
Hamidreza Kasaei -
2022 : Representing Interactions for Robot Navigation - Katherine Driggs-Campbell & Zhe Huang »
Hamidreza Kasaei -
2022 : Towards Certifiably Safe Nonlinear Control with Sensor and Dynamics Uncertainties - Sarah Dean & Andrew Taylor »
Hamidreza Kasaei -
2022 : A Vision for Certifiable Perception: from Outlier-Robust Estimation to Self-Supervised Learning - Luca Carlone & Rajat Talak »
Hamidreza Kasaei -
2022 : Self-driving Cars - Matthew Johnson-Roberson »
Hamidreza Kasaei -
2022 : Trustworthy AI Robotics for Real-world Logistics - Anusha Nagabandi »
Hamidreza Kasaei -
2022 : A Generalist Agent (GATO) - Scott Reed & Gabriel Barth-Maron & Jackie Kay »
Hamidreza Kasaei -
2022 : Language as a Connective Tissue for Robotics - Karol Hausman & Brian Ichter »
Hamidreza Kasaei -
2022 : Aligning Robot Representations with Humans »
Andreea Bobu · Andi Peng · Pulkit Agrawal · Julie A Shah · Anca Dragan -
2022 : PARTNR: Pick and place Ambiguity Resolving by Trustworthy iNteractive leaRning »
Jelle Luijkx · Zlatan Ajanovic · Laura Ferranti · Jens Kober -
2022 : Safety-Guaranteed Skill Discovery for Robot Manipulation Tasks »
Sunin Kim · Jaewoon Kwon · Taeyoon Lee · Younghyo Park · Julien PEREZ -
2022 : Interactive Language: Talking to Robots in Real Time »
Corey Lynch · Pete Florence · Jonathan Tompson · Ayzaan Wahid · Tianli Ding · James Betker · Robert Baruch · Travis Armstrong -
2022 : Robotic Skill Acquistion via Instruction Augmentation with Vision-Language Models »
Ted Xiao · Harris Chan · Pierre Sermanet · Ayzaan Wahid · Anthony Brohan · Karol Hausman · Sergey Levine · Jonathan Tompson -
2022 Spotlight: Batch Bayesian optimisation via density-ratio estimation with guarantees »
Rafael Oliveira · Louis Tiao · Fabio Ramos -
2022 : Anca Dragan: Learning human preferences from language »
Anca Dragan -
2022 : What to learn from humans? »
Danica Kragic -
2022 : Interactive Imitation Learning in Robotics »
Jens Kober -
2022 Poster: First Contact: Unsupervised Human-Machine Co-Adaptation via Mutual Information Maximization »
Siddharth Reddy · Sergey Levine · Anca Dragan -
2022 Poster: Uni[MASK]: Unified Inference in Sequential Decision Problems »
Micah Carroll · Orr Paradise · Jessy Lin · Raluca Georgescu · Mingfei Sun · David Bignell · Stephanie Milani · Katja Hofmann · Matthew Hausknecht · Anca Dragan · Sam Devlin -
2022 Poster: Batch Bayesian optimisation via density-ratio estimation with guarantees »
Rafael Oliveira · Louis Tiao · Fabio Ramos -
2022 Poster: Improving Zero-Shot Generalization in Offline Reinforcement Learning using Generalized Similarity Functions »
Bogdan Mazoure · Ilya Kostrikov · Ofir Nachum · Jonathan Tompson -
2021 : Panel II: Machine decisions »
Anca Dragan · Karen Levy · Himabindu Lakkaraju · Ariel Rosenfeld · Maithra Raghu · Irene Y Chen -
2021 : DLO@Scale: A Large-scale Meta Dataset for Learning Non-rigid Object Pushing Dynamics »
Robert Gieselmann · Danica Kragic · Florian T. Pokorny · Alberta Longhini -
2021 Workshop: 4th Robot Learning Workshop: Self-Supervised and Lifelong Learning »
Alex Bewley · Masha Itkina · Hamidreza Kasaei · Jens Kober · Nathan Lambert · Julien PEREZ · Ransalu Senanayake · Vincent Vanhoucke · Markus Wulfmeier · Igor Gilitschenski -
2021 : Implicit Behavioral Cloning Q&A »
Pete Florence · Corey Lynch · Andy Zeng · Oscar Ramirez · Ayzaan Wahid · Laura Downs · Adrian Wong · Igor Mordatch · Jonathan Tompson -
2021 : Implicit Behavioral Cloning »
Pete Florence · Corey Lynch · Andy Zeng · Oscar Ramirez · Ayzaan Wahid · Laura Downs · Adrian Wong · Igor Mordatch · Jonathan Tompson -
2021 : BASALT: A MineRL Competition on Solving Human-Judged Task + Q&A »
Rohin Shah · Cody Wild · Steven Wang · Neel Alex · Brandon Houghton · William Guss · Sharada Mohanty · Stephanie Milani · Nicholay Topin · Pieter Abbeel · Stuart Russell · Anca Dragan -
2021 Poster: Evidential Softmax for Sparse Multimodal Distributions in Deep Generative Models »
Phil Chen · Masha Itkina · Ransalu Senanayake · Mykel J Kochenderfer -
2021 Poster: Pragmatic Image Compression for Human-in-the-Loop Decision-Making »
Sid Reddy · Anca Dragan · Sergey Levine -
2021 Poster: Is Bang-Bang Control All You Need? Solving Continuous Control with Bernoulli Policies »
Tim Seyde · Igor Gilitschenski · Wilko Schwarting · Bartolomeo Stellato · Martin Riedmiller · Markus Wulfmeier · Daniela Rus -
2020 : Keynote: Anca Dragan »
Anca Dragan -
2020 : Mini-panel discussion 3 - Prioritizing Real World RL Challenges »
Chelsea Finn · Thomas Dietterich · Angela Schoellig · Anca Dragan · Anusha Nagabandi · Doina Precup -
2020 : Invited Talk - "RL with Sim2Real in the Loop / Online Domain Adaptation for Mapping" »
Fabio Ramos · Anthony Tompkins -
2020 : Discussion Panel »
Pete Florence · Dorsa Sadigh · Carolina Parada · Jeannette Bohg · Roberto Calandra · Peter Stone · Fabio Ramos -
2020 : Bayesian optimization by density ratio estimation »
Louis Tiao · Aaron Klein · Cedric Archambeau · Edwin Bonilla · Matthias W Seeger · Fabio Ramos -
2020 Workshop: 3rd Robot Learning Workshop »
Masha Itkina · Alex Bewley · Roberto Calandra · Igor Gilitschenski · Julien PEREZ · Ransalu Senanayake · Markus Wulfmeier · Vincent Vanhoucke -
2020 : Introduction »
Masha Itkina -
2020 : Contributed Talk 2: Witness Autoencoder: Shaping the Latent Space with Witness Complexes »
Anastasiia Varava · Danica Kragic · Simon Schönenberger · Jen Jen Chung · Roland Siegwart · Vladislav Polianskii -
2020 Poster: Sparse Spectrum Warped Input Measures for Nonstationary Kernel Learning »
Anthony Tompkins · Rafael Oliveira · Fabio Ramos -
2020 : Q&A for invited speaker, Anca Dragan »
Anca Dragan -
2020 : Getting human-robot interaction strategies to emerge from first principles »
Anca Dragan -
2020 Poster: AvE: Assistance via Empowerment »
Yuqing Du · Stas Tiomkin · Emre Kiciman · Daniel Polani · Pieter Abbeel · Anca Dragan -
2020 Poster: Reward-rational (implicit) choice: A unifying formalism for reward learning »
Hong Jun Jeon · Smitha Milli · Anca Dragan -
2020 Poster: Preference learning along multiple criteria: A game-theoretic perspective »
Kush Bhatia · Ashwin Pananjady · Peter Bartlett · Anca Dragan · Martin Wainwright -
2020 Poster: Evidential Sparsification of Multimodal Latent Spaces in Conditional Variational Autoencoders »
Masha Itkina · Boris Ivanovic · Ransalu Senanayake · Mykel J Kochenderfer · Marco Pavone -
2019 : Poster Presentations »
Rahul Mehta · Andrew Lampinen · Binghong Chen · Sergio Pascual-Diaz · Jordi Grau-Moya · Aldo Faisal · Jonathan Tompson · Yiren Lu · Khimya Khetarpal · Martin Klissarov · Pierre-Luc Bacon · Doina Precup · Thanard Kurutach · Aviv Tamar · Pieter Abbeel · Jinke He · Maximilian Igl · Shimon Whiteson · Wendelin Boehmer · Raphaël Marinier · Olivier Pietquin · Karol Hausman · Sergey Levine · Chelsea Finn · Tianhe Yu · Lisa Lee · Benjamin Eysenbach · Emilio Parisotto · Eric Xing · Ruslan Salakhutdinov · Hongyu Ren · Anima Anandkumar · Deepak Pathak · Christopher Lu · Trevor Darrell · Alexei Efros · Phillip Isola · Feng Liu · Bo Han · Gang Niu · Masashi Sugiyama · Saurabh Kumar · Janith Petangoda · Johan Ferret · James McClelland · Kara Liu · Animesh Garg · Robert Lange -
2019 Workshop: Machine Learning for Autonomous Driving »
Rowan McAllister · Nicholas Rhinehart · Fisher Yu · Li Erran Li · Anca Dragan -
2019 Workshop: Robot Learning: Control and Interaction in the Real World »
Roberto Calandra · Markus Wulfmeier · Kate Rakelly · Sanket Kamthe · Danica Kragic · Stefan Schaal · Markus Wulfmeier -
2019 : Poster Session »
Lili Yu · Aleksei Kroshnin · Alex Delalande · Andrew Carr · Anthony Tompkins · Aram-Alexandre Pooladian · Arnaud Robert · Ashok Vardhan Makkuva · Aude Genevay · Bangjie Liu · Bo Zeng · Charlie Frogner · Elsa Cazelles · Esteban G Tabak · Fabio Ramos · François-Pierre PATY · Georgios Balikas · Giulio Trigila · Hao Wang · Hinrich Mahler · Jared Nielsen · Karim Lounici · Kyle Swanson · Mukul Bhutani · Pierre Bréchet · Piotr Indyk · samuel cohen · Stefanie Jegelka · Tao Wu · Thibault Sejourne · Tudor Manole · Wenjun Zhao · Wenlin Wang · Wenqi Wang · Yonatan Dukler · Zihao Wang · Chaosheng Dong -
2019 : Poster session »
Sebastian Farquhar · Erik Daxberger · Andreas Look · Matt Benatan · Ruiyi Zhang · Marton Havasi · Fredrik Gustafsson · James A Brofos · Nabeel Seedat · Micha Livne · Ivan Ustyuzhaninov · Adam Cobb · Felix D McGregor · Patrick McClure · Tim R. Davidson · Gaurush Hiranandani · Sanjeev Arora · Masha Itkina · Didrik Nielsen · William Harvey · Matias Valdenegro-Toro · Stefano Peluchetti · Riccardo Moriconi · Tianyu Cui · Vaclav Smidl · Taylan Cemgil · Jack Fitzsimons · He Zhao · · mariana vargas vieyra · Apratim Bhattacharyya · Rahul Sharma · Geoffroy Dubourg-Felonneau · Jonathan Warrell · Slava Voloshynovskiy · Mihaela Rosca · Jiaming Song · Andrew Ross · Homa Fashandi · Ruiqi Gao · Hooshmand Shokri Razaghi · Joshua Chang · Zhenzhong Xiao · Vanessa Boehm · Giorgio Giannone · Ranganath Krishnan · Joe Davison · Arsenii Ashukha · Jeremiah Liu · Sicong (Sheldon) Huang · Evgenii Nikishin · Sunho Park · Nilesh Ahuja · Mahesh Subedar · · Artyom Gadetsky · Jhosimar Arias Figueroa · Tim G. J. Rudner · Waseem Aslam · Adrián Csiszárik · John Moberg · Ali Hebbal · Kathrin Grosse · Pekka Marttinen · Bang An · Hlynur Jónsson · Samuel Kessler · Abhishek Kumar · Mikhail Figurnov · Omesh Tickoo · Steindor Saemundsson · Ari Heljakka · Dániel Varga · Niklas Heim · Simone Rossi · Max Laves · Waseem Gharbieh · Nicholas Roberts · Luis Armando Pérez Rey · Matthew Willetts · Prithvijit Chakrabarty · Sumedh Ghaisas · Carl Shneider · Wray Buntine · Kamil Adamczewski · Xavier Gitiaux · Suwen Lin · Hao Fu · Gunnar Rätsch · Aidan Gomez · Erik Bodin · Dinh Phung · Lennart Svensson · Juliano Tusi Amaral Laganá Pinto · Milad Alizadeh · Jianzhun Du · Kevin Murphy · Beatrix Benkő · Shashaank Vattikuti · Jonathan Gordon · Christopher Kanan · Sontje Ihler · Darin Graham · Michael Teng · Louis Kirsch · Tomas Pevny · Taras Holotyak -
2019 Poster: On the Utility of Learning about Humans for Human-AI Coordination »
Micah Carroll · Rohin Shah · Mark Ho · Tom Griffiths · Sanjit Seshia · Pieter Abbeel · Anca Dragan -
2018 : Anca Dragan »
Anca Dragan -
2018 : Opening Remark »
Li Erran Li · Anca Dragan -
2018 Workshop: NIPS Workshop on Machine Learning for Intelligent Transportation Systems 2018 »
Li Erran Li · Anca Dragan · Juan Carlos Niebles · Silvio Savarese -
2018 : Fabio Ramos (Uni. of Sydney): Learning and Planning in Spatial-Temporal Data »
Fabio Ramos -
2018 : Anca Dragan »
Anca Dragan -
2018 : Coffee Break and Poster Session I »
Pim de Haan · Bin Wang · Dequan Wang · Aadil Hayat · Ibrahim Sobh · Muhammad Asif Rana · Thibault Buhet · Nicholas Rhinehart · Arjun Sharma · Alex Bewley · Michael Kelly · Lionel Blondé · Ozgur S. Oguz · Vaibhav Viswanathan · Jeroen Vanbaar · Konrad Żołna · Negar Rostamzadeh · Rowan McAllister · Sanjay Thakur · Alexandros Kalousis · Chelsea Sidrane · Sujoy Paul · Daphne Chen · Michal Garmulewicz · Henryk Michalewski · Coline Devin · Hongyu Ren · Jiaming Song · Wen Sun · Hanzhang Hu · Wulong Liu · Emilie Wirbel -
2018 Workshop: Modeling and decision-making in the spatiotemporal domain »
Ransalu Senanayake · Neal Jean · Fabio Ramos · Girish Chowdhary -
2018 Poster: Discovery of Latent 3D Keypoints via End-to-end Geometric Reasoning »
Supasorn Suwajanakorn · Noah Snavely · Jonathan Tompson · Mohammad Norouzi -
2018 Oral: Discovery of Latent 3D Keypoints via End-to-end Geometric Reasoning »
Supasorn Suwajanakorn · Noah Snavely · Jonathan Tompson · Mohammad Norouzi -
2018 Poster: Integrated accounts of behavioral and neuroimaging data using flexible recurrent neural network models »
Amir Dezfouli · Richard Morris · Fabio Ramos · Peter Dayan · Bernard Balleine -
2018 Oral: Integrated accounts of behavioral and neuroimaging data using flexible recurrent neural network models »
Amir Dezfouli · Richard Morris · Fabio Ramos · Peter Dayan · Bernard Balleine -
2018 Poster: Where Do You Think You're Going?: Inferring Beliefs about Dynamics from Behavior »
Sid Reddy · Anca Dragan · Sergey Levine -
2017 : 6 Spotlight Talks (3 min each) »
Mennatullah Siam · Mohit Prabhushankar · Priyam Parashar · Mustafa Mukadam · hengshuai yao · Ransalu Senanayake -
2017 : Morning panel discussion »
Jürgen Schmidhuber · Noah Goodman · Anca Dragan · Pushmeet Kohli · Dhruv Batra -
2017 : "Communication via Physical Action" »
Anca Dragan -
2017 Workshop: 2017 NIPS Workshop on Machine Learning for Intelligent Transportation Systems »
Li Erran Li · Anca Dragan · Juan Carlos Niebles · Silvio Savarese -
2017 : Invited talk: Robot Transparency as Optimal Control »
Anca Dragan -
2017 Workshop: Acting and Interacting in the Real World: Challenges in Robot Learning »
Ingmar Posner · Raia Hadsell · Martin Riedmiller · Markus Wulfmeier · Rohan Paul -
2017 Poster: Hierarchical Attentive Recurrent Tracking »
Adam Kosiorek · Alex Bewley · Ingmar Posner -
2016 : Learning Reliable Objectives »
Anca Dragan -
2016 : Invited Talk: Autonomous Cars that Coordinate with People (Anca Dragan, Berkeley) »
Anca Dragan -
2016 Poster: Spatio-Temporal Hilbert Maps for Continuous Occupancy Representation in Dynamic Environments »
Ransalu Senanayake · Lionel Ott · Simon O'Callaghan · Fabio Ramos -
2016 Poster: Cooperative Inverse Reinforcement Learning »
Dylan Hadfield-Menell · Stuart J Russell · Pieter Abbeel · Anca Dragan -
2014 Poster: On Integrated Clustering and Outlier Detection »
Lionel Ott · Linsey Pang · Fabio Ramos · Sanjay Chawla