I’ve finished my PhD
@berkeley_ai
!
This wraps up 8 wonderful years and 3 degrees from
@Berkeley_EECS
, starting as a freshman undergrad in 2016. Thanks everyone and go bears! 🐻
I’ll still be in the Bay, I’m headed to Apple next as a Research Scientist - very excited about it!
Companies like
@Waymo
&
@Amazon
use remote humans to manage fleets of robots for self-driving & logistics. We introduce new formalism, algorithms, and open-source
@NVIDIA
Isaac Gym benchmarks for “Interactive Fleet Learning” (IFL).
w/
@ken_goldberg
,
@pabbeel
,
@berkeley_ai
(1/8)
How can robots learn skills from many different teachers? Especially when they must succeed outside the training distribution?
Our
#CoRL2023
paper tackles both multimodality and distribution shift with IIFL: Implicit Interactive Fleet Learning.
@berkeley_ai
@UCBerkeley
(1/7)
Tired of collecting interventions all day to train with DAgger?
Introducing IntervenGen from
@NVIDIARobotics
+
@berkeley_ai
. From just 10 corrective human interventions, IntervenGen generates 1000+ to cover broad robot mistake distributions 👇
🧵 1/
Evaluating progress in robotic manipulation is challenging due to the cost & diversity of robots—but what if we had shared access to a remote robot over the Internet?
I’m in Japan all week and will be presenting our work w/
@GoogleAI
tomorrow at 11:20 AM JST at
#IROS2022
! (1/3)
Thrilled and surprised to see
@OpenAI
cofounder and brilliant researcher
@johnschulman2
reference my work in his highly anticipated keynote...! Appreciate the shoutout and thanks for the great talk! 😄
Is your robot pestering you? Robots learning a new task should only ask humans for help when needed. Our new algorithm ThriftyDAgger will be presented at
#CORL2021
from
@AUTOLab_Cal
@berkeley_ai
(1/8) Website w/ paper, code, videos:
In Seattle for the summer as a research scientist intern at the
@nvidia
Seattle Robotics Lab; very excited to continue my PhD research on scaling imitation learning w/ talented roboticists
@AjayMandlekar
@CaelanGarrett
@ankurhandos
and Dieter Fox. LMK if you’re in the city! ☀️
Between the AI researchers I already follow for work and the rest of Twitter posting AI hype every day (ironically, like a bot would), my feed has been unbearably homogeneous for months. Super low SNR and feels like closing a fridge dissatisfied. I miss the pre-chatGPT Twitter
Lizards and many other reptilians are famous for walking on vertical surfaces or even upside down. They use van der Waals forces.
Yet, when the friction is very low, even lizards experience huge issues.
[📹 oryzae1824]
It was awesome to be a part of this!! X-embodiment Fleet Learning with an Avengers-level crossover of 173 roboticists in 34 labs around the world, and all data is open sourced!
Excited to release this today! The “bitter lesson” of LLMs has been that supervised learning at scale is all you need — IFL may be the recipe for scaling up supervised learning in robotics 🦾
Seems like a good time to retweet this ;) Recent results are amazing but let's not conclude that robotics is solved... lots of great research remains to be done!
CoRL: 6th Conference on Robot Learning will be in Auckland, New Zealand, 14-18 Dec 2022 (where it'll be summertime ;):
@CoRL_Conf
Deadline 15 June: We welcome submissions addressing the theory and practice of machine learning for robots and automation.
MimicGen is an exceptionally intuitive and effective way to scale up human data collection. Stop by the OOD workshop at CoRL tomorrow for a sneak peek of an extension of MimicGen I’ve been working on with
@AjayMandlekar
@CaelanGarrett
@NVIDIAAI
👀
Tired of collecting demonstrations all day to train your robot?
Introducing MimicGen, an autonomous data generation system for robotics. Using just 200 human demos we generated a large multi-task dataset of 50K demos!
#CoRL2023
#NVIDIAResearch
👇
🧵 1/
Delighted to share that Interactive Fleet Learning has been accepted for oral presentation at
@corl_conf
!! Looking forward to presenting our work at
#CoRL2022
in New Zealand this December!
Companies like
@Waymo
&
@Amazon
use remote humans to manage fleets of robots for self-driving & logistics. We introduce new formalism, algorithms, and open-source
@NVIDIA
Isaac Gym benchmarks for “Interactive Fleet Learning” (IFL).
w/
@ken_goldberg
,
@pabbeel
,
@berkeley_ai
(1/8)
Great to be in New Zealand for
#CoRL2022
@corl_conf
! 🤖 2 things: (1) stop by the “Learning from Humans” oral session tomorrow at 3:35 PM to catch my talk on fleet learning, (2) join the 1st ever *CoRL Puzzle Hunt* which
@ashwinb96
@brthananjeyan
and I have organized for you all!
An excerpt from Godel, Escher, Bach (1979, before NNs) — perhaps the “black box” (uninterpretable) aspect of deep neural nets, often cited as one of the cons, is actually unavoidable in any sufficiently powerful approximation of intelligence…!
We finally jumped on the LLM hype train in this project, applying PaLM to challenging robot search problems. Fun collaboration with
@brian_ichter
, one of the folks behind Google SayCan, led by
@19kaushiks
@satviks107
@RavenHuang4
Wouldn’t it be nice if ChatGPT could find your missing keys for you? Our latest research from
@berkeley_ai
+
@GoogleAI
suggests that robots can use large language models (LLMs) to find hidden objects faster. 🧵👇
Well, I finally gave into FOMO and got on twitter, so prepare for a strange combination of work-related things and random musings masquerading as insight
The world would be a better place if people kept *probability distributions* of their beliefs instead of absolute beliefs, and adjusted them over time proportional to evidence. Like 80% certainty for a political position, or 99.9% for a scientific law
In the last two years, large foundation models have proven capable of perceiving and reasoning about the world around us unlocking a key possibility for scaling robotics.
We introduce a AutoRT, a framework for orchestrating robotic agents in the wild using foundation models!
For those who couldn't make it all the way to New Zealand ;)... the video from my talk at CoRL on interactive fleet learning is now available on YouTube!
Delighted to share that Interactive Fleet Learning has been accepted for oral presentation at
@corl_conf
!! Looking forward to presenting our work at
#CoRL2022
in New Zealand this December!
We present the first systematic benchmarking of fabric manipulation algorithms on physical hardware, with 4 new learning algorithms and 4 baselines for T-shirt folding. All data collection, training, experiments conducted remotely—I’ve never seen the robot in person! (2/3)
We also execute 1000 real-world trials of physical block-pushing with a fleet of 4 ABB YuMi arms & 2 human supervisors performing teleoperation remotely over the Internet. (7/8)
We introduce the IFL Benchmark, an open-source Python benchmark and toolkit built on top of
@NVIDIA
Isaac Gym for rapid prototyping + standardized evaluation of IFL algorithms with 100s of robots in simulation. (4/8)
We propose Fleet-DAgger, a novel family of IFL algorithms including fleet adaptations of existing 1-robot, 1-human interactive learning algorithms like ThriftyDAgger and EnsembleDAgger. (5/8)
@yacineMTB
Douglas Hofstadter’s notion of a “strange loop”: a paradoxical loop where you think you’re making progress but somehow end up back at the start. He writes about how both Godel’s theorem and human consciousness are strange loops. Check out this video:
Robots make mistakes! Robotics companies use remote human supervision of robot fleets in a range of applications. Human interventions during task execution improve reliability + provide more data for the robots to improve w/ continual learning. (2/8)
Work done in collaboration with co-first author Gaurav Datta, Anrui Gu, Eugen Solowjow,
@Ken_Goldberg
@AUTOLab_Cal
@berkeley_ai
. See you in Atlanta! ✈️
arXiv:
OpenReview: (7/7)
Today,
@TIME
revealed its annual list of the Best Inventions, which features extraordinary innovations changing our lives.
@PhantomAuto_
is proud to be included in
#TIMEBestInventions
of 2022 list for its Remote Operation Platform for Logistics.
We can tell our robots what we want them to do, but language can be underspecified. Goal images are worth 1,000 words, but can be overspecified.
Hand-drawn sketches are a happy medium for communicating goals to robots!
🤖✏️Introducing RT-Sketch:
🧵1/11
Key question: How can we optimally allocate M humans to N robots (M << N) for learning + execution? To our knowledge, we formalize multi-robot, multi-human interactive learning for the first time. (3/8)
@ericjang11
Agree, but also something being purely fiat doesn’t mean it can be safely ignored—almost everything we consider to be true is a collective imagination (morality, Enlightenment ideals like the value of empirical science, the existence of corporations and nations, etc)
@kscottz
I feel like machines are robots that “graduated,” they do their jobs so well that we take their autonomy for granted and they don’t seem like robots anymore. What was once smart starts to look dumb. Something similar is happening with the Roomba right now.
Check out the HMC22 workshop tomorrow (in Paris or free online), where myself and
@Ken_Goldberg
will be presenting a keynote on our recent work and perspectives on human-robot collaboration for an interdisciplinary audience!
The program for our next workshop,
#AFPLhmc22
, is how live: -- and we have an amazing lineup of speakers, an interactive workshop on human-machine collaboration and
#sustainability
, and an interactive
#art
exhibit. Registration is open and free!
Real-time language unlocks some interesting new capabilities, e.g.
one operator controlling 4 robots simultaneously using only spoken commands.
This is potentially an interesting way to scale up future collection.
@xiao_ted
+1, I’d argue robotics (decision making under uncertainty) is much more in line with the AI problem than ML (statistical function approximation) is. ML is an approach to AI
In IFL Benchmark simulation environments with 100 robots & 10 humans, a novel algorithm outperforms baselines w.r.t. return on human effort, throughput, hard failures, & idle time. (6/8)
@Altimor
Idk, personally I think it’s kinda nice that our research doesn’t become obsolete every few months and we don’t have to train with a million GPUs
My book What's Our Problem? is now available.
The book introduces a new framework for thinking about our chaotic political environment. With 303 drawings, it's a toolbox for understanding our societies, our group dynamics, and our own minds.
Get it here:
It was a fun time. Locomotion and manipulation software was underwhelming (not new or SOTA), but impressive progress in a short time and may show results in the future if their warehouses are sufficiently structured. Good to see the industry and press interest in robotics
The key insight of IIFL is a novel application of Jeffreys Divergence. Due to symmetry, the intractable partition function cancels out! I’m very excited about this technique: it can estimate uncertainty for *ANY* energy-based model (not just for interactive IL). (5/7)
Imitation learning is brittle outside the training distribution. Online interventions help but take significant human data and effort to show how to recover from all the possible mistakes a robot may make. Can we instead automatically generate them?
🧵 2/
@GlenBerseth
It essentially allows authenticated users to establish a network connection with the robot server and starts a big observation-action loop. (Note it’s not publicly available at the moment) Setup/maintenance cost is imposed on the host in this model.
We build on NVIDIA MimicGen. By executing the robot policy during both *data collection* and *data generation,* the robot encounters novel mistake states, from which we can apply transformed recovery segments.
🧵 3/
Tired of collecting demonstrations all day to train your robot?
Introducing MimicGen, an autonomous data generation system for robotics. Using just 200 human demos we generated a large multi-task dataset of 50K demos!
#CoRL2023
#NVIDIAResearch
👇
🧵 1/
We evaluate IntervenGen in contact-rich tasks and observe that it increases robustness up to 39x over existing baselines. IntervenGen with 10 source interventions outperforms a policy trained with 100 human interventions by 24%, with just 12% of the data collection effort.
🧵 4/
For those who say AI isn’t creative as it can’t have original thought—
Humans are no better. Creativity is the act of mimetic recombination. We combine the set of ideas we’re exposed to (“influences”) to form new ones
And like AI, humans often don’t realize when they’re copying
The policy adapted to dynamic pose changes in the environment without object tracking and was robust to both physical perturbations of the end effector as well as visual distractors in the scene.
🧵 7/
@zkirby2020
It’s more the engineering for sure. Open world autonomous manipulation isn’t nearly reliable enough yet, our current paradigm is collect a bunch of data then train on it, but it’s not clear yet (1) where we will get all that data and (2) even if we do, if it’ll be 99.9% reliable
Simulation and physical robot experiments suggest that compared with baselines, ThriftyDAgger can both increase task success rate and reduce human supervision, both during training and during execution. (6/8)