Byron Boots

Byron Boots

University of Washington

H-index: 45

North America-United States

About Byron Boots

Byron Boots, With an exceptional h-index of 45 and a recent h-index of 38 (since 2020), a distinguished researcher at University of Washington, specializes in the field of Machine Learning, Artificial Intelligence, Robotics.

His recent articles reflect a diverse array of research interests and contributions to the field:

Robotic System Performing Dynamic Interaction in Human-Robot Cooperative Work for Assembly Operation

LocoMan: Advancing Versatile Quadrupedal Dexterity with Lightweight Loco-Manipulators

Multi-Sample Long Range Path Planning under Sensing Uncertainty for Off-Road Autonomous Driving

Scalable measurement error mitigation via iterative bayesian unfolding

Adversarial model for offline reinforcement learning

Model predictive control techniques for autonomous systems

Controlling position of robot by determining goal proposals by using neural networks

Motion policy networks

Byron Boots Information

University

University of Washington

Position

Associate Professor

Citations(all)

7746

Citations(since 2020)

6142

Cited By

3843

hIndex(all)

45

hIndex(since 2020)

38

i10Index(all)

110

i10Index(since 2020)

99

Email

University Profile Page

University of Washington

Byron Boots Skills & Research Interests

Machine Learning

Artificial Intelligence

Robotics

Top articles of Byron Boots

Robotic System Performing Dynamic Interaction in Human-Robot Cooperative Work for Assembly Operation

Authors

Satoshi Nakamura,Carolina Higuera Arias,Mohak Bhardwaj,Byron Boots

Published Date

2024/1/8

In order to realize human-robot cooperative work, it is necessary to have dynamic interaction for unexpected human's motion that are difficult to define in advance. In this research, we propose a robotic system that supports operator work without interfering with human work by reacting and avoiding operator motion through real-time motion generation and activation re-planning. This system is included 3D recognition of the robot's surroundings and real-time motion generation technology. We implemented in the setup of assembly work assuming a high-mix, low-volume production site, and was demonstrated in a support task that robot distribute the parts for operator's assembly work. Experimental results show that the robot can conduct the cooperative work without interfering with operator work by performing appropriate avoidance actions against operator motion.

LocoMan: Advancing Versatile Quadrupedal Dexterity with Lightweight Loco-Manipulators

Authors

Changyi Lin,Xingyu Liu,Yuxiang Yang,Yaru Niu,Wenhao Yu,Tingnan Zhang,Jie Tan,Byron Boots,Ding Zhao

Journal

arXiv preprint arXiv:2403.18197

Published Date

2024/3/27

Quadrupedal robots have emerged as versatile agents capable of locomoting and manipulating in complex environments. Traditional designs typically rely on the robot's inherent body parts or incorporate top-mounted arms for manipulation tasks. However, these configurations may limit the robot's operational dexterity, efficiency and adaptability, particularly in cluttered or constrained spaces. In this work, we present LocoMan, a dexterous quadrupedal robot with a novel morphology to perform versatile manipulation in diverse constrained environments. By equipping a Unitree Go1 robot with two low-cost and lightweight modular 3-DoF loco-manipulators on its front calves, LocoMan leverages the combined mobility and functionality of the legs and grippers for complex manipulation tasks that require precise 6D positioning of the end effector in a wide workspace. To harness the loco-manipulation capabilities of LocoMan, we introduce a unified control framework that extends the whole-body controller (WBC) to integrate the dynamics of loco-manipulators. Through experiments, we validate that the proposed whole-body controller can accurately and stably follow desired 6D trajectories of the end effector and torso, which, when combined with the large workspace from our design, facilitates a diverse set of challenging dexterous loco-manipulation tasks in confined spaces, such as opening doors, plugging into sockets, picking objects in narrow and low-lying spaces, and bimanual manipulation.

Multi-Sample Long Range Path Planning under Sensing Uncertainty for Off-Road Autonomous Driving

Authors

Matt Schmittle,Rohan Baijal,Brian Hou,Siddhartha Srinivasa,Byron Boots

Journal

arXiv preprint arXiv:2403.11298

Published Date

2024/3/17

We focus on the problem of long-range dynamic replanning for off-road autonomous vehicles, where a robot plans paths through a previously unobserved environment while continuously receiving noisy local observations. An effective approach for planning under sensing uncertainty is determinization, where one converts a stochastic world into a deterministic one and plans under this simplification. This makes the planning problem tractable, but the cost of following the planned path in the real world may be different than in the determinized world. This causes collisions if the determinized world optimistically ignores obstacles, or causes unnecessarily long routes if the determinized world pessimistically imagines more obstacles. We aim to be robust to uncertainty over potential worlds while still achieving the efficiency benefits of determinization. We evaluate algorithms for dynamic replanning on a large real-world dataset of challenging long-range planning problems from the DARPA RACER program. Our method, Dynamic Replanning via Evaluating and Aggregating Multiple Samples (DREAMS), outperforms other determinization-based approaches in terms of combined traversal time and collision cost. https://sites.google.com/cs.washington.edu/dreams/

Scalable measurement error mitigation via iterative bayesian unfolding

Authors

Bibek Pokharel,Siddarth Srinivasan,Gregory Quiroz,Byron Boots

Journal

Physical Review Research

Published Date

2024/2/21

Measurement errors are a significant obstacle to achieving scalable quantum computation. To counteract systematic readout errors, researchers have developed postprocessing techniques known as measurement error mitigation methods. However, these methods face a tradeoff between scalability and returning nonnegative probabilities. In this paper, we present a solution to overcome this challenge. Our approach focuses on iterative Bayesian unfolding, a standard mitigation technique used in high-energy physics experiments, and implements it in a scalable way. We demonstrate our method on experimental Greenberger-Horne-Zeilinger state preparation on up to 127 qubits and on the Bernstein-Vazirani algorithm on up to 26 qubits. Compared to state-of-the-art methods (such as M3), our implementation guarantees valid probability distributions, returns comparable or better-mitigated results, and does so …

Adversarial model for offline reinforcement learning

Authors

Mohak Bhardwaj,Tengyang Xie,Byron Boots,Nan Jiang,Ching-An Cheng

Journal

arXiv preprint arXiv:2302.11048

Published Date

2023/2/21

We propose a novel model-based offline Reinforcement Learning (RL) framework, called Adversarial Model for Offline Reinforcement Learning (ARMOR), which can robustly learn policies to improve upon an arbitrary reference policy regardless of data coverage. ARMOR is designed to optimize policies for the worst-case performance relative to the reference policy through adversarially training a Markov decision process model. In theory, we prove that ARMOR, with a well-tuned hyperparameter, can compete with the best policy within data coverage when the reference policy is supported by the data. At the same time, ARMOR is robust to hyperparameter choices: the policy learned by ARMOR, with any admissible hyperparameter, would never degrade the performance of the reference policy, even when the reference policy is not covered by the dataset. To validate these properties in practice, we design a scalable implementation of ARMOR, which by adversarial training, can optimize policies without using model ensembles in contrast to typical model-based methods. We show that ARMOR achieves competent performance with both state-of-the-art offline model-free and model-based RL algorithms and can robustly improve the reference policy over various hyperparameter choices.

Model predictive control techniques for autonomous systems

Published Date

2023/5/9

Apparatuses, systems, and techniques to infer a sequence of actions to perform using one or more neural networks trained, at least in part, by optimizing a probability distribution function using a cost function, wherein the probability distribution represents different sequences of actions that can be performed. In at least one embodiment, a model predictive control problem is formulated as a Bayesian inference task to infer a set of solutions.

Controlling position of robot by determining goal proposals by using neural networks

Published Date

2024/4/16

A framework for offline learning from a set of diverse and suboptimal demonstrations operates by selectively imitating local sequences from the dataset. At least one embodiment recovers performant policies from large manipulation datasets by decomposing the problem into a goal-conditioned imitation and a high-level goal selection mechanism.

Motion policy networks

Authors

Adam Fishman,Adithyavairan Murali,Clemens Eppner,Bryan Peele,Byron Boots,Dieter Fox

Journal

arXiv preprint arXiv:2210.12209

Published Date

2022/10/21

Collision-free motion generation in unknown environments is a core building block for robot manipulation. Generating such motions is challenging due to multiple objectives; not only should the solutions be optimal, the motion generator itself must be fast enough for real-time performance and reliable enough for practical deployment. A wide variety of methods have been proposed ranging from local controllers to global planners, often being combined to offset their shortcomings. We present an end-to-end neural model called Motion Policy Networks (M Nets) to generate collision-free, smooth motion from just a single depth camera observation. M Nets are trained on over 3 million motion planning problems in more than 500,000 environments. Our experiments show that M Nets are significantly faster than global planners while exhibiting the reactivity needed to deal with dynamic scenes. They are 46% better than prior neural planners and more robust than local control policies. Despite being only trained in simulation, M Nets transfer well to the real robot with noisy partial point clouds. Videos and code are available at https://mpinets. github. io

Stackelberg games for learning emergent behaviors during competitive autocurricula

Authors

Boling Yang,Liyuan Zheng,Lillian J Ratliff,Byron Boots,Joshua R Smith

Published Date

2023/5/29

Autocurricular training is an important sub-area of multi-agent reinforcement learning (MARL) that allows multiple agents to learn emergent skills in an unsupervised co-evolving scheme. The robotics community has experimented auto-curricular training with physically grounded problems, such as robust control and interactive manipulation tasks. However, the asymmetric nature of these tasks makes the generation of sophisticated policies challenging. Indeed, the asymmetry in the environment may implicitly or explicitly provide an advantage to a subset of agents which could, in turn, lead to a low-quality equilibrium. This paper proposes a novel game-theoretic algorithm, Stackelberg Multi-Agent Deep Deterministic Policy Gradient (ST-MADDPG), which formulates a two-player MARL problem as a Stackelberg game with one player as the ‘leader’ and the other as the ‘follower’ in a hierarchical interaction structure …

Deep Model Predictive Optimization

Authors

Jacob Sacks,Rwik Rana,Kevin Huang,Alex Spitzer,Guanya Shi,Byron Boots

Journal

arXiv preprint arXiv:2310.04590

Published Date

2023/10/6

A major challenge in robotics is to design robust policies which enable complex and agile behaviors in the real world. On one end of the spectrum, we have model-free reinforcement learning (MFRL), which is incredibly flexible and general but often results in brittle policies. In contrast, model predictive control (MPC) continually re-plans at each time step to remain robust to perturbations and model inaccuracies. However, despite its real-world successes, MPC often under-performs the optimal strategy. This is due to model quality, myopic behavior from short planning horizons, and approximations due to computational constraints. And even with a perfect model and enough compute, MPC can get stuck in bad local optima, depending heavily on the quality of the optimization algorithm. To this end, we propose Deep Model Predictive Optimization (DMPO), which learns the inner-loop of an MPC optimization algorithm directly via experience, specifically tailored to the needs of the control problem. We evaluate DMPO on a real quadrotor agile trajectory tracking task, on which it improves performance over a baseline MPC algorithm for a given computational budget. It can outperform the best MPC algorithm by up to 27% with fewer samples and an end-to-end policy trained with MFRL by 19%. Moreover, because DMPO requires fewer samples, it can also achieve these benefits with 4.3X less memory. When we subject the quadrotor to turbulent wind fields with an attached drag plate, DMPO can adapt zero-shot while still outperforming all baselines. Additional results can be found at https://tinyurl.com/mr2ywmnw.

V-STRONG: Visual Self-Supervised Traversability Learning for Off-road Navigation

Authors

Sanghun Jung,JoonHo Lee,Xiangyun Meng,Byron Boots,Alexander Lambert

Journal

arXiv preprint arXiv:2312.16016

Published Date

2023/12/26

Reliable estimation of terrain traversability is critical for the successful deployment of autonomous systems in wild, outdoor environments. Given the lack of large-scale annotated datasets for off-road navigation, strictly-supervised learning approaches remain limited in their generalization ability. To this end, we introduce a novel, image-based self-supervised learning method for traversability prediction, leveraging a state-of-the-art vision foundation model for improved out-of-distribution performance. Our method employs contrastive representation learning using both human driving data and instance-based segmentation masks during training. We show that this simple, yet effective, technique drastically outperforms recent methods in predicting traversability for both on- and off-trail driving scenarios. We compare our method with recent baselines on both a common benchmark as well as our own datasets, covering a diverse range of outdoor environments and varied terrain types. We also demonstrate the compatibility of resulting costmap predictions with a model-predictive controller. Finally, we evaluate our approach on zero- and few-shot tasks, demonstrating unprecedented performance for generalization to new environments. Videos and additional material can be found here: \url{https://sites.google.com/view/visual-traversability-learning}.

Learning semantics-aware locomotion skills from human demonstration

Authors

Yuxiang Yang,Xiangyun Meng,Wenhao Yu,Tingnan Zhang,Jie Tan,Byron Boots

Journal

Conference on Robot Learning, 2022

Published Date

2022/6/27

The semantics of the environment, such as the terrain type and property, reveals important information for legged robots to adjust their behaviors. In this work, we present a framework that learns semantics-aware locomotion skills from perception for quadrupedal robots, such that the robot can traverse through complex offroad terrains with appropriate speeds and gaits using perception information. Due to the lack of high-fidelity outdoor simulation, our framework needs to be trained directly in the real world, which brings unique challenges in data efficiency and safety. To ensure sample efficiency, we pre-train the perception model with an off-road driving dataset. To avoid the risks of real-world policy exploration, we leverage human demonstration to train a speed policy that selects a desired forward speed from camera image. For maximum traversability, we pair the speed policy with a gait selector, which selects a robust locomotion gait for each forward speed. Using only 40 minutes of human demonstration data, our framework learns to adjust the speed and gait of the robot based on perceived terrain semantics, and enables the robot to walk over 6km without failure at close-to-optimal speed

Neural contact fields: Tracking extrinsic contact with tactile sensing

Authors

Carolina Higuera,Siyuan Dong,Byron Boots,Mustafa Mukadam

Published Date

2023/5/29

We present Neural Contact Fields, a method that brings together neural fields and tactile sensing to address the problem of tracking extrinsic contact between object and environment. Knowing where the external contact occurs is a first step towards methods that can actively control it in facilitating downstream manipulation tasks. Prior work for localizing environmental contacts typically assume a contact type (e.g. point or line), does not capture contact/no-contact transitions, and only works with basic geometric-shaped objects. Neural Contact Fields are the first method that can track arbitrary multi-modal extrinsic contacts without making any assumptions about the contact type. Our key insight is to estimate the probability of contact for any 3D point in the latent space of object's shapes, given vision-based tactile inputs that sense the local motion resulting from the external contact. In experiments, we find that Neural …

Perceiving Extrinsic Contacts from Touch Improves Learning Insertion Policies

Authors

Carolina Higuera,Joseph Ortiz,Haozhi Qi,Luis Pineda,Byron Boots,Mustafa Mukadam

Journal

arXiv preprint arXiv:2309.16652

Published Date

2023/9/28

Robotic manipulation tasks such as object insertion typically involve interactions between object and environment, namely extrinsic contacts. Prior work on Neural Contact Fields (NCF) use intrinsic tactile sensing between gripper and object to estimate extrinsic contacts in simulation. However, its effectiveness and utility in real-world tasks remains unknown. In this work, we improve NCF to enable sim-to-real transfer and use it to train policies for mug-in-cupholder and bowl-in-dishrack insertion tasks. We find our model NCF-v2, is capable of estimating extrinsic contacts in the real-world. Furthermore, our insertion policy with NCF-v2 outperforms policies without it, achieving 33% higher success and 1.36x faster execution on mug-in-cupholder, and 13% higher success and 1.27x faster execution on bowl-in-dishrack.

Dynamo-grasp: Dynamics-aware optimization for grasp point detection in suction grippers

Authors

Boling Yang,Soofiyan Layakalli Atar,Markus Grotz,Byron Boots,Joshua Smith

Published Date

2023/8/30

In this research, we introduce a novel approach to the challenge of suction grasp point detection. Our method, exploiting the strengths of physics-based simulation and data-driven modeling, accounts for object dynamics during the grasping process, markedly enhancing the robot’s capability to handle previously unseen objects and scenarios in real-world settings. We benchmark DYNAMO-GRASP against established approaches via comprehensive evaluations in both simulated and real-world environments. DYNAMO-GRASP delivers improved grasping performance with greater consistency in both simulated and real-world settings. Remarkably, in real-world tests with challenging scenarios, our method demonstrates a success rate improvement of up to $48% $ over SOTA methods. Demonstrating a strong ability to adapt to complex and unexpected object dynamics, our method offers robust generalization to real-world challenges. The results of this research set the stage for more reliable and resilient robotic manipulation in intricate real-world situations. Experiment videos, dataset, model, and code are available at: https://sites. google. com/view/dynamo-grasp.

LiDAR-UDA: Self-ensembling Through Time for Unsupervised LiDAR Domain Adaptation

Authors

Amirreza Shaban*,JoonHo Lee*,Sanghun Jung*,Xiangyun Meng,Byron Boots

Published Date

2023

We introduce LiDAR-UDA, a novel two-stage self-training-based Unsupervised Domain Adaptation (UDA) method for LiDAR segmentation. Existing self-training methods use a model trained on labeled source data to generate pseudo labels for target data and refine the predictions via fine-tuning the network on the pseudo labels. These methods suffer from domain shifts caused by different LiDAR sensor configurations in the source and target domains. We propose two techniques to reduce sensor discrepancy and improve pseudo label quality: 1) LiDAR beam subsampling, which simulates different LiDAR scanning patterns by randomly dropping beams; 2) cross-frame ensembling, which exploits temporal consistency of consecutive frames to generate more reliable pseudo labels. Our method is simple, generalizable, and does not incur any extra inference cost. We evaluate our method on several public LiDAR datasets and show that it outperforms the state-of-the-art methods by more than 3.9% mIoU on average for all scenarios. Code will be available at https://github. com/JHLee0513/lidar_uda.

Learning to read braille: Bridging the tactile reality gap with diffusion models

Authors

Carolina Higuera,Byron Boots,Mustafa Mukadam

Journal

arXiv preprint arXiv:2304.01182

Published Date

2023/4/3

Simulating vision-based tactile sensors enables learning models for contact-rich tasks when collecting real world data at scale can be prohibitive. However, modeling the optical response of the gel deformation as well as incorporating the dynamics of the contact makes sim2real challenging. Prior works have explored data augmentation, fine-tuning, or learning generative models to reduce the sim2real gap. In this work, we present the first method to leverage probabilistic diffusion models for capturing complex illumination changes from gel deformations. Our tactile diffusion model is able to generate realistic tactile images from simulated contact depth bridging the reality gap for vision-based tactile sensing. On real braille reading task with a DIGIT sensor, a classifier trained with our diffusion model achieves 75.74% accuracy outperforming classifiers trained with simulation and other approaches. Project page: https://github.com/carolinahiguera/Tactile-Diffusion

Mahalo: Unifying offline reinforcement learning and imitation learning from observations

Authors

Anqi Li,Byron Boots,Ching-An Cheng

Published Date

2023/7/3

We study a new paradigm for sequential decision making, called offline policy learning from observations (PLfO). Offline PLfO aims to learn policies using datasets with substandard qualities: 1) only a subset of trajectories is labeled with rewards, 2) labeled trajectories may not contain actions, 3) labeled trajectories may not be of high quality, and 4) the data may not have full coverage. Such imperfection is common in real-world learning scenarios, and offline PLfO encompasses many existing offline learning setups, including offline imitation learning (IL), offline IL from observations (ILfO), and offline reinforcement learning (RL). In this work, we present a generic approach to offline PLfO, called Modality-agnostic Adversarial Hypothesis Adaptation for Learning from Observations (MAHALO). Built upon the pessimism concept in offline RL, MAHALO optimizes the policy using a performance lower bound that accounts for uncertainty due to the dataset’s insufficient coverage. We implement this idea by adversarially training data-consistent critic and reward functions, which forces the learned policy to be robust to data deficiency. We show that MAHALO consistently outperforms or matches specialized algorithms across a variety of offline PLfO tasks in theory and experiments. Our code is available at https://github. com/AnqiLi/mahalo.

Cajun: Continuous adaptive jumping using a learned centroidal controller

Authors

Yuxiang Yang,Guanya Shi,Xiangyun Meng,Wenhao Yu,Tingnan Zhang,Jie Tan,Byron Boots

Published Date

2023/12/2

We present CAJun, a novel hierarchical learning and control framework that enables legged robots to jump continuously with adaptive jumping distances. CAJun consists of a high-level centroidal policy and a low-level leg controller. In particular, we use reinforcement learning (RL) to train the centroidal policy, which specifies the gait timing, base velocity, and swing foot position for the leg controller. The leg controller optimizes motor commands for the swing and stance legs according to the gait timing to track the swing foot target and base velocity commands. Additionally, we reformulate the stance leg optimizer in the leg controller to speed up policy training by an order of magnitude. Our system combines the versatility of learning with the robustness of optimal control. We show that after 20 minutes of training on a single GPU, CAJun can achieve continuous, long jumps with adaptive distances on a Go1 robot with small sim-to-real gaps. Moreover, the robot can jump across gaps with a maximum width of 70cm, which is over $40% $ wider than existing methods.

Cafa: Class-aware feature alignment for test-time adaptation

Authors

Sanghun Jung,Jungsoo Lee,Nanhee Kim,Amirreza Shaban,Byron Boots,Jaegul Choo

Published Date

2023

Despite recent advancements in deep learning, deep neural networks continue to suffer from performance degradation when applied to new data that differs from training data. Test-time adaptation (TTA) aims to address this challenge by adapting a model to unlabeled data at test time. TTA can be applied to pretrained networks without modifying their training procedures, enabling them to utilize a well-formed source distribution for adaptation. One possible approach is to align the representation space of test samples to the source distribution (ie, feature alignment). However, performing feature alignment in TTA is especially challenging in that access to labeled source data is restricted during adaptation. That is, a model does not have a chance to learn test data in a class-discriminative manner, which was feasible in other adaptation tasks (eg, unsupervised domain adaptation) via supervised losses on the source data. Based on this observation, we propose a simple yet effective feature alignment loss, termed as Class-Aware Feature Alignment (CAFA), which simultaneously 1) encourages a model to learn target representations in a class-discriminative manner and 2) effectively mitigates the distribution shifts at test time. Our method does not require any hyper-parameters or additional losses, which are required in previous approaches. We conduct extensive experiments on 6 different datasets and show our proposed method consistently outperforms existing baselines.

See List of Professors in Byron Boots University(University of Washington)

Byron Boots FAQs

What is Byron Boots's h-index at University of Washington?

The h-index of Byron Boots has been 38 since 2020 and 45 in total.

What are Byron Boots's top articles?

The articles with the titles of

Robotic System Performing Dynamic Interaction in Human-Robot Cooperative Work for Assembly Operation

LocoMan: Advancing Versatile Quadrupedal Dexterity with Lightweight Loco-Manipulators

Multi-Sample Long Range Path Planning under Sensing Uncertainty for Off-Road Autonomous Driving

Scalable measurement error mitigation via iterative bayesian unfolding

Adversarial model for offline reinforcement learning

Model predictive control techniques for autonomous systems

Controlling position of robot by determining goal proposals by using neural networks

Motion policy networks

...

are the top articles of Byron Boots at University of Washington.

What are Byron Boots's research interests?

The research interests of Byron Boots are: Machine Learning, Artificial Intelligence, Robotics

What is Byron Boots's total number of citations?

Byron Boots has 7,746 citations in total.

What are the co-authors of Byron Boots?

The co-authors of Byron Boots are Dieter Fox, Le Song, Siddhartha Srinivasa, Frank Dellaert, J. Andrew Bagnell, Geoff Gordon.

    Co-Authors

    H-index: 128
    Dieter Fox

    Dieter Fox

    University of Washington

    H-index: 83
    Le Song

    Le Song

    Georgia Institute of Technology

    H-index: 76
    Siddhartha Srinivasa

    Siddhartha Srinivasa

    University of Washington

    H-index: 75
    Frank Dellaert

    Frank Dellaert

    Georgia Institute of Technology

    H-index: 73
    J. Andrew Bagnell

    J. Andrew Bagnell

    Carnegie Mellon University

    H-index: 65
    Geoff Gordon

    Geoff Gordon

    Carnegie Mellon University

    academic-engine

    Useful Links