Haochen Shi

I am a first-year CS Ph.D. student at Stanford, advised by Karen Liu and Shuran Song. During my Master's at Stanford, my research focused on robotics and computer vision, more specificly deformable object manipulation, in collaboration with Jiajun Wu, Huazhe Xu, and Yunzhu Li. Previously, during undergraduate at UW-Madison, my work with Michael Gleicher and Danny Rakita explored motion planning algorithms for robots.

I'm interested in challenging robotic manipulation and locomotion tasks. In my leisure time, I love playing tennis, Go, chess, and mahjong.

Email  /  Google Scholar  /  Twitter  /  Github

Haochen Shi

Research

FürElise: Capturing and Physically Synthesizing Hand Motions of Piano Performance

Ruocheng Wang*, Pei Xu*, Haochen Shi, Elizabeth Schumann, C. Karen Liu

SIGGRAPH Asia, 2024

Piano playing requires agile, precise, and coordinated hand control that stretches the limits of dexterity. Hand motion models with the sophistication to accurately recreate piano playing have a wide range of applications in character animation, embodied AI, biomechanics, and VR/AR. In this paper, we construct a first-of-its-kind large-scale dataset that contains approximately 10 hours of 3D hand motion and audio from 15 elite-level pianists playing 153 pieces of classical music.

DexCap: Scalable and Portable Mocap Data Collection System for Dexterous Manipulation

Chen Wang, Haochen Shi, Weizhuo Wang, Ruohan Zhang, Li Fei-Fei, C. Karen Liu

RSS, 2024

Imitation learning from human hand motion data presents a promising avenue for imbuing robots with human-like dexterity in real-world manipulation tasks. Despite this potential, substantial challenges persist, particularly with the portability of existing hand motion capture (mocap) systems and the difficulty of translating mocap data into effective control policies. To tackle these issues, we introduce DexCap, a portable hand motion capture system, alongside DexIL, a novel imitation algorithm for training dexterous robot skills directly from human hand mocap data.

RoboPack: Learning Tactile-Informed Dynamics Models for Dense Packing

Bo Ai*, Stephan Tian*, Haochen Shi, Yixuan Wang, Cheston Tan, Yunzhu Li, Jiajun Wu

RSS, 2024

Abridged in ICRA 2024 workshops: ViTac, 3DVRM, and Future Roadmap for Manipulation Skills

Tactile feedback is critical for understanding the dynamics of both rigid and deformable objects in many manipulation tasks, such as non-prehensile manipulation and dense packing. We introduce an approach that combines visual and tactile sensing for robotic manipulation by learning a neural, tactile-informed dynamics model. Our proposed framework, RoboPack, employs a recurrent graph neural network to estimate object states, including particles and object-level latent physics information, from historical visuo-tactile observations and to perform future state predictions.

RoboCraft: Learning to see, simulate, and shape elasto-plastic objects in 3D with graph networks

Haochen Shi*, Huazhe Xu*, Zhiao Huang, Yunzhu Li, Jiajun Wu

The International Journal of Robotics Research (IJRR)

Modeling and manipulating elasto-plastic objects are essential capabilities for robots to perform complex industrial and household interaction tasks (e.g., stuffing dumplings, rolling sushi, and making pottery). However, due to the high degrees of freedom of elasto-plastic objects, significant challenges exist in virtually every aspect of the robotic manipulation pipeline, e.g., representing the states, modeling the dynamics, and synthesizing the control signals. We propose to tackle these challenges by employing a particle-based representation for elasto-plastic objects in a model-based planning framework.

RoboCook: Long-Horizon Elasto-Plastic Object Manipulation with Diverse Tools

Haochen Shi*, Huazhe Xu*, Samuel Clarke, Yunzhu Li, Jiajun Wu

CoRL, 2023

Best System Paper

Humans excel in complex long-horizon soft body manipulation tasks via flexible tool use: bread baking requires a knife to slice the dough and a rolling pin to flatten it. Often regarded as a hallmark of human cognition, tool use in autonomous robots remains limited due to challenges in understanding tool-object interactions. Here we develop an intelligent robotic system, RoboCook, which perceives, models, and manipulates elasto-plastic objects with various tools.

RoboCraft: Learning to See, Simulate, and Shape Elasto-Plastic Objects with Graph Networks

Haochen Shi*, Huazhe Xu*, Zhiao Huang, Yunzhu Li, Jiajun Wu

RSS, 2022

Abridged in ICRA 2022 workshop on Representing and Manipulating Deformable Objects

Covered by [MIT News] [Stanford HAI] [New Scientist]

Modeling and manipulating elasto-plastic objects are essential capabilities for robots to perform complex industrial and household interaction tasks. However, significant challenges exist in virtually every aspect of the robotic manipulation pipeline, e.g., representing the states, modeling the dynamics, and synthesizing the control signals. We propose to tackle these challenges by employing a particle-based representation for elasto-plastic objects in a model-based planning framework.

CollisionIK: A per-instant pose optimization method for generating robot motions with environment collision avoidance

Daniel Rakita, Haochen Shi, Bilge Mutlu, Michael Gleicher

ICRA, 2021

In this work, we present a per-instant pose optimization method that can generate configurations that achieve specified pose or motion objectives as best as possible over a sequence of solutions, while also simultaneously avoiding collisions with static or dynamic obstacles in the environment.

Teaching

Talks

Professional Service