|
Robots Learning from Humans, Humans Learning from Robots
Chaitanya Chawla, Ananya Bal, Laszlo Jeni, Guanya Shi
In Progress
Code (coming soon)
Bridging Generative Diffusion and Reinforcement Learning for Dexterous Robot Manipulation:
Developed a framework for learning contact-rich manipulation by retargeting human-object
interactions from the ARCTIC dataset to a humanoid robot using RL environments.
To enhance generalizability beyond the dataset, we are integrating a text-conditioned Diffusion
Transformer with IK-guidance and Critic-informed sampling, ensuring the generative model
accounts for robot kinematics and task-success metrics. Stay tuned!
|
|
Data Engine for Web-Scale Robot Training
Chaitanya Chawla*, Jeremy Collins*, Rutav Shah, Krishnan Srinivasan, Homanga Bharadwaj
In Progress
Code (coming soon)
Building a data engine for web-scale robot training. Stay tuned!
|
|
SkillGraph: An Ontology-Based Framework for Intelligent Multi-Robot Assembly and Planning
Peiqi Liu*, Philip Huang*, Chaitanya Chawla*, Guanya Shi, Jiaoyang Li, Changliu Liu
In Preparation for Robotics and Automation Letters (RA-L)
Technical Report (Request Access) /
Code (coming soon)
Developed SkillGraph, an ontology-driven framework that unifies semantic knowledge representation
with automated task planning to enable flexible multi-robot assembly.
By bridging high-level reasoning with low-level control, the system dynamically maps abstract task
goals into executable skill primitives, ensuring robust manipulation in complex environments.
The SkillGraph comprises of multiple user-interfaces to interact with the system, one of which is a human-video demonstration
—where the system is able to learn the skill from the human-video demonstration and then use it to plan and execute the task.
|
|
Humanoid Policy ~ Human Policy
Ri-Zhao Qiu*, Shiqi Yang*, Xuxin Cheng*, Chaitanya Chawla*,
Jialong Li, Tairan He, Ge Yan, David J. Yoon, Ryan Hoque, Jian Zhang, Sha Yi, Guanya Shi, Xiaolong Wang
Conference on Robot Learning, CoRL 2025
Robot Data Workshop @ CoRL 2025
Paper /
Website /
Code /
Dataset
Developed HAT, an end-to-end policy that models human demonstrators as a distinct humanoid embodiment
to enable scalable imitation learning. The framework utilizes a unified kinematic representation to
directly map egocentric human video to robot actions, bypassing the need for complex intermediate
affordance representations. This approach allows for efficient learning from diverse human demonstrations
—significantly improving policy robustness and out-of-distribution generalization for bimanual manipulation.
|
|
Translating Agent-Environment Interactions across Humans and Robots
Tanmay Shankar, Chaitanya Chawla, Almut Wakel, Jean Oh
International Conference on Intelligent Robots and Systems, IROS 2024
Paper /
Website /
Code /
Video
We developed an unsupervised approach to learn temporal abstractions of skills incorporating
agent-environment interactions. We hope to learn representations of patterns of motion of objects
in the environment, or patterns of change of state. Our approach is able to learn semantically meaningful
skill segments across robot and human demonstrations, despite being completely unsupervised.
|
|
Robot-Agnostic Framework for One-Shot Intrinsic Feature Extraction
Chaitanya Chawla,
Andrei Costinescu,
Darius Burschka
Arxiv 2023
Report /
Presentation /
Code
We developed an algorithmic framework to extract different intrinsic features from human
demonstrations. We are studying various features, including interactions with objects
along the trajectory, analyzing the environment for interactions with the background
(e.g., wiping or writing), and classifying the type of motion within a trajectory
segment (e.g., shaking, rotating, or transporting).
|
|
Visual Teleoperation using Dynamic Motion Primitives
Chaitanya Chawla,
Dongheui Lee
Independent Research Project
Report /
Presentation /
Code
We presented a method to learn human motions using a Learning-From-Demonstration approach.
Using Dynamic Motion Primitives, we were able to teleoperate a Franka Panda Arm using the learned trajectories.
|
Projects
|
Multi-GPU SpeedUp with Custom-implementation in Needle Framework (DDP and ZeRO-3)
Yifu Yuan, Yuanhang Zhang, Chaitanya Chawla
10-714: Deep Learning Systems
Report /
Code (Request Access) /
Video
Multi-GPU SpeedUp with Custom-implementation in Needle Framework (DDP and ZeRO-3)
|
|
Self-supervised fine-tuning Pre-Grasps through 3D Object Generation
Chaitanya Chawla,
Almut Wakel, Eyob Dagnachew
10-623: Generative AI
Report /
Code /
Presentation
Comparing different methods including Autoencoders and PCA, for feature representation in face reconstruction
|
|
Learning Dexterous Manipulation from Human Video Pretraining using 3D Point Tracks
Chaitanya Chawla,
Sungjae Park, Lucas Wu, Junkai Huang, Yanbo Xu
16-831: Introduction to Robot Learning
Report /
Presentation
We proposed a pipeline to benchmark pre-training methods using different state representations.
Our method consisted of extracting sensorimotor information from videos by lifting the human hand and the manipulated object in a
shared 3D space in simulation (IsaacGym), i.e. either 3D point-tracks or 3D meshes.
Then, we retarget hand-trajectories to a Franka with a Shadow hand.
Finally, we fine-tune on various tasks.
|
|
Candy Throwing Robot for Halloween 2023!
C. Chawla
2 hours long Project on Halloween Eve, Bot Intelligence Group
Distributing candies during Halloween at the Robotics Institute, Carnegie Mellon University
|
Work Experience
Robotics Engineer | July 2023 - April 2024
- Developed a perception pipeline for detecting and reporting measurements from analog gauges.
Set up data-annotation, post-processing, and real-time inference of gauge measurements.
Dockerized the application to integrate with Boston Dynamics' Spot to autonomously collect data
in a factory environment.
- Developed and deployed a sludge-detection system based on Yolo-v8 to detect amount of sludge present in outdoor water-storage tanks.
Achieved an accuracy of 95% by fine-tuning a trained RESNET backbone with custom data as well as data-augmentation to account for various lighting,
time of the day, as well as adversarial objects.
- Created a webRTC pipeline using gRPC to transfer point cloud data from Spot's
LIDAR sensor to Oculus VR Headset, enabling a remote user to observe Spot's immediate
environment in real time.
- Migrated the company's robotic framework from ROS to ROS2.
Achievements
- German National Scholarship - Deutschland Stipendium (2021-2024)
- Heinrich and Lotte Muhlfenzl Scholarship - undergraduate research scholarship (2023)
- TUM PROMOS - merit scholarship for stay-abroad research (2023)
- Max Weber Program - nominated by the university (2022)
Reviewer
- Robotics and Automation Letters (RA-L)
- International Conference on Learning Representations (ICLR)
- Conference on Robot Learning (CoRL)
- International Conference on Intelligent Robots and Systems (IROS)
Teaching Assistant
- Introduction to Robot Learning (16-831), Carnegie Mellon University, 2025
- Robotic Control Laboratory (6-931), Technical University of Munich, 2023
- Mathematical Analysis (9-411), Technical University of Munich, 2022
|