I am Zhengtong Xu, and I go by Tong. I'm a fourth-year PhD candidate in robot learning at Purdue University, advised by Professor Yu She.
I'm also a part-time student researcher at Meta Reality Labs, where I work on dexterous manipulation policy learning. Before that, I did research internships at Meta Reality Labs and MERL.
[Jan. 2026]I received the Bilsland Dissertation Fellowship ($62,513.22) from Purdue University to support my final year of PhD studies.
[Oct. 2025]One paper accepted by IEEE T-RO.
[Apr. 2025]I received the Magoon Graduate Student Research Excellence Award at Purdue University (awarded to only 25 PhD students across the entire College of Engineering).
[Nov. 2024]I have passed my PhD preliminary exam and officially become a PhD candidate.
My research focuses on developing learning frameworks to enable robots to perform everyday manipulation tasks with human-level proficiency and dexterity. Specifically, I investigate:
Scalable robot learning paradigms based on vision, tactile, 3D, and their multimodal fusion.
Neural-symbolic approaches for building generalizable and interpretable robot learning systems, including differentiable optimization and VLM-based robot agents.
Contact-Grounded Policy: Dexterous Visuotactile Policy with Generative Contact Grounding Zhengtong Xu, Yeping Wang, Ben Abbatematteo, Jom Preechayasomboon, Sonny Chan, Nick Colonnese, Amirhossein H. Memar
Under Review, 2026
@misc{xu2026cgp,
title={Contact-Grounded Policy: Dexterous Visuotactile Policy with Generative Contact Grounding},
author={Zhengtong Xu and Yeping Wang and Ben Abbatematteo and Jom Preechayasomboon and Sonny Chan and Nick Colonnese and Amirhossein H. Memar},
year={2026},
eprint={2603.05687},
archivePrefix={arXiv},
primaryClass={cs.RO},
url={https://arxiv.org/abs/2603.05687},
}
@article{xu2024letac,
author={Xu, Zhengtong and She, Yu},
journal={IEEE Transactions on Robotics},
title={{LeTac-MPC}: Learning Model Predictive Control for Tactile-Reactive Grasping},
year={2024},
volume={40},
number={},
pages={4376-4395},
doi={10.1109/TRO.2024.3463470}
}
A generalizable end-to-end tactile-reactive grasping controller with differentiable MPC, combining learning and model-based approaches.
Stiffness Copilot: An Impedance Policy for Contact-Rich Teleoperation
Yeping Wang, Zhengtong Xu, Jom Preechayasomboon, Ben Abbatematteo, Amirhossein H. Memar, Nick Colonnese, Sonny Chan
Under Review, 2026
A stiffness generation policy for contact-rich teleoperation via zero-shot sim-to-real transfer.
MuxGel: Simultaneous Dual-Modal Visuo-Tactile Sensing via Spatially Multiplexing and Deep Reconstruction
Zhixian Hu, Zhengtong Xu, Sheeraz Athar, Juan Wachs, Yu She
Under Review, 2026
Enriches pretrained VLAs with tactile perception through contact-aware multimodal fusion for contact-rich manipulation.
AgenticLab: A Real-World Robot Agent Platform that Can See, Think, and Act
Pengyuan Guo*, Zhonghao Mai*, Zhengtong Xu*, Kaidi Zhang, Heng Zhang, Zichen Miao, Arash Ajoudani, Zachary Kingston, Qiang Qiu, Yu She
Under Review, 2026
@misc{guo2026agenticlab,
title={AgenticLab: A Real-World Robot Agent Platform that Can See, Think, and Act},
author={Pengyuan Guo and Zhonghao Mai and Zhengtong Xu and Kaidi Zhang and Heng Zhang and Zichen Miao and Arash Ajoudani and Zachary Kingston and Qiang Qiu and Yu She},
year={2026},
eprint={2602.01662},
archivePrefix={arXiv},
primaryClass={cs.RO},
url={https://arxiv.org/abs/2602.01662},
}
A model-agnostic robot agent platform and benchmark for open-world manipulation.
ManiFeel: Benchmarking and Understanding Visuotactile Manipulation Policy Learning
Quan Khanh Luu*, Pokuang Zhou*, Zhengtong Xu*, Zhiyuan Zhang, Qiang Qiu, Yu She
Under Review, 2026
New England Manipulation Symposium (Oral), 2025
@misc{zhang2025canonical,
title={Canonical Policy: Learning Canonical 3D Representation for SE(3)-Equivariant Policy},
author={Zhiyuan Zhang and Zhengtong Xu and Jai Nanda Lakamsani and Yu She},
year={2025},
eprint={2505.18474},
archivePrefix={arXiv},
primaryClass={cs.RO},
url={https://arxiv.org/abs/2505.18474},
}
Canonical Policy enables equivariant observation-to-action mappings by grouping both in-distribution and out-of-distribution point clouds to a canonical 3D representation.
UniT: Data Efficient Tactile Representation with Generalization to Unseen Objects Zhengtong Xu, Raghava Uppuluri, Xinwei Zhang, Cael Fitch, Philip Glen Crandall, Wan Shou, Dongyi Wang, Yu She
IEEE Robotics and Automation Letters (RA-L), 2025
@misc{xu2025unit,
title={{UniT}: Data Efficient Tactile Representation with Generalization to Unseen Objects},
author={Zhengtong Xu and Raghava Uppuluri and Xinwei Zhang and Cael Fitch and Philip Glen Crandall and Wan Shou and Dongyi Wang and Yu She},
year={2025},
eprint={2408.06481},
archivePrefix={arXiv},
primaryClass={cs.RO},
url={https://arxiv.org/abs/2408.06481},
}
VILP integrates the video generation model into policies, enabling the representation of multi-modal action distributions while reducing reliance on extensive high-quality robot action data.
Safe Human-Robot Collaboration with Risk-tunable Control Barrier Functions
Vipul K. Sharma*, Pokuang Zhou*, Zhengtong Xu*, Yu She, S. Sivaranjani
IEEE/ASME Transactions on Mechatronics (TMECH), 2025
@ARTICLE{xu2025rtcbf,
author={Sharma, Vipul K. and Zhou, Pokuang and Xu, Zhengtong and She, Yu and Sivaranjani, S.},
journal={IEEE/ASME Transactions on Mechatronics},
title={Safe Human–Robot Collaboration With Risk Tunable Control Barrier Functions},
year={2025},
doi={10.1109/TMECH.2025.3572047}}
We address safety in human-robot collaboration with uncertain human positions by formulating a chance-constrained problem using uncertain control barrier functions.
LeTO: Learning Constrained Visuomotor Policy with Differentiable Trajectory Optimization Zhengtong Xu, Yu She
IEEE Transactions on Automation Science and Engineering (T-ASE), 2024
@article{athar2023vistac,
title={Vistac towards a unified multi-modal sensing finger for robotic manipulation},
author={Athar, Sheeraz and Patel, Gaurav and Xu, Zhengtong and Qiu, Qiang and She, Yu},
journal={IEEE Sensors Journal},
year={2023},
publisher={IEEE}
}
VisTac seamlessly combines high-resolution tactile and visual perception in a single unified device.
Awards
Bilsland Dissertation Fellowship, Purdue University, 2026 (valued at $62,513.22, supporting final year of PhD studies)
Magoon Graduate Student Research Excellence Award, Purdue University, 2025 (awarded to only 25 PhD students across the entire Purdue College of Engineering)
Dr. Theodore J. and Isabel M. Williams Fellowship, Purdue University, 2022
National Scholarship, Ministry of Education of China, 2017
Reviewer Service
Conference on Robot Learning (CoRL)
IEEE Robotics and Automation Letters (RA-L)
IEEE Transactions on Robotics (T-RO)
IEEE International Conference on Robotics and Automation (ICRA)
Teaching
Vertically Integrated Projects (VIP)-GE Robotics and Autonomous Systems, Grad Mentor, Spring 2024/Fall 2023/Summer 2023
IE 474-Industrial Control Systems, Teaching Assistant, Fall 2022