Zhengtong Xu 徐政通

Email: xu1703 AT purdue.edu

I am Zhengtong Xu, and I go by Tong. I'm a fourth-year PhD candidate in robot learning at Purdue LogoPurdue University, advised by Professor Yu She.

I'm also a part-time student researcher at Meta LogoMeta Reality Labs, where I work on dexterous manipulation policy learning. Before that, I did research internships at Meta LogoMeta Reality Labs and MERL LogoMERL.

I received my Bachelor's degree in mechanical engineering at HUST LogoHuazhong University of Science and Technology.

G. Scholar  /  Twitter  /  Github  /  LinkedIn

profile photo

News

  • [Jan. 2026]I received the Bilsland Dissertation Fellowship ($62,513.22) from Purdue University to support my final year of PhD studies.
  • [Oct. 2025]One paper accepted by IEEE T-RO.
  • [Apr. 2025]I received the Magoon Graduate Student Research Excellence Award at Purdue University (awarded to only 25 PhD students across the entire College of Engineering).
  • [Nov. 2024]I have passed my PhD preliminary exam and officially become a PhD candidate.
  • [Sep. 2024]One paper accepted by IEEE T-RO.

Research

* indicates equal contribution. indicates equal advising. Selected papers are highlighted.

My research focuses on developing learning frameworks to enable robots to perform everyday manipulation tasks with human-level proficiency and dexterity. Specifically, I investigate:

  • Scalable robot learning paradigms based on vision, tactile, 3D, and their multimodal fusion.
  • Neural-symbolic approaches for building generalizable and interpretable robot learning systems, including differentiable optimization and VLM-based robot agents.

Contact-Grounded Policy: Dexterous Visuotactile Policy with Generative Contact Grounding
Zhengtong Xu, Yeping Wang, Ben Abbatematteo, Jom Preechayasomboon, Sonny Chan, Nick Colonnese, Amirhossein H. Memar
Under Review, 2026

website / arXiv / video / bibtex

A visuotactile policy that empowers robot dexterity via multimodal contact grounding.

DiffOG: Differentiable Policy Trajectory Optimization with Generalizability
Zhengtong Xu, Zichen Miao, Qiang Qiu, Zhe Zhang, Yu She
IEEE Transactions on Robotics (T-RO), 2025

website / arXiv / video / code / bibtex

DiffOG introduces a transformer-based differentiable trajectory optimization framework for action refinement in imitation learning.

LeTac-MPC: Learning Model Predictive Control for Tactile-reactive Grasping
Zhengtong Xu, Yu She
IEEE Transactions on Robotics (T-RO), 2024

arXiv / video / code / bibtex

A generalizable end-to-end tactile-reactive grasping controller with differentiable MPC, combining learning and model-based approaches.

Stiffness Copilot: An Impedance Policy for Contact-Rich Teleoperation
Yeping Wang, Zhengtong Xu, Jom Preechayasomboon, Ben Abbatematteo, Amirhossein H. Memar, Nick Colonnese, Sonny Chan
Under Review, 2026

arXiv(soon) / video(soon)

A stiffness generation policy for contact-rich teleoperation via zero-shot sim-to-real transfer.

MuxGel: Simultaneous Dual-Modal Visuo-Tactile Sensing via Spatially Multiplexing and Deep Reconstruction
Zhixian Hu, Zhengtong Xu, Sheeraz Athar, Juan Wachs, Yu She
Under Review, 2026

website / arXiv

A single sensor achieving simultaneous visual and tactile perception via spatial multiplexing and cross-modal generation.

Learning Tactile-Aware Quadrupedal Loco-Manipulation Policies
Pokuang Zhou, Yuhao Zhou, Quan Luu, Seungho Han, Heng Zhang, Binghao Huang, Yunzhu Li, Arash Ajoudani, Zhengtong Xu, Yu She
Under Review, 2026

arXiv(soon) / video(soon)

A hierarchical training framework for learning tactile-aware quadrupedal loco-manipulation policies.

TacVLA: Contact-Aware Tactile Fusion for Robust Vision-Language-Action Manipulation
Kaidi Zhang*, Heng Zhang*, Zhengtong Xu, Zhiyuan Zhang, MD Rakibul Islam Prince, Li Xiang, Xiaojing Han, Yuhao Zhou, Arash Ajoudani, Yu She
Under Review, 2026

arXiv(soon) / video(soon)

Enriches pretrained VLAs with tactile perception through contact-aware multimodal fusion for contact-rich manipulation.

AgenticLab: A Real-World Robot Agent Platform that Can See, Think, and Act
Pengyuan Guo*, Zhonghao Mai*, Zhengtong Xu*, Kaidi Zhang, Heng Zhang, Zichen Miao, Arash Ajoudani, Zachary Kingston, Qiang Qiu, Yu She
Under Review, 2026

website / arXiv / video / code(soon) / bibtex

A model-agnostic robot agent platform and benchmark for open-world manipulation.

ManiFeel: Benchmarking and Understanding Visuotactile Manipulation Policy Learning
Quan Khanh Luu*, Pokuang Zhou*, Zhengtong Xu*, Zhiyuan Zhang, Qiang Qiu, Yu She
Under Review, 2026
New England Manipulation Symposium (Oral), 2025

website / arXiv / code / bibtex

ManiFeel is a reproducible and scalable simulation benchmark for studying supervised visuotactile policy learning.

UNIC: Learning Unified Multimodal Extrinsic Contact Estimation
Zhengtong Xu, Yuki Shirai
IEEE International Conference on Robotics and Automation (ICRA), 2026

arXiv / video / code(soon) / bibtex

A unified multimodal framework for extrinsic contact estimation that operates without any prior knowledge or camera calibration.

Canonical Policy: Learning Canonical 3D Representation for SE(3)-Equivariant Policy
Zhiyuan Zhang*, Zhengtong Xu*, Jai Nanda Lakamsani, Yu She
IEEE Transactions on Robotics (T-RO), conditionally accepted, 2026

website / arXiv / code / bibtex

Canonical Policy enables equivariant observation-to-action mappings by grouping both in-distribution and out-of-distribution point clouds to a canonical 3D representation.

UniT: Data Efficient Tactile Representation with Generalization to Unseen Objects
Zhengtong Xu, Raghava Uppuluri, Xinwei Zhang, Cael Fitch, Philip Glen Crandall, Wan Shou, Dongyi Wang, Yu She
IEEE Robotics and Automation Letters (RA-L), 2025

website / arXiv / video / code / bibtex

Learn a tactile representation with generalizability only by a single simple object.

VILP: Imitation Learning with Latent Video Planning
Zhengtong Xu, Qiang Qiu, Yu She
IEEE Robotics and Automation Letters (RA-L), 2025

arXiv / video / code / bibtex

VILP integrates the video generation model into policies, enabling the representation of multi-modal action distributions while reducing reliance on extensive high-quality robot action data.

Safe Human-Robot Collaboration with Risk-tunable Control Barrier Functions
Vipul K. Sharma*, Pokuang Zhou*, Zhengtong Xu*, Yu She, S. Sivaranjani
IEEE/ASME Transactions on Mechatronics (TMECH), 2025

paper / video / bibtex

We address safety in human-robot collaboration with uncertain human positions by formulating a chance-constrained problem using uncertain control barrier functions.

LeTO: Learning Constrained Visuomotor Policy with Differentiable Trajectory Optimization
Zhengtong Xu, Yu She
IEEE Transactions on Automation Science and Engineering (T-ASE), 2024

arXiv / video / code / bibtex

LeTO is a "gray box" method which marries optimization-based safety and interpretability with representational abilities of neural networks.

sym

VisTac: Toward a Unified Multimodal Sensing Finger for Robotic Manipulation
Sheeraz Athar*, Gaurav Patel*, Zhengtong Xu, Qiang Qiu, Yu She
IEEE Sensors Journal, 2023

paper / video / bibtex

VisTac seamlessly combines high-resolution tactile and visual perception in a single unified device.

Awards

  • Bilsland Dissertation Fellowship, Purdue University, 2026
    (valued at $62,513.22, supporting final year of PhD studies)
  • Magoon Graduate Student Research Excellence Award, Purdue University, 2025
    (awarded to only 25 PhD students across the entire Purdue College of Engineering)
  • Dr. Theodore J. and Isabel M. Williams Fellowship, Purdue University, 2022
  • National Scholarship, Ministry of Education of China, 2017

Reviewer Service

  • Conference on Robot Learning (CoRL)
  • IEEE Robotics and Automation Letters (RA-L)
  • IEEE Transactions on Robotics (T-RO)
  • IEEE International Conference on Robotics and Automation (ICRA)

Teaching

  • Vertically Integrated Projects (VIP)-GE Robotics and Autonomous Systems, Grad Mentor, Spring 2024/Fall 2023/Summer 2023
  • IE 474-Industrial Control Systems, Teaching Assistant, Fall 2022

Website template from Jon Barron's website.