• Home
  • Uncategorized
  • Sim-and-Human Co-training for Data-Efficient and Generalizable Robotic Manipulation

arXiv:2601.19406v1 Announce Type: cross
Abstract: Synthetic simulation data and real-world human data provide scalable alternatives to circumvent the prohibitive costs of robot data collection. However, these sources suffer from the sim-to-real visual gap and the human-to-robot embodiment gap, respectively, which limits the policy’s generalization to real-world scenarios. In this work, we identify a natural yet underexplored complementarity between these sources: simulation offers the robot action that human data lacks, while human data provides the real-world observation that simulation struggles to render. Motivated by this insight, we present SimHum, a co-training framework to simultaneously extract kinematic prior from simulated robot actions and visual prior from real-world human observations. Based on the two complementary priors, we achieve data-efficient and generalizable robotic manipulation in real-world tasks. Empirically, SimHum outperforms the baseline by up to $mathbf40%$ under the same data collection budget, and achieves a $mathbf62.5%$ OOD success with only 80 real data, outperforming the real only baseline by $7.1times$. Videos and additional information can be found at hrefhttps://kaipengfang.github.io/sim-and-humanproject website.

Subscribe for Updates

Copyright 2025 dijee Intelligence Ltd.   dijee Intelligence Ltd. is a private limited company registered in England and Wales at Media House, Sopers Road, Cuffley, Hertfordshire, EN6 4RY, UK registeration number 16808844