Recent industrial paradigms have increasingly adopted human-centric approaches that prioritize worker well-being in collaborative robotics, where humans and robots share workspaces and coordinate their actions in real time. Understanding these interactions requires tools capable of capturing the embodied, dynamic nature of human-robot collaboration while remaining non-intrusive and ecologically valid. Markerless pose estimation offers a promising solution, yet existing approaches track humans and robots separately, missing the collaborative dynamics that emerge from their synchronized movements. We present an integrated pipeline for concurrent 3D pose estimation of humans and the UR10e cobot using a single RGB-D camera. The system combines 2D keypoint detection with depth-based reconstruction to generate synchronized 3D trajectories for both agents within a unified reference frame. Quantitative validation was conducted on a custom-trained UR10e model using encoder-based ground truth, demonstrating the system’s reliability across both collaborative assembly tasks and controlled trajectories. By capturing human and robot as a unified dyadic system rather than independent entities, the pipeline enables quantitative analysis of interaction fluency, embodied coordination, and shared action dynamics, fundamental aspects for advancing safe, efficient, and human-centered collaborative robotics. Link Github: https://github.com/egleorl/Markerless-3D-Human-Cobot-Pose-Estimation
Capturing Shared Actions and Coordination in Collaborative Workspaces: A Markerless Pipeline for Synchronized Human-Robot 3D Pose Estimation
Federica Nenna
;Michele Mingardi;Giulia Buodo;Luciano Gamberini
2026
Abstract
Recent industrial paradigms have increasingly adopted human-centric approaches that prioritize worker well-being in collaborative robotics, where humans and robots share workspaces and coordinate their actions in real time. Understanding these interactions requires tools capable of capturing the embodied, dynamic nature of human-robot collaboration while remaining non-intrusive and ecologically valid. Markerless pose estimation offers a promising solution, yet existing approaches track humans and robots separately, missing the collaborative dynamics that emerge from their synchronized movements. We present an integrated pipeline for concurrent 3D pose estimation of humans and the UR10e cobot using a single RGB-D camera. The system combines 2D keypoint detection with depth-based reconstruction to generate synchronized 3D trajectories for both agents within a unified reference frame. Quantitative validation was conducted on a custom-trained UR10e model using encoder-based ground truth, demonstrating the system’s reliability across both collaborative assembly tasks and controlled trajectories. By capturing human and robot as a unified dyadic system rather than independent entities, the pipeline enables quantitative analysis of interaction fluency, embodied coordination, and shared action dynamics, fundamental aspects for advancing safe, efficient, and human-centered collaborative robotics. Link Github: https://github.com/egleorl/Markerless-3D-Human-Cobot-Pose-Estimation| File | Dimensione | Formato | |
|---|---|---|---|
|
Capturing_Shared_Actions_and_Coordination_in_Collaborative_Workspaces_A_Markerless_Pipeline_for_Synchronized_HumanRobot_3D_Pose_Estimation.pdf
accesso aperto
Tipologia:
Published (Publisher's Version of Record)
Licenza:
Creative commons
Dimensione
2.92 MB
Formato
Adobe PDF
|
2.92 MB | Adobe PDF | Visualizza/Apri |
Pubblicazioni consigliate
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.




