Volltext-Downloads (blau) und Frontdoor-Views (grau)

A Novel Approach for Leveraging Object Detection for 3D Human Pose Estimation in Complex Human-Robot Collaboration Environments

  • Although 3D Human Pose Estimation has major breakthroughs in recent years, 3D pose estimation in complex scenarios remains difficult. One of the reasons is the lack of diverse 3D datasets for training and generalizing the models. This issue is counteracted by acquiring a dataset of Human-Robot Collaboration scenes featuring different objects, such as a cobot. We propose a novel two-step method,Although 3D Human Pose Estimation has major breakthroughs in recent years, 3D pose estimation in complex scenarios remains difficult. One of the reasons is the lack of diverse 3D datasets for training and generalizing the models. This issue is counteracted by acquiring a dataset of Human-Robot Collaboration scenes featuring different objects, such as a cobot. We propose a novel two-step method, where first a 3D Object detection task with VoteNet is performed to identify the human in the scenario and claim it as a region of interest for the pose estimation task. Second, this region of interest is cropped and passed into the 3D Human Pose Estimation algorithm SPiKE, which locates 15 keypoints of the human. Based on this procedure, our method improves detection in complex scenarios. Furthermore, this article compares the benefits of training the algorithm additionally on the obtained Human-Robot Collaboration dataset compared to training it with the standard ITOP dataset. While the SPiKE algorithm makes no correct prediction on the Human-Robot Collaboration scenario, the results of the two-step SPiKEVN approach with mAP of 41.17 % is significantly lower as the benchmark model on the ITOP dataset. Nonetheless, the SPiKEVN model exhibits similar performance to SPiKEman with a difference of 2.43 % mAP indicating the method is effectively functioning.show moreshow less

Export metadata

Additional Services

Search Google Scholar

Statistics

frontdoor_oas
Metadaten
Document Type:Conference Proceeding
State of review:Begutachtet (reviewed)
Conference Type:Konferenzartikel
Zitierlink: https://opus.hs-offenburg.de/11589
Bibliografische Angaben
Title (English):A Novel Approach for Leveraging Object Detection for 3D Human Pose Estimation in Complex Human-Robot Collaboration Environments
Conference:IEEE International Conference on Automation Science and Engineering (21. : 17-21 August 2025 : Los Angeles, CA, USA)
Author:Sinan SümeStaff MemberORCiD, Amal Kaithavalappil AjayStaff MemberORCiD, Thomas WendtStaff MemberORCiDGND, Stefan RupitschORCiD
Year of Publication:2025
Publisher:IEEE
First Page:1134
Last Page:1139
Parent Title (English):2025 IEEE 21st International Conference on Automation Science and Engineering (CASE)
ISBN:979-8-3315-2246-9 (Elektronisch)
ISBN:979-8-3315-2247-6 (Print on Demand)
ISSN:2161-8089 (Elektronisch)
ISSN:2161-8070 (Print on Demand)
DOI:https://doi.org/10.1109/CASE58245.2025.11164071
Language:English
Inhaltliche Informationen
Institutes:Fakultät Wirtschaft (W)
Research:WLRI - Work-Life Robotics Institute
Collections of the Offenburg University:Bibliografie
Tag:3D Human Pose Detection; Human-Robot Collaboration; Human-Robot Interaction; Object Detection; Time of Flight Sensor
Formale Angaben
Relevance for "Jahresbericht über Forschungsleistungen":5-fach | Konferenzbeitrag
Open Access: Closed 
Licence (German):License LogoUrheberrechtlich geschützt