A Novel Approach for Leveraging Object Detection for 3D Human Pose Estimation in Complex Human-Robot Collaboration Environments
- Although 3D Human Pose Estimation has major breakthroughs in recent years, 3D pose estimation in complex scenarios remains difficult. One of the reasons is the lack of diverse 3D datasets for training and generalizing the models. This issue is counteracted by acquiring a dataset of Human-Robot Collaboration scenes featuring different objects, such as a cobot. We propose a novel two-step method,Although 3D Human Pose Estimation has major breakthroughs in recent years, 3D pose estimation in complex scenarios remains difficult. One of the reasons is the lack of diverse 3D datasets for training and generalizing the models. This issue is counteracted by acquiring a dataset of Human-Robot Collaboration scenes featuring different objects, such as a cobot. We propose a novel two-step method, where first a 3D Object detection task with VoteNet is performed to identify the human in the scenario and claim it as a region of interest for the pose estimation task. Second, this region of interest is cropped and passed into the 3D Human Pose Estimation algorithm SPiKE, which locates 15 keypoints of the human. Based on this procedure, our method improves detection in complex scenarios. Furthermore, this article compares the benefits of training the algorithm additionally on the obtained Human-Robot Collaboration dataset compared to training it with the standard ITOP dataset. While the SPiKE algorithm makes no correct prediction on the Human-Robot Collaboration scenario, the results of the two-step SPiKEVN approach with mAP of 41.17 % is significantly lower as the benchmark model on the ITOP dataset. Nonetheless, the SPiKEVN model exhibits similar performance to SPiKEman with a difference of 2.43 % mAP indicating the method is effectively functioning.…


| Document Type: | Conference Proceeding |
|---|---|
| State of review: | Begutachtet (reviewed) |
| Conference Type: | Konferenzartikel |
| Zitierlink: | https://opus.hs-offenburg.de/11589 | Bibliografische Angaben |
| Title (English): | A Novel Approach for Leveraging Object Detection for 3D Human Pose Estimation in Complex Human-Robot Collaboration Environments |
| Conference: | IEEE International Conference on Automation Science and Engineering (21. : 17-21 August 2025 : Los Angeles, CA, USA) |
| Author: | Sinan SümeStaff MemberORCiD, Amal Kaithavalappil AjayStaff MemberORCiD, Thomas WendtStaff MemberORCiDGND, Stefan RupitschORCiD |
| Year of Publication: | 2025 |
| Publisher: | IEEE |
| First Page: | 1134 |
| Last Page: | 1139 |
| Parent Title (English): | 2025 IEEE 21st International Conference on Automation Science and Engineering (CASE) |
| ISBN: | 979-8-3315-2246-9 (Elektronisch) |
| ISBN: | 979-8-3315-2247-6 (Print on Demand) |
| ISSN: | 2161-8089 (Elektronisch) |
| ISSN: | 2161-8070 (Print on Demand) |
| DOI: | https://doi.org/10.1109/CASE58245.2025.11164071 |
| Language: | English | Inhaltliche Informationen |
| Institutes: | Fakultät Wirtschaft (W) |
| Research: | WLRI - Work-Life Robotics Institute |
| Collections of the Offenburg University: | Bibliografie |
| Tag: | 3D Human Pose Detection; Human-Robot Collaboration; Human-Robot Interaction; Object Detection; Time of Flight Sensor | Formale Angaben |
| Relevance for "Jahresbericht über Forschungsleistungen": | 5-fach | Konferenzbeitrag |
| Open Access: | Closed |
| Licence (German): | Urheberrechtlich geschützt |



