Autonomous Hand Segmentation and Pose Estimation in-the-Wild

Abstract

This paper proposes an autonomous hand pose estimation method for in-the-wild datasets by cascading a pretrained YOLO image segmentation module with Keypoint Transformer. Precise segmentation of hands in-the-wild is achieved via pre-trained YOLO model. The segmented hand images are then centered in the input image, effectively reducing background interference and providing clearer and more accurate inputs for subsequent keypoint detection and pose estimation. Experimental results show that the method significantly improves the accuracy and efficiency of pose estimation, with the accurate segmentation of hand positions playing a crucial role in the entire image processing pipeline.

Publication
In 2025 International Conference on Pattern Recognition, Machine Vision and Artificial Intelligence (PRMVAI)
Mingqi Chen
Mingqi Chen
Postdoctoral Researcher
Qiang Li
Qiang Li
Professor and header of AG

My research interests include underwater robot, collaborative robots, humanoid robots.