GenHMR: A Breakthrough in 3D Human Pose and Shape Estimation – Accepted at AAAI 2025
![](https://research.charlotte.edu/wp-content/uploads/sites/631/2025/02/Squating.gif)
We are pleased to announce that our paper, “GenHMR: Generative Human Mesh Recovery for Accurate 3D Pose and Shape Estimation,” has been accepted for presentation at AAAI 2025. GenHMR introduces a novel generative framework that surpasses existing methods in parametric 3D human pose and shape (HPS) estimation, offering state-of-the-art accuracy and robustness.
Advancing 3D Human Mesh Recovery
Traditional deterministic and probabilistic methods for 3D pose estimation struggle with depth ambiguity, occlusions, and uncertainty in 2D-to-3D mappings. GenHMR redefines the problem by explicitly modeling uncertainties and leveraging generative modeling techniques to achieve high-fidelity 3D reconstructions.
Key Innovations of GenHMR:
- Generative HMR Reformulation: Moves beyond deterministic pipelines by modeling uncertainties in 2D-to-3D mapping, effectively addressing depth ambiguities and occlusions.
- Pose Tokenizer (VQ-VAE): Encodes 3D poses into discrete tokens, ensuring compact, kinematically valid representations for high-quality reconstructions.
- Image-Conditional Masked Transformer: Learns probabilistic pose distributions by predicting masked pose tokens, capturing complex joint dependencies for enhanced 3D accuracy.
- Uncertainty-Guided Sampling: Iteratively refines pose predictions, progressively improving accuracy by re-predicting low-confidence tokens during inference.
- 2D Pose-Guided Refinement: Ensures precise alignment between the reconstructed 3D mesh and 2D pose cues, enhancing final outputs.
Setting a New Standard for 3D Human Mesh Recovery
By combining generative modeling, iterative refinement, and uncertainty quantification, GenHMR outperforms existing methods in both accuracy and robustness, establishing a new benchmark for 3D human pose estimation.
Acknowledgments
This research was made possible through the contributions of Pu Wang, Ekkasit Pinyoanuntapong, Hongfei Xue, Srijan Das, and Chen Chen.
Read the Full Paper: Link
Project Page: Link
Stay connected for more updates on advancements in human-centric AI and 3D vision research.