Animatable 3D human reconstruction from a single image is a challenging problem due to the ambiguity in decoupling geometry, appearance, and deformation.
Recent advances in 3D human reconstruction mainly focus on static human modeling, and the reliance of using synthetic 3D scans for training limits their generalization ability.
Conversely, optimization-based video methods achieve higher fidelity but demand controlled capture conditions and computationally intensive refinement processes.
Motivated by the emergence of large reconstruction models for efficient static reconstruction, we propose LHM (Large Animatable Human Reconstruction Model) to infer high-fidelity avatars represented as 3D Gaussian splatting in a feed-forward pass.
Our model leverages a multimodal transformer architecture to effectively encode the human body positional features and image features with attention mechanism, enabling detailed preservation of clothing geometry and texture.
To further boost the face identity preservation and fine detail recovery, we propose a head feature pyramid encoding scheme to aggregate multi-scale features of the head regions.
Extensive experiments demonstrate that our LHM generates plausible animatable human in seconds without post-processing for face and hands,
outperforming existing methods in both reconstruction accuracy and generalization ability.
Video
For netizens in China, considering the problem of Internet restrictions, we provide a video link to bilibili.
Methodology
Overview of the proposed LHM. Our method extracts body and head image tokens from the input image,
and utilizes the proposed Multimodal Body-Head Transformer (MBHT) to fuse the 3D geometric body tokens with the image tokens.
After the attention-based fusion process, the geometric body tokens are decoded into Gaussian parameters.
Animation Results
Citation
If you find our approach helpful, you may consider citing our work.
@article{qiu2025LHM,
title={LHM: Large Animatable Human Reconstruction Model for Single Image to 3D in One Second},
author={Lingteng Qiu and Xiaodong Gu and Peihao Li and Qi Zuo
and Weichao Shen and Junfei Zhang and Kejie Qiu and Weihao Yuan
and Guanying Chen and Zilong Dong and Liefeng Bo
},
booktitle={arXiv preprint arXiv:xxxxx},
year={2025}
}