NeurIPS 2026 Submission

FactorizedHMR: A Hybrid Framework for Video Human Mesh Recovery

Human Mesh Recovery (HMR) is fundamentally ambiguous: under occlusion or weak depth cues, multiple 3D bodies can explain the same image evidence. Building on this observation, we propose FactorizedHMR, a two-stage framework that treats these two regimes differently.

University of Central Florida

Abstract

Human Mesh Recovery (HMR) is fundamentally ambiguous: under occlusion or weak depth cues, multiple 3D bodies can explain the same image evidence. This ambiguity is not uniform across the body, as torso pose and root structure are often relatively well constrained, whereas distal articulations such as the arms and legs are more uncertain. Building on this observation, we propose FactorizedHMR, a two-stage framework that treats these two regimes differently. A deterministic regression module first recovers a stable torso-root anchor, and a probabilistic flow-matching module then completes the remaining non-torso articulation. To make this completion reliable, we combine a composite target representation with geometry-aware supervision and feature-aware classifier-free guidance, preserving the torso-root anchor while improving single-reference recovery of ambiguity-prone articulation. We also introduce a synthetic data pipeline that provides the paired image-camera-motion supervision under diverse viewpoints. Across camera-space and world-space benchmarks, FactorizedHMR remains competitive with strong baselines, with the clearest gains in occlusion-heavy recovery and drift-sensitive world-space metrics.

Motivation

Human Mesh Recovery (HMR) reconstructs 3D human pose, shape, and motion from visual observations. Yet video HMR remains difficult under partial observation, especially with occlusion, truncation, and depth ambiguity, where multiple plausible 3D bodies can explain the same image evidence.

Deterministic methods are efficient, but under ambiguity they often regress toward an average solution. Probabilistic methods are more expressive, but modeling the full state stochastically can waste capacity on well-constrained torso, root, and camera variables that are often better handled deterministically.

This motivates our central claim: HMR benefits from uncertainty-aware completion, in which stable variables are estimated deterministically while ambiguity-prone variables are refined probabilistically.

Contributions

  • We propose FactorizedHMR, a hybrid video HMR framework that decouples stable torso-root estimation from ambiguity-prone motion completion.
  • We formulate selective probabilistic HMR as masked conditional flow matching, keeping anchor variables fixed while generating only non-torso and world-motion variables.
  • We introduce geometry-aware completion objectives that combine composite rotation-joint targets, representation-aware noising, joint-bone consistency, projection loss, and feature-aware CFG.
  • We develop a camera-aware synthetic training pipeline and demonstrate competitive overall performance, with the strongest gains under severe occlusion.
(a) GT
(b) GVHMR
(c) Ours
Ground-truth example under occlusion GVHMR qualitative example FactorizedHMR qualitative example
Qualitative comparison on EMDB. The left arm and hand are heavily occluded by the tree, so multiple poses are plausible. GVHMR predicts a left hand that is not visible in the image (red square), illustrating how deterministic pipelines can commit to an implausible average solution under ambiguity. Our method instead preserves the visible pose while producing a more plausible completion for the occluded limb.
Per-joint confidence consistency visualization
Visualization of per-joint confidence consistency across the test datasets, where higher values correspond to more reliable observation. Visual evidence is concentrated in the torso, while distal limbs are less consistently detected.
FactorizedHMR example result 1 FactorizedHMR example result 2
Example results for FactorizedHMR. Stage 1 (green) deterministically estimates a torso-root anchor for the less ambiguous body parts and Stage 2 (red) uses conditional flow matching to complete the non-torso body poses, both for camera and world space.

Method

Given an input video, our goal is to recover temporally consistent human motion in both camera space and world space. Torso pose, body shape, and camera-space root motion are usually well constrained and form a structural anchor, while non-torso articulation and world motion are more ambiguous because they are more sensitive to occlusion, truncation, and camera-body disentanglement.

FactorizedHMR pipeline overview
Overview of FactorizedHMR. Input video frames are preprocessed into ray-embedded keypoints, image embeddings, and camera-motion features. Stage 1 uses a deterministic regressor to predict a torso-root structural anchor, 𝒜 = [θtorso, β, Γc, τc]. Conditioned on the shared video features and this anchor, Stage 2 uses masked flow matching to complete the non-torso articulation and world-motion variables, yielding the final human motion in camera and world space. The two stages are trained sequentially, and the training losses are described in Sec. 3.4.

Stage 1: Deterministic Structural Estimation

Stage 1 uses a deterministic transformer with rotary position embeddings, motivated by GVHMR, to estimate the structural anchor. It predicts only torso pose, body shape, and camera-space trajectory variables, deferring the more ambiguous world-motion components to Stage 2.

Inputs include bounding-box features, 2D keypoint observations, image features, and relative camera-motion features. Keypoints are converted to camera-ray directions with the intrinsics, embedded with sinusoidal encodings, projected by per-modality MLPs, and summed into per-frame tokens for the relative transformer.

Stage 2: Non-Torso Articulation and World Completion via Masked Flow Matching

Stage 2 augments local rotations with 3D camera-space joint positions, since rotations alone do not directly expose resulting limb geometry. Variables copied from Stage 1 are fixed by the mask, while Stage 2 generates the non-torso and world-motion subset; the rotation branch defines the final body and the joint-position branch provides auxiliary geometric supervision.

We split the latent into known coordinates from the structural anchor and unknown coordinates for ambiguous non-torso and world-motion variables. At inference, only the generated coordinates start from Gaussian noise, and an ODE sampler re-imposes the known coordinates after each update to produce conditional completion around a fixed structural anchor.

Anchor Conditioning and CFG

Stage 2 shares the visual and camera conditions used by Stage 1 and also receives a compact torso-pose condition from the Stage 1 anchor. We apply classifier-free guidance by dropping observation-heavy conditions during training while keeping the structural-anchor conditions, strengthening observation-conditioned completion without discarding the anchor.

Geometry-Aware Training

In addition to the generative objective, we use geometry-aware losses similar to GVHMR. For Stage 2, these supervise pose, joints, root motion, projected 2D joints and vertices, camera-frame translation, and recovered world translation. We also add joint-bone consistency and a direct projection loss for ambiguous non-torso joints to keep the completed motion consistent with SMPL-X geometry, image evidence, and camera/world trajectory constraints.

Synthetic Dataset Generation Pipeline

Motion-only corpora such as AMASS contain rich human dynamics, but they do not provide paired realistic videos and are therefore insufficient for training models for recovering human motion from imagery. Real videos offer much richer appearance variation, but their annotations are often noisy or incomplete. To address this, we construct a synthetic pipeline that preserves exact motion-capture supervision while using Uni3C as a video synthesis engine, which unifies camera and human-motion control from a single-view image across broader visual domains.

Given a motion sequence and its SMPL-X parameters, we sample a camera trajectory with known camera parameters. We render geometric control signals, including SMPL-X body renderings, hand renderings, masks, and an OpenPose-style first-frame pose. The pose rendering is used to synthesize a reference image with identity and scene prompts using depth-conditioned FLUX.1. We then estimate monocular metric depth for the reference image using Depth Pro and unproject it into a scene point cloud, which serves as the environment coordinate system. The SMPL-X motion and scene point cloud are rendered under the same target camera trajectory, producing aligned human-motion and camera-control conditions for Uni3C. The final RGB video is then generated by Uni3C from the reference image, the rendered point-cloud trajectory, and the rendered human motion.

Synthetic dataset generation example row 1 column 2
Synthetic dataset generation example row 2 column 2
Input and output examples from the synthetic dataset generation pipeline.

Results

Both stages use 12-layer transformers with rotary position embeddings, 8 attention heads, and hidden size 512. We train on AMASS, BEDLAM, H36M, and 3DPW, plus Uni3C-based synthetic sequences for additional camera-aware supervision.

We report camera-space metrics on 3DPW, RICH, and EMDB-1; world-space metrics on RICH and EMDB-2; and occlusion-focused results on 3DPW-XOCC.

Camera-Space Motion Recovery

Our method remains highly competitive on standard camera-space benchmarks. MPJPE improves consistently across datasets, while PA-MPJPE and PVE remain broadly competitive, indicating that the factorization improves ambiguous articulation without sacrificing torso stability.

Models 3DPW RICH EMDB
PA-MPJPE ↓ MPJPE ↓ PVE ↓ PA-MPJPE ↓ MPJPE ↓ PVE ↓ PA-MPJPE ↓ MPJPE ↓ PVE ↓
Per-frame
CLIFF*43.069.081.256.6102.6115.068.1103.3128.0
HybrIK*41.871.682.356.496.8110.465.6103.0122.2
HMR2.044.469.882.248.196.0110.960.698.0120.3
ReFit*40.565.375.147.980.792.958.688.0104.5
Temporal
VIBE*51.982.998.468.4120.5140.281.4125.9146.8
TRACE*50.979.195.4------70.9109.9127.4
SLAHMR55.9----52.5----69.593.5110.7
PACE------49.3----------
WHAM*35.957.868.744.380.091.250.479.794.4
GVHMR*37.156.968.839.566.074.444.674.086.0
TRAM*35.659.369.6------45.774.486.6
GENMO*42.962.382.8------47.681.294.6
Ours*36.956.268.339.165.874.545.773.686.5
Camera-space metrics. We evaluate the camera-space motion quality on the 3DPW, RICH and EMDB-1 datasets. * denotes models trained with the 3DPW training set.

Regional Breakdown

The torso/non-torso split highlights the intended division of labor: Stage 1 improves the torso subset over GVHMR, while Stage 2 achieves the best MPJPE on the non-torso subset across all three datasets.

Torso Setup

Models 3DPW RICH EMDB
PA-MPJPE ↓ MPJPE ↓ PA-MPJPE ↓ MPJPE ↓ PA-MPJPE ↓ MPJPE ↓
GVHMR5.9627.319.0336.7914.9651.56
Ours5.7227.219.0035.8214.2650.03

Non-Torso Setup

Models 3DPW RICH EMDB
PA-MPJPE ↓ MPJPE ↓ PA-MPJPE ↓ MPJPE ↓ PA-MPJPE ↓ MPJPE ↓
GVHMR39.564.945.680.751.685.4
Ours39.664.745.479.653.084.6
Regional camera-space evaluation against GVHMR. The left table reports the torso region predicted by Stage 1. The right table reports the non-torso region refined by Stage 2.

World-Space Motion Recovery and Occlusion-Specific Benchmark

In world space, the clearest gain appears on drift-sensitive W-MPJPE, suggesting better long-horizon coherence. On 3DPW-XOCC, our method achieves the best performance on all reported metrics, showing the largest benefits under severe occlusion.

Models EMDB RICH
WA-MPJPE ↓ W-MPJPE ↓ Root translation ↓ Jitter ↓ Foot-sliding ↓ WA-MPJPE ↓ W-MPJPE ↓ Root translation ↓ Jitter ↓ Foot-sliding ↓
TRACE529.01702.317.72987.6370.7238.1925.4610.41578.6230.7
GLAMR280.8726.611.446.320.7129.4236.23.849.718.1
SLAHMR326.9776.110.231.314.598.1186.428.934.35.1
WHAM135.6354.86.022.54.4109.9184.64.119.73.3
GVHMR111.0276.52.016.73.578.8126.32.412.83.0
TRAM76.4222.41.418.111.0----------
GENMO65.1210.91.09.588.380.7127.22.68.62.7
Ours70.5192.51.517.79.386.6123.32.517.78.6
World-space reconstruction on EMDB and RICH datasets.
Methods MPJPE ↓ PA-MPJPE ↓ PVE ↓
HybrIK148.398.7164.5
PARE114.267.7133.0
PARE + VIBE97.360.2114.9
NIKI (frame-based)110.760.5128.6
NIKI (temporal)88.952.198.0
GENMO76.248.494.2
Ours66.345.181.2
Occlusion-specific evaluation on the 3DPW-XOCC dataset.

Ablation Studies

Adding Stage 2 yields the largest improvement, showing that the torso-root anchor alone is not enough for full-body recovery. Composite representation, geometry-aware losses, CFG scaling, and synthetic data each provide further gains.

Variant PA-MPJPE ↓ MPJPE ↓ PVE ↓
Stage 1 Model65.077.0122.4
+ Stage 2 Model39.960.373.3
+ Composite Representation39.459.271.8
+ Geometry-aware losses37.257.269.2
+ CFG scaling37.156.568.5
+ synthetic dataset36.956.268.3
Progressive ablation on camera-space reconstruction metrics, compared on 3DPW.
Input
GT
GVHMR
Ours
Body-shape recovery input Body-shape recovery ground truth Body-shape recovery GVHMR Body-shape recovery ours
Body-shape recovery under heavy occlusion. FactorizedHMR better preserves body volume.

Qualitative Results

These examples show common failure modes of deterministic pipelines under ambiguity. Compared with GVHMR and GENMO, our method produces distal poses that better match the visible evidence.

GT
GVHMR
GENMO
Ours
Qualitative pose comparison GT 1 Qualitative pose comparison GVHMR 1 Qualitative pose comparison GENMO 1 Qualitative pose comparison Ours 1
GT
GVHMR
Ours
Qualitative pose comparison GT 2 Qualitative pose comparison GVHMR 2 Qualitative pose comparison Ours 2
GT
GVHMR
Ours
Qualitative pose comparison GT 3 Qualitative pose comparison GVHMR 3 Qualitative pose comparison Ours 3
Body-pose recovery in ambiguity-heavy scenes. The red squares in the GVHMR and GENMO results indicate incorrectly predicted body poses.

Runtime Analysis

On a 1193-frame sequence, the full two-stage model requires 3.92 ± 0.03 seconds versus 0.41 ± 0.01 seconds for GVHMR and uses 1057.0 MB versus 303.6 MB of peak GPU memory on an RTX A5000. With 50 denoising steps, it still runs at 304.45 FPS, and the cost can be reduced with fewer steps.

Appendix Figures and Tables

ODE sampling steps ablation figure
Effect of the number of ODE sampling steps on MPJPE and WA2-MPJPE metrics. Both metrics improve rapidly from very small step counts and then saturate, with only minor differences beyond roughly 20 steps. We use 50 steps as a conservative near-converged default.
3DPW-XOCC example figure
Example of 3DPW-XOCC's occluded frames (right) compared to the original 3DPW dataset (left).

Per-Joint Baseline Error Analysis

Joint group / joint 3DPW EMDB-1 RICH Mean
Torso joints
Pelvis / root--14.7110.7412.73
Left hip16.979.328.3711.55
Right hip16.979.328.3711.55
Spine / torso--55.0435.7145.38
Neck47.9878.4862.6763.04
Left shoulder52.2076.6759.8962.92
Right shoulder48.1775.6055.6659.81
Non-torso joints
Left elbow69.6484.9073.8976.14
Right elbow61.8278.7269.6970.08
Left wrist84.70109.2789.1794.38
Right wrist82.3499.3986.9989.57
Left hand--125.99101.72113.86
Right hand--115.8899.08107.48
Left knee46.5355.6666.9856.39
Right knee44.5756.0762.7154.45
Left ankle83.4289.24108.9193.86
Right ankle80.0692.56106.8793.16
Left foot--104.42115.53109.98
Right foot--107.93113.44110.69
Anchor mean36.4643.3339.739.83
Distal mean69.1493.3491.2584.58
Distal - anchor32.6850.0151.5544.75
Per-joint MPJPE calculated by GVHMR across test datasets. Lower values indicate joints that are more reliably recovered by a strong deterministic baseline.

Additional Qualitative Comparisons

Additional qualitative comparisons in the appendix. Each example is shown left-to-right as GT, baseline, and Ours. Across a broader set of ambiguity-heavy scenes, FactorizedHMR produces distal-limb completions that remain more consistent with the visible evidence under occlusion, truncation, and clutter.
Input
GVHMR
Ours
Additional body-pose recovery comparisons between our method and GVHMR.
Input
GVHMR
Ours
Additional body-pose recovery comparisons between our method and GVHMR.
Input
GENMO
Ours
Additional body-pose recovery comparisons between our method and GENMO.

Conclusion and Limitations

We presented FactorizedHMR, a video HMR framework that uses an uncertainty-aware factorization: a deterministic stage estimates a stable torso-root anchor, and a probabilistic flow-matching stage completes ambiguity-prone non-torso articulation and world motion. This design, together with geometry-aware refinement and camera-aware synthetic supervision, improves reconstruction while preserving stability. Across camera-space and world-space benchmarks, FactorizedHMR remains strong against both deterministic and generative baselines and displays clear qualitative advantages in ambiguous scenes.

The current uncertainty partition between Stage 1 and Stage 2 is fixed, which can still fail under extreme truncation or rapid camera motion: if the Stage 1 structural anchor is biased, Stage 2 may inherit that error rather than correct it. In addition, probabilistic refinement increases inference cost, which limits practical deployment. Future work could make the factorization adaptive, improve sampling efficiency, and develop ambiguity-aware benchmarks for calibrated multi-hypothesis video HMR.

Citation

If you find this work useful, please consider citing:


@article{Kwon2026FactorizedHMR,
  title={FactorizedHMR: A Hybrid Framework for Video Human Mesh Recovery},
  author={Patrick Kwon and Chen Chen},
  journal={arXiv preprint arXiv:2605.14854},
  year={2026}
}