Fall Risk and Gait Analysis in Community-Dwelling Older Adults using World-Spaced 3D Human Mesh Recovery

Chitra Banarjee1, Patrick Kwon2, Ania Lipat3, Rui Xie4, Chen Chen2, Ladda Thiamwong3
1College of Medicine, UCF 2Institute of AI, UCF 3College of Nursing, UCF 4College of Sciences, UCF
Overview of the GVHMR-based fall risk and gait analysis pipeline.

The overall pipeline of our study. Basing on the world trajectory extraction results from the TUG test using GVHMR, we analyze the gait factors and conduct statistical analysis.

Abstract

Gait assessment is a key clinical indicator of fall risk and overall health in older adults. However, standard clinical practice is largely limited to stopwatch-measured gait speed. We present a pipeline that leverages a 3D Human Mesh Recovery (HMR) model to extract gait parameters from recordings of older adults completing the Timed Up and Go (TUG) test.

From videos recorded across different community centers, we extract and analyze spatiotemporal gait parameters, including step time, sit-to-stand duration, and step length. We found that video-derived step time was significantly correlated with IMU-based insole measurements.

Using linear mixed effects models, we confirmed that shorter, more variable step lengths and longer sit-to-stand durations were predicted by higher self-rated fall risk and fear of falling. These findings demonstrate that our pipeline can enable accessible and ecologically valid gait analysis in community settings.

Introduction

Movement is central to human interaction and well-being across the lifespan. Especially for older adults, movement is an indicator of overall health in the physical, mental, and cognitive domains. As such, gait evaluation is a key clinical measure for identifying older adults at risk for several physiological and pathological aging conditions, commonly used in primary care visits and geriatric assessments to screen participants at risk for frailty or falls.

However, clinical gait evaluation remains largely limited to stopwatch-measured gait speed, as comprehensive assessments are constrained by limited access to technology and specialized training. Although the biomechanical correlates of fall risk are well-established, existing methodologies, such as wearable inertial sensors, optoelectronic marker-based systems, and multi-camera markerless motion capture, require dedicated infrastructure and considerable technical training, restricting their deployment beyond controlled clinical or research settings.

To solve this, we present a novel computer vision-informed approach that evaluates 3D gait measures using a Human Mesh Recovery method, from a single monocular video camera that records the Timed Up and Go test for older adults. We leverage Ground View HMR as the foundation of our pipeline to recover full human body motion anchored in world coordinates.

Contributions

  • A gait acquisition protocol that uses a single camera to record older adults completing the TUG test across multiple community centers, only requiring limited clinical infrastructure and supervision.
  • A novel application of world-grounded 3D HMR to clinical gait analysis, leveraging GVHMR to reconstruct full-body trajectories and spatiotemporal gait parameters.
  • An automated pipeline extracting spatiotemporal parameters of gait and TUG transitions via signal processing and peak detection.
  • A correlation analysis comparing video-derived step time against insole measurements, and a linear mixed effects framework evaluating how fall risk factors predict spatiotemporal gait in community-dwelling older adults.

Method Overview

Participants were instructed to complete the Timed Up and Go test three times. The trials were recorded using a conventional monocular video camera, GoPro Hero 12, set up approximately 280 cm from the 3-meter mark.

In addition to the video, we used the wearable XSENSOR insole system. These sensors served as a comparative reference for evaluating the video-derived gait measures. Participants were also given questionnaires to evaluate subjective fall risk factors and a postural sway assessment to quantify balance.

Each TUG video is processed through GVHMR, which recovers human pose separately for camera-space and world-space. Here, we mainly focus on the world-space trajectory information. The gait parameters are then derived from the 3D joint positions regressed from the SMPL-X kinematic model in world space.

Timed Up and Go setup.

Schematic of Timed Up and Go (TUG) test set up.

Data Collection

The final sample included 207 video recordings of the vTUG from 52 older adults, after removing participants without 3 valid trials. Of those 207 trials, 30 assessments with 90 trials contained simultaneously collected iTUG.

The sample consisted of community-dwelling older adults. Participants were recruited within Orlando, Florida using various strategies, including flyers, word-of-mouth, and collaboration with community partners.

Gait Analysis

Using the world-space joint trajectories recovered by GVHMR, we derive a set of spatiotemporal gait parameters that characterize participants' mobility during the TUG test. These include temporal and spatial gait measures, and transition durations between subtasks such as sit-to-stand and turning.

Prior to analysis, we reduce high-frequency noise by applying Gaussian smoothing to each time series of 3D body keypoints from video data and to the wearable sensor signals. Step time was calculated using GVHMR joint positions, while step length and step width were extracted by detecting local minima of the ankle joint positions to identify steps.

Peak detection for sit-to-stand, stand-to-sit, and turning was computed by calculating a composite sit-to-stand signal and hip line signal, which utilize the hip and shoulder coordinates to segment subtasks.

XSENSOR insole and IMU system.

XSENSOR Insole and IMU System.

Snapshots of research assistant completing the TUG with GVHMR outputs.

Snapshots of research assistant completing the TUG, with the camera-centric view overlay of the HMR and the global-centric view.

Detection of heel strike using peak detection.

Detection of heel strike using peak detection, according to the detected joints using GVHMR.

Automated subtask segmentation for TUG trials.

Automated Subtask Segmentation for three trials. Solid line refers to the hip line velocity and dashed line refers to sit-to-stand and stand-to-sit signal.

Key Results

The Spearman's rank correlation analysis indicates moderate agreement between insole-derived and video-derived step time (n = 90, rho = 0.673, p < 0.001). However, video-derived step time systematically underestimated the insole measurements.

Self-rated fall risk significantly predicted sit-to-stand duration. Spatial measures of gait involved step length and step width, but only step length and step length variability were predicted by a fixed effect of self-rated fall risk and fear of falling. Step width variability was predicted by age but not fall risk factors.

Step length and step length variability showed far stronger between-participant consistency and a substantially better model fit than sit-to-stand durations. These results suggest that older adults with poorer balance or reduced confidence tend to take smaller, more inconsistent steps and require more time to complete sit-to-stand transitions.

Relationship between step length, step length variability, and sit-to-stand duration with fall risk.

Discussion

We demonstrate the novel application of GVHMR to extract clinically meaningful gait metrics from monocular videos of older adults completing the TUG, with derived measures showing moderately strong correlations with simultaneously collected wearable-sensor data.

Shorter and more variable step length was associated with greater self-rated fall risk and fear of falling, consistent with literature linking cautious gait patterns to both perceived and physiological instability. Greater between-participant consistency and model fit indicate that gait characteristics such as step length may be more stable and more strongly tied to fall-risk status than sit-to-stand performance.

Conclusion

We offer a feasible and clinically relevant method for extracting gait and transitional movement metrics from a single TUG recording. GVHMR's ability to segment short, meaningful subtasks and produce measures that align with wearable-derived metrics highlights its potential as an accessible assessment tool.

Associations between fall risk factors and specific gait features, particularly step length variability and sit-to-stand duration, underscore the value of video-based analysis for identifying subtle mobility changes. Future work should evaluate whether the derived metrics can enhance fall risk screening, support remote assessments, or predict prospective functional decline.

BibTeX

@article{banarjee2026fallriskgvhmr,
  title     = {Fall Risk and Gait Analysis in Community-Dwelling Older Adults using World-Spaced 3D Human Mesh Recovery},
  author    = {Banarjee, Chitra and Kwon, Patrick and Lipat, Ania and Xie, Rui and Chen, Chen and Thiamwong, Ladda},
  journal   = {arXiv preprint arXiv:2604.11961},
  year      = {2026},
  note      = {Presented at the CVPR 2026 Computer Vision for Biomechanics Workshop (CVBW)}
}