Vol. 19, No. 12, December 31, 2025
10.3837/tiis.2025.12.009,
Download Paper (Free):
Abstract
The leading cause of injury among the elderly is falls, although video-based detection is constrained by the lack of labeled data, environmental diversity, and most often false alarms. This paper proposes a deep learning model that combines realistic data augmentation with spatiotemporal motion modeling to overcome them. A Wasserstein generative adversarial network (WGAN) is used to build 3D sequences of human skeletons to reduce the issue of class imbalance and enhance generalization of the model. The analyzed augmented data is then analyzed with a hybrid 3D convolutional-recurrent model which is simultaneously trained on spatial configurations and temporal motion cues. It includes a single-shot detector (YOLOv8) that is used to localize persons; it guarantees the accuracy of region proposals and minimizes the background noise prior to the pose estimation. The suggested method is tested on Le2i fall detection benchmark, with a test accuracy of 97.33, which is much higher than the baseline that is trained on real data only (89.70). The findings of experiments prove the versatility of the model and its capability to separate falls and everyday actions in the complicated real-life situations. A scalable and reliable fall detection based on the generative data expansion and efficient use of spatiotemporal learning designed in the framework makes it feasible to apply to healthcare and surveillance. In general, the study has shown that incorporation of synthetic motion generation with sophisticated deep learning models can significantly improve the performance, generalization, and functionality of vision-based fall detection systems.
Statistics
Show / Hide Statistics
Statistics (Cumulative Counts from December 1st, 2015)
Multiple requests among the same browser session are counted as one view.
If you mouse over a chart, the values of data points will be shown.
Cite this article
[IEEE Style]
Q. M. Haref, J. Long, Z. Yang, "Enhanced Video-Based Fall Detection Using 3D Human Pose Estimation and Advanced Generative Adversarial Networks: Integrating WGAN Data Augmentation with a Hybrid 3D CNN–GRU Architecture," KSII Transactions on Internet and Information Systems, vol. 19, no. 12, pp. 4346-4371, 2025. DOI: 10.3837/tiis.2025.12.009.
[ACM Style]
Qasim Mahdi Haref, Jun Long, and Zhan Yang. 2025. Enhanced Video-Based Fall Detection Using 3D Human Pose Estimation and Advanced Generative Adversarial Networks: Integrating WGAN Data Augmentation with a Hybrid 3D CNN–GRU Architecture. KSII Transactions on Internet and Information Systems, 19, 12, (2025), 4346-4371. DOI: 10.3837/tiis.2025.12.009.
[BibTeX Style]
@article{tiis:105402, title="Enhanced Video-Based Fall Detection Using 3D Human Pose Estimation and Advanced Generative Adversarial Networks: Integrating WGAN Data Augmentation with a Hybrid 3D CNN–GRU Architecture", author="Qasim Mahdi Haref and Jun Long and Zhan Yang and ", journal="KSII Transactions on Internet and Information Systems", DOI={10.3837/tiis.2025.12.009}, volume={19}, number={12}, year="2025", month={December}, pages={4346-4371}}