EldCare: Real-Time Monitoring System for Safe and Independent Elderly Living using a Novel Vision-Based Framework with Spatio-Temporal Action Localization.
Aryaman Khanna, Ryan Park, Shubhrangshu Debsarkar
Thomas Jefferson High School for Science and Technology
Introduction
Figure 1: Predicted population by age group from 1950-2050
Figure 2: Picture of elderly person falling
Existing Work
Marker-based systems
Current markerless systems
Figure 3: Marker-based fall detection bracelet
Figure 4: Markerless motion capture with pose estimation
Goal and Constraints
This includes:
Figure 5: EldCare V1
Figure 6: EldCare V2
Methods - Summary of System
Figure 7: Flowchart demonstrating EldCare pose estimation and spatio-temporal action classification
Figure 9: Action Summary Interface and Live Action Classification Performed by IOS App
Figure 8: EldCare Prototype using Nvidia Jetson Nano and webcam
Methods - Data
Methods - Pose Estimation
Figure 10: Tracking network architecture: regression with heatmap supervision
Figure 11: Data flow for real-time video capture and rendered output
Methods - Spatio-Temporal Action Localization
Figure 12: Performance of EfficientNet and other state-of-the-art networks on the ImageNet database.
Figure 13: EfficientNet-GRU hybrid architecture.
Results
Figure 14: Confusion Matrix Representing our Model Classification Accuracy.
Figure 15: Accuracy of training and testing data plotted over iterations
Figure 16: Training and testing loss plotted over iterations
Table 1: Table comparison of accuracies of previous elderly care pose-detection models. EldCare displays the greatest accuracy.
Data Analysis
Table 2: Table comparison of accuracies of various machine learning architectures on common publicly available action recognition datasets.
Conclusion
EldCare V1
EldCare V2
Future Goal
References
Buzzelli, M., Albé, A., & Ciocca, G. (2020). A Vision-based system for monitoring elderly people at home. Applied Sciences, 10(1). https://doi.org/10.3390/app10010374
Fan, K., Wang, P., Hu, Y., & Dou, B. (2017). Fall detection via human posture representation and support vector machine. International Journal of Distributed Sensor Networks, 13(5), 155014771770741. https://doi.org/10.1177/1550147717707418
Harrou, F., Zerrouki, N., Sun, Y., & Houacine, A. (2019). An Integrated Vision-Based Approach for Efficient Human Fall Detection in a Home Environment. IEEE Access, 7, 114966–114974. https://doi.org/10.1109/access.2019.2936320
Hellsten, T., Karlsson, J., Shamsuzzaman, M., & Pulkkis, G. (2021). The Potential of Computer Vision-Based Marker-Less Human Motion Analysis for Rehabilitation. Rehabilitation Process and Outcome. https://doi.org/10.1177/11795727211022330
Martinez, G. H., Kitani , K., & Bansal, A. (2019). (rep.). (Y. Sheikh, Ed.)OpenPose: Whole-Body Pose Estimation. Carnegie Mellon University. Retrieved January 18, 2022, from https://www.ri.cmu.edu/wp-content/uploads/2019/05/MS_Thesis___Gines_Hidalgo___latest_compressed.pdf.
Oudah, M., Al-Naji, A., & Chahl, J. (2021). Computer Vision for Elderly Care Based on Deep Learning CNN and SVM. IOP Conference Series: Materials Science and Engineering, 1105(1), 012070. https://doi.org/10.1088/1757-899x/1105/1/012070
Pham, H.-H., Khoudour, L., Crouzil, A., Zegers, P., & Velastin, S. A. (2015). Video-based human action recognition using deep learning. Video-Based Human Action Recognition Using Deep Learning: a Review. https://doi.org/https://e-archivo.uc3m.es/bitstream/handle/10016/26542/videobased_2015.pdf
Song, L., Yu, G., Yuan, J., & Liu, Z. (2021). Human pose estimation and its application to action recognition: A survey. Journal of Visual Communication and Image Representation, 76, 103055. https://doi.org/10.1016/j.jvcir.2021.103055
Tang, J. (2020). MediaPipe Pose. MediaPipe. Retrieved January 6, 2022, from https://google.github.io/mediapipe/solutions/pose.html