An Interpretable Deep Learning-based Feature Reduction in Video-Based Human Activity Recognition
Peer reviewed, Journal article
Published version
View/ Open
Date
2024Metadata
Show full item recordCollections
Original version
Dutt, M., Goodwin, M. & Omlin, C. W. P. (2024). An Interpretable Deep Learning-based Feature Reduction in Video-Based Human Activity Recognition. IEEE Access, 3432776. https://doi.org/10.1109/ACCESS.2024.3432776Abstract
This paper presents a human activity recognition framework tailored for healthcare applications, emphasizing the essential balance between accuracy and interpretability required for medical monitoring. The model utilizes MediaPipe to capture the complex dynamics of human movements and introduce an interpretable feature reduction function. This method improves traditional dimensionality reduction techniques like principal component analysis. Our feature engineering is based on the importance of feature permutations; it selectively retains salient features, thus enhancing the interpretability essential for the medical domain. We validated our method on the “NTU RGB+D” dataset; it improves the recognition accuracy for a range of human activities that may be relevant for elderly care. However, the recognition of subtler activities like neck pain and headaches requires further investigation. This study underscores the potential to advance patient monitoring and sets the stage for its expanded application in various medical contexts.