Explaining Human Activity Recognition with SHAP: Validating Insights with Perturbation and Quantitative Measures
Explaining Human Activity Recognition with SHAP: Validating Insights with Perturbation and Quantitative Measures
Felix Tempel,E. A. F. Ihlen,Lars Adde,Inga Strümke
2024 · DOI: 10.1016/j.compbiomed.2025.109838
引用数 3
TLDR
This study uses SHapley Additive exPlanations (SHAP) to explain the decision-making process of Graph Convolution Networks (GCNs) when classifying activities with skeleton data and highlights that SHAP can provide granular insights into the input feature contribution to the prediction outcome of GCNs in HAR tasks.
