Citation
Abstract
Human Activity Recognition (HAR) has been a highly debated topic in recent years. Due to privacy concerns, especially in smart home environments, sensor-based HAR is commonly employed. This method involves users carrying smart devices, such as smartwatches or smartphones, that calculate tri-axial acceleration through gyroscopes to determine specific activities. However,Many users and developers do not fully comprehend deep learning and other algorithms. Users often worry about the unknown, and the development of HAR based on wearable devices will become increasingly challenging for developers. Consequently, the concept of Explainable AI (XAI) has been introduced to allow human users to understand and trust the machine learning algorithms and the results they produce. In this paper, we used the UCI dataset to analyze the factors that significantly influence the machine learning model with a CNN-LSTM architecture, employing XAI techniques to provide a clear demonstration of these impacts.
Download File
Official URL or Download Paper: https://ieeexplore.ieee.org/document/10791196/
|
Additional Metadata
| Item Type: | Conference or Workshop Item (Paper) |
|---|---|
| Divisions: | Faculty of Computer Science and Information Technology School of Business and Economics |
| DOI Number: | https://doi.org/10.1109/ISCT62336.2024.10791196 |
| Publisher: | Institute of Electrical and Electronics Engineers Inc. |
| Keywords: | HAR; XAI; UCI dataset; CNN-LSTM; Machine learning |
| Depositing User: | Mr. Mohamad Syahrul Nizam Md Ishak |
| Date Deposited: | 30 Oct 2025 06:43 |
| Last Modified: | 30 Oct 2025 06:43 |
| Altmetrics: | http://www.altmetric.com/details.php?domain=psasir.upm.edu.my&doi=10.1109/ISCT62336.2024.10791196 |
| URI: | http://psasir.upm.edu.my/id/eprint/121323 |
| Statistic Details: | View Download Statistic |
Actions (login required)
![]() |
View Item |
