Abstract
Indoor localization using deep learning has emerged as a promising approach due to its high accuracy in mapping and predicting user locations for complex datasets. However, the inherent complexity of deep learning models often limits their interpretability, creating a gap in user trust and understanding. This paper introduces eXLoc, a novel framework that integrates Explainable AI, Class Activation Mapping (CAM), into deep learning models for indoor localization to enhance model transparency and interpretability. We introduce a new metric called Impact Score to identify significant APs that affect model predictions. This enhances model interpretability and allows a model to identify the influential APs via their impact on localization performance. We have extensively evaluated eXLoc over eight different places from two real-world RSSI datasets. We gained insights into how the model generates predictions, and identified the reasons for the model's poor performance. These results demonstrate that our approach can be effectively utilized in enabling users to have more trust and understanding of the model in many real-world scenarios.
Original language | English |
---|---|
Title of host publication | GLOBECOM 2024 - 2024 IEEE Global Communications Conference |
Publisher | Institute of Electrical and Electronics Engineers Inc. |
Pages | 277-282 |
Number of pages | 6 |
ISBN (Electronic) | 9798350351255 |
DOIs | |
State | Published - 2024 |
Event | 2024 IEEE Global Communications Conference, GLOBECOM 2024 - Cape Town, South Africa Duration: 8 Dec 2024 → 12 Dec 2024 |
Publication series
Name | Proceedings - IEEE Global Communications Conference, GLOBECOM |
---|---|
ISSN (Print) | 2334-0983 |
ISSN (Electronic) | 2576-6813 |
Conference
Conference | 2024 IEEE Global Communications Conference, GLOBECOM 2024 |
---|---|
Country/Territory | South Africa |
City | Cape Town |
Period | 8/12/24 → 12/12/24 |
Bibliographical note
Publisher Copyright:© 2024 IEEE.