Projects per year
Abstract
Load Disaggregation, or Non-intrusive Load Monitoring (NILM), refers to the process of estimating energy consumption of individual domestic appliances from aggregated household consumption. Recently, Deep Learning (DL) approaches have seen increased adoption in NILM community. However, DL NILM models are often treated as black-box algorithms, which introduces algorithmic transparency and explainability concerns, hindering wider adoption. Recent works have investigated explainability of DL NILM, however they are limited to computationally expensive methods or simple classification problems. In this work, we present a methodology for explainability of regression-based DL NILM with visual explanations, using explainable AI (XAI). Two explainability levels are provided. Sequence-level explanations highlight important features of predicted time-series sequence of interest, while point-level explanations enable visualising explanations at a point in time. To facilitate wider adoption of XAI, we define desirable properties of NILM explanations -faithfulness, robustness and effective complexity. Addressing the limitation of existing XAI NILM approaches that don’t assess the quality of explanations, desirable properties of explanations are used for quantitative evaluation of explainability. We show that proposed framework enables better understanding of NILM outputs and helps improve design by providing a visualization strategy and rigorous evaluation of quality of XAI methods, leading to transparency of outcomes.
Original language | English |
---|---|
Pages (from-to) | 4345-4356 |
Number of pages | 12 |
Journal | IEEE Transactions on Consumer Electronics |
Volume | 70 |
Issue number | 1 |
Early online date | 1 Aug 2023 |
DOIs | |
Publication status | Published - 1 Feb 2024 |
Funding
This work was supported by the European Union's Horizon 2020 Research and Innovation Programme under the Marie Sk\u0142odowska-Curie under Grant 955422.
Keywords
- deep neural network
- explainable AI
- load disaggregation
Fingerprint
Dive into the research topics of 'Towards transparent load disaggregation – a framework for quantitative evaluation of explainability using explainable AI'. Together they form a unique fingerprint.Projects
- 1 Active
-
building GrEener and more sustainable soCieties by filling the Knowledge gap in social science and engineering responsible artificial intelligence co-creatiOn (GECKO) MSCA-ITN-2020
Stankovic, V. (Principal Investigator) & Stankovic, L. (Co-investigator)
European Commission - Horizon Europe + H2020
1/01/21 → 30/06/25
Project: Research
Datasets
-
REFIT: Electrical Load Measurements (Cleaned)
Murray, D. (Creator) & Stankovic, L. (Supervisor), Zenodo, 1 Mar 2023
Dataset