Objectives
Recently, a number of scientific papers have shown that it is possible to extract very detailed information about a household’s routine, with no knowledge besides its smart meter data. Whilst this can be useful if the householder, and data owner, has agreed to it, there are also implications of trust, identity and privacy violation if not the case. The amount and value of information, i.e., utility, that can be extracted depends on the granularity of the data; especially at risk are measurements at sampling rates of 1-60secs or higher. This has led to recent policy changes in granularity of data available to third parties, including energy suppliers, to ensure privacy of users. Additionally, in order to ensure trustworthiness of inference on the data, explainability of the AI inference algorithms is critical. This PhD will focus on the privacy, utility and trustworthiness nexus within the context of sustainability inferences and recommendations from smart home technologies. This can include obfuscation methods for protecting low rate (> 0.02 Hz) meter data with the constraint that it is still possible to extract billing information, as well as other more detailed information as co-agreed with the data user. Differential privacy has been proposed to protect appliance usage information in smart metering but can suffer from complexity. Prior to developing the tools, it is important to understand the tasks one may wish to perform on data (e.g., the inferences one may wish to derive from data; the recommendations one may wish to derive from data; level of explainability to maximise trust) along with the tasks one may wish to prevent to be performed on data (e.g., certain inferences that stakeholders may deem to be privacy-invasive).
ESR2 will address the challenging task of co-designing information extraction methods, such as NILM, with stakeholders such that the inference or recommender systems are safe, reliable, trustworthy, legally compliant and ethically sound. Particular objectives are: (1) Develop a deep understanding of utility, i.e., what information (e.g., activities and behaviour patterns) can be extracted from household data (e.g., smart meter, demographics) at different sampling rates and by which AI methods; (2) Develop data obfuscation ML methods bearing in mind utility and privacy trade-offs; (3) Co-design with a range of data owners, from individual householders to building management services, inference methods or recommender systems that address utility, privacy and trust requirements (4) Test and break the designs with full awareness of Responsible Research and Innovation (RRI).
Expected Results
(1) Report on what information can be mined and how, clustering categories of information (utility); (2) Novel approaches for changing the statistical properties of data that ensure that trade-off utility and privacy (utility vs privacy); (3) a suite of explainable information mining methods on ‘secured’ data co-designed with living labs, virtual labs, focus groups and industry that meet RRI criteria (utility vs privacy vs trust).
Hosting institution: Strathclyde University
Supervision: Dr. Lina Stankovic, Dr. Vladimir Stankovic