Difference between revisions of "SCOTT:BB26.G"

From its-wiki.no

Jump to: navigation, search
Line 129: Line 129:
 
## Anonymization techniques
 
## Anonymization techniques
 
## De-anonymization methods, like machine learning algorithms
 
## De-anonymization methods, like machine learning algorithms
## Results about how anonymization of data influences desired learning results, e.g., before anonymization some learning and profiling can be done, whereas after anonymization such profiling cannot be done any more and thus the <<personalised pricing scheme>> cannot be applied any more. (see survey chapter [https://link.springer.com/chapter/10.1007%2F978-0-387-70992-5_2 A General Survey of Privacy-Preserving Data Mining Models and Algorithms] and the book that it comes from [https://link.springer.com/book/10.1007/978-0-387-70992-5] along with the standard book [https://books.google.no/books?id=pQws07tdpjoC&lpg=PP1&ots=tzFqZXiC0W&lr&pg=PP1#v=onepage&q&f=false Data Mining: Cncepts and Techniques 2011])
+
## Results about how anonymization of data influences desired learning results, e.g., before anonymization some learning and profiling can be done, whereas after anonymization such profiling cannot be done any more and thus the <<personalized pricing scheme>> cannot be applied any more. (see survey chapter [https://link.springer.com/chapter/10.1007%2F978-0-387-70992-5_2 A General Survey of Privacy-Preserving Data Mining Models and Algorithms] and the book that it comes from [https://link.springer.com/book/10.1007/978-0-387-70992-5] along with the standard book [https://books.google.no/books?id=pQws07tdpjoC&lpg=PP1&ots=tzFqZXiC0W&lr&pg=PP1#v=onepage&q&f=false Data Mining: Concepts and Techniques 2011])
 
## Differential privacy (see survey [https://link.springer.com/chapter/10.1007/978-3-540-79228-4_1] and book [https://dl.acm.org/citation.cfm?id=2693053 The Algorithmic Foundations of Differential Privacy] and more resources from a simple search)
 
## Differential privacy (see survey [https://link.springer.com/chapter/10.1007/978-3-540-79228-4_1] and book [https://dl.acm.org/citation.cfm?id=2693053 The Algorithmic Foundations of Differential Privacy] and more resources from a simple search)
 
## Survey of Privacy-by-Design concepts and current works. Examples include applications
 
## Survey of Privacy-by-Design concepts and current works. Examples include applications

Revision as of 10:25, 7 March 2018

Title Privacy labels (A-F)
Page Title BB26.G Privacy labels (A-F)
Technology Line Reference Architecture/Implementation
Lead partner UiO
Leader Christian Johansen
Contributors UiO, Smart Innovation Norway
Related to Use Cases SCOTT:WP7, SCOTT:WP8, SCOTT:WP11, SCOTT:WP14"<s>SCOTT:WP14</s>" cannot be used as a page name in this wiki., SCOTT:WP21, SCOTT:WP12"<s>SCOTT:WP12</s>" cannot be used as a page name in this wiki.
Description In order to allow authorities or standardization and authorization bodies to evaluate a product with respect to privacy aspects, before attaching a Privacy Label, we need to provide both a methodology and suggest tools to be used in the assessment. Moreover, we need to study closely how the Privacy labels should “look and feel” to the end customers. For this we wil apply interaction design techniques, including surveys and other user analysis techniques.

The work has to focus on several aspects:

  • Privacy Label content which will study what kind of information should the Privacy Label contain and how this information is perceived by the end customer. This work is essential for the end- user adoption.
  • Privacy evaluation tools are essential in order for a standardization company/authority to be able to evaluate the respective IoT product. Tools that will be investigated in this BB include, complex verification tools, but also simple testing tools and techniques.
  • Methodology will be designed and prepared for demonstration and application inside the various Use Cases of SCOTT where this BB is applicable. The methodology is something that certification bodies are used with. This includes thorough documentation (and specific to us we will investigate what kind of documentation is needed for Privacy aspects) as well as testing results (for Privacy such testings are not immediately available, and they will have to be investigated part of this BB).
Main output Methodology for privacy evaluation and standardisation to be used in assessing products.

The methodology will be tested and developed together with the Use Cases of SCOTT. Scales and recommendations for Privacy Labeling ranges for different sectors that SCOTT has use Cases in. Recommendations for how to achieve the standard required by each specific label for a specific domain. These recommendations would be tested together with the Industry partners in the respective use Cases to assess their feasibility.

BB category Methodology (for SW/HW development), Profile, Standard, Means for establishing cross-domain interoperability, Process, Other
Baseline We would like to introduce privacy labels for applications and components, similar to the energy labels (A++, A+, A, B,...F), see IoTSec:Privacy_Label. Customers in Europe have an understanding of these labels for white goods, and thus we should use a similar technology to introduce "privacy" labeling. E.g. You would like to buy yourself a sports device (Fitbit, Google watch,...) or application (Endomondo, Strava,...). A potential difference between the tools might be expressed through the privacy label, e.g. a Polar device having an A-privacy, while a Garmin device having a B-privacy. - Our analysis can then show the relation between application goals and system capabilities (configuration of components) to achieve the required privacy level.
Current TRL TRL 1-2 for the ideas of Privacy Labels
Target TRL TRL 6

Overview

  • WPs of interest
    • WP7 can be a core WP for Privacy Labels BB
    • WP21 is also good for applying Privacy Labels
    • WP11 mentions Privacy Labels.
It is also interesting for applying Privacy Labels because it works with complex systems that manipulate data of various kinds. Fine-grained access control is also applicable, like our 5th step in S-ABAC "Query-based AC", which can also be good to achieve better privacy.
    • We will not be involved in WP12 nor WP14
WPs in focus
Core Extended Future Cancelled
WP7 WP21 WP11 and WP8 WP12 and WP14

Activities

  • Related activities include those started in BB26.F Measurable on Multi-Metrics and Measurable aspects of Privacy
  • Privacy evaluation of the TellU Diabetics app demonstrator from WP21 is planned and under way, with deadline in Spring 2019.
Activities chart
Title Status Responsible Deadlines
Privacy evaluation of the TellU Diabetics app demonstrator from WP21. Planned and Initial MSc discussions Simen Dagfinrud and Christian Johansen tentative final in Spring 2019
User presentation of Privacy Label. Planned Christian Johansen start in Summer 2018
User Studies. Planned Christian Johansen and others (Heidi, ++) start in Autumn 2018

Division of Work and Research Directions

Planned Outcomes

The work in the Privacy Labelling aims to provide the following tangible results.

Privacy Labelling for Decision Makers

The purpose is to have a lightweight version useful for persons that make decisions where privacy aspects are involved. Examples of use:

  1. Example of the CEO or Project manager, like from Statsbygg, that need to decide on whether to include many sensors in a Smart office building for monitoring the indoor air quality. What are the privacy implications of this? How should she take a decisions, based on what information?
    • This Result should give here the needed tools to help with the decision.
  2. Before doing any expensive and time-consuming PL certification actions, a CEO or investor needs to have initial indications of Privacy Aspects regarding the new App or IoT device or technology that is planned.
    • This Result should enable economical and feasibility calculations of the privacy implications on the business development.
    • This work will be done in collaboration with Smart Innovation Norway.

These examples clarify the Requirement RQ561.

Privacy Labelling for Technical privacy engineers

This is similar to any other certification methods, like for the SIL levels, for the ANSSI security classifications, etc. The goal is two-fold: (i) to have a methodology, i.e., a decision-tree to guide various technological decisions; as well as (ii) to have guidelines for what technologies can achieve what privacy guarantees. This includes a good overview, semantically meaningful, over the various privacy relevant concepts, like anonymization, personal identifiable data, machine learning information extraction, aggregation techniques, deletion, etc. This also includes methods and tools for measuring the various privacy relevant techniques and how much they achieve their purposes.

NOTE that all the above aspects are not related to security, i.e., the confidentiality is assumed provided by other means. For otherwise, if confidentiality of data in transit or stored in broken, then no privacy guarantees can be met any more, thus making any Privacy Labelling irrelevant.

Security evaluation should be done independently of the privacy evaluation.

Privacy Labelling for the Users

This is not present in classical standardisation methods, because those, like the goal above, are meant only for technical people, and have a very simple outcome like a number or a yes/no answer.

The main purpose of Privacy Labelling is to present the outcome of the privacy certification to Users. However, privacy is highly difficult to present, compared to classical aspects like the Energy Consumption labels where the range is the number of consumed KW/hour. Moreover, privacy is also highly personal, i.e., what for some person is highly private for another may not be, and thus does not impact her decisions. The goal is to study how to best present the various concepts involved in technically deciding a privacy label. This means that a User can understand first the privacy implications for some device/service/system that she wants to acquire, as well as how to compare with other similar products based on the privacy aspects. Examples of use:

  1. An older lady wishes to buy some Smart Home Energy Management system, but she is concerned that all these new gadgets are too intrusive into her rather classical style of living. Even more so since several of her family members, or friends at the Bridge Club tell her about <<how the TV listens to what people talk in the house>>, or how <<medical persons watch over people taking a bath>>. She wants some simple indication given by some authority that she can trust (and she mostly trusts when government associations are involved) about how much this new system would expose her, so she can decide what to buy, restricted by here limited budget.
  2. A young adult who likes to have various new electronic devices, is interested in installing a new app on his phone for tracking his weekend hiking trips. He wants to know about how his location data is being used by the app provider so that he can make an informed decision regarding choosing the various different functionalities that the app promises against the privacy exposure that he would have to accept.

The first example clarifies the Requirement RQ558.

Privacy Labelling for certifying experts and certification bodies

These include minimal requirements, alignment with existing regulations like GDPR, adoption and relation to existing standards of relevance. Examples of use:

  1. An expert needs to carry out a certification job for a specific new IoT device. She needs both methodological as well as tool support.
  2. A regulatory body, like a GDPR national authority, needs to make a compliance evaluation.

Sub-components

The work in the Privacy Labelling is divided into several sub-components, each trying to achieve one of the above goals.

PL4Decisions

This sub-component works towards the following measurable outcomes.

  1. A simplified decision process from PL-CERT, with different domain-specific versions
  2. Tools easy to use by decision making people, like questionnaires

This sub-component works to attain the Requirement RQ561.

PL-Methods

This sub-component works towards the following measurable outcomes. This sub-component works to attain the Requirement RQ560.

  1. A decision process, with associated tools like UIs and databases of resources
  2. Surveys of existing techniques that can be applied to answer the questions at any of the decision nodes. Include:
    1. Anonymization techniques
    2. De-anonymization methods, like machine learning algorithms
    3. Results about how anonymization of data influences desired learning results, e.g., before anonymization some learning and profiling can be done, whereas after anonymization such profiling cannot be done any more and thus the <<personalized pricing scheme>> cannot be applied any more. (see survey chapter A General Survey of Privacy-Preserving Data Mining Models and Algorithms and the book that it comes from [1] along with the standard book Data Mining: Concepts and Techniques 2011)
    4. Differential privacy (see survey [2] and book The Algorithmic Foundations of Differential Privacy and more resources from a simple search)
    5. Survey of Privacy-by-Design concepts and current works. Examples include applications
      1. in Programming,
      2. in System architecture,
      3. in Databases,
      4. in Machine Learning
    6. Privacy for Location data (see old book Privacy, security and trust within the context of pervasive computing or newer survey A survey of computational location privacy)
    7. Privacy Patterns (starting from A Literature Study on Privacy Patterns Research) and question on finding which such patterns involve measurable aspects
    8. Privacy in Biometrics (see book Biometrics: theory, methods, and applications)
    9. Clustering techniques for Privacy (see [3] along with 2017 Privacy and utility preserving data clustering for data anonymization and distribution on Hadoop or An overview of the use of clustering for data privacy or The Effect of Clustering on Data Privacy)
  3. Techniques for measuring various relevant privacy aspects, including tools to automatically do the measuring
    1. Include the multi-metrics framework
  4. Tools for aggregation of the privacy measurements
  5. Examples of use

PL-UX

This sub-component works towards the following measurable outcomes.

  1. Privacy Labelling color range and visual cues, including icons
  2. Privacy Label user friendly explanations, i.e., details
  3. Usability studies to evaluate the User response

This sub-component works to attain the Requirement RQ559.

PL-CERT

This sub-component works towards the following measurable outcomes.

  1. A certification process with all decision points identified and methods for making and evaluation and decision at the respective point.
  2. Clearly described relationships with existing standards
  3. Method of aligning Privacy Labelling to existing regulations including GDPR and Norwegian National regulations from Datatilsynet

This sub-component works to attain the Requirement RQ560.

RoadMap

Deliverables and Documents

Practical Aspects

Implementations and User Testing

  • See the RoadMap

Demonstrations and Use Cases

  • In WP7 in an initial preliminary phase at M14
  • In WP21 in a more concrete phase at M24

SCOTT status

From Ramiro: An overview of the instructions for updating the building blocks and the collection of the requirements can be found in this presentation (slide 19-24). https://projects.avl.com/16/0094/WP26/Documents/02_Meetings%20and%20WebEx/20170703_SCOTT_Presentation_WP26.pptx?Web=1


The official and complete instructions can be found in the following presentation from SP1 requirements management. https://projects.avl.com/16/0094/WP01/Documents/03_Deliverables/SCOTT%20REQM%20Approach_Guidance_June2017.pptx?Web=1