PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

50
PriBots Hamza Harkous 1 , Kassem Fawaz 2 , Kang. G. Shin 2 , Karl Aberer 1 1 EPFL; 2 University of Michigan Conversational Privacy with Chatbots Workshop on the Future of Privacy Notices and Indicators, SOUPS 2016

Transcript of PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

Page 1: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

PriBots

Hamza Harkous1, Kassem Fawaz2, Kang. G. Shin2, Karl Aberer1 1EPFL; 2University of Michigan

Conversational Privacy with Chatbots

Workshop on the Future of Privacy Notices and Indicators, SOUPS 2016

Page 2: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

2

Privacy Notice: Current State

Page 3: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

legally binding

human understandable

Dual Role:

2

Privacy Notice: Current State

Page 4: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

Can we split these roles?

legally binding

human understandable

Dual Role:

2

Privacy Notice: Current State

Page 5: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

Figure 1. An example of a standardized table is shown on the left, and a standardized short table on the right. The comparison highlights the rows

deleted to “shorten” this version. These deleted rows are listed directly below the table. While both formats contain the legend (bottom right), it is

displayed only on the right here due to space constraints.

have already been deployed by major corporations, makingthem a viable, real world summary format for privacy poli-cies. These policies were stripped of brand information, butthe formatting and styles were retained.

METHODOLOGYWe conducted an online user study in summer 2009 usingAmazon’s Mechanical Turk and a tool we developed calledSurveyor’s Point. Mechanical Turk offers workers the abil-ity to perform short tasks and get compensated. People canplace jobs through Mechanical Turk, specifying the numberof workers they are looking for, necessary qualifications, theamount they are willing to pay, and details about the task.Mechanical Turk payments are generally calibrated for thelength of the task. For our approximately 15-minute study,we paid $0.75 on successful completion.

We developed a custom survey-management tool called Sur-veyor’s Point to facilitate our data collection. Our implemen-tation allows us to show respondents a single question on thescreen along with links for switching back and forth betweentwo policies within a single browser window. This allowedus to track the number of users who looked at each policyand the number of times they switched between them. Addi-tionally, Surveyor’s Point allowed us to collect the amount oftime that users spent reading the policies, as well as informa-tion about whether they clicked through to opt-out forms, toadditional policy information links, or from a layered noticethrough to the full text policy.

In preparation for this study we first performed three smallerpilot tests of our survey framework. We ran our pilot studieswith approximately thirty users each, across 2-3 conditions.Our pilot studies helped us to finalize remaining design de-cisions surrounding the standardized short table, refine ourquestionnaire, and test the integration of Surveyor’s Pointwith Mechanical Turk.2

We then conducted our large-scale study and completed theanalysis with 764 participants (409 female, 355 male), ran-domly assigned to five conditions (see Table 1): full policytext, standardized table, standardized short table, standard-ized short text, and layered text. We dropped 25 additionalparticipants from the study prior to analysis due to incom-plete data or for completing the study in an amount of timethat indicated inadequate attention to the task (defined astime on task that was two standard deviations lower than themean). We chose a between-subjects design to remove learn-ing effects and ensure the study could be completed withinabout 15 minutes. Participants in each condition followedthe same protocol; only the policy format differed.

PoliciesWe selected policies for the study from companies that con-sumers would conceivably interact with. We narrowed our

2The two systems are linked using a shared key that Surveyor’sPoint generates on the completion of our survey, which a participantthen enters back into Mechanical Turk. This allows us to link anentry in Mechanical Turk with an entry in Surveyor’s Point andverify the worker completed the survey before payment.

3

Standardization

3

Page 6: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

Figure 1. An example of a standardized table is shown on the left, and a standardized short table on the right. The comparison highlights the rows

deleted to “shorten” this version. These deleted rows are listed directly below the table. While both formats contain the legend (bottom right), it is

displayed only on the right here due to space constraints.

have already been deployed by major corporations, makingthem a viable, real world summary format for privacy poli-cies. These policies were stripped of brand information, butthe formatting and styles were retained.

METHODOLOGYWe conducted an online user study in summer 2009 usingAmazon’s Mechanical Turk and a tool we developed calledSurveyor’s Point. Mechanical Turk offers workers the abil-ity to perform short tasks and get compensated. People canplace jobs through Mechanical Turk, specifying the numberof workers they are looking for, necessary qualifications, theamount they are willing to pay, and details about the task.Mechanical Turk payments are generally calibrated for thelength of the task. For our approximately 15-minute study,we paid $0.75 on successful completion.

We developed a custom survey-management tool called Sur-veyor’s Point to facilitate our data collection. Our implemen-tation allows us to show respondents a single question on thescreen along with links for switching back and forth betweentwo policies within a single browser window. This allowedus to track the number of users who looked at each policyand the number of times they switched between them. Addi-tionally, Surveyor’s Point allowed us to collect the amount oftime that users spent reading the policies, as well as informa-tion about whether they clicked through to opt-out forms, toadditional policy information links, or from a layered noticethrough to the full text policy.

In preparation for this study we first performed three smallerpilot tests of our survey framework. We ran our pilot studieswith approximately thirty users each, across 2-3 conditions.Our pilot studies helped us to finalize remaining design de-cisions surrounding the standardized short table, refine ourquestionnaire, and test the integration of Surveyor’s Pointwith Mechanical Turk.2

We then conducted our large-scale study and completed theanalysis with 764 participants (409 female, 355 male), ran-domly assigned to five conditions (see Table 1): full policytext, standardized table, standardized short table, standard-ized short text, and layered text. We dropped 25 additionalparticipants from the study prior to analysis due to incom-plete data or for completing the study in an amount of timethat indicated inadequate attention to the task (defined astime on task that was two standard deviations lower than themean). We chose a between-subjects design to remove learn-ing effects and ensure the study could be completed withinabout 15 minutes. Participants in each condition followedthe same protocol; only the policy format differed.

PoliciesWe selected policies for the study from companies that con-sumers would conceivably interact with. We narrowed our

2The two systems are linked using a shared key that Surveyor’sPoint generates on the completion of our survey, which a participantthen enters back into Mechanical Turk. This allows us to link anentry in Mechanical Turk with an entry in Surveyor’s Point andverify the worker completed the survey before payment.

3

Standardization

3

Summarization

Page 7: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

Figure 1. An example of a standardized table is shown on the left, and a standardized short table on the right. The comparison highlights the rows

deleted to “shorten” this version. These deleted rows are listed directly below the table. While both formats contain the legend (bottom right), it is

displayed only on the right here due to space constraints.

have already been deployed by major corporations, makingthem a viable, real world summary format for privacy poli-cies. These policies were stripped of brand information, butthe formatting and styles were retained.

METHODOLOGYWe conducted an online user study in summer 2009 usingAmazon’s Mechanical Turk and a tool we developed calledSurveyor’s Point. Mechanical Turk offers workers the abil-ity to perform short tasks and get compensated. People canplace jobs through Mechanical Turk, specifying the numberof workers they are looking for, necessary qualifications, theamount they are willing to pay, and details about the task.Mechanical Turk payments are generally calibrated for thelength of the task. For our approximately 15-minute study,we paid $0.75 on successful completion.

We developed a custom survey-management tool called Sur-veyor’s Point to facilitate our data collection. Our implemen-tation allows us to show respondents a single question on thescreen along with links for switching back and forth betweentwo policies within a single browser window. This allowedus to track the number of users who looked at each policyand the number of times they switched between them. Addi-tionally, Surveyor’s Point allowed us to collect the amount oftime that users spent reading the policies, as well as informa-tion about whether they clicked through to opt-out forms, toadditional policy information links, or from a layered noticethrough to the full text policy.

In preparation for this study we first performed three smallerpilot tests of our survey framework. We ran our pilot studieswith approximately thirty users each, across 2-3 conditions.Our pilot studies helped us to finalize remaining design de-cisions surrounding the standardized short table, refine ourquestionnaire, and test the integration of Surveyor’s Pointwith Mechanical Turk.2

We then conducted our large-scale study and completed theanalysis with 764 participants (409 female, 355 male), ran-domly assigned to five conditions (see Table 1): full policytext, standardized table, standardized short table, standard-ized short text, and layered text. We dropped 25 additionalparticipants from the study prior to analysis due to incom-plete data or for completing the study in an amount of timethat indicated inadequate attention to the task (defined astime on task that was two standard deviations lower than themean). We chose a between-subjects design to remove learn-ing effects and ensure the study could be completed withinabout 15 minutes. Participants in each condition followedthe same protocol; only the policy format differed.

PoliciesWe selected policies for the study from companies that con-sumers would conceivably interact with. We narrowed our

2The two systems are linked using a shared key that Surveyor’sPoint generates on the completion of our survey, which a participantthen enters back into Mechanical Turk. This allows us to link anentry in Mechanical Turk with an entry in Surveyor’s Point andverify the worker completed the survey before payment.

3

Standardization

3

Summarization

Challenges

one size fits all

user education

Page 8: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

4

Privacy Choice: Current State

Page 9: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

fragmented ecosystem

difficult to find

4

Privacy Choice: Current State

Page 10: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

5

Conversation-first Interfaces

Page 11: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

The Rise of Conversational UI

6

Page 12: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

PriBots: Conversational Privacy Bots

Message|

7

Page 13: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

PriBots: Conversational Privacy Bots

Message|

7

Page 14: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

PriBots: Conversational Privacy Bots

Message|

Appeal to new tech adopters

7

Page 15: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

PriBots: Conversational Privacy Bots

Message|

Appeal to new tech adopters

Appeal to existing users

7

Page 16: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

PriBots: Conversational Privacy Bots

Message|

Appeal to new tech adopters

Appeal to existing users

An intuitive way to 1. communicate privacy policies 2. adjust privacy preferences

7

Page 17: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

1- Communicating Privacy Policies

8

Page 18: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

Channel

9

Page 19: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

Channel

Primary9

Page 20: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

Channel

Primary Secondary9

Page 21: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

Timing

10

Page 22: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

Timing

At-setup

10

Page 23: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

Timing

At-setup On-demand

10

Page 24: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

Feedback

implicit: sentiment analysis

explicit: structured messages

gathering users’ concerns

11

Page 25: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

Voicing User Concerns

Providers traditionally

say what they want

12

Page 26: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

Voicing User Concerns

Providers traditionally

say what they want

Users’ concerns might not

be covered

12

Page 27: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

Voicing User Concerns

Providers traditionally

say what they want

Users’ concerns might not

be covered

PriBots activate the

two-way channel

12

Page 28: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

2- Setting Privacy Preferences

13

Page 29: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

14

Page 30: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

Service and platform-dependent interface

14

Page 31: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

Service and platform-dependent interface Tradeoffs for simplicity: try finding this setting on Mobile Web version

14

Page 32: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

15

Page 33: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

Unique interface with all functionalities Ability to suggest adjustments to the user (combining notice and choice/preferences )

15

Page 34: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

User Input

Analysis & Classification

System Architecture

16

Page 35: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

User Input

Analysis & Classification

Query Structured Query

Statement

Yes

No

System Architecture

16

Page 36: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

User Input

Analysis & Classification

Query Structured Query

Statement

Yes

No

Retrieval Module

Result

Knowledge Base

System Architecture

16

Page 37: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

User Input

Analysis & Classification

Query Structured Query

Statement

Yes

No

Confident?Answer

Formulation in NL

Fallback Answer

Generation

Yes

No

Retrieval Module

Result

Knowledge Base

System Architecture

16

Page 38: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

PriBot Reply

User Input

Analysis & Classification

Query Structured Query

Statement

Yes

No

Confident?Answer

Formulation in NL

Fallback Answer

Generation

Yes

No

Retrieval Module

Result

Knowledge Base

System Architecture

16

Page 39: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

PriBot Reply

User Input

Analysis & Classification

Query Structured Query

Statement

Yes

No

Confident?Answer

Formulation in NL

Fallback Answer

Generation

Yes

No

Retrieval Module

Result

Knowledge Base

System Architecture

16

Feedback DB

AnalyticsAmendments/ Improvements

Augment the Knowledge Base

• unanswered queries • frequent questions • user sentiments

Feedback DB

Page 40: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

Challenges

17

Page 41: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

Mature User Understanding Text processing

Question answering

Domain-specific datasets and ontologies

Graceful fallback

18

Page 42: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

Legal Challenges

Inherently error prone: are they legally binding?

Accounting for false-positives and false-negatives

The case of 3rd party PriBots: defamation possibilities?

19

Page 43: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

Trusting the Machine rule-based vs. AI-based

user backlash?

regulate the confidence level

20

Page 44: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

PriBots’ Personality

positive tone → higher trust

diversified content → reduced habituation

21

Page 45: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

Deployment

22

Page 46: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

provider

3rd parties

Deployment

22

Page 47: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

provider

3rd parties

Deployment

22

Suitable for Voice Assistants

Page 48: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

Rule-based Prototype

System Implementation

User studies

What’s Next?

23

Privacy as a Dialogue

Page 49: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

24

Questions/Feedback?

[email protected]

Page 50: PriBots: Conversational Privacy with Chatbots (at SOUPS 2016)

Image/Media Credits Zara Picken: slide 12 Egor Kosten: slide 24 Alex Prokhoda: slide 6 Freepik: slide 23 Geoff Keough: slide 14 Victor: slide 12