TY - GEN
T1 - Crowdsourcing annotations for websites' privacy policies
T2 - 25th International World Wide Web Conference, WWW 2016
AU - Wilson, Shomir
AU - Schaub, Florian
AU - Ramanath, Rohan
AU - Sadeh, Norman
AU - Liu, Fei
AU - Smith, Noah A.
AU - Liu, Frederick
PY - 2016
Y1 - 2016
N2 - Website privacy policies are often long and difficult to understand. While research shows that Internet users care about their privacy, they do not have time to understand the policies of every website they visit, and most users hardly ever read privacy policies. Several recent efforts aim to crowdsource the interpretation of privacy policies and use the resulting annotations to build more effective user interfaces that provide users with salient policy summaries. However, very little attention has been devoted to studying the accuracy and scalability of crowdsourced privacy policy annotations, the types of questions crowdworkers can effectively answer, and the ways in which their productivity can be enhanced. Prior research indicates that most Internet users often have great difficulty understanding privacy policies, suggesting limits to the effectiveness of crowdsourcing approaches. In this paper, we assess the viability of crowdsourcing privacy policy annotations. Our results suggest that, if carefully deployed, crowdsourcing can indeed result in the generation of non-Trivial annotations and can also help identify elements of ambiguity in policies. We further introduce and evaluate a method to improve the annotation process by predicting and highlighting paragraphs relevant to specific data practices.
AB - Website privacy policies are often long and difficult to understand. While research shows that Internet users care about their privacy, they do not have time to understand the policies of every website they visit, and most users hardly ever read privacy policies. Several recent efforts aim to crowdsource the interpretation of privacy policies and use the resulting annotations to build more effective user interfaces that provide users with salient policy summaries. However, very little attention has been devoted to studying the accuracy and scalability of crowdsourced privacy policy annotations, the types of questions crowdworkers can effectively answer, and the ways in which their productivity can be enhanced. Prior research indicates that most Internet users often have great difficulty understanding privacy policies, suggesting limits to the effectiveness of crowdsourcing approaches. In this paper, we assess the viability of crowdsourcing privacy policy annotations. Our results suggest that, if carefully deployed, crowdsourcing can indeed result in the generation of non-Trivial annotations and can also help identify elements of ambiguity in policies. We further introduce and evaluate a method to improve the annotation process by predicting and highlighting paragraphs relevant to specific data practices.
UR - http://www.scopus.com/inward/record.url?scp=85025813224&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85025813224&partnerID=8YFLogxK
U2 - 10.1145/2872427.2883035
DO - 10.1145/2872427.2883035
M3 - Conference contribution
AN - SCOPUS:85025813224
T3 - 25th International World Wide Web Conference, WWW 2016
SP - 133
EP - 143
BT - 25th International World Wide Web Conference, WWW 2016
PB - International World Wide Web Conferences Steering Committee
Y2 - 11 April 2016 through 15 April 2016
ER -