Automated generation of privacy policy using deep models

Nastaran Bateni, Rozita Dara
{"title":"Automated generation of privacy policy using deep models","authors":"Nastaran Bateni, Rozita Dara","doi":"10.1109/istas52410.2021.9629155","DOIUrl":null,"url":null,"abstract":"Personal information protection and compliance with privacy regulations are becoming increasingly important due to a large number of security and privacy breaches. These privacy breaches can harm individuals in both personal and social contexts. Privacy policies are the primary means of communication with which service providers could inform users about the data collection and sharing practices. The content and transparency of these legal documents are of importance as they can help users make decisions about the service providers’ data privacy practices and can build trust with the users. Although many regulations and best practices have provided recommendations and guidelines on the content of privacy policies, research has shown that the content of these documents is usually incomplete and miss important topics. To address this issue, we propose and validate the use of automated generative models for creating the content of privacy policies. These generative approaches use deep learning models to generate enriched data practices and text for privacy policies. In this study, we trained two generative models (Long Short-Term Memory (LSTM) and bidirectional Long Short-Term Memory bi-LSTM) on annotated privacy policies to automatically generate privacy data practices. Training and testing were performed on three levels of paragraph, sentence, and data practice. Our findings have shown promising results and have suggested that models trained on legal data practices using bi-LSTM algorithm create more accurate results.","PeriodicalId":314239,"journal":{"name":"2021 IEEE International Symposium on Technology and Society (ISTAS)","volume":"20 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-10-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE International Symposium on Technology and Society (ISTAS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/istas52410.2021.9629155","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

Abstract

Personal information protection and compliance with privacy regulations are becoming increasingly important due to a large number of security and privacy breaches. These privacy breaches can harm individuals in both personal and social contexts. Privacy policies are the primary means of communication with which service providers could inform users about the data collection and sharing practices. The content and transparency of these legal documents are of importance as they can help users make decisions about the service providers’ data privacy practices and can build trust with the users. Although many regulations and best practices have provided recommendations and guidelines on the content of privacy policies, research has shown that the content of these documents is usually incomplete and miss important topics. To address this issue, we propose and validate the use of automated generative models for creating the content of privacy policies. These generative approaches use deep learning models to generate enriched data practices and text for privacy policies. In this study, we trained two generative models (Long Short-Term Memory (LSTM) and bidirectional Long Short-Term Memory bi-LSTM) on annotated privacy policies to automatically generate privacy data practices. Training and testing were performed on three levels of paragraph, sentence, and data practice. Our findings have shown promising results and have suggested that models trained on legal data practices using bi-LSTM algorithm create more accurate results.
使用深度模型自动生成隐私策略
由于大量的安全和隐私泄露,个人信息保护和遵守隐私法规变得越来越重要。这些侵犯隐私的行为会对个人和社会环境造成伤害。隐私政策是服务提供商告知用户有关数据收集和共享做法的主要通信手段。这些法律文件的内容和透明度非常重要,因为它们可以帮助用户对服务提供商的数据隐私做法做出决策,并可以与用户建立信任。尽管许多法规和最佳实践都对隐私政策的内容提供了建议和指导方针,但研究表明,这些文件的内容通常是不完整的,并且遗漏了重要的主题。为了解决这个问题,我们提出并验证了使用自动生成模型来创建隐私策略的内容。这些生成方法使用深度学习模型为隐私政策生成丰富的数据实践和文本。在这项研究中,我们训练了两个生成模型(长短期记忆(LSTM)和双向长短期记忆bi-LSTM)在标注隐私策略上自动生成隐私数据实践。训练和测试在段落、句子和数据练习三个层次上进行。我们的研究结果显示了有希望的结果,并表明使用bi-LSTM算法在法律数据实践中训练的模型可以产生更准确的结果。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信