{"title":"Automated generation of privacy policy using deep models","authors":"Nastaran Bateni, Rozita Dara","doi":"10.1109/istas52410.2021.9629155","DOIUrl":null,"url":null,"abstract":"Personal information protection and compliance with privacy regulations are becoming increasingly important due to a large number of security and privacy breaches. These privacy breaches can harm individuals in both personal and social contexts. Privacy policies are the primary means of communication with which service providers could inform users about the data collection and sharing practices. The content and transparency of these legal documents are of importance as they can help users make decisions about the service providers’ data privacy practices and can build trust with the users. Although many regulations and best practices have provided recommendations and guidelines on the content of privacy policies, research has shown that the content of these documents is usually incomplete and miss important topics. To address this issue, we propose and validate the use of automated generative models for creating the content of privacy policies. These generative approaches use deep learning models to generate enriched data practices and text for privacy policies. In this study, we trained two generative models (Long Short-Term Memory (LSTM) and bidirectional Long Short-Term Memory bi-LSTM) on annotated privacy policies to automatically generate privacy data practices. Training and testing were performed on three levels of paragraph, sentence, and data practice. Our findings have shown promising results and have suggested that models trained on legal data practices using bi-LSTM algorithm create more accurate results.","PeriodicalId":314239,"journal":{"name":"2021 IEEE International Symposium on Technology and Society (ISTAS)","volume":"20 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-10-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE International Symposium on Technology and Society (ISTAS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/istas52410.2021.9629155","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
Personal information protection and compliance with privacy regulations are becoming increasingly important due to a large number of security and privacy breaches. These privacy breaches can harm individuals in both personal and social contexts. Privacy policies are the primary means of communication with which service providers could inform users about the data collection and sharing practices. The content and transparency of these legal documents are of importance as they can help users make decisions about the service providers’ data privacy practices and can build trust with the users. Although many regulations and best practices have provided recommendations and guidelines on the content of privacy policies, research has shown that the content of these documents is usually incomplete and miss important topics. To address this issue, we propose and validate the use of automated generative models for creating the content of privacy policies. These generative approaches use deep learning models to generate enriched data practices and text for privacy policies. In this study, we trained two generative models (Long Short-Term Memory (LSTM) and bidirectional Long Short-Term Memory bi-LSTM) on annotated privacy policies to automatically generate privacy data practices. Training and testing were performed on three levels of paragraph, sentence, and data practice. Our findings have shown promising results and have suggested that models trained on legal data practices using bi-LSTM algorithm create more accurate results.