{"title":"Web Site Management System through Private Information Extraction","authors":"Myung Sil Choi, Y. Park, K. Ahn","doi":"10.1109/WAINA.2009.114","DOIUrl":null,"url":null,"abstract":"This paper presents a management system to effectively extract private information embedded in certain websites. To protect the target private information, it is necessary to first collect all private information from the target website, because vulnerable information leakages should be analyzed so as to come up with appropriate protection means. We use a crawling method in order to collect private information from the website. And a website is ordered based on collected information with structured. This way, we use a directed graph, determine web documents as a kind of nodes, and assign a weighting to documents containing private information, thus addressing time and economic problems for every crawling. We experimented with actual websites, demonstrating that our crawling method was superior in extracting and analyzing private information from websites.","PeriodicalId":159465,"journal":{"name":"2009 International Conference on Advanced Information Networking and Applications Workshops","volume":"74 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2009-05-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2009 International Conference on Advanced Information Networking and Applications Workshops","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/WAINA.2009.114","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
This paper presents a management system to effectively extract private information embedded in certain websites. To protect the target private information, it is necessary to first collect all private information from the target website, because vulnerable information leakages should be analyzed so as to come up with appropriate protection means. We use a crawling method in order to collect private information from the website. And a website is ordered based on collected information with structured. This way, we use a directed graph, determine web documents as a kind of nodes, and assign a weighting to documents containing private information, thus addressing time and economic problems for every crawling. We experimented with actual websites, demonstrating that our crawling method was superior in extracting and analyzing private information from websites.