机器学习和社交网络风险的其他信息分析。

Jun Zhuang, Chen Wang, James H Lambert
{"title":"机器学习和社交网络风险的其他信息分析。","authors":"Jun Zhuang, Chen Wang, James H Lambert","doi":"10.1111/risa.13945","DOIUrl":null,"url":null,"abstract":"The world has entered a digital era where people and machines leave digital footprints in websites, social media, cameras, sensor logs, and mobile devices. For example, manufacturing and operating systems collect streaming data through sensors and the Internet of Things (IoT); vehicles generate vast amounts of trajectory and sensor log data in aviation and surface transportation. Whether from natural, technological, or adversarial hazards, risks arise from various root causes, including human errors. Risk analysis has opportunities to integrate big data, natural language processing, computer vision, and machine learning methods in the digital era. This special issue features papers presented at the Conference on Risk Analysis, Decision Analysis and Security, Buffalo/Niagara Falls, NY, July 30–August 2, 2019 (organized by Drs. Jun Zhuang and Chen Wang). Collectively, the papers describe cutting-edge research on the possibilities of harnessing high-volume, high-dimensional, multisource, and multimodal data to give insights for risk assessment, communication, and management, as well as cover perspectives discussing the scope and limitations of big data risk analytics. The applications include food safety, cyber security, disaster mitigation, and recovery, misinformation and disinformation in social media, aviation safety, insurance fraud detection, health risk for emergency responders, autonomous driving, privacy risk management, and service failure prediction in transportation. The data utilized by these studies range from streaming data (e.g., from online social media, flight data recorders, and sensors of operating equipment), event data (e.g., fraud records of insurance, logs of stress leaves of emergency responders, and food safety incidents), to expert judgments. The methodologies cover a broad spectrum of analytic modeling, statistical inference, machine learning, expert elicitation, and combinations. Welburn and Strong propose an analytic framework to describe the systemic cyber risk resulting from cascading common cause or independent failures following a cyber incident. They apply the sector-level input–output analysis in economics to assess the aggregate losses associated with firm-level cyber incidents. Their model is validated using a cyber-attack case with known damages. The model can help determine cyber insurance premiums and make cybersecurity policies. Allodi et al. propose a model of a “work averse” attacker in a cybersecurity setting where the attacker is inclined to adopt existing toolkits if they can cause enough harm to the systems rather than develop exploits for new vulnerabilities. The authors build hypotheses based on this model, and evaluate these hypotheses using a large-scale dataset involving two million attack signatures recorded by Symantec against an extensive collection of information systems. The analytic and empirical analyses provide an example of work-aversion tendencies of cyber attackers using mass (but imperfect) data. Cains et al. conduct expert elicitation by interviewing cyber researchers from multiple disciplines and academia and government-military sectors. They perform a data-driven thematic analysis of the interview responses. They identify “context-driven,” “resilient system functionality,” and “maintenance of CIA (confidentiality, integrity, availability)” as salient themes for defining cyber security, and found “impacts of CIA vulnerabilities,” “probabilities of outcomes,” and “context-driven” as the most salient themes for cyber security risk. The resulting definitions encompass interdisciplinary perspectives and explicitly incorporate human factors as a significant cyber security risk factor. Yeo et al. examine the Twitter communication data corresponding to the 2016 Southern Louisiana flood recovery process. They discover fading communication volumes, a lack of local voices, and prolonging negative sentiments over the recovery period. The findings highlight the need for investing in risk communication during the disaster recovery process by better utilizing the platform of online social media and supporting local voices. Gongora-Svartzman and Ramirez-Marquez measure social cohesion in disasters by connections and interactions of people through social media outlets. They implement text processing techniques and graph network analysis to understand the relationships between nine types of participants (e.g., citizen, media, politics-government, entertainment, business, charity-NGOs-volunteering, etc.) during hurricanes Harvey, Irma, and Maria. The social cohesion measure revealed the various actors and their dynamics before, during, and after a disaster strikes. Byrd and John conduct experiments to investigate how the public distinguishes between true and false information in online social media following extreme events. They present the subjects with actual social media posts following a natural disaster or soft-target terrorist attack in the United States, and deliberately controlled the error penalties and base rate of true posts. They observe that the subjects performed","PeriodicalId":517072,"journal":{"name":"Risk analysis : an official publication of the Society for Risk Analysis","volume":" ","pages":"1603-1605"},"PeriodicalIF":0.0000,"publicationDate":"2022-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Machine learning and other information analyses for risk in social networks.\",\"authors\":\"Jun Zhuang, Chen Wang, James H Lambert\",\"doi\":\"10.1111/risa.13945\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The world has entered a digital era where people and machines leave digital footprints in websites, social media, cameras, sensor logs, and mobile devices. For example, manufacturing and operating systems collect streaming data through sensors and the Internet of Things (IoT); vehicles generate vast amounts of trajectory and sensor log data in aviation and surface transportation. Whether from natural, technological, or adversarial hazards, risks arise from various root causes, including human errors. Risk analysis has opportunities to integrate big data, natural language processing, computer vision, and machine learning methods in the digital era. This special issue features papers presented at the Conference on Risk Analysis, Decision Analysis and Security, Buffalo/Niagara Falls, NY, July 30–August 2, 2019 (organized by Drs. Jun Zhuang and Chen Wang). Collectively, the papers describe cutting-edge research on the possibilities of harnessing high-volume, high-dimensional, multisource, and multimodal data to give insights for risk assessment, communication, and management, as well as cover perspectives discussing the scope and limitations of big data risk analytics. The applications include food safety, cyber security, disaster mitigation, and recovery, misinformation and disinformation in social media, aviation safety, insurance fraud detection, health risk for emergency responders, autonomous driving, privacy risk management, and service failure prediction in transportation. The data utilized by these studies range from streaming data (e.g., from online social media, flight data recorders, and sensors of operating equipment), event data (e.g., fraud records of insurance, logs of stress leaves of emergency responders, and food safety incidents), to expert judgments. The methodologies cover a broad spectrum of analytic modeling, statistical inference, machine learning, expert elicitation, and combinations. Welburn and Strong propose an analytic framework to describe the systemic cyber risk resulting from cascading common cause or independent failures following a cyber incident. They apply the sector-level input–output analysis in economics to assess the aggregate losses associated with firm-level cyber incidents. Their model is validated using a cyber-attack case with known damages. The model can help determine cyber insurance premiums and make cybersecurity policies. Allodi et al. propose a model of a “work averse” attacker in a cybersecurity setting where the attacker is inclined to adopt existing toolkits if they can cause enough harm to the systems rather than develop exploits for new vulnerabilities. The authors build hypotheses based on this model, and evaluate these hypotheses using a large-scale dataset involving two million attack signatures recorded by Symantec against an extensive collection of information systems. The analytic and empirical analyses provide an example of work-aversion tendencies of cyber attackers using mass (but imperfect) data. Cains et al. conduct expert elicitation by interviewing cyber researchers from multiple disciplines and academia and government-military sectors. They perform a data-driven thematic analysis of the interview responses. They identify “context-driven,” “resilient system functionality,” and “maintenance of CIA (confidentiality, integrity, availability)” as salient themes for defining cyber security, and found “impacts of CIA vulnerabilities,” “probabilities of outcomes,” and “context-driven” as the most salient themes for cyber security risk. The resulting definitions encompass interdisciplinary perspectives and explicitly incorporate human factors as a significant cyber security risk factor. Yeo et al. examine the Twitter communication data corresponding to the 2016 Southern Louisiana flood recovery process. They discover fading communication volumes, a lack of local voices, and prolonging negative sentiments over the recovery period. The findings highlight the need for investing in risk communication during the disaster recovery process by better utilizing the platform of online social media and supporting local voices. Gongora-Svartzman and Ramirez-Marquez measure social cohesion in disasters by connections and interactions of people through social media outlets. They implement text processing techniques and graph network analysis to understand the relationships between nine types of participants (e.g., citizen, media, politics-government, entertainment, business, charity-NGOs-volunteering, etc.) during hurricanes Harvey, Irma, and Maria. The social cohesion measure revealed the various actors and their dynamics before, during, and after a disaster strikes. Byrd and John conduct experiments to investigate how the public distinguishes between true and false information in online social media following extreme events. They present the subjects with actual social media posts following a natural disaster or soft-target terrorist attack in the United States, and deliberately controlled the error penalties and base rate of true posts. They observe that the subjects performed\",\"PeriodicalId\":517072,\"journal\":{\"name\":\"Risk analysis : an official publication of the Society for Risk Analysis\",\"volume\":\" \",\"pages\":\"1603-1605\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-08-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Risk analysis : an official publication of the Society for Risk Analysis\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://doi.org/10.1111/risa.13945\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Risk analysis : an official publication of the Society for Risk Analysis","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1111/risa.13945","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

本文章由计算机程序翻译,如有差异,请以英文原文为准。
Machine learning and other information analyses for risk in social networks.
The world has entered a digital era where people and machines leave digital footprints in websites, social media, cameras, sensor logs, and mobile devices. For example, manufacturing and operating systems collect streaming data through sensors and the Internet of Things (IoT); vehicles generate vast amounts of trajectory and sensor log data in aviation and surface transportation. Whether from natural, technological, or adversarial hazards, risks arise from various root causes, including human errors. Risk analysis has opportunities to integrate big data, natural language processing, computer vision, and machine learning methods in the digital era. This special issue features papers presented at the Conference on Risk Analysis, Decision Analysis and Security, Buffalo/Niagara Falls, NY, July 30–August 2, 2019 (organized by Drs. Jun Zhuang and Chen Wang). Collectively, the papers describe cutting-edge research on the possibilities of harnessing high-volume, high-dimensional, multisource, and multimodal data to give insights for risk assessment, communication, and management, as well as cover perspectives discussing the scope and limitations of big data risk analytics. The applications include food safety, cyber security, disaster mitigation, and recovery, misinformation and disinformation in social media, aviation safety, insurance fraud detection, health risk for emergency responders, autonomous driving, privacy risk management, and service failure prediction in transportation. The data utilized by these studies range from streaming data (e.g., from online social media, flight data recorders, and sensors of operating equipment), event data (e.g., fraud records of insurance, logs of stress leaves of emergency responders, and food safety incidents), to expert judgments. The methodologies cover a broad spectrum of analytic modeling, statistical inference, machine learning, expert elicitation, and combinations. Welburn and Strong propose an analytic framework to describe the systemic cyber risk resulting from cascading common cause or independent failures following a cyber incident. They apply the sector-level input–output analysis in economics to assess the aggregate losses associated with firm-level cyber incidents. Their model is validated using a cyber-attack case with known damages. The model can help determine cyber insurance premiums and make cybersecurity policies. Allodi et al. propose a model of a “work averse” attacker in a cybersecurity setting where the attacker is inclined to adopt existing toolkits if they can cause enough harm to the systems rather than develop exploits for new vulnerabilities. The authors build hypotheses based on this model, and evaluate these hypotheses using a large-scale dataset involving two million attack signatures recorded by Symantec against an extensive collection of information systems. The analytic and empirical analyses provide an example of work-aversion tendencies of cyber attackers using mass (but imperfect) data. Cains et al. conduct expert elicitation by interviewing cyber researchers from multiple disciplines and academia and government-military sectors. They perform a data-driven thematic analysis of the interview responses. They identify “context-driven,” “resilient system functionality,” and “maintenance of CIA (confidentiality, integrity, availability)” as salient themes for defining cyber security, and found “impacts of CIA vulnerabilities,” “probabilities of outcomes,” and “context-driven” as the most salient themes for cyber security risk. The resulting definitions encompass interdisciplinary perspectives and explicitly incorporate human factors as a significant cyber security risk factor. Yeo et al. examine the Twitter communication data corresponding to the 2016 Southern Louisiana flood recovery process. They discover fading communication volumes, a lack of local voices, and prolonging negative sentiments over the recovery period. The findings highlight the need for investing in risk communication during the disaster recovery process by better utilizing the platform of online social media and supporting local voices. Gongora-Svartzman and Ramirez-Marquez measure social cohesion in disasters by connections and interactions of people through social media outlets. They implement text processing techniques and graph network analysis to understand the relationships between nine types of participants (e.g., citizen, media, politics-government, entertainment, business, charity-NGOs-volunteering, etc.) during hurricanes Harvey, Irma, and Maria. The social cohesion measure revealed the various actors and their dynamics before, during, and after a disaster strikes. Byrd and John conduct experiments to investigate how the public distinguishes between true and false information in online social media following extreme events. They present the subjects with actual social media posts following a natural disaster or soft-target terrorist attack in the United States, and deliberately controlled the error penalties and base rate of true posts. They observe that the subjects performed
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信