Military Artificial Intelligence and the Principle of Distinction: A State Responsibility Perspective

IF 0.7 Q2 Social Sciences
Magdalena Pacholska
{"title":"Military Artificial Intelligence and the Principle of Distinction: A State Responsibility Perspective","authors":"Magdalena Pacholska","doi":"10.1017/S0021223722000188","DOIUrl":null,"url":null,"abstract":"Abstract Military artificial intelligence (AI)-enabled technology might still be in the relatively fledgling stages but the debate on how to regulate its use is already in full swing. Much of the discussion revolves around autonomous weapons systems (AWS) and the ‘responsibility gap’ they would ostensibly produce. This contribution argues that while some military AI technologies may indeed cause a range of conceptual hurdles in the realm of individual responsibility, they do not raise any unique issues under the law of state responsibility. The following analysis considers the latter regime and maps out crucial junctions in applying it to potential violations of the cornerstone of international humanitarian law (IHL) – the principle of distinction – resulting from the use of AI-enabled military technologies. It reveals that any challenges in ascribing responsibility in cases involving AWS would not be caused by the incorporation of AI, but stem from pre-existing systemic shortcomings of IHL and the unclear reverberations of mistakes thereunder. The article reiterates that state responsibility for the effects of AWS deployment is always retained through the commander's ultimate responsibility to authorise weapon deployment in accordance with IHL. It is proposed, however, that should the so-called fully autonomous weapon systems – that is, machine learning-based lethal systems that are capable of changing their own rules of operation beyond a predetermined framework – ever be fielded, it might be fairer to attribute their conduct to the fielding state, by conceptualising them as state agents, and treat them akin to state organs.","PeriodicalId":44911,"journal":{"name":"ISRAEL LAW REVIEW","volume":null,"pages":null},"PeriodicalIF":0.7000,"publicationDate":"2022-12-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"ISRAEL LAW REVIEW","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1017/S0021223722000188","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"Social Sciences","Score":null,"Total":0}
引用次数: 0

Abstract

Abstract Military artificial intelligence (AI)-enabled technology might still be in the relatively fledgling stages but the debate on how to regulate its use is already in full swing. Much of the discussion revolves around autonomous weapons systems (AWS) and the ‘responsibility gap’ they would ostensibly produce. This contribution argues that while some military AI technologies may indeed cause a range of conceptual hurdles in the realm of individual responsibility, they do not raise any unique issues under the law of state responsibility. The following analysis considers the latter regime and maps out crucial junctions in applying it to potential violations of the cornerstone of international humanitarian law (IHL) – the principle of distinction – resulting from the use of AI-enabled military technologies. It reveals that any challenges in ascribing responsibility in cases involving AWS would not be caused by the incorporation of AI, but stem from pre-existing systemic shortcomings of IHL and the unclear reverberations of mistakes thereunder. The article reiterates that state responsibility for the effects of AWS deployment is always retained through the commander's ultimate responsibility to authorise weapon deployment in accordance with IHL. It is proposed, however, that should the so-called fully autonomous weapon systems – that is, machine learning-based lethal systems that are capable of changing their own rules of operation beyond a predetermined framework – ever be fielded, it might be fairer to attribute their conduct to the fielding state, by conceptualising them as state agents, and treat them akin to state organs.
军事人工智能与区别原则:国家责任视角
军事人工智能(AI)技术可能仍处于相对羽翼未丰的阶段,但关于如何规范其使用的辩论已经全面展开。大部分讨论都围绕着自主武器系统(AWS)及其表面上可能产生的“责任鸿沟”展开。这篇文章认为,虽然一些军用人工智能技术确实可能在个人责任领域造成一系列概念障碍,但在国家责任法下,它们并没有引起任何独特的问题。以下分析考虑了后一种制度,并指出了将其应用于因使用人工智能军事技术而可能违反国际人道法基石(区分原则)的关键节点。它表明,在涉及人工智能的案件中,责任归属方面的任何挑战都不是由人工智能的纳入引起的,而是源于国际人道法先前存在的系统性缺陷以及其中错误的不明确影响。该条重申,根据国际人道法,指挥官负有授权部署武器的最终责任,因此始终保留国家对部署空战部队的影响的责任。然而,有人建议,如果所谓的完全自主武器系统——即基于机器学习的致命系统,能够在预定框架之外改变自己的操作规则——被部署,那么通过将其概念化为国家代理人,并将其视为类似于国家机关,将其行为归因于部署国可能更公平。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
1.00
自引率
14.30%
发文量
19
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信