‘A cannon's burst discharged against a ruinated wall’: A Critique of Quantitative Methods in Shakespearean Authorial Attribution

Q2 Arts and Humanities
Authorship Pub Date : 2018-12-13 DOI:10.21825/AJ.V7I2.9737
D. Auerbach
{"title":"‘A cannon's burst discharged against a ruinated wall’: A Critique of Quantitative Methods in Shakespearean Authorial Attribution","authors":"D. Auerbach","doi":"10.21825/AJ.V7I2.9737","DOIUrl":null,"url":null,"abstract":"Authorship studies has, over the last two decades, absorbed a number of quantitative methods only made possible through the use of computers. The New Oxford Shakespeare Authorship Companion presents a number of studies that utilize such methods, including some based in machine learning or “deep learning” models.This paper focuses on the specific application of three such methods in Jack Elliott and Brett Greatley-Hirsch’s “Arden of Faversham and the Print of Many.” It finds that their attribution of the authorship of Arden to William Shakespeare is suspect under all three such methods: Delta, Nearest Shrunken Centroid, and Random Forests. The underlying models do not sufficiently justify the attributions, the data provided are insufficiently specific, and the internals of the methods are too opaque to bear up to scrutiny. This article attempts to depict the internal flaws of the methods, with a particular focus on Nearest Shrunken Centroid.These methodological flaws arguably arise in part from a lack of rigor, but also from an impoverished treatment of the available data, focusing exclusively on comparative word frequencies within and across authors. A number of potentially fruitful directions that authorship studies are suggested, that could increase the robustness and accuracy of quantitative methods, as well as warn of the potential limits of such methods.","PeriodicalId":30455,"journal":{"name":"Authorship","volume":" ","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2018-12-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"7","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Authorship","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.21825/AJ.V7I2.9737","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"Arts and Humanities","Score":null,"Total":0}
引用次数: 7

Abstract

Authorship studies has, over the last two decades, absorbed a number of quantitative methods only made possible through the use of computers. The New Oxford Shakespeare Authorship Companion presents a number of studies that utilize such methods, including some based in machine learning or “deep learning” models.This paper focuses on the specific application of three such methods in Jack Elliott and Brett Greatley-Hirsch’s “Arden of Faversham and the Print of Many.” It finds that their attribution of the authorship of Arden to William Shakespeare is suspect under all three such methods: Delta, Nearest Shrunken Centroid, and Random Forests. The underlying models do not sufficiently justify the attributions, the data provided are insufficiently specific, and the internals of the methods are too opaque to bear up to scrutiny. This article attempts to depict the internal flaws of the methods, with a particular focus on Nearest Shrunken Centroid.These methodological flaws arguably arise in part from a lack of rigor, but also from an impoverished treatment of the available data, focusing exclusively on comparative word frequencies within and across authors. A number of potentially fruitful directions that authorship studies are suggested, that could increase the robustness and accuracy of quantitative methods, as well as warn of the potential limits of such methods.
“大炮的爆裂声击中了被毁坏的墙壁”:莎士比亚权威归因的定量方法批判
在过去的二十年里,作者研究吸收了许多只有通过使用计算机才能实现的定量方法。《新牛津莎士比亚作家指南》介绍了许多利用这种方法的研究,包括一些基于机器学习或“深度学习”模型的研究。本文重点研究了Jack Elliott和Brett Greatley Hirsch的《Faversham的阿登和许多人的版画》中这三种方法的具体应用。发现他们将阿登的作者归属于William Shakespeare在所有三种方法下都是可疑的:三角洲、最近收缩质心和随机森林。基础模型没有充分证明归因的合理性,提供的数据不够具体,方法的内部过于不透明,无法经得起审查。本文试图描述这些方法的内部缺陷,特别关注最近收缩质心。这些方法上的缺陷可以说部分源于缺乏严谨性,但也源于对现有数据的处理不力,只关注作者内部和作者之间的比较词频。作者研究提出了一些可能富有成效的方向,这些方向可以提高定量方法的稳健性和准确性,并警告这些方法的潜在局限性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
0.20
自引率
0.00%
发文量
0
审稿时长
24 weeks
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信