Towards the automatic detection and identification of English puns

Q2 Social Sciences
Tristan Miller, M. Turkovic
{"title":"Towards the automatic detection and identification of English puns","authors":"Tristan Miller, M. Turkovic","doi":"10.7592/EJHR2016.4.1.MILLER","DOIUrl":null,"url":null,"abstract":"Lexical polysemy, a fundamental characteristic of all human languages, has long been regarded as a major challenge to machine translation, human–computer interaction, and other applications of computational natural language processing (NLP). Traditional approaches to automatic word sense disambiguation (WSD) rest on the assumption that there exists a single, unambiguous communicative intention underlying every word in a document. However, writers sometimes intend for a word to be interpreted as simultaneously carrying multiple distinct meanings. This deliberate use of lexical ambiguity — i.e. punning — is a particularly common source of humour, and therefore has important implications for how NLP systems process documents and interact with users. In this paper we make a case for research into computational methods for the detection of puns in running text and for the isolation of the intended meanings. We discuss the challenges involved in adapting principles and techniques from WSD to humorously ambiguous text, and outline our plans for evaluating WSD-inspired systems in a dedicated pun identification task. We describe the compilation of a large manually annotated corpus of puns and present an analysis of its properties. While our work is principally concerned with simple puns which are monolexemic and homographic (i.e. exploiting single words which have different meanings but are spelled identically), we touch on the challenges involved in processing other types .","PeriodicalId":37540,"journal":{"name":"European Journal of Humour Research","volume":"4 1","pages":"59-75"},"PeriodicalIF":0.0000,"publicationDate":"2016-01-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"29","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"European Journal of Humour Research","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.7592/EJHR2016.4.1.MILLER","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"Social Sciences","Score":null,"Total":0}
引用次数: 29

Abstract

Lexical polysemy, a fundamental characteristic of all human languages, has long been regarded as a major challenge to machine translation, human–computer interaction, and other applications of computational natural language processing (NLP). Traditional approaches to automatic word sense disambiguation (WSD) rest on the assumption that there exists a single, unambiguous communicative intention underlying every word in a document. However, writers sometimes intend for a word to be interpreted as simultaneously carrying multiple distinct meanings. This deliberate use of lexical ambiguity — i.e. punning — is a particularly common source of humour, and therefore has important implications for how NLP systems process documents and interact with users. In this paper we make a case for research into computational methods for the detection of puns in running text and for the isolation of the intended meanings. We discuss the challenges involved in adapting principles and techniques from WSD to humorously ambiguous text, and outline our plans for evaluating WSD-inspired systems in a dedicated pun identification task. We describe the compilation of a large manually annotated corpus of puns and present an analysis of its properties. While our work is principally concerned with simple puns which are monolexemic and homographic (i.e. exploiting single words which have different meanings but are spelled identically), we touch on the challenges involved in processing other types .
实现英语双关语的自动检测与识别
词汇多义是所有人类语言的基本特征,长期以来一直被认为是机器翻译、人机交互和其他计算自然语言处理(NLP)应用的主要挑战。传统的自动词义消歧(WSD)方法基于一个假设,即文档中的每个单词都存在一个单一的、明确的交际意图。然而,作家有时希望一个词被解释为同时具有多种不同的含义。这种刻意使用的词汇歧义——即双关语——是幽默的一种特别常见的来源,因此对NLP系统如何处理文档和与用户交互具有重要意义。在本文中,我们对运行文本中双关语检测和意图意义分离的计算方法进行了研究。我们讨论了将WSD的原则和技术应用于幽默的歧义文本所涉及的挑战,并概述了我们在一个专门的双关语识别任务中评估WSD启发系统的计划。我们描述了一个大型人工标注双关语语料库的编译,并对其属性进行了分析。虽然我们的工作主要涉及单音节和同形双关语(即利用具有不同含义但拼写相同的单个单词)的简单双关语,但我们触及了处理其他类型双关语所涉及的挑战。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
European Journal of Humour Research
European Journal of Humour Research Social Sciences-Cultural Studies
CiteScore
1.10
自引率
0.00%
发文量
34
审稿时长
6 weeks
期刊介绍: The European Journal of Humour Research (EJHR) is a peer-reviewed quarterly journal with an international multidisciplinary editorial board. Although geographically-oriented towards the ˋold continentˊ, the European perspective aims at an international readership and contributors. EJHR covers the full range of work being done on all aspects of humour phenomenon. EJHR is designed to respond to the important changes that have affected the study of humour but particular predominance is given to the past events and current developments in Europe.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信