arXiv - CS - Computation and Language最新文献

筛选
英文 中文
The CLC-UKET Dataset: Benchmarking Case Outcome Prediction for the UK Employment Tribunal CLC-UKET 数据集:英国就业法庭案件结果预测基准
arXiv - CS - Computation and Language Pub Date : 2024-09-12 DOI: arxiv-2409.08098
Huiyuan Xie, Felix Steffek, Joana Ribeiro de Faria, Christine Carter, Jonathan Rutherford
{"title":"The CLC-UKET Dataset: Benchmarking Case Outcome Prediction for the UK Employment Tribunal","authors":"Huiyuan Xie, Felix Steffek, Joana Ribeiro de Faria, Christine Carter, Jonathan Rutherford","doi":"arxiv-2409.08098","DOIUrl":"https://doi.org/arxiv-2409.08098","url":null,"abstract":"This paper explores the intersection of technological innovation and access\u0000to justice by developing a benchmark for predicting case outcomes in the UK\u0000Employment Tribunal (UKET). To address the challenge of extensive manual\u0000annotation, the study employs a large language model (LLM) for automatic\u0000annotation, resulting in the creation of the CLC-UKET dataset. The dataset\u0000consists of approximately 19,000 UKET cases and their metadata. Comprehensive\u0000legal annotations cover facts, claims, precedent references, statutory\u0000references, case outcomes, reasons and jurisdiction codes. Facilitated by the\u0000CLC-UKET data, we examine a multi-class case outcome prediction task in the\u0000UKET. Human predictions are collected to establish a performance reference for\u0000model comparison. Empirical results from baseline models indicate that\u0000finetuned transformer models outperform zero-shot and few-shot LLMs on the UKET\u0000prediction task. The performance of zero-shot LLMs can be enhanced by\u0000integrating task-related information into few-shot examples. We hope that the\u0000CLC-UKET dataset, along with human annotations and empirical findings, can\u0000serve as a valuable benchmark for employment-related dispute resolution.","PeriodicalId":501030,"journal":{"name":"arXiv - CS - Computation and Language","volume":"157 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142184413","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
TravelAgent: An AI Assistant for Personalized Travel Planning TravelAgent:个性化旅行规划的人工智能助手
arXiv - CS - Computation and Language Pub Date : 2024-09-12 DOI: arxiv-2409.08069
Aili Chen, Xuyang Ge, Ziquan Fu, Yanghua Xiao, Jiangjie Chen
{"title":"TravelAgent: An AI Assistant for Personalized Travel Planning","authors":"Aili Chen, Xuyang Ge, Ziquan Fu, Yanghua Xiao, Jiangjie Chen","doi":"arxiv-2409.08069","DOIUrl":"https://doi.org/arxiv-2409.08069","url":null,"abstract":"As global tourism expands and artificial intelligence technology advances,\u0000intelligent travel planning services have emerged as a significant research\u0000focus. Within dynamic real-world travel scenarios with multi-dimensional\u0000constraints, services that support users in automatically creating practical\u0000and customized travel itineraries must address three key objectives:\u0000Rationality, Comprehensiveness, and Personalization. However, existing systems\u0000with rule-based combinations or LLM-based planning methods struggle to fully\u0000satisfy these criteria. To overcome the challenges, we introduce TravelAgent, a\u0000travel planning system powered by large language models (LLMs) designed to\u0000provide reasonable, comprehensive, and personalized travel itineraries grounded\u0000in dynamic scenarios. TravelAgent comprises four modules: Tool-usage,\u0000Recommendation, Planning, and Memory Module. We evaluate TravelAgent's\u0000performance with human and simulated users, demonstrating its overall\u0000effectiveness in three criteria and confirming the accuracy of personalized\u0000recommendations.","PeriodicalId":501030,"journal":{"name":"arXiv - CS - Computation and Language","volume":"26 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142184532","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Source2Synth: Synthetic Data Generation and Curation Grounded in Real Data Sources Source2Synth:基于真实数据源的合成数据生成和整理
arXiv - CS - Computation and Language Pub Date : 2024-09-12 DOI: arxiv-2409.08239
Alisia Lupidi, Carlos Gemmell, Nicola Cancedda, Jane Dwivedi-Yu, Jason Weston, Jakob Foerster, Roberta Raileanu, Maria Lomeli
{"title":"Source2Synth: Synthetic Data Generation and Curation Grounded in Real Data Sources","authors":"Alisia Lupidi, Carlos Gemmell, Nicola Cancedda, Jane Dwivedi-Yu, Jason Weston, Jakob Foerster, Roberta Raileanu, Maria Lomeli","doi":"arxiv-2409.08239","DOIUrl":"https://doi.org/arxiv-2409.08239","url":null,"abstract":"Large Language Models still struggle in challenging scenarios that leverage\u0000structured data, complex reasoning, or tool usage. In this paper, we propose\u0000Source2Synth: a new method that can be used for teaching LLMs new skills\u0000without relying on costly human annotations. Source2Synth takes as input a\u0000custom data source and produces synthetic data points with intermediate\u0000reasoning steps grounded in real-world sources. Source2Synth improves the\u0000dataset quality by discarding low-quality generations based on their\u0000answerability. We demonstrate the generality of this approach by applying it to\u0000two challenging domains: we test reasoning abilities in multi-hop question\u0000answering (MHQA), and tool usage in tabular question answering (TQA). Our\u0000method improves performance by 25.51% for TQA on WikiSQL and 22.57% for MHQA on\u0000HotPotQA compared to the fine-tuned baselines.","PeriodicalId":501030,"journal":{"name":"arXiv - CS - Computation and Language","volume":"57 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142184410","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Supporting Online Discussions: Integrating AI Into the adhocracy+ Participation Platform To Enhance Deliberation 支持在线讨论:将人工智能融入adhocracy+参与平台,以提高议事能力
arXiv - CS - Computation and Language Pub Date : 2024-09-12 DOI: arxiv-2409.07780
Maike Behrendt, Stefan Sylvius Wagner, Stefan Harmeling
{"title":"Supporting Online Discussions: Integrating AI Into the adhocracy+ Participation Platform To Enhance Deliberation","authors":"Maike Behrendt, Stefan Sylvius Wagner, Stefan Harmeling","doi":"arxiv-2409.07780","DOIUrl":"https://doi.org/arxiv-2409.07780","url":null,"abstract":"Online spaces allow people to discuss important issues and make joint\u0000decisions, regardless of their location or time zone. However, without proper\u0000support and thoughtful design, these discussions often lack structure and\u0000politeness during the exchanges of opinions. Artificial intelligence (AI)\u0000represents an opportunity to support both participants and organizers of\u0000large-scale online participation processes. In this paper, we present an\u0000extension of adhocracy+, a large-scale open source participation platform, that\u0000provides two additional debate modules that are supported by AI to enhance the\u0000discussion quality and participant interaction.","PeriodicalId":501030,"journal":{"name":"arXiv - CS - Computation and Language","volume":"27 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142184436","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An Unsupervised Dialogue Topic Segmentation Model Based on Utterance Rewriting 基于语句重写的无监督对话主题分割模型
arXiv - CS - Computation and Language Pub Date : 2024-09-12 DOI: arxiv-2409.07672
Xia Hou, Qifeng Li, Tongliang Li
{"title":"An Unsupervised Dialogue Topic Segmentation Model Based on Utterance Rewriting","authors":"Xia Hou, Qifeng Li, Tongliang Li","doi":"arxiv-2409.07672","DOIUrl":"https://doi.org/arxiv-2409.07672","url":null,"abstract":"Dialogue topic segmentation plays a crucial role in various types of dialogue\u0000modeling tasks. The state-of-the-art unsupervised DTS methods learn topic-aware\u0000discourse representations from conversation data through adjacent discourse\u0000matching and pseudo segmentation to further mine useful clues in unlabeled\u0000conversational relations. However, in multi-round dialogs, discourses often\u0000have co-references or omissions, leading to the fact that direct use of these\u0000discourses for representation learning may negatively affect the semantic\u0000similarity computation in the neighboring discourse matching task. In order to\u0000fully utilize the useful cues in conversational relations, this study proposes\u0000a novel unsupervised dialog topic segmentation method that combines the\u0000Utterance Rewriting (UR) technique with an unsupervised learning algorithm to\u0000efficiently utilize the useful cues in unlabeled dialogs by rewriting the\u0000dialogs in order to recover the co-referents and omitted words. Compared with\u0000existing unsupervised models, the proposed Discourse Rewriting Topic\u0000Segmentation Model (UR-DTS) significantly improves the accuracy of topic\u0000segmentation. The main finding is that the performance on DialSeg711 improves\u0000by about 6% in terms of absolute error score and WD, achieving 11.42% in terms\u0000of absolute error score and 12.97% in terms of WD. on Doc2Dial the absolute\u0000error score and WD improves by about 3% and 2%, respectively, resulting in SOTA\u0000reaching 35.17% in terms of absolute error score and 38.49% in terms of WD.\u0000This shows that the model is very effective in capturing the nuances of\u0000conversational topics, as well as the usefulness and challenges of utilizing\u0000unlabeled conversations.","PeriodicalId":501030,"journal":{"name":"arXiv - CS - Computation and Language","volume":"38 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142184439","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
WhisperNER: Unified Open Named Entity and Speech Recognition WhisperNER:统一开放式命名实体和语音识别
arXiv - CS - Computation and Language Pub Date : 2024-09-12 DOI: arxiv-2409.08107
Gil Ayache, Menachem Pirchi, Aviv Navon, Aviv Shamsian, Gill Hetz, Joseph Keshet
{"title":"WhisperNER: Unified Open Named Entity and Speech Recognition","authors":"Gil Ayache, Menachem Pirchi, Aviv Navon, Aviv Shamsian, Gill Hetz, Joseph Keshet","doi":"arxiv-2409.08107","DOIUrl":"https://doi.org/arxiv-2409.08107","url":null,"abstract":"Integrating named entity recognition (NER) with automatic speech recognition\u0000(ASR) can significantly enhance transcription accuracy and informativeness. In\u0000this paper, we introduce WhisperNER, a novel model that allows joint speech\u0000transcription and entity recognition. WhisperNER supports open-type NER,\u0000enabling recognition of diverse and evolving entities at inference. Building on\u0000recent advancements in open NER research, we augment a large synthetic dataset\u0000with synthetic speech samples. This allows us to train WhisperNER on a large\u0000number of examples with diverse NER tags. During training, the model is\u0000prompted with NER labels and optimized to output the transcribed utterance\u0000along with the corresponding tagged entities. To evaluate WhisperNER, we\u0000generate synthetic speech for commonly used NER benchmarks and annotate\u0000existing ASR datasets with open NER tags. Our experiments demonstrate that\u0000WhisperNER outperforms natural baselines on both out-of-domain open type NER\u0000and supervised finetuning.","PeriodicalId":501030,"journal":{"name":"arXiv - CS - Computation and Language","volume":"24 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142184412","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-object event graph representation learning for Video Question Answering 用于视频问题解答的多对象事件图表示学习
arXiv - CS - Computation and Language Pub Date : 2024-09-12 DOI: arxiv-2409.07747
Yanan Wang, Shuichiro Haruta, Donghuo Zeng, Julio Vizcarra, Mori Kurokawa
{"title":"Multi-object event graph representation learning for Video Question Answering","authors":"Yanan Wang, Shuichiro Haruta, Donghuo Zeng, Julio Vizcarra, Mori Kurokawa","doi":"arxiv-2409.07747","DOIUrl":"https://doi.org/arxiv-2409.07747","url":null,"abstract":"Video question answering (VideoQA) is a task to predict the correct answer to\u0000questions posed about a given video. The system must comprehend spatial and\u0000temporal relationships among objects extracted from videos to perform causal\u0000and temporal reasoning. While prior works have focused on modeling individual\u0000object movements using transformer-based methods, they falter when capturing\u0000complex scenarios involving multiple objects (e.g., \"a boy is throwing a ball\u0000in a hoop\"). We propose a contrastive language event graph representation\u0000learning method called CLanG to address this limitation. Aiming to capture\u0000event representations associated with multiple objects, our method employs a\u0000multi-layer GNN-cluster module for adversarial graph representation learning,\u0000enabling contrastive learning between the question text and its relevant\u0000multi-object event graph. Our method outperforms a strong baseline, achieving\u0000up to 2.2% higher accuracy on two challenging VideoQA datasets, NExT-QA and\u0000TGIF-QA-R. In particular, it is 2.8% better than baselines in handling causal\u0000and temporal questions, highlighting its strength in reasoning multiple\u0000object-based events.","PeriodicalId":501030,"journal":{"name":"arXiv - CS - Computation and Language","volume":"24 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142184443","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Stable Language Model Pre-training by Reducing Embedding Variability 通过减少嵌入变异性实现稳定的语言模型预训练
arXiv - CS - Computation and Language Pub Date : 2024-09-12 DOI: arxiv-2409.07787
Woojin Chung, Jiwoo Hong, Na Min An, James Thorne, Se-Young Yun
{"title":"Stable Language Model Pre-training by Reducing Embedding Variability","authors":"Woojin Chung, Jiwoo Hong, Na Min An, James Thorne, Se-Young Yun","doi":"arxiv-2409.07787","DOIUrl":"https://doi.org/arxiv-2409.07787","url":null,"abstract":"Stable pre-training is essential for achieving better-performing language\u0000models. However, tracking pre-training stability by calculating gradient\u0000variance at every step is impractical due to the significant computational\u0000costs. We explore Token Embedding Variability (TEV) as a simple and efficient\u0000proxy for assessing pre-training stability in language models with pre-layer\u0000normalization, given that shallower layers are more prone to gradient explosion\u0000(section 2.2). Moreover, we propose Multi-head Low-Rank Attention (MLRA) as an\u0000architecture to alleviate such instability by limiting the exponential growth\u0000of output embedding variance, thereby preventing the gradient explosion\u0000(section 3.2). Empirical results on GPT-2 with MLRA demonstrate increased\u0000stability and lower perplexity, particularly in deeper models.","PeriodicalId":501030,"journal":{"name":"arXiv - CS - Computation and Language","volume":"49 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142184435","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
On the Role of Context in Reading Time Prediction 论语境在阅读时间预测中的作用
arXiv - CS - Computation and Language Pub Date : 2024-09-12 DOI: arxiv-2409.08160
Andreas Opedal, Eleanor Chodroff, Ryan Cotterell, Ethan Gotlieb Wilcox
{"title":"On the Role of Context in Reading Time Prediction","authors":"Andreas Opedal, Eleanor Chodroff, Ryan Cotterell, Ethan Gotlieb Wilcox","doi":"arxiv-2409.08160","DOIUrl":"https://doi.org/arxiv-2409.08160","url":null,"abstract":"We present a new perspective on how readers integrate context during\u0000real-time language comprehension. Our proposals build on surprisal theory,\u0000which posits that the processing effort of a linguistic unit (e.g., a word) is\u0000an affine function of its in-context information content. We first observe that\u0000surprisal is only one out of many potential ways that a contextual predictor\u0000can be derived from a language model. Another one is the pointwise mutual\u0000information (PMI) between a unit and its context, which turns out to yield the\u0000same predictive power as surprisal when controlling for unigram frequency.\u0000Moreover, both PMI and surprisal are correlated with frequency. This means that\u0000neither PMI nor surprisal contains information about context alone. In response\u0000to this, we propose a technique where we project surprisal onto the orthogonal\u0000complement of frequency, yielding a new contextual predictor that is\u0000uncorrelated with frequency. Our experiments show that the proportion of\u0000variance in reading times explained by context is a lot smaller when context is\u0000represented by the orthogonalized predictor. From an interpretability\u0000standpoint, this indicates that previous studies may have overstated the role\u0000that context has in predicting reading times.","PeriodicalId":501030,"journal":{"name":"arXiv - CS - Computation and Language","volume":"34 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142184441","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Learning Rules from KGs Guided by Language Models 在语言模型指导下从幼稚园学习规则
arXiv - CS - Computation and Language Pub Date : 2024-09-12 DOI: arxiv-2409.07869
Zihang Peng, Daria Stepanova, Vinh Thinh Ho, Heike Adel, Alessandra Russo, Simon Ott
{"title":"Learning Rules from KGs Guided by Language Models","authors":"Zihang Peng, Daria Stepanova, Vinh Thinh Ho, Heike Adel, Alessandra Russo, Simon Ott","doi":"arxiv-2409.07869","DOIUrl":"https://doi.org/arxiv-2409.07869","url":null,"abstract":"Advances in information extraction have enabled the automatic construction of\u0000large knowledge graphs (e.g., Yago, Wikidata or Google KG), which are widely\u0000used in many applications like semantic search or data analytics. However, due\u0000to their semi-automatic construction, KGs are often incomplete. Rule learning\u0000methods, concerned with the extraction of frequent patterns from KGs and\u0000casting them into rules, can be applied to predict potentially missing facts. A\u0000crucial step in this process is rule ranking. Ranking of rules is especially\u0000challenging over highly incomplete or biased KGs (e.g., KGs predominantly\u0000storing facts about famous people), as in this case biased rules might fit the\u0000data best and be ranked at the top based on standard statistical metrics like\u0000rule confidence. To address this issue, prior works proposed to rank rules not\u0000only relying on the original KG but also facts predicted by a KG embedding\u0000model. At the same time, with the recent rise of Language Models (LMs), several\u0000works have claimed that LMs can be used as alternative means for KG completion.\u0000In this work, our goal is to verify to which extent the exploitation of LMs is\u0000helpful for improving the quality of rule learning systems.","PeriodicalId":501030,"journal":{"name":"arXiv - CS - Computation and Language","volume":"29 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142184433","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信