{"title":"Prompt-Based Out-of-Distribution Intent Detection","authors":"Rudolf Chow;Albert Y. S. Lam","doi":"10.1109/TETCI.2024.3372440","DOIUrl":null,"url":null,"abstract":"Recent rapid advances in pre-trained language models, such as BERT and GPT, in natural language processing (NLP) have greatly improved the efficacy of text classifiers, easily surpassing human level performance in standard datasets like GLUE. However, most of these standard tasks implicitly assume a closed-world situation, where all testing data are supposed to lie in the same scope or distribution of the training data. Out-of-distribution (OOD) detection is the task of detecting when an input data point lies beyond the scope of the seen training set. This is becoming increasingly important as NLP agents, such as chatbots or virtual assistants, have been being deployed ubiquitously in our daily lives, thus attracting more attention from the research community to make it more accurate and robust at the same time. Recent work can be broadly categorized into two orthogonal approaches – data generative/augmentative methods and threshold/boundary learning. In this work, we follow the former and propose a method for the task based on prompting, which is known for its zero and few-shot capabilities. Generating synthetic outliers in terms of prompts allows the model to more efficiently learn OOD samples than the existing methods. Testing on nine different settings across three standard datasets used for OOD detection, our method with adaptive decision boundary is able to achieve competitive or superior performances compared with the current state-of-the-art in all cases. We also provide extensive analysis on each dataset as well as perform comprehensive ablation studies on each component of our model.","PeriodicalId":13135,"journal":{"name":"IEEE Transactions on Emerging Topics in Computational Intelligence","volume":"9 3","pages":"2371-2382"},"PeriodicalIF":5.3000,"publicationDate":"2025-03-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Emerging Topics in Computational Intelligence","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/11016146/","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Recent rapid advances in pre-trained language models, such as BERT and GPT, in natural language processing (NLP) have greatly improved the efficacy of text classifiers, easily surpassing human level performance in standard datasets like GLUE. However, most of these standard tasks implicitly assume a closed-world situation, where all testing data are supposed to lie in the same scope or distribution of the training data. Out-of-distribution (OOD) detection is the task of detecting when an input data point lies beyond the scope of the seen training set. This is becoming increasingly important as NLP agents, such as chatbots or virtual assistants, have been being deployed ubiquitously in our daily lives, thus attracting more attention from the research community to make it more accurate and robust at the same time. Recent work can be broadly categorized into two orthogonal approaches – data generative/augmentative methods and threshold/boundary learning. In this work, we follow the former and propose a method for the task based on prompting, which is known for its zero and few-shot capabilities. Generating synthetic outliers in terms of prompts allows the model to more efficiently learn OOD samples than the existing methods. Testing on nine different settings across three standard datasets used for OOD detection, our method with adaptive decision boundary is able to achieve competitive or superior performances compared with the current state-of-the-art in all cases. We also provide extensive analysis on each dataset as well as perform comprehensive ablation studies on each component of our model.
期刊介绍:
The IEEE Transactions on Emerging Topics in Computational Intelligence (TETCI) publishes original articles on emerging aspects of computational intelligence, including theory, applications, and surveys.
TETCI is an electronics only publication. TETCI publishes six issues per year.
Authors are encouraged to submit manuscripts in any emerging topic in computational intelligence, especially nature-inspired computing topics not covered by other IEEE Computational Intelligence Society journals. A few such illustrative examples are glial cell networks, computational neuroscience, Brain Computer Interface, ambient intelligence, non-fuzzy computing with words, artificial life, cultural learning, artificial endocrine networks, social reasoning, artificial hormone networks, computational intelligence for the IoT and Smart-X technologies.