{"title":"对时间序列数据中的空值进行全数据集上下文感知预测,以实现低复杂度的更快推理","authors":"Sharmen Akhter , Nosin Ibna Mahbub , Junyoung Park , Eui-Nam Huh","doi":"10.1016/j.ipm.2025.104370","DOIUrl":null,"url":null,"abstract":"<div><div>Null value handling in time series datasets demands heavy and explicit data imputation models, which add latency during inferences for various tasks. This concern raises the following question: <em>is it possible to avoid these explicit data imputation models to perform task predictions directly during inference without imputation?</em> As a pioneer, this paper proposes <strong>ZeroTIP</strong>, a knowledge distillation (KD)-based <strong>Zero T</strong>ime <strong>I</strong>mputation for <strong>P</strong>rediction strategy for workload predictions without having additional data-imputation models. During the training period, a student network is forced to reason the missing or null values implicitly and mimic the inference (workload prediction task) while taking synthetically corrupted data as input and being supervised by the pretrained teacher network that contains representations of the original dataset. Only the student network is used during inference. ZeroTIP reduced the inference time by almost 99.9% by avoiding explicit data imputation. A version of ZeroTIP, called ZeroTIP-DI, was deployed for the data imputation task to evaluate the significance of ZeroTIP in reasoning data context and pattern. For a prediction length of 48 and 96, ZeroTIP-DI achieved an average improvement of 38.37 (97.08%) and 21.67 (95.08%) times the baseline, highlighting its superiority.</div></div>","PeriodicalId":50365,"journal":{"name":"Information Processing & Management","volume":"63 2","pages":"Article 104370"},"PeriodicalIF":6.9000,"publicationDate":"2025-09-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Whole dataset context-aware prediction on the null values in time series data for faster inferencing with low complexity\",\"authors\":\"Sharmen Akhter , Nosin Ibna Mahbub , Junyoung Park , Eui-Nam Huh\",\"doi\":\"10.1016/j.ipm.2025.104370\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Null value handling in time series datasets demands heavy and explicit data imputation models, which add latency during inferences for various tasks. This concern raises the following question: <em>is it possible to avoid these explicit data imputation models to perform task predictions directly during inference without imputation?</em> As a pioneer, this paper proposes <strong>ZeroTIP</strong>, a knowledge distillation (KD)-based <strong>Zero T</strong>ime <strong>I</strong>mputation for <strong>P</strong>rediction strategy for workload predictions without having additional data-imputation models. During the training period, a student network is forced to reason the missing or null values implicitly and mimic the inference (workload prediction task) while taking synthetically corrupted data as input and being supervised by the pretrained teacher network that contains representations of the original dataset. Only the student network is used during inference. ZeroTIP reduced the inference time by almost 99.9% by avoiding explicit data imputation. A version of ZeroTIP, called ZeroTIP-DI, was deployed for the data imputation task to evaluate the significance of ZeroTIP in reasoning data context and pattern. For a prediction length of 48 and 96, ZeroTIP-DI achieved an average improvement of 38.37 (97.08%) and 21.67 (95.08%) times the baseline, highlighting its superiority.</div></div>\",\"PeriodicalId\":50365,\"journal\":{\"name\":\"Information Processing & Management\",\"volume\":\"63 2\",\"pages\":\"Article 104370\"},\"PeriodicalIF\":6.9000,\"publicationDate\":\"2025-09-09\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Information Processing & Management\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0306457325003115\",\"RegionNum\":1,\"RegionCategory\":\"管理学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, INFORMATION SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Processing & Management","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0306457325003115","RegionNum":1,"RegionCategory":"管理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
Whole dataset context-aware prediction on the null values in time series data for faster inferencing with low complexity
Null value handling in time series datasets demands heavy and explicit data imputation models, which add latency during inferences for various tasks. This concern raises the following question: is it possible to avoid these explicit data imputation models to perform task predictions directly during inference without imputation? As a pioneer, this paper proposes ZeroTIP, a knowledge distillation (KD)-based Zero Time Imputation for Prediction strategy for workload predictions without having additional data-imputation models. During the training period, a student network is forced to reason the missing or null values implicitly and mimic the inference (workload prediction task) while taking synthetically corrupted data as input and being supervised by the pretrained teacher network that contains representations of the original dataset. Only the student network is used during inference. ZeroTIP reduced the inference time by almost 99.9% by avoiding explicit data imputation. A version of ZeroTIP, called ZeroTIP-DI, was deployed for the data imputation task to evaluate the significance of ZeroTIP in reasoning data context and pattern. For a prediction length of 48 and 96, ZeroTIP-DI achieved an average improvement of 38.37 (97.08%) and 21.67 (95.08%) times the baseline, highlighting its superiority.
期刊介绍:
Information Processing and Management is dedicated to publishing cutting-edge original research at the convergence of computing and information science. Our scope encompasses theory, methods, and applications across various domains, including advertising, business, health, information science, information technology marketing, and social computing.
We aim to cater to the interests of both primary researchers and practitioners by offering an effective platform for the timely dissemination of advanced and topical issues in this interdisciplinary field. The journal places particular emphasis on original research articles, research survey articles, research method articles, and articles addressing critical applications of research. Join us in advancing knowledge and innovation at the intersection of computing and information science.