Whole dataset context-aware prediction on the null values in time series data for faster inferencing with low complexity

IF 6.9 1区 管理学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS
Sharmen Akhter , Nosin Ibna Mahbub , Junyoung Park , Eui-Nam Huh
{"title":"Whole dataset context-aware prediction on the null values in time series data for faster inferencing with low complexity","authors":"Sharmen Akhter ,&nbsp;Nosin Ibna Mahbub ,&nbsp;Junyoung Park ,&nbsp;Eui-Nam Huh","doi":"10.1016/j.ipm.2025.104370","DOIUrl":null,"url":null,"abstract":"<div><div>Null value handling in time series datasets demands heavy and explicit data imputation models, which add latency during inferences for various tasks. This concern raises the following question: <em>is it possible to avoid these explicit data imputation models to perform task predictions directly during inference without imputation?</em> As a pioneer, this paper proposes <strong>ZeroTIP</strong>, a knowledge distillation (KD)-based <strong>Zero T</strong>ime <strong>I</strong>mputation for <strong>P</strong>rediction strategy for workload predictions without having additional data-imputation models. During the training period, a student network is forced to reason the missing or null values implicitly and mimic the inference (workload prediction task) while taking synthetically corrupted data as input and being supervised by the pretrained teacher network that contains representations of the original dataset. Only the student network is used during inference. ZeroTIP reduced the inference time by almost 99.9% by avoiding explicit data imputation. A version of ZeroTIP, called ZeroTIP-DI, was deployed for the data imputation task to evaluate the significance of ZeroTIP in reasoning data context and pattern. For a prediction length of 48 and 96, ZeroTIP-DI achieved an average improvement of 38.37 (97.08%) and 21.67 (95.08%) times the baseline, highlighting its superiority.</div></div>","PeriodicalId":50365,"journal":{"name":"Information Processing & Management","volume":"63 2","pages":"Article 104370"},"PeriodicalIF":6.9000,"publicationDate":"2025-09-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Processing & Management","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0306457325003115","RegionNum":1,"RegionCategory":"管理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
引用次数: 0

Abstract

Null value handling in time series datasets demands heavy and explicit data imputation models, which add latency during inferences for various tasks. This concern raises the following question: is it possible to avoid these explicit data imputation models to perform task predictions directly during inference without imputation? As a pioneer, this paper proposes ZeroTIP, a knowledge distillation (KD)-based Zero Time Imputation for Prediction strategy for workload predictions without having additional data-imputation models. During the training period, a student network is forced to reason the missing or null values implicitly and mimic the inference (workload prediction task) while taking synthetically corrupted data as input and being supervised by the pretrained teacher network that contains representations of the original dataset. Only the student network is used during inference. ZeroTIP reduced the inference time by almost 99.9% by avoiding explicit data imputation. A version of ZeroTIP, called ZeroTIP-DI, was deployed for the data imputation task to evaluate the significance of ZeroTIP in reasoning data context and pattern. For a prediction length of 48 and 96, ZeroTIP-DI achieved an average improvement of 38.37 (97.08%) and 21.67 (95.08%) times the baseline, highlighting its superiority.

Abstract Image

对时间序列数据中的空值进行全数据集上下文感知预测,以实现低复杂度的更快推理
时间序列数据集中的空值处理需要大量显式的数据输入模型,这会在各种任务的推断期间增加延迟。这种担忧引发了以下问题:是否有可能避免这些显式数据输入模型在推理过程中直接执行任务预测而不输入?作为先驱,本文提出了ZeroTIP,这是一种基于知识蒸馏(KD)的预测零时间插入策略,用于工作负载预测,而无需额外的数据插入模型。在训练期间,学生网络被迫隐式地推断缺失值或空值并模仿推理(工作量预测任务),同时将综合损坏的数据作为输入,并由包含原始数据集表示的预训练教师网络进行监督。在推理过程中只使用学生网络。ZeroTIP通过避免显式的数据插入,将推理时间减少了近99.9%。ZeroTIP的一个版本称为ZeroTIP- di,用于数据输入任务,以评估ZeroTIP在推理数据上下文和模式方面的重要性。在预测长度为48和96的情况下,ZeroTIP-DI平均提高基线38.37倍(97.08%)和21.67倍(95.08%),突出了其优越性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Information Processing & Management
Information Processing & Management 工程技术-计算机:信息系统
CiteScore
17.00
自引率
11.60%
发文量
276
审稿时长
39 days
期刊介绍: Information Processing and Management is dedicated to publishing cutting-edge original research at the convergence of computing and information science. Our scope encompasses theory, methods, and applications across various domains, including advertising, business, health, information science, information technology marketing, and social computing. We aim to cater to the interests of both primary researchers and practitioners by offering an effective platform for the timely dissemination of advanced and topical issues in this interdisciplinary field. The journal places particular emphasis on original research articles, research survey articles, research method articles, and articles addressing critical applications of research. Join us in advancing knowledge and innovation at the intersection of computing and information science.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信