PFunk-H: approximate query processing using perceptual models

HILDA '16 Pub Date : 2016-06-26 DOI:10.1145/2939502.2939512
Daniel Alabi, Eugene Wu
{"title":"PFunk-H: approximate query processing using perceptual models","authors":"Daniel Alabi, Eugene Wu","doi":"10.1145/2939502.2939512","DOIUrl":null,"url":null,"abstract":"Interactive visualization tools (e.g., crossfilter) are critical to many data analysts by making the discovery and verification of hypotheses quick and seamless. Increasing data sizes has made the scalability of these tools a necessity. To bridge the gap between data sizes and interactivity, many visualization systems have turned to sampling-based approximate query processing frameworks. However, these systems are currently oblivious to human perceptual visual accuracy. This could either lead to overly aggressive sampling when the approximation accuracy is higher than needed or an incorrect visual rendering when the accuracy is too lax. Thus, for both correctness and efficiency, we propose to use empirical knowledge of human perceptual limitations to automatically bound the error of approximate answers meant for visualization.\n This paper explores a preliminary model of sampling-based approximate query processing that uses perceptual models (encoded as functions) to construct approximate answers intended for visualization. We present initial results that show that the approximate and non-approximate answers for a given query differ by a perceptually indiscernible amount, as defined by perceptual functions.","PeriodicalId":356971,"journal":{"name":"HILDA '16","volume":"56 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2016-06-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"30","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"HILDA '16","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/2939502.2939512","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 30

Abstract

Interactive visualization tools (e.g., crossfilter) are critical to many data analysts by making the discovery and verification of hypotheses quick and seamless. Increasing data sizes has made the scalability of these tools a necessity. To bridge the gap between data sizes and interactivity, many visualization systems have turned to sampling-based approximate query processing frameworks. However, these systems are currently oblivious to human perceptual visual accuracy. This could either lead to overly aggressive sampling when the approximation accuracy is higher than needed or an incorrect visual rendering when the accuracy is too lax. Thus, for both correctness and efficiency, we propose to use empirical knowledge of human perceptual limitations to automatically bound the error of approximate answers meant for visualization. This paper explores a preliminary model of sampling-based approximate query processing that uses perceptual models (encoded as functions) to construct approximate answers intended for visualization. We present initial results that show that the approximate and non-approximate answers for a given query differ by a perceptually indiscernible amount, as defined by perceptual functions.
PFunk-H:使用感知模型的近似查询处理
交互式可视化工具(例如交叉过滤器)对于许多数据分析师来说至关重要,因为它可以快速无缝地发现和验证假设。不断增长的数据量使得这些工具的可伸缩性成为必要。为了弥合数据大小和交互性之间的差距,许多可视化系统转向基于抽样的近似查询处理框架。然而,这些系统目前忽略了人类感知视觉的准确性。当近似精度高于所需时,这可能会导致过度激进的采样,或者当精度过于松散时导致不正确的视觉渲染。因此,为了正确性和效率,我们建议使用人类感知局限性的经验知识来自动约束用于可视化的近似答案的误差。本文探讨了一个基于抽样的近似查询处理的初步模型,该模型使用感知模型(编码为函数)来构建用于可视化的近似答案。我们提出的初步结果表明,给定查询的近似和非近似答案的差异在感知上是不可分辨的,由感知函数定义。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信