{"title":"Statistical projections for multi-dimensional visual data exploration","authors":"H. Nguyen, D. Stone, E. W. Bethel","doi":"10.1109/LDAV.2016.7874338","DOIUrl":"https://doi.org/10.1109/LDAV.2016.7874338","url":null,"abstract":"When working with large, multidimensional and multivariate data, science users are frequently interested in understanding variation in data, as opposed to the actual data values. Our work focuses on exploring how a simple statistical metric, the Coefficient of Variation (or Cv), can be used in several different ways to facilitate understanding variation in large data. As a statistical measure, it offers a key advantage over more widely accepted measures like standard deviation, namely to its ability to capture local variation properties. As a multidimensional projection operator, Cv is an effective way of reducing data size while preserving the key variational signal. Visualizations produced from Cv that target conveying variation in data are highly informative, especially compared to those produced with more widely known methods. We demonstrate these ideas within the context of a two-part application case study focusing on understanding long-term trends in the the changes in precipitation and winds in large-scale climate model ensemble output.","PeriodicalId":148570,"journal":{"name":"2016 IEEE 6th Symposium on Large Data Analysis and Visualization (LDAV)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114960313","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Contour forests: Fast multi-threaded augmented contour trees","authors":"Charles Gueunet, P. Fortin, J. Jomier","doi":"10.1109/LDAV.2016.7874333","DOIUrl":"https://doi.org/10.1109/LDAV.2016.7874333","url":null,"abstract":"This paper presents a new algorithm for the fast, shared memory multi-threaded computation of contour trees on tetrahedral meshes. In contrast to previous multi-threaded algorithms, our technique computes the augmented contour tree. Such an augmentation is required to enable the full extent of contour tree based applications, including for instance data segmentation. Our approach relies on a range-driven domain partitioning. We show how to exploit such a partitioning to rapidly compute contour forests. We also show how such forests can be efficiently turned into the output contour tree. We report performance numbers that compare our approach to a reference sequential implementation for the computation of augmented contour trees. These experiments demonstrate the run-time efficiency of our approach. We demonstrate the utility of our approach with several data segmentation tasks. We also provide a lightweight VTK-based C++ implementation of our approach for reproduction purposes.","PeriodicalId":148570,"journal":{"name":"2016 IEEE 6th Symposium on Large Data Analysis and Visualization (LDAV)","volume":"103 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123494683","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A study of scientific visualization on heterogeneous processors using Legion","authors":"Lina Yu, Hongfeng Yu","doi":"10.1109/LDAV.2016.7874341","DOIUrl":"https://doi.org/10.1109/LDAV.2016.7874341","url":null,"abstract":"We present a study of scientific visualization on heterogeneous processors using the Legion runtime system. We describe the main functions in our approach to conduct scientific visualization that can consist of multiple operations with different data requirements. Our approach can help users simplify programming on the data partition, data organization and data movement for distributed-memory heterogeneous architectures, thereby facilitating a simultaneous execution of multiple operations on modern and future supercomputers. We demonstrate the scalable performance and the easy usage of our approach by a hybrid data partitioning and distribution scheme for different data types using both CPUs and GPUs on a heterogeneous system.","PeriodicalId":148570,"journal":{"name":"2016 IEEE 6th Symposium on Large Data Analysis and Visualization (LDAV)","volume":"127 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134633756","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Tom Vierjahn, Marc-André Hermanns, B. Mohr, Matthias S. Müller, T. Kuhlen, B. Hentschel
{"title":"Correlating sub-phenomena in performance data in the frequency domain","authors":"Tom Vierjahn, Marc-André Hermanns, B. Mohr, Matthias S. Müller, T. Kuhlen, B. Hentschel","doi":"10.1109/LDAV.2016.7874340","DOIUrl":"https://doi.org/10.1109/LDAV.2016.7874340","url":null,"abstract":"Finding and understanding correlated performance behaviour of the individual functions of massively parallel high-performance computing (HPC) applications is a time-consuming task. In this poster, we propose filtered correlation analysis for automatically locating interdependencies in call-path performance profiles. Transforming the data into the frequency domain splits a performance phenomenon into sub-phenomena to be correlated separately. We provide the mathematical framework and an overview over the visualization, and we demonstrate the effectiveness of our technique.","PeriodicalId":148570,"journal":{"name":"2016 IEEE 6th Symposium on Large Data Analysis and Visualization (LDAV)","volume":"871 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121513841","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Karsten Schatz, C. Müller, M. Krone, J. Schneider, G. Reina, T. Ertl
{"title":"Interactive visual exploration of a trillion particles","authors":"Karsten Schatz, C. Müller, M. Krone, J. Schneider, G. Reina, T. Ertl","doi":"10.1109/LDAV.2016.7874310","DOIUrl":"https://doi.org/10.1109/LDAV.2016.7874310","url":null,"abstract":"We present a method for the interactive exploration of tera-scale particle data sets. Such data sets arise from molecular dynamics, particle-based fluid simulation, and astrophysics. Our visualization technique provides a focus+context view of the data that runs interactively on commodity hardware. The method is based on a hybrid multi-scale rendering architecture, which renders the context as a hierarchical density volume. Fine details in the focus are visualized using direct particle rendering. In addition, clusters like dark matter halos can be visualized as semi-transparent spheres enclosing the particles. Since the detail data is too large to be stored in main memory, our approach uses an out-of-core technique that streams data on demand. Our technique is designed to take advantage of a dual-GPU configuration, in which the workload is split between the GPUs based on the type of data. Structural features in the data are visually enhanced using advanced rendering and shading techniques. To allow users to easily identify interesting locations even in overviews, both the focus and context view use color tables to show data attributes on the respective scale. We demonstrate that our technique achieves interactive performance on a one trillionpar-ticle data set from the DarkSky simulation.","PeriodicalId":148570,"journal":{"name":"2016 IEEE 6th Symposium on Large Data Analysis and Visualization (LDAV)","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134439336","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
H. Carr, G. Weber, Christopher M. Sewell, J. Ahrens
{"title":"Parallel peak pruning for scalable SMP contour tree computation","authors":"H. Carr, G. Weber, Christopher M. Sewell, J. Ahrens","doi":"10.1109/LDAV.2016.7874312","DOIUrl":"https://doi.org/10.1109/LDAV.2016.7874312","url":null,"abstract":"As data sets grow to exascale, automated data analysis and visualisation are increasingly important, to intermediate human understanding and to reduce demands on disk storage via in situ analysis. Trends in architecture of high performance computing systems necessitate analysis algorithms to make effective use of combinations of massively multicore and distributed systems. One of the principal analytic tools is the contour tree, which analyses relationships between contours to identify features of more than local importance. Unfortunately, the predominant algorithms for computing the contour tree are explicitly serial, and founded on serial metaphors, which has limited the scalability of this form of analysis. While there is some work on distributed contour tree computation, and separately on hybrid GPU-CPU computation, there is no efficient algorithm with strong formal guarantees on performance allied with fast practical performance. We report the first shared SMP algorithm for fully parallel contour tree computation, withfor-mal guarantees of O(lgnlgt) parallel steps and O(n lgn) work, and implementations with up to 10x parallel speed up in OpenMP and up to 50x speed up in NVIDIA Thrust.","PeriodicalId":148570,"journal":{"name":"2016 IEEE 6th Symposium on Large Data Analysis and Visualization (LDAV)","volume":"5 3","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114059161","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Min Shih, S. Rizzi, J. Insley, T. Uram, V. Vishwanath, M. Hereld, M. Papka, K. Ma
{"title":"Parallel distributed, GPU-accelerated, advanced lighting calculations for large-scale volume visualization","authors":"Min Shih, S. Rizzi, J. Insley, T. Uram, V. Vishwanath, M. Hereld, M. Papka, K. Ma","doi":"10.1109/LDAV.2016.7874309","DOIUrl":"https://doi.org/10.1109/LDAV.2016.7874309","url":null,"abstract":"The benefits of applying advanced illumination models to volume visualization have been demonstrated by many researchers. For a parallel distributed, GPU computing environment, however, there is no efficient algorithm for scalable global illumination calculations. This paper presents a parallel, data-distributed and GPU-accelerated algorithm for volume rendering with advanced lighting. Our approach features tunable soft shadows for enhancing perception of complex spatial structures and relationships. For lighting calculations, our design effectively avoids data exchange among GPUs. Performance evaluation on a GPU cluster using up to 128 GPUs shows scalable rendering performance, with both the number of GPUs and volume data size.","PeriodicalId":148570,"journal":{"name":"2016 IEEE 6th Symposium on Large Data Analysis and Visualization (LDAV)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128786341","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Yucong Ye, Tyson Neuroth, F. Sauer, K. Ma, G. Borghesi, Aditya Konduri, H. Kolla, Jacqueline H. Chen
{"title":"In situ generated probability distribution functions for interactive post hoc visualization and analysis","authors":"Yucong Ye, Tyson Neuroth, F. Sauer, K. Ma, G. Borghesi, Aditya Konduri, H. Kolla, Jacqueline H. Chen","doi":"10.1109/LDAV.2016.7874311","DOIUrl":"https://doi.org/10.1109/LDAV.2016.7874311","url":null,"abstract":"The growing power and capacity of supercomputers enable scientific simulations at extreme scale, leading to not only more accurate modeling and greater predictive ability but also massive quantities of data to analyze. New approaches to data analysis and visualization are this needed to support interactive exploration through selective data access for gaining insights into terabytes and petabytes of data. In this paper, we present an in situ data processing method for both generating probability distribution functions (PDFs) from field data and reorganizing particle data using a single spatial organization scheme. This coupling between PDFs and particles allows for the interactive post hoc exploration of both data types simultaneously. Scientists can explore trends in large-scale data through the PDFs and subsequently extract desired particle subsets for further analysis. We evaluate the usability of our in situ method using a petascale combustion simulation and demonstrate the increases in task efficiency and accuracy that the resulting workflow provides to scientists.","PeriodicalId":148570,"journal":{"name":"2016 IEEE 6th Symposium on Large Data Analysis and Visualization (LDAV)","volume":"84 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116075024","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
William Halsey, C. Steed, R. Dehoff, V. Paquit, Sean L. Yoder
{"title":"Segmented time series visualization tool for additive manufacturing","authors":"William Halsey, C. Steed, R. Dehoff, V. Paquit, Sean L. Yoder","doi":"10.1109/LDAV.2016.7874336","DOIUrl":"https://doi.org/10.1109/LDAV.2016.7874336","url":null,"abstract":"Additive manufacturing promises to deliver the ability to build complex shapes and parts while using raw materials more efficiently than traditional manufacturing approaches. However, material scientists are continually striving to understand how complex build parameters affect the 3D printing process and the quality of the final product. Understanding the intricate relationships between parameters and final product will yield the opportunity for automatic tuning of variables to ensure consistency of quality across build iterations.","PeriodicalId":148570,"journal":{"name":"2016 IEEE 6th Symposium on Large Data Analysis and Visualization (LDAV)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128173706","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Josua Krause, Aritra Dasgupta, Jean-Daniel Fekete, E. Bertini
{"title":"SeekAView: An intelligent dimensionality reduction strategy for navigating high-dimensional data spaces","authors":"Josua Krause, Aritra Dasgupta, Jean-Daniel Fekete, E. Bertini","doi":"10.1109/LDAV.2016.7874305","DOIUrl":"https://doi.org/10.1109/LDAV.2016.7874305","url":null,"abstract":"Dealing with the curse of dimensionality is a key challenge in high-dimensional data visualization. We present SeekAView to address three main gaps in the existing research literature. First, automated methods like dimensionality reduction or clustering suffer from a lack of transparency in letting analysts interact with their outputs in real-time to suit their exploration strategies. The results often suffer from a lack of interpretability, especially for domain experts not trained in statistics and machine learning. Second, exploratory visualization techniques like scatter plots or parallel coordinates suffer from a lack of visual scalability: it is difficult to present a coherent overview of interesting combinations of dimensions. Third, the existing techniques do not provide a flexible workflow that allows for multiple perspectives into the analysis process by automatically detecting and suggesting potentially interesting subspaces. In SeekAView we address these issues using suggestion based visual exploration of interesting patterns for building and refining multidimensional subspaces. Compared to the state-of-the-art in subspace search and visualization methods, we achieve higher transparency in showing not only the results of the algorithms, but also interesting dimensions calibrated against different metrics. We integrate a visually scalable design space with an iterative workflow guiding the analysts by choosing the starting points and letting them slice and dice through the data to find interesting subspaces and detect correlations, clusters, and outliers. We present two usage scenarios for demonstrating how SeekAView can be applied in real-world data analysis scenarios.","PeriodicalId":148570,"journal":{"name":"2016 IEEE 6th Symposium on Large Data Analysis and Visualization (LDAV)","volume":"62 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131219847","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}