{"title":"Accelerating Accuracy Improvement for Floating Point Programs via Memory Based Pruning","authors":"Anxiang Xiao, Enyi Tang, Xin Chen, Linzhang Wang","doi":"10.1145/3457913.3457918","DOIUrl":"https://doi.org/10.1145/3457913.3457918","url":null,"abstract":"Floating point programs are widely used in science and engineering. These programs approximate real arithmetic with floating point arithmetic, which introduces rounding errors and leads to inaccurate results. In life-critical software systems, inaccuracy can lead to catastrophes. To avoid these tragedies, program rewriting techniques have been proposed to automatically improve accuracy for floating point programs. Unfortunately, efficiency can be the bottleneck of such techniques due to the lack of effective pruning strategy. We introduce a heuristic pruning strategy called Memory Based Pruning to accelerate these rewrite-based approaches. The heuristic comes from an observation that floating point programs with similar computation steps usually suffer from similar inaccuracy problems, therefore they can be improved by the same or similar rewriting rules. With Memory Based Pruning, rewriting rules used in accuracy improvement are memorized in a knowledge memory. While improving accuracy for similar programs, these rules can help to guide the search and prune away probably useless rewriting rules. Consequently, a large amount of rule examinations can be avoided and the improving efficiency is significantly promoted. Based on the above approach, we have implemented a tool called HMBP and evaluated it on FPBench. In the evaluation, we compare the efficiency of HMBP and the well-known accuracy-improving tool Herbie. The results show that HMBP is 2.32x faster than Herbie in average. For the best case, the speedup is up to 29.64x.","PeriodicalId":194449,"journal":{"name":"Proceedings of the 12th Asia-Pacific Symposium on Internetware","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128469005","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Overwhelming Uncertainty in Self-adaptation: An Empirical Study on PLA and CobRA","authors":"Jingxin Fan, Yanxiang Tong, Yi Qin, Xiaoxing Ma","doi":"10.1145/3457913.3457943","DOIUrl":"https://doi.org/10.1145/3457913.3457943","url":null,"abstract":"Self-adaptation is a promising approach to enable software systems to address the challenge of uncertainty. Different from traditional reactive adaptation mechanisms that focus on the system’s current environment state only, proactive adaptation mechanisms predict the potential environmental changes and make better adaptation plan accordingly. Proactive Latency-aware Adaptation (PLA for shot) and Control-based Requirements-oriented Adaptation (CobRA for short) are two representative approaches to build proactive self-adaptation mechanisms. Despite their different design and implementation details, PLA and CobRA are reported to have a very similar performance in supporting self-adaptation. In this paper, we conduct an in-depth comparison between these two approaches, trying to explain their effectiveness. We separate a proactive self-adaptation mechanism into three modules, namely system modelling, environment predicting, and uncertainty filtering. We identify the design choices of PLA and CobRA approaches, in terms of these three modules. We performed an ablation study on the three modules of PLA and compared their performance with CobRA. Our study reveals the very important role of uncertainty filtering in supporting self-adaptation, as well as the huge impact of a fluctuant environment on a self-adaptation mechanism. Based on this observation, we briefly discuss a conceptual self-adaptation mechanism, MAPE-U (monitoring, analyzing, planning, executing with uncertainty).","PeriodicalId":194449,"journal":{"name":"Proceedings of the 12th Asia-Pacific Symposium on Internetware","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115356555","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Best Answerers Prediction With Topic Based GAT In Q&A Sites","authors":"Yuexin Huang, Hailong Sun","doi":"10.1145/3457913.3457935","DOIUrl":"https://doi.org/10.1145/3457913.3457935","url":null,"abstract":"Q&A communities are playing an important role in online knowledge sharing, where a large number of users with various knowledge background make tremendous contributions to solving many technical problems based on crowd intelligence. However, as new questions are increasingly posted, it is a non-trivial issue to find a matching answerer for each question. As a result, many questions fail to receive satisfying answers in time. This paper addresses the problem by predicting the best answerer for the new question. Many existing efforts are devoted to predicting the best answerer mainly by calculating the textual similarity between questions and a user’s historical post documents. Some works consider other features, such as the similarity of tags between questions and users, the average quality of a user’s historical answers, and so on. But few works consider interaction within the community. In recent years, works that take account of the interaction between community items (such as GCN and GAT) have made considerable progress in graph mining tasks like item recommendation, node representation, node classification, and link prediction. This kind of graph mining method can easily leverage interactive information in the community and encode it in an easy-to-use way which is very helpful for downstream tasks such as recommendation. However, questions that need to be recommended to answerers are new coming ones and with no interaction with any other node in the community yet. How to make reasonable use of collaborative information to improve recommendation performance is a real challenge. In this paper, we use the interactive information between candidate answerers and combine text information to make our best answerer recommendation. There are two main parts, LDA(Latent Dirichlet Allocation) topic model is used to capture the text information and graph attention networks (GATs) for interaction. We evaluated our approach on a real dataset from Stack Exchange. The result shows that our approach outperforms all the baseline methods.","PeriodicalId":194449,"journal":{"name":"Proceedings of the 12th Asia-Pacific Symposium on Internetware","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128538356","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A Feature Table approach to decomposing monolithic applications into microservices","authors":"Yuyang Wei, Yijun Yu, Minxue Pan, Tian Zhang","doi":"10.1145/3457913.3457939","DOIUrl":"https://doi.org/10.1145/3457913.3457939","url":null,"abstract":"Microservice architecture refers to the use of numerous small-scale and independently deployed services, instead of encapsulating all functions into one monolith. It has been a challenge in software engineering to decompose a monolithic system into smaller parts. In this paper, we propose the Feature Table approach, a structured approach to service decomposition based on the correlation between functional features and microservices: (1) we defined the concept of Feature Cards and 12 instances of such cards; (2) we formulated Decomposition Rules to decompose monolithic applications; (3) we designed the Feature Table Analysis Tool to provide semi-automatic analysis for identification of microservices; and (4) we formulated Mapping Rules to help developers implement microservice candidates. We performed a case study on Cargo Tracking System to validate our microservice-oriented decomposition approach. Cargo Tracking System is a typical case that has been decomposed by other related methods (dataflow-driven approach, Service Cutter, and API Analysis). Through comparison with the related methods in terms of specific coupling and cohesion metrics, the results show that the proposed Feature Table approach can deliver more reasonable microservice candidates, which are feasible in implementation with semi-automatic support.","PeriodicalId":194449,"journal":{"name":"Proceedings of the 12th Asia-Pacific Symposium on Internetware","volume":"119 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116387595","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Yinyuan Zhang, Yang Zhang, Yiwen Wu, Yao Lu, Tao Wang, Xinjun Mao
{"title":"Exploring the Dependency Network of Docker Containers: Structure, Diversity, and Relationship","authors":"Yinyuan Zhang, Yang Zhang, Yiwen Wu, Yao Lu, Tao Wang, Xinjun Mao","doi":"10.1145/3457913.3457927","DOIUrl":"https://doi.org/10.1145/3457913.3457927","url":null,"abstract":"Container technologies are being widely used in large scale production cloud environments, of which Docker has become the de-facto industry standard. As a key step, containers need to define their dependent base image, which makes complex dependencies exist in a large number of containers. Prior studies have shown that references between software packages could form technical dependencies, thus forming a dependency network. However, little is known about the details of docker container dependency networks. In this paper, we perform an empirical study on the dependency network of docker containers from more than 120,000 dockerfiles. We construct the container dependency network and analyze its network structure. Further, we focus on the Top-100 dominant containers and investigate their subnetworks, including diversity and relationships. Our findings help to characterize and understand the container dependencies in the docker community and motivate the need for developing container dependency management tools.","PeriodicalId":194449,"journal":{"name":"Proceedings of the 12th Asia-Pacific Symposium on Internetware","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126791338","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Enhancing FSCS-ART through Test Input Quantization and Inverted Lists","authors":"Muhammad Ashfaq, Rubing Huang, Michael Omari","doi":"10.1145/3457913.3457916","DOIUrl":"https://doi.org/10.1145/3457913.3457916","url":null,"abstract":"Fixed-size-candidate-set adaptive random testing (FSCS-ART) is an ART technique well-known for its best failure-detection effectiveness and usages in testing many real-life applications. However, it faces substantial computational overhead in terms of O(n2) time cost for generating n test inputs, which becomes worse for high dimensional input domains (number of inputs a software takes). As real-life programs generally have low failure-rates and have high dimensional input domains, it is vital to reduce the computational overhead while preserving the failure-detection effectiveness for efficient software testing. In this work, we adopted Quantization and InVerted File structure approach to enhance the original FSCS-ART, called QIVFSCS-ART. The proposed method preprocesses the software input domain by partitioning it into discrete cells by using K-means clustering using a uniform random dataset. After this, the quantized form of each executed test input is stored in the inverted list of its cell’s center, called centroid. Results show that the proposed method significantly relieves the computational overhead of FSCS-ART while preserving its failure-detection effectiveness, especially for the high-dimensional software input domains.","PeriodicalId":194449,"journal":{"name":"Proceedings of the 12th Asia-Pacific Symposium on Internetware","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128424768","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"SEDL: A Symmetric Encryption Method Based on Deep Learning","authors":"Xiang Li, Peng Wang","doi":"10.1145/3457913.3457921","DOIUrl":"https://doi.org/10.1145/3457913.3457921","url":null,"abstract":"Recent years have seen an increasing emphasis on information security, and various encryption methods have been proposed. However, for symmetric encryption methods, the well-known encryption techniques still rely on the key space to guarantee security and suffer from frequent key updating. Aiming to solve those problems, this paper proposes a novel general symmetry-key encryption method based on deep learning called SEDL, where the secret key includes hyperparameters in deep learning model and the core step of encryption is processing input data with weights trained under hyperparameters. Firstly, both communication parties establish a weight vector table by training a deep learning model on the constructed synthetic training sets according to specified hyperparameters. Then, a self-update codebook is constructed on the weight vector table with the SHA-256 function and other tricks. When communication starts, encryption and decryption are equivalent to indexing the corresponding value on the codebook to obtain ciphertext or plaintext, respectively. Results of experiments and relevant analyses show that SEDL performs well for security, efficiency, generality, and has a lower demand for the frequency of key redistribution. Especially, as a supplement to current encryption methods, the time-consuming process of constructing a codebook increases the difficulty of brute-force attacks, meanwhile, it does not degrade the efficiency of communications.","PeriodicalId":194449,"journal":{"name":"Proceedings of the 12th Asia-Pacific Symposium on Internetware","volume":"167 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122869619","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Tianyi Wu, Liwei Shen, Xin Peng, Biao Shen, Zhengjie Li
{"title":"Group Activity Matching with Blockchain Backed Credible Commitment","authors":"Tianyi Wu, Liwei Shen, Xin Peng, Biao Shen, Zhengjie Li","doi":"10.1145/3457913.3457923","DOIUrl":"https://doi.org/10.1145/3457913.3457923","url":null,"abstract":"Humans are social creatures which enjoy participating in group activities. Existing platforms such as event-based social networks and social-matching applications empower people to organize and participate in different kinds of interest-based activities. However, credibility issues are inevitable since the participants’ commitment to participate in activities on time can hardly be guaranteed. As a result, many activities are canceled due to lack of participation, which impairs people’s will for attending activities and increases the difficulty of coalescing activity groups. In this work, we propose an approach, called BC-GAM, for matching group activities with blockchain backed credible commitment. Based on a formalization of the matching problem, BC-GAM works as follows. During user requesting (for certain type of activity), the user is required to pay a variable deposit which reflects his/her level of commitment for participating in the activity. Our matching algorithm then automatically coalesces the users according to the user requests and availability of facilities. Our algorithm is designed to maximally allow the enrolled users to participate in activities based on their commitment. Furthermore, BC-GAM utilizes blockchain techniques and smart contracts so that the user requesting, commitment and participation are automatically executed in a distributed and trusted way. We implemented BC-GAM on a hyperledger and developed a user interface for requesting and inquiry of the activities. Based on the blockchain platform, we performed experiments with not only simulated data but also actual user studies. The experiment results show that the matching algorithm is effective and efficient, and BC-GAM can be potentially applied in practice.","PeriodicalId":194449,"journal":{"name":"Proceedings of the 12th Asia-Pacific Symposium on Internetware","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128647633","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Scheduling Distributed Deep Learning Jobs in Heterogeneous Cluster with Placement Awareness","authors":"Qingping Li, Jingwei Xu, Chun Cao","doi":"10.1145/3457913.3457936","DOIUrl":"https://doi.org/10.1145/3457913.3457936","url":null,"abstract":"Deep Neural Network models are integrated as parts of many real-world software applications. Due to the huge model size and complex computation, distributed deep learning (DDL) framework aims to provide a high-quality cluster scheduler to manage DDL training jobs from both resource allocation and job scheduling. However, existing schedulers either allocate a fixed amount of resources, or lack the control over task placement, which lead less efficient training. In this paper, we propose DeepSys, a GPU cluster scheduler tailored for DDL jobs. For single model, DeepSys builds a speed model to predict accurate training speed, and a memory model for high-quality resource utilization. For job scheduling, DeepSys considers resource allocation and task placement to provide efficient job scheduling in cluster. Experiments implemented on Kubernetes in two clusters show the advantage to the compared methods by 20% - 25% and 10% - 15% on average job completion time and makespan, respectively.","PeriodicalId":194449,"journal":{"name":"Proceedings of the 12th Asia-Pacific Symposium on Internetware","volume":"51 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114243843","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Huaiwei Yang, Shuang Liu, Lin Gui, Yongxin Zhao, Jun Sun, Junjie Chen
{"title":"What Makes Open Source Software Projects Impactful: A Data-Driven Approach","authors":"Huaiwei Yang, Shuang Liu, Lin Gui, Yongxin Zhao, Jun Sun, Junjie Chen","doi":"10.1145/3457913.3457932","DOIUrl":"https://doi.org/10.1145/3457913.3457932","url":null,"abstract":"With the wide adoption and acceptance of open source version control and hosting systems, more and more companies, including Google, Microsoft, Apple and Facebook, are putting their projects on such platforms, e.g., GitHub. It is very important for open source projects to be impactful, i.e., to attract attentions from the open source development community, so as to gain support on development, testing as well as maintenance from the community. However, the question of what factors affect open source project impact, remains largely open. Given the numerous confounding factors and the complex correlations among the factors, it is a challenge to answer the question. In this study, we gather a large dataset from GitHub and provide empirical insights on this question base on a data-driven approach. We randomly collect 146,286 projects from GitHub and then adopt data analysis techniques to automatically analyze the correlations of different features with the software project impact. We also provide suggestions on how to potentially make open source projects impactful base on our analysis results.","PeriodicalId":194449,"journal":{"name":"Proceedings of the 12th Asia-Pacific Symposium on Internetware","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115821301","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}