{"title":"Evacuee Flow Optimisation Using G-Network with Multiple Classes of Positive Customers","authors":"Huibo Bi","doi":"10.1109/MASCOTS.2016.62","DOIUrl":"https://doi.org/10.1109/MASCOTS.2016.62","url":null,"abstract":"Previous queueing theory based emergency navigation algorithms in built environments normally treat each significant location (such as doorways and staircases) as an \"independent\" queue and all the evacuees in a homogeneous manner. Hence, the interactions among linked queues caused by the re-routing instructions generated by the emergency navigation system, the panic behaviours such as evacuees not following the evacuation instructions, as well as the influence of diverse mobilities of evacuees are ignored. In this paper, we employ a Cognitive Packet Network based algorithm to customise distinct paths for diverse categories of evacuees. A G-network based model is used to analyse the latency on a path via capturing the dynamics of diverse categories of evacuees under the influence of panic and re-routing decisions from the navigation system. Moreover, by modelling the probabilistic choices of evacuees towards all the linked queues, the G-network model closely approximates the movement of the evacuees under the instructions of the Cognitive Packet Network based algorithm. The simulation results indicate that the use of the G-network model can improve the survival rates and ease the congestion during an evacuation process when there is a certain likelihood that evacuees do not follow evacuation instructions due to panic.","PeriodicalId":129389,"journal":{"name":"2016 IEEE 24th International Symposium on Modeling, Analysis and Simulation of Computer and Telecommunication Systems (MASCOTS)","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123786936","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Stefano Martina, Marco Paolieri, T. Papini, E. Vicario
{"title":"Performance Evaluation of Fischer's Protocol through Steady-State Analysis of Markov Regenerative Processes","authors":"Stefano Martina, Marco Paolieri, T. Papini, E. Vicario","doi":"10.1109/MASCOTS.2016.72","DOIUrl":"https://doi.org/10.1109/MASCOTS.2016.72","url":null,"abstract":"Fischer's protocol is a well-known timed mechanism through which a set of processes can synchronize access to a critical section without relying on atomic test-and-set operations, as might occur in a distributed environment or on a low-level computing platform. The protocol is based on a deterministic waiting time that can be defined so as to guarantee that possible interference due to concurrent accesses with random bounded delays be resolved with certainty. While protocol correctness descends from firm lower and upper bounds on waiting times and random delays, performance attained in synchronization also depends on continuous distributions of delays. Performance evaluation of a correct implementation thus requires the solution of a non-Markovian model whose underlying stochastic process falls in the class of Markov regenerative processes (MRPs) with multiple concurrent delays with non-exponential duration. Numerical solution of this class of models is to a large extent still an open problem. We provide a twofold contribution. We first introduce a novel method for the steady-state analysis of MRPs where regenerations are reached in a bounded number of discrete events, which enlarges the class amenable to numerical solution by allowing multiple concurrent timers with non-exponential distributions. The proposed technique is then applied to Fischer's protocol by characterizing the latency overhead due to synchronization, which comprises the first case where performance of the protocol is quantitatively assessed by jointly accounting for firm bounds and continuous distributions of delays.","PeriodicalId":129389,"journal":{"name":"2016 IEEE 24th International Symposium on Modeling, Analysis and Simulation of Computer and Telecommunication Systems (MASCOTS)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122352440","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Achieving Efficiency without Sacrificing Model Accuracy: Network Calculus on Compact Domains","authors":"Kai Lampka, Steffen Bondorf, J. Schmitt","doi":"10.1109/MASCOTS.2016.9","DOIUrl":"https://doi.org/10.1109/MASCOTS.2016.9","url":null,"abstract":"Messages traversing a network commonly experience waiting times due to sharing the forwarding resources. During those times, the crossed systems must provide sufficient buffer space for queueing messages. Network Calculus (NC) is a mathematical methodology for bounding flow delays and system buffer requirements. The accuracy of these performance bounds depends mainly on two factors: the principles manifesting in the NC flow equation and the functions describing the system. We focus on the latter aspect. Common implementations of NC overapproximate these functions in order to keep the analysis computationally feasible. However, overapproximation often results in a loss of accuracy of the performance bounds. In this paper, we make such compromising tradeoffs between model accuracy and computational effort obsolete. We limit the accurate system description to functions of a compact domain, such that the accuracy of the NC analysis is preserved. Tying the domain bound to the algebraic operators of NC instead of the operational semantics of components, allows us to directly apply our solution to algebraic NC analyses that implement principles such as pay burst only once and pay multiplexing only once.","PeriodicalId":129389,"journal":{"name":"2016 IEEE 24th International Symposium on Modeling, Analysis and Simulation of Computer and Telecommunication Systems (MASCOTS)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130513313","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"I/O Scheduling Schemes for Better I/O Proportionality on Flash-Based SSDs","authors":"Jaeho Kim, Eunjae Lee, S. Noh","doi":"10.1109/MASCOTS.2016.53","DOIUrl":"https://doi.org/10.1109/MASCOTS.2016.53","url":null,"abstract":"In cloud computing, multiple servers are consolidated into a physical machine in order to reduce the cost of deploying the servers. Guaranteeing the service level objective (SLO) of each server is one of the most important factors in a virtualization system. Particularly, isolating the I/O resources among VMs competing for a shared storage system is challenging. Recently, use of flash based Solid State Drives (SSDs) is being extended to enterprise systems. However, there are few studies for guaranteeing SLOs on such systems. In this paper, we empirically analyze the I/O behavior of a shared SSD. We show that performance SLOs of storage systems employing SSDs being shared by VMs or tasks are not satisfactory. We analyze and show that components of SSDs such as channels, DRAM buffer, and Native Command Queuing (NCQ) are the reasons behind this problem. Based on these analysis and observations, we propose two SSD-aware host level I/O schedulers that we call A+CFQ and H+BFQ, which are extensions of state-of-the-art I/O schedulers CFQ and BFQ, respectively. Through implementation and experiments on Linux, we show that the proposed I/O schedulers improve proportionality without sacrifice to performance.","PeriodicalId":129389,"journal":{"name":"2016 IEEE 24th International Symposium on Modeling, Analysis and Simulation of Computer and Telecommunication Systems (MASCOTS)","volume":"65 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121685954","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Autoscaling Effects in Speed Scaling Systems","authors":"B. Elahi, C. Williamson","doi":"10.1109/MASCOTS.2016.48","DOIUrl":"https://doi.org/10.1109/MASCOTS.2016.48","url":null,"abstract":"In this paper, we study the dynamics of coupled speed scaling systems, in which service rate is a function of system occupancy. We focus on both Processor Sharing (PS) and Shortest Remaining Processing Time (SRPT) as scheduling disciplines, and study their speed scaling dynamics under heavy load. Using a combination of Markov chain analysis and discrete-event simulation, we identify several important properties of speed scaling systems, which we call the autoscaling effect, the α effect, and the saturation effect. We also identify different overload regimes for PS and SRPT. In particular, SRPT exhibits a starvation effect that differs from the compensation effect of PS. These dynamics lead to different stability, fairness, and robustness properties for PS and SRPT under heavy load.","PeriodicalId":129389,"journal":{"name":"2016 IEEE 24th International Symposium on Modeling, Analysis and Simulation of Computer and Telecommunication Systems (MASCOTS)","volume":"138 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131132201","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Characterizing Performance and Power towards Efficient Synchronization of GPU Kernels","authors":"Islam Harb, W. Feng","doi":"10.1109/MASCOTS.2016.58","DOIUrl":"https://doi.org/10.1109/MASCOTS.2016.58","url":null,"abstract":"There is a lack of support for explicit synchronization in GPUs between the streaming multiprocessors (SMs) adversely impacts the performance of the GPUs to efficiently perform inter-block communication. In this paper, we present several approaches to inter-block synchronization using explicit/implicit CPU-based and dynamic parallelism (DP) mechanisms. Although this topic has been addressed in previous research studies, there has been neither a solid quantification of such overhead, nor guidance on when to use each of the different approaches. Therefore, we quantify the synchronization overhead relative to the number of kernel launches and the input data sizes. The quantification, in turn, provides insight as to when to use each of the aforementioned synchronization mechanisms in a target application. Our results show that implicit CPU synchronization has a significant overhead that hurts the application performance when using medium to large data sizes with relatively large number of kernel launches (i.e. ~1100-5000). Hence, it is recommended to use explicit CPU synchronization with these configurations. In addition, among the three different approaches, we conclude that dynamic parallelism (DP) is the most efficient with small data sizes (i.e., ~128k bytes), regardless of the number of kernel launches. Also, Dynamic Parallelism (DP), implicitly, performs inter-block (i.e. global) synchronization with no CPU intervention. Therefore, DP significantly reduces the power consumed by the CPU and PCIe for global synchronization. Our findings show that DP reduces the power consumption by ~8-10%. However, DP-based synchronization is a trade-off, in which it is accompanied by ~2-5% performance loss.","PeriodicalId":129389,"journal":{"name":"2016 IEEE 24th International Symposium on Modeling, Analysis and Simulation of Computer and Telecommunication Systems (MASCOTS)","volume":"61 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130837223","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Prospects for Shaping User-Centric Mobile Application Workloads to Benefit the Cloud","authors":"Maciej Swiech, Huaqian Cai, P. Dinda, Gang Huang","doi":"10.1109/MASCOTS.2016.68","DOIUrl":"https://doi.org/10.1109/MASCOTS.2016.68","url":null,"abstract":"Approaches to making cloud operation more efficient, for example through scheduling and power management, largely assume that the workload offered from mobile, user-facing applications is a given and that the cloud must simply adapt to it. We flip this assumption 180 degrees and ask to what extent can we instead shape the user-centric workload into a form that would benefit such approaches. Using a toolchain hat allows us to interpose on frontend/backend interactions in popular Android applications, we add the ability to introduce delays and collect information about user satisfaction. We conduct an \"in the wild\" user study using this capability, and report on its results. Delays of up to 750 ms can be introduced with little effect on most users, although this is very much user and application dependent. Finally, given our study results, we consider reshaping the application requests by selective delays to have exponential interarrival times (Poisson arrivals), and find that we are often able to do so without exceeding the user's delay tolerance.","PeriodicalId":129389,"journal":{"name":"2016 IEEE 24th International Symposium on Modeling, Analysis and Simulation of Computer and Telecommunication Systems (MASCOTS)","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115537728","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Vertical Integration of CDN and Network Operator: Model and Analysis","authors":"P. Maillé, G. Simon, B. Tuffin","doi":"10.1109/MASCOTS.2016.24","DOIUrl":"https://doi.org/10.1109/MASCOTS.2016.24","url":null,"abstract":"Content Delivery Network (CDN) services largely affect the delivery quality perceived by users. While those services were initially offered by independent entities, some large ISP now develop their own CDN activities to control costs and delivery quality. But this new activity is also a new source of revenues for those vertically integrated ISP-CDNs, which can sell those services to content providers. In this paper, we investigate the impact of having an ISP and a vertically-integrated CDN, on the main actors of the ecosystem (users, competing ISPs). Our approach is based on an economic model of revenues and costs, and a multilevel game-theoretic formulation of the interactions among actors. Our model incorporates the possibility for the vertically-integrated ISP to partially offer CDN services to competitors in order to optimize the trade-off between CDN revenue (if fully offered) and competitive advantage on subscriptions at the ISP level (if not offered to competitors). Our results highlight two counterintuitive phenomena: an ISP may prefer an independent CDN over controlling (integrating) a CDN, and from the user point of view vertical integration is preferable to an independent CDN or a no-CDN configuration. Hence, a regulator may want to elicit such CDN-ISP vertical integrations rather than prevent them.","PeriodicalId":129389,"journal":{"name":"2016 IEEE 24th International Symposium on Modeling, Analysis and Simulation of Computer and Telecommunication Systems (MASCOTS)","volume":"57 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125116549","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
M. HoseinyFarahabady, Young Choon Lee, Albert Y. Zomaya, Z. Tari, A. Song
{"title":"A Model Predictive Controller for Contention-Aware Resource Allocation in Virtualized Data Centers","authors":"M. HoseinyFarahabady, Young Choon Lee, Albert Y. Zomaya, Z. Tari, A. Song","doi":"10.1109/MASCOTS.2016.59","DOIUrl":"https://doi.org/10.1109/MASCOTS.2016.59","url":null,"abstract":"Data center efficiency is primarily sought by sharing physical resources, such as processors, memory, and disks in the form of virtual machines or containers among multiple users, i.e., workload consolidation. However, the reality is co-located applications in these virtual platforms compete for resources and interfere with each others' performance, resulting in performance variability/degradation. In this paper, we present the contentionaware resource allocation (CARA) solution, which optimizes data center efficiency. It is essentially devised based on a model predictive control that enables to make judicious consolidation decisions with future system states. CARA consolidates workloads explicitly taking into account the correlation between shared and isolated resource usage patterns. Based on our experimental results, CARA improves the overall resource utilization by 32%, without a significant impact on the quality-of-service (QoS) enforcement level. Such improvement results in a fewer number of active servers and in turn contributes to an overall energy saving by 33%.","PeriodicalId":129389,"journal":{"name":"2016 IEEE 24th International Symposium on Modeling, Analysis and Simulation of Computer and Telecommunication Systems (MASCOTS)","volume":"931 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132284884","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Exploiting Replication for Energy Efficiency of Heterogeneous Storage Systems","authors":"Everett Neil Rush, Nihat Altiparmak","doi":"10.1109/MASCOTS.2016.70","DOIUrl":"https://doi.org/10.1109/MASCOTS.2016.70","url":null,"abstract":"As a result of immense growth of digital data in the last decade, energy consumption has become an important issue in data storage systems. In the US alone, data centers were projected to consume $4 billion (40 TWh) yearly electricity in 2005. This cost had reached to $10 billion (100 TWh) in 2011, and expected to be around $20 billion (200 TWh) in 2016 by doubling itself every 5 years. In addition to the economic burden on companies and research institutions, these large scale data storage systems also have a negative impact on the environment. According to the EPA, generating 1 KWh of electricity in the US results in an average of 1.55 pounds of carbon dioxide emissions. Considering a projected 200 TWh energy requirement for 2016, energy-efficient data storage systems can have a huge economic and environmental impacts on society. This project exploits replication and heterogeneity existing in modern multi-disk storage systems and proposes an energy-efficient and performance-aware replica selection technique to reduce the energy consumption of data storage systems without negatively affecting their performance. Our proposed technique exploits the difference between active and idle energy consumption in heterogeneous disks holding the same replica and selects replicas by balancing energy and performance.","PeriodicalId":129389,"journal":{"name":"2016 IEEE 24th International Symposium on Modeling, Analysis and Simulation of Computer and Telecommunication Systems (MASCOTS)","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133767180","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}