ACM Transactions on

Modeling and Performance Evaluation of Computing Systems (TOMPECS)

Latest Articles

Efficiency and Optimality of Largest Deficit First Prioritization: Dynamic User Prioritization for Soft Real-Time Applications

An increasing number of real-time applications with compute and/or communication deadlines are being supported on a shared infrastructure. Such... (more)

CloudHeat: An Efficient Online Market Mechanism for Datacenter Heat Harvesting

Datacenters are major energy consumers and dissipate an enormous amount of waste heat. Simple outdoor discharging of datacenter heat is energy-consuming and environmentally unfriendly. By harvesting datacenter waste heat and selling to the district heating system (DHS), both energy cost compensation and environment protection can be achieved. To... (more)

GPSonflow: Geographic Positioning of Storage for Optimal Nice Flow

This article evaluates the maximum data flow from a sender to a receiver via the internet when all transmissions are scheduled for early morning hours. The significance of early morning hours is that internet congestion is low while users sleep. When the sender and receiver lie in proximal time zones, a direct transmission from sender to receiver... (more)

Searching for a Single Community in a Graph

In standard graph clustering/community detection, one is interested in partitioning the graph into more densely connected subsets of nodes. In contrast, the search problem of this article aims to only find the nodes in a single such community, the target, out of the many communities that may exist. To do so, we are given suitable side information... (more)



ACM Transactions on Modeling and Performance Evaluation of Computing Systems (ToMPECS) is a new ACM journal that publishes refereed articles on all aspects of the modeling, analysis, and performance evaluation of computing and communication systems.

The target areas for the application of these performance evaluation methodologies are broad, and include traditional areas such as computer networks, computer systems, storage systems, telecommunication networks, and Web-based systems, as well as new areas such as data centers, green computing/communications, energy grid networks, and on-line social networks.

Issues of the journal will be published on a quarterly basis, appearing both in print form and in the ACM Digital Library. The first issue will likely be released in late 2015 or early 2016.

Forthcoming Articles
Ensuring Persistent Content in Opportunistic Networks via Stochastic Stability Analysis

The emerging device-to-device communication solutions and the abundance of mobile applications and services make opportunistic networking not only a feasible solution, but also an important component of future wireless networks. Specifically, the distribution of locally relevant content could be based on the community of mobile users visiting an area, if long term content survival can be ensured this way. In this paper we establish the conditions of content survival in such opportunistic networks, considering the user mobility patterns, as well as the time users keep forwarding the content, as the controllable system parameter. We demonstrate that a tractable epidemic model adequately characterizes the content spreading process, and derive the necessary user contribution to ensure content survival. We show that the required contribution from the users depends significantly on the size of the population, that users need to redistribute content only in a short period within their stay, and that they can decrease their contribution significantly in crowded areas. Hence, with the appropriate control of the system parameters, opportunistic content sharing can be both reliable and sustainable.

Scale-out vs Scale-up: A Study of ARM-based SoCs on Server-class workloads

ARM 64-bit processing has generated enthusiasm to develop ARM-based servers that are targeted for both data centers and supercomputers. In addition to the server-class components and hardware advancements, ARM software environment has been growing substantially over the past decade. Major development ecosystems and libraries are ported and optimized to run on ARM environment, making ARM suitable for server-class workloads. There are two trends in available ARM SoCs in the market: the mobile-class ARM SoCs rely on heterogeneous integration of a mix of CPU cores, GPGPU streaming multiprocessors (SMs), and other accelerators, whereas the server-class SoCs instead rely on integrating a larger number of CPU cores with no GPGPU support and a number of IO accelerators. For scaling the number of processing cores, there are two different paradigm: mobile-class SoCs uses scale-out architecture in the form of a cluster of simpler systems connected with the network, whereas sever-class ARM SoCs uses the scale-up solution and leverage symmetric multiprocessing (SMP) to pack large number of cores on the chip. In this work, we present ScaleSoC cluster which is a scale-out solution based on mobile class ARM SoCs. ScaleSoC leverage fast network connectivity and GPGPU acceleration to improve performance and energy efficiency compared to previous ARM clusters. We consider a wide range of modern server-class parallel workloads including latency-sensitive transactional workloads, MPI-based CPU and GPGPU accelerated scientific applications, and emerging artificial intelligence workloads. We study the performance and energy efficiency of ScaleSoC compared to server-class ARM SoCs and discrete GPGPUs in depth for each type of server-class workloads. We quantify the network overhead on the performance of ScaleSoC and show packing a large number of ARM cores on a single chip does not necessarily guarantee better performance due to shared resources such as last level cache become the performance bottleneck. We characterize the GPGPU accelerated workloads and demonstrate for applications that can leverage the better CPU-GPGPU balance of ScaleSoC cluster, performance and energy efficiency both get improved compared to discrete GPGPUs. We also analyze the scalability and performance limitations of the proposed ScaleSoC cluster.

Quantifying Cloud Performance and Dependability: Taxonomy, Metric Design, and Emerging Challenges

In the past decade, cloud computing has emerged from a pursuit for a service-driven information and communication technology (ICT), into a significant fraction of the ICT market. Responding to the growth of the market, many alternative cloud services and their underlying systems are currently vying for the attention of cloud users and providers. Thus, benchmarking them is needed, to enable cloud users to make an informed choice, and to enable system DevOps to tune, design, and evaluate their systems. This requires focusing on old and new system properties, possibly leading to the re-design of classic benchmarking metrics, such as expressing performance as throughput and latency (response time), and the design of new, cloud-specific metrics. Addressing this requirement, in this work we focus on four system properties: (i) elasticity of the cloud service, to accommodate large variations in the amount of service requested, (ii) performance isolation between the tenants of shared cloud systems, (iii) availability of cloud services and systems, and the (iv) operational risk of running a production system in a cloud environment. Focusing on key metrics, for each of these properties we review the state-of-the-art, then select or propose new metrics together with measurement approaches. We see the presented metrics as a foundation towards upcoming, industry-standard, cloud benchmarks.

Mean Field Games in Nudge Systems for Societal Networks

We consider the general problem of resource sharing in societal networks, consisting of interconnected communication, transportation, energy and other networks important to the functioning of society. Participants in such network need to take decisions daily, both on the quantity of resources to use as well as the periods of usage. With this in mind we discuss the problem of incentivizing users to behave in such a way that society as a whole benefits. In order to perceive societal level impact such incentives may take the form of rewarding users with lottery tickets based on good behavior, and periodically conducting a lottery to translate these tickets into real rewards. We will pose the user decision problem as a mean field game (MFG), and the incentives question as one of trying to select a good mean field equilibrium (MFE). In such a framework, each agent (a participant in the societal network) takes a decision based on an assumed distribution of actions of his/her competitors, and the incentives provided by the social planner. The system is said to be at MFE if the agent's action is a sample drawn from the assumed distribution. We will show the existence of such an MFE under different settings, and also illustrate how to choose an attractive equilibrium using as an example demand-response in energy networks.

On the Convergence of the TTL Approximation for an LRU Cache under Independent Stationary Request Processes

In this paper we focus on the LRU cache where requests for distinct contents are described by independent stationary and ergodic processes. We extend a TTL-based approximation of the cache hit probability first proposed by Fagin \cite{fagin1977asymptotic} for the independence reference model to this more general workload model. We show that under very general conditions this approximation is exact as the cache size and the number of contents go to infinity. Moreover, we establish this not only for the aggregate cache hit probability as in \cite{fagin1977asymptotic} but also for every individual content. Last, we obtain a rate of convergence.

System and Architecture Level Characterization of Big Data Applications on Big and Little Core Server Architectures

The rapid growth in the data yields challenges to process data efficiently using current high performance server architectures such as big Xeon cores. Furthermore, physical design constraints, such as power and density, have become the dominant limiting factor for scaling out servers. Low-power embedded cores in servers such as little Atom have emerged as a promising solution to enhance energy-efficiency to address these challenges. Therefore, the question of whether to process the big data applications on big Xeon or Little Atom based servers becomes important. In this work, through methodical investigation of power and performance measurements, and comprehensive applicationlevel, system-level and micro-architectural level analysis, we characterize dominant big data applications on big Xeon and little Atom-based server architectures. The characterization results across a wide range of real-world big data applications and various software stacks demonstrate how the choice of big vs little core-based server for energyefficiency is significantly influenced by the size of data, performance constraints, and presence of accelerator. In addition, we analyze processor resource utilization of this important class of applications such as memory footprints, CPU utilization and disk bandwidth to understand their run-time behavior. Furthermore, we perform micro-architecture level analysis to highlight where improvement is needed in big and little core microarchitectures to address their performance bottlenecks.

QMLE: a Methodology for Statistical Inference of Service Demands from Queueing Data

Estimating the demands placed by services on physical resources is an essential step for the definition of performance models. For example, scalability analysis relies on these parameters to predict queueing delays under increasing loads. In this paper, we investigate maximum likelihood (ML) estimators for demands at load-independent and load-dependent resources in systems with parallelism constraints. We define a likelihood function based on state measurements and derive necessary conditions for its maximization. We then obtain novel estimators that accurately and inexpensively obtain service demands using only aggregate state data. With our approach, confidence intervals can be rigorously derived, explicitly taking into account both topology and concurrency levels of the services. Our estimators and their confidence intervals are validated against simulations and real system measurements for two multi-tier applications, showing high accuracy also in the presence of load-dependent resources.

All ACM Journals | See Full Journal Index

enter search term and/or author name