No abstract available.
Proceeding Downloads
An efficient multi-tier tablet server storage architecture
Distributed, structured data stores such as Big Table, HBase, and Cassandra use a cluster of machines, each running a database-like software system called the Tablet Server Storage Layer or TSSL. A TSSL's performance on each node directly impacts the ...
Small-world datacenters
In this paper, we propose an unorthodox topology for datacenters that eliminates all hierarchical switches in favor of connecting nodes at random according to a small-world-inspired distribution. Specifically, we examine topologies where the underlying ...
Modeling and synthesizing task placement constraints in Google compute clusters
Evaluating the performance of large compute clusters requires benchmarks with representative workloads. At Google, performance benchmarks are used to obtain performance metrics such as task scheduling delays and machine resource utilizations to assess ...
DOT: a matrix model for analyzing, optimizing and deploying software for big data analytics in distributed systems
Traditional parallel processing models, such as BSP, are "scale up" based, aiming to achieve high performance by increasing computing power, interconnection network bandwidth, and memory/storage capacity within dedicated systems, while big data ...
CloudScale: elastic resource scaling for multi-tenant cloud systems
Elastic resource scaling lets cloud systems meet application service level objectives (SLOs) with minimum resource provisioning costs. In this paper, we present CloudScale, a system that automates fine-grained elastic resource scaling for multi-tenant ...
ALIAS: scalable, decentralized label assignment for data centers
Modern data centers can consist of hundreds of thousands of servers and millions of virtualized end hosts. Managing address assignment while simultaneously enabling scalable communication is a challenge in such an environment. We present ALIAS, an ...
Incoop: MapReduce for incremental computations
Many online data sets evolve over time as new entries are slowly added and existing entries are deleted or modified. Taking advantage of this, systems for incremental bulk data processing, such as Google's Percolator, can achieve efficient updates. To ...
CloudNaaS: a cloud networking platform for enterprise applications
Enterprises today face several challenges when hosting line-of-business applications in the cloud. Central to many of these challenges is the limited support for control over cloud network functions, such as, the ability to ensure security, performance ...
YCSB++: benchmarking and performance debugging advanced features in scalable table stores
- Swapnil Patil,
- Milo Polte,
- Kai Ren,
- Wittawat Tantisiriroj,
- Lin Xiao,
- Julio López,
- Garth Gibson,
- Adam Fuchs,
- Billie Rinaldi
Inspired by Google's BigTable, a variety of scalable, semi-structured, weak-semantic table stores have been developed and optimized for different priorities such as query speed, ingest speed, availability, and interactivity. As these systems mature, ...
Silverline: toward data confidentiality in storage-intensive cloud applications
By offering high availability and elastic access to resources, third-party cloud infrastructures such as Amazon EC2 are revolutionizing the way today's businesses operate. Unfortunately, taking advantage of their benefits requires businesses to accept a ...
CoScan: cooperative scan sharing in the cloud
We present CoScan, a scheduling framework that eliminates redundant processing in workflows that scan large batches of data in a map-reduce computing environment. CoScan merges Pig programs from multiple users at runtime to reduce I/O contention while ...
Query optimization for massively parallel data processing
MapReduce has been widely recognized as an efficient tool for large-scale data analysis. It achieves high performance by exploiting parallelism among processing nodes while providing a simple interface for upper-layer applications. Some vendors have ...
PrIter: a distributed framework for prioritized iterative computations
Iterative computations are pervasive among data analysis applications in the cloud, including Web search, online social network analysis, recommendation systems, and so on. These cloud applications typically involve data sets of massive scale. Fast ...
Policy expressivity in the Anzere personal cloud
We present a technique for partially replicating data items at scale according to expressive policy specifications. Recent projects have addressed the challenge of policy-based replication of personal data (photos, music, etc.) within a network of ...
ActiveSLA: a profit-oriented admission control framework for database-as-a-service providers
The system overload is a common problem in a Database-as-a-Serice (DaaS) environment because of unpredictable and bursty workloads from various clients. Due to the service delivery nature of DaaS, such system overload usually has direct economic impact ...
Orleans: cloud computing for everyone
Cloud computing is a new computing paradigm, combining diverse client devices -- PCs, smartphones, sensors, single-function, and embedded -- with computation and data storage in the cloud. As with every advance in computing, programming is a fundamental ...
PipeCloud: using causality to overcome speed-of-light delays in cloud-based disaster recovery
Disaster Recovery (DR) is a desirable feature for all enterprises, and a crucial one for many. However, adoption of DR remains limited due to the stark tradeoffs it imposes. To recover an application to the point of crash, one is limited by financial ...
No one (cluster) size fits all: automatic cluster sizing for data-intensive analytics
Infrastructure-as-a-Service (IaaS) cloud platforms have brought two unprecedented changes to cluster provisioning practices. First, any (nonexpert) user can provision a cluster of any size on the cloud within minutes to run her data-processing jobs. The ...
Pesto: online storage performance management in virtualized datacenters
Virtualized datacenters strive to reduce costs through workload consolidation. Workloads exhibit a diverse set of IO behaviors and varying IO load that makes it difficult to estimate the IO performance on shared storage. As a result, system ...
Making time-stepped applications tick in the cloud
Scientists are currently evaluating the cloud as a new platform. Many important scientific applications, however, perform poorly in the cloud. These applications proceed in highly parallel discrete time-steps or "ticks," using logical synchronization ...
Trojan data layouts: right shoes for a running elephant
MapReduce is becoming ubiquitous in large-scale data analysis. Several recent works have shown that the performance of Hadoop MapReduce could be improved, for instance, by creating indexes in a non-invasive manner. However, they ignore the impact of the ...
Cuanta: quantifying effects of shared on-chip resource interference for consolidated virtual machines
Workload consolidation is very attractive for cloud platforms due to several reasons including reduced infrastructure costs, lower energy consumption, and ease of management. Advances in virtualization hardware and software continue to improve resource ...
Small cache, big effect: provable load balancing for randomly partitioned cluster services
Load balancing requests across a cluster of back-end servers is critical for avoiding performance bottlenecks and meeting service-level objectives (SLOs) in large-scale cloud computing services. This paper shows how a small, fast popularity-based front-...
Opportunistic flooding to improve TCP transmit performance in virtualized clouds
Virtualization is a key technology that powers cloud computing platforms such as Amazon EC2. Virtual machine (VM) consolidation, where multiple VMs share a physical host, has seen rapid adoption in practice with increasingly large number of VMs per ...
Improving per-node efficiency in the datacenter with new OS abstractions
We believe datacenters can benefit from more focus on per-node efficiency, performance, and predictability, versus the more common focus so far on scalability to a large number of nodes. Improving per-node efficiency decreases costs and fault recovery ...
Declarative automated cloud resource orchestration
As cloud computing becomes widely deployed, one of the challenges faced involves the ability to orchestrate a highly complex set of subsystems (compute, storage, network resources) that span large geographic areas serving diverse clients. To ease this ...
Automatic management of partitioned, replicated search services
Low-latency, high-throughput web services are typically achieved through partitioning, replication, and caching. Although these strategies and the general design of large-scale distributed search systems are well known, the academic literature provides ...
Scaling the mobile millennium system in the cloud
- Timothy Hunter,
- Teodor Moldovan,
- Matei Zaharia,
- Samy Merzgui,
- Justin Ma,
- Michael J. Franklin,
- Pieter Abbeel,
- Alexandre M. Bayen
We report on our experience scaling up the Mobile Millennium traffic information system using cloud computing and the Spark cluster computing framework. Mobile Millennium uses machine learning to infer traffic conditions for large metropolitan areas ...
To cloud or not to cloud?: musings on costs and viability
In this paper we aim to understand the types of applications for which cloud computing is economically tenable, i.e., for which the cost savings associated with cloud placement outweigh any associated deployment costs.
We discover two scenarios. (i) In ...
Switching the optical divide: fundamental challenges for hybrid electrical/optical datacenter networks
- Hamid Hajabdolali Bazzaz,
- Malveeka Tewari,
- Guohui Wang,
- George Porter,
- T. S. Eugene Ng,
- David G. Andersen,
- Michael Kaminsky,
- Michael A. Kozuch,
- Amin Vahdat
Recent proposals to build hybrid electrical (packet-switched) and optical (circuit switched) data center interconnects promise to reduce the cost, complexity, and energy requirements of very large data center networks. Supporting realistic traffic ...