Arif Merchant

ORCID: 0000-0002-0913-1459
Publications
Citations
Views
---
Saved
---
About
Contact & Profiles
Research Areas
  • Advanced Data Storage Technologies
  • Distributed systems and fault tolerance
  • Cloud Computing and Resource Management
  • Caching and Content Delivery
  • Parallel Computing and Optimization Techniques
  • Distributed and Parallel Computing Systems
  • Software System Performance and Reliability
  • Interconnection Networks and Systems
  • Peer-to-Peer Network Technologies
  • IoT and Edge/Fog Computing
  • Network Traffic and Congestion Control
  • Advanced Queuing Theory Analysis
  • VLSI and FPGA Design Techniques
  • Advanced Database Systems and Queries
  • Software-Defined Networks and 5G
  • Wireless Communication Networks Research
  • Advanced Wireless Network Optimization
  • graph theory and CDMA systems
  • Cellular Automata and Applications
  • Advanced MIMO Systems Optimization
  • Constraint Satisfaction and Optimization
  • Optimization and Search Problems
  • Personal Information Management and User Behavior
  • Magnetic properties of thin films
  • Low-power high-performance VLSI design

Google (United States)
2011-2024

University of California, Irvine
2012

Hewlett-Packard (United States)
1998-2010

Princeton University
1994-2002

NEC (United States)
2002

NEC (Japan)
1995

IBM (United States)
1992

Stanford University
1991

Data centers are often under-utilized due to over-provisioning as well time-varying resource demands of typical enterprise applications. One approach increase utilization is consolidate applications in a shared infrastructure using virtualization. Meeting application-level quality service (QoS) goals becomes challenge consolidated environment application needs differ. Furthermore, for multi-tier applications, the amount resources needed achieve their QoS might be different at each tier and...

10.1145/1272996.1273026 article EN 2007-03-21

Virtualized data centers enable sharing of resources among hosted applications. However, it is difficult to satisfy service-level objectives(SLOs) applications on shared infrastructure, as application workloads and resource consumption patterns change over time. In this paper, we present AutoControl, a control system that automatically adapts dynamic workload changes achieve SLOs. AutoControl combination an online model estimator novel multi-input, multi-output (MIMO) controller. The...

10.1145/1519065.1519068 article EN 2009-04-01

We propose a new paradigm for building scalable distributed systems. Our approach does not require dealing with message-passing protocols -- major complication in existing Instead, developers just design and manipulate data structures within our service called Sinfonia. Sinfonia keeps applications on set of memory nodes, each exporting linear address space. At the core is novel minitransaction primitive that enables efficient consistent access to data, while hiding complexities arise from...

10.1145/1294261.1294278 article EN 2007-10-14

Enterprise-scale storage systems, which can contain hundreds of host computers and devices up to tens thousands disks logical volumes, are difficult design. The volume choices that need be made is massive, many have unforeseen interactions. Storage system design tedious complicated do by hand, usually leading solutions grossly over-provisioned, substantially under-performing or, in the worst case, both.To solve configuration nightmare, we present minerva: a suite tools for designing systems...

10.1145/502912.502915 article EN ACM Transactions on Computer Systems 2001-11-01

This paper describes the design, implementation, and evaluation of a Federated Array Bricks (FAB), distributed disk array that provides reliability traditional enterprise arrays with lower cost better scalability. FAB is built from collection bricks, small storage appliances containing commodity disks, CPU, NVRAM, network interface cards. deploys new majority-voting-based algorithm to replicate or erasure-code logical blocks across bricks reconfiguration move data in background when are...

10.1145/1024393.1024400 article EN 2004-10-07

Virtualized servers run a diverse set of virtual machines (VMs), ranging from interactive desktops to test and development environments even batch workloads. Hypervisors are responsible for multiplexing the underlying hardware resources among VMs while providing them desired degree isolation using resource management controls. Existing methods provide many knobs allocating CPU memory VMs, but support control IO allocation has been quite limited. in hypervisor introduces significant new...

10.5555/1924943.1924974 article EN Operating Systems Design and Implementation 2010-10-04

Considers a problem of network design personal communication services (PCS). The is to assign cells the switches PCS in an optimum manner. authors consider two types costs. One cost handoffs between cells. other cabling (or trunking) cell site and its associated switch. constrained by call volume that each switch can handle. formulate exactly as integer programming problem. They also propose heuristic solution for this show it performs extremely well.< <ETX...

10.1109/90.469954 article EN IEEE/ACM Transactions on Networking 1995-01-01

Modern disk drives read-ahead data and reorder incoming requests in a workload-dependent fashion. This improves their performance, but makes simple analytical models of them inadequate for performance prediction, capacity planning, workload balancing, so on. To address this problem we have developed new analytic model that do readahead request reordering. We did by developing the drive components (queues, caches, mechanism) transformation technique composing them. Our includes effects...

10.1145/277851.277906 article EN 1998-06-01

Data centers are often under-utilized due to over-provisioning as well time-varying resource demands of typical enterprise applications. One approach increase utilization is consolidate applications in a shared infrastructure using virtualization. Meeting application-level quality service (QoS) goals becomes challenge consolidated environment application needs differ. Furthermore, for multi-tier applications, the amount resources needed achieve their QoS might be different at each tier and...

10.1145/1272998.1273026 article EN ACM SIGOPS Operating Systems Review 2007-03-21

Storage consolidation is becoming an attractive paradigm for data organization because of the economies sharing and ease centralized management. However, resources viable only if applications can be isolated from each other. This work targets problem providing performance guarantees to application irrespective behavior other workloads. Application requirements are represented in terms average throughput, latency maximum burst size. Most earlier schemes do weighted bandwidth allocation; that...

10.1145/1254882.1254885 article EN 2007-06-12

Feedback mechanisms can help today's increasingly complex computer systems adapt to changes in workloads or operating conditions. Control theory offers a principled way for designing feedback loops deal with unpredictable changes, uncertainties, and disturbances systems. We provide an overview of the joint research at HP Labs University Michigan past few years, where control was applied automated resource service level management data centers. highlight key benefits control-theoretic...

10.1145/1496909.1496922 article EN ACM SIGOPS Operating Systems Review 2009-01-01

The performance modeling and analysis of disk arrays is challenging due to the presence multiple disks, large array caches, sophisticated controllers. Moreover, storage manufacturers may not reveal internal algorithms implemented in their devices, so real are effectively black-boxes. We use standard techniques develop an integrated model that incorporates some complexities arrays. show how measurement data baseline models can be used extract information about various features a array. In...

10.1109/tpds.2004.9 article EN IEEE Transactions on Parallel and Distributed Systems 2004-05-06

Enterprise storage systems depend on disk arrays for their capacity and availability needs. To design maintain that efficiently satisfy evolving requirements, it is critical to be able evaluate configuration alternatives without having physically implement them. In this paper, we describe an analytical model predict array throughput, based a hierarchical decomposition of the internal architecture. We validate against state-of-the-art variety synthetic workloads configurations. our knowledge,...

10.1109/mascot.2001.948868 article EN 2002-11-13

Most of the data in large-scale storage clusters is erasure coded. At exascale, optimizing codes for low overhead, efficient reconstruction, and easy deployment critical importance. Locally recoverable (LRCs) have deservedly gained central importance this field, because they can balance many these requirements. In our work, we study wide LRCs; LRCs with large number blocks per stripe overhead. These are a natural next step practitioners to unlock higher savings, but come their own...

10.1145/3626198 article EN cc-by ACM Transactions on Storage 2023-10-05

We propose a new paradigm for building scalable distributed systems. Our approach does not require dealing with message-passing protocols, major complication in existing Instead, developers just design and manipulate data structures within our service called Sinfonia. Sinfonia keeps applications on set of memory nodes, each exporting linear address space. At the core is minitransaction primitive that enables efficient consistent access to data, while hiding complexities arise from...

10.1145/1629087.1629088 article EN ACM Transactions on Computer Systems 2009-11-01

Solid-state drives (SSDs) based on NAND flash are making deep inroads into data centers as well the consumer market. In 2016, manufacturers shipped more than 130 million units totaling around 50 Exabytes of storage capacity. As amount stored solid state keeps increasing, it is important to understand reliability characteristics these devices. For a long time, our knowledge about was derived from controlled experiments in lab environments under synthetic workloads, often using methods for...

10.1109/jproc.2017.2735969 article EN Proceedings of the IEEE 2017-08-18

Sharding is a fundamental building block of large-scale applications, but most have their own custom, ad-hoc implementations. Our goal to make sharding as easily reusable filesystem or lock manager. Slicer Google's general purpose service. It monitors signals such load hotspots and server health dynamically shard work over set servers. Its goals are maintain high availability reduce imbalance while minimizing churn from moved work.In this paper, we describe Slicer's design implementation....

10.5555/3026877.3026934 article EN Operating Systems Design and Implementation 2016-11-02

Storage systems account for a major portion of the total cost ownership (TCO) warehouse-scale computers, and thus have impact on overall system's efficiency. Machine learning (ML)-based methods solving key problems in storage system efficiency, such as data placement, shown significant promise. However, there are few known practical deployments methods. Studying this problem context real-world hyperscale center at Google, we identify number challenges that believe cause lack adoption....

10.48550/arxiv.2501.05651 preprint EN arXiv (Cornell University) 2025-01-09

A Redundant Array of Independent Disks (RAID) G disks provides protection against single disk failures by adding one parity block for each G-1 data blocks. In a clustered RAID, the data/parity blocks are distributed over cluster C (C>G), thus reducing additional load on due to failure. However, most methods proposed implementing such mapping do not work general and values. this paper, we describe fast algorithm based almost-random permutations. An analytical model is constructed, queue with...

10.1109/12.485575 article EN IEEE Transactions on Computers 1996-03-01
Coming Soon ...