Robert Latham

ORCID: 0000-0002-5285-6375
Publications
Citations
Views
---
Saved
---
About
Contact & Profiles
Research Areas
  • Advanced Data Storage Technologies
  • Parallel Computing and Optimization Techniques
  • Distributed and Parallel Computing Systems
  • Distributed systems and fault tolerance
  • Particle physics theoretical and experimental studies
  • High-Energy Particle Collisions Research
  • Utopian, Dystopian, and Speculative Fiction
  • Gothic Literature and Media Analysis
  • Digital Games and Media
  • Cloud Computing and Resource Management
  • Environmental Conservation and Management
  • Particle Detector Development and Performance
  • Scientific Computing and Data Management
  • Algorithms and Data Compression
  • Caching and Content Delivery
  • Embedded Systems Design Techniques
  • Quantum Chromodynamics and Particle Interactions
  • Sexuality, Behavior, and Technology
  • Asian Culture and Media Studies
  • Computer Graphics and Visualization Techniques
  • Political Economy and Marxism
  • Remote Sensing and LiDAR Applications
  • Gender, Feminism, and Media
  • Service-Oriented Architecture and Web Services
  • Simulation Techniques and Applications

Argonne National Laboratory
2015-2025

Illinois Institute of Technology
2016

Office of Scientific and Technical Information
2012

Alpha Star (United States)
2000

Oklahoma State University
1982

University of Minnesota
1973

Dataset storage, exchange, and access play a critical role in scientific applications. For such purposes netCDF serves as portable, efficient file format programming interface, which is popular numerous application domains. However, the original interface does not provide an mechanism for parallel data storage access. In this work, we present new writing reading datasets. This derived with minimal changes from serial but defines semantics tailored high performance. The underlying I/O...

10.1145/1048935.1050189 article EN 2003-11-15

Developing and tuning computational science applications to run on extreme scale systems are increasingly complicated processes. Challenges such as managing memory access message-passing behavior made easier by tools designed specifically aid in these Tools that can help users better understand the of their application with respect I/O have not yet reached level utility necessary play a central role development tuning. This deficiency tool set means we poor understanding how specific...

10.1109/clustr.2009.5289150 article EN 2009-01-01

Computational science applications are driving a demand for increasingly powerful storage systems. While many techniques available capturing the I/O behavior of individual application trial runs and specific components system, continuous characterization production system remains daunting challenge systems with hundreds thousands compute cores multiple petabytes storage. As result, these often designed without clear understanding diverse computational workloads they will support. In this...

10.1145/2027066.2027068 article EN ACM Transactions on Storage 2011-10-01

Current leadership-class machines suffer from a significant imbalance between their computational power and I/O bandwidth. While Moore's law ensures that the of high-performance computing systems increases with every generation, same is not true for subsystems. The scalability challenges faced by existing parallel file respect to increasing number clients, coupled minimalistic compute node kernels running on these machines, call new paradigm meet requirements data-intensive scientific...

10.1109/clustr.2009.5289188 article EN 2009-01-01

Today's top high performance computing systems run applications with hundreds of thousands processes, contain storage nodes, and must meet massive I/O requirements for capacity performance. These leadership-class face daunting challenges to deploying scalable systems. In this paper we present a case study the scalability on Intrepid, IBM Blue Gene/P system at Argonne Leadership Computing Facility. Listed in 5 fastest supercomputers 2008, Intrepid runs computational science intensive demands...

10.1145/1654059.1654100 article EN 2009-11-14

Computational science applications are driving a demand for increasingly powerful storage systems. While many techniques available capturing the I/O behavior of individual application trial runs and specific components system, continuous characterization production system remains daunting challenge systems with hundreds thousands compute cores multiple petabytes storage. As result, these often designed without clear understanding diverse computational workloads they will support.In this...

10.1109/msst.2011.5937212 article EN 2011-05-01

SUMMARY Exploding dataset sizes from extreme‐scale scientific simulations necessitates efficient data management and reduction schemes to mitigate I/O costs. With the discrepancy between bandwidth computational power, scientists are forced capture infrequently, thereby making collection an inherently lossy process. Although compression can be effective solution, random nature of real‐valued datasets renders lossless routines ineffective. These techniques also impose significant overhead...

10.1002/cpe.2887 article EN Concurrency and Computation Practice and Experience 2012-07-11

Efficient handling of large volumes data is a necessity for exascale scientific applications and database systems. To address the growing imbalance between amount available storage being produced by high speed (FLOPS) processors on system, must be compressed to reduce total placed file General-purpose loss less compression frameworks, such as zlib bzlib2, are commonly used datasets requiring compression. Quite often, however, many sets compress poorly, referred hard-to-compress datasets, due...

10.1109/icde.2012.114 article EN 2012-04-01

The advent of exascale supercomputers heralds a new era scientific discovery, yet it introduces significant architectural challenges that must be overcome for MPI applications to fully exploit its potential. Among these is the adoption heterogeneous architectures, particularly integration GPUs accelerate computation. Additionally, complexity multithreaded programming models has also become critical factor in achieving performance at scale. efficient utilization hardware acceleration...

10.1177/10943420241311608 article EN The International Journal of High Performance Computing Applications 2025-01-09

Efficient analytics of scientific data from extreme-scale simulations is quickly becoming a top-notch priority. The increasing simulation output sizes demand for paradigm shift in how conducted. In this paper, we argue that query-driven over compressed---rather than original, full-size---data promising strategy order to meet storage-and-I/O-bound application challenges. As proof-of-principle, propose parallel query processing engine, called ISABELA-QA designed and optimized knowledge priors...

10.1145/2063384.2063425 article EN 2011-11-08

Parallel I/O plays a crucial role for most data-intensive applications running on massively parallel systems like Blue Gene/L that provides the promise of delivering enormous computational capability. We designed and implemented highly scalable file architecture Gene/L, which leverages benefit hierarchical functional partitioning design system software with separate cores. The exploits scalability aspect GPFS (General File System) at backend, while using MPI as an interface between...

10.1109/hpca.2006.1598125 article EN 2006-03-21

In addition to their role as simulation engines, modern supercomputers can be harnessed for scientific visualization. Their extensive concurrency, parallel storage systems, and high-performance interconnects mitigate the expanding size complexity of datasets prepare in situ visualization these data. ongoing research into testing volume rendering on IBM Blue Gene/P (BG/P), we measure performance disk I/O, rendering, compositing large datasets, evaluate bottlenecks with respect system-specific...

10.1109/icpp.2009.27 article EN International Conference on Parallel Processing 2009-09-01

Concurrent accesses to the shared storage resources in current HPC machines lead severe performance degradation caused by I/O contention. In this study, we identify some key challenges efficiently handling interleaved data accesses, and propose a system-wide solution optimize global performance. We implemented tested several scheduling policies, including prioritizing specific applications leveraging burst buffers defer conflicting from another application and/or directing requests different...

10.1109/escience.2016.7870922 article EN 2016-10-01

We present a scalable technique for the simulation of collider events with multijet final states, based on an improved parton-level event file format and $\mathrm{I}/\mathrm{O}$. The method is implemented both leading- next-to-leading-order QCD calculations. perform comprehensive analysis computing performance validate our new framework using Higgs-boson plus production up to seven jets. make resulting code base available public use.

10.1103/physrevd.109.014013 article EN cc-by Physical review. D/Physical review. D. 2024-01-17

From the novels of Anne Rice to The Lost Boys, from Terminator cyberpunk science fiction, vampires and cyborgs have become strikingly visible figures within American popular culture, especially youth culture. In Consuming Youth, Rob Latham explains why, showing how film, other media deploy these ambiguous monsters embody work through implications a capitalist system in which both consume are consumed. Inspired by Marx's use cyborg vampire as metaphor for objectification physical labor...

10.2307/3738782 article EN The Modern Language Review 2004-04-01

The FLASH code is a computational science tool for simulating and studying thermonuclear reactions. program periodically outputs large checkpoint files (to resume calculation from particular point in time) smaller plot (for visualization analysis). Initial experiments on BlueGene/P spent excessive time input/output (I/O), making it difficult to do actual science. Our investigation of I/O revealed several locations the software stack where we could make improvements. Fixing data corruption...

10.1088/1749-4699/5/1/015001 article EN Computational Science & Discovery 2012-03-20

The optimization of parallel I/O has become challenging because the increasing storage hierarchy, performance variability shared systems, and number factors in hardware software stacks that impact performance. In this paper, we perform an in-depth study complexity involved autotuning modeling, including architecture, stack, noise. We propose a novel hybrid model combining analytical models for communication operations black-box individual operations. experimental results show approach...

10.1109/cluster.2015.29 article EN 2015-09-01

Accurate analysis of HPC storage system designs is contingent on the use I/O workloads that are truly representative expected use. However, analyses generally bound to specific workload modeling techniques such as synthetic benchmarks or trace replay mechanisms, despite fact no single technique appropriate for all cases. In this work, we present design IOWA, a novel abstraction allows arbitrary consumer components obtain from range diverse input sources. Thus, researchers can choose...

10.1145/2832087.2832091 article EN 2015-11-11

There are two popular parallel I/O programming styles used by modern scientific computational applications: unique-file and shared-file. Unique-file usually gives satisfactory performance, but its major drawback is that managing a large number of files can overwhelm the task post-simulation data processing. Shared-file produces fewer allows arrays partitioned among processes to be saved in canonical order. As processors on machines increases into thousands more, problem size turn global...

10.1109/icpp.2009.68 article EN International Conference on Parallel Processing 2009-09-01
Coming Soon ...