Miguel Gila

ORCID: 0009-0006-9363-7060
Publications
Citations
Views
---
Saved
---
About
Contact & Profiles
Research Areas
  • Distributed and Parallel Computing Systems
  • Advanced Data Storage Technologies
  • Cloud Computing and Resource Management
  • Scientific Computing and Data Management
  • Parallel Computing and Optimization Techniques
  • Particle Detector Development and Performance

ETH Zurich
2018-2024

Swisscom (Switzerland)
2014-2024

CSCS - Swiss National Supercomputing Centre
2014-2023

Supercomputers have been driving innovations for performance and scaling benefiting several scientific applications the past few decades. Yet their ecosystems remain virtually unchanged when it comes to integrating distributed data-driven workflows, primarily due rather rigid access methods restricted configuration management options. X-as-a-Service model of cloud has introduced, among other features, a developer-centric DevOps approach empowering developers infrastructure, platform software...

10.1177/10943420231167811 article EN cc-by The International Journal of High Performance Computing Applications 2023-04-11

After the successful first run of LHC, data taking is scheduled to restart in Summer 2015 with experimental conditions leading increased volumes and event complexity. In order process generated such scenario exploit multicore architectures current CPUs, LHC experiments have developed parallelized software for reconstruction simulation. However, a good fraction their computing effort still expected be executed as single-core tasks. Therefore, jobs diverse resources requirements will...

10.1088/1742-6596/664/6/062016 article EN Journal of Physics Conference Series 2015-12-23

Leading hybrid and heterogeneous supercomputing systems process hundreds of thousands jobs using complex scheduling algorithms parameters. The centers operating these aim to achieve higher levels resource utilization while being restricted by compliance with policy constraints. There is a critical need for high-fidelity, high-performance tool familiar interfaces that allows not only tuning optimization the operational job scheduler but also enables exploration new management algorithms. We...

10.1109/sc.2018.00028 article EN 2018-11-01

Abstract The prompt reconstruction of the data recorded from Large Hadron Collider (LHC) detectors has always been addressed by dedicated resources at CERN Tier-0. Such workloads come in spikes due to nature operation accelerator and special high load occasions experiments have commissioned methods distribute (spill-over) a fraction sites outside CERN. present work demonstrates new way supporting Tier-0 environment provisioning elastically for such spilled-over workflows onto Piz Daint...

10.1007/s41781-020-00052-w article EN cc-by Computing and Software for Big Science 2021-02-08

This paper introduces the versatile software-defined cluster, a novel framework that integrates HPC and Cloud technologies offering service-oriented approach for computing resources instead of hardware focused one, maintaining infrastructure independence avoiding vendor lock-in. It addresses challenges rigidity lack customizability in conventional systems, facilitating more efficient use shared infrastructures. The core concept revolves around three-tiered structure—Infrastructure, Service...

10.1109/mcse.2024.3394164 article EN Computing in Science & Engineering 2024-04-26

Complex applications and workflows needs are often exclusively expressed in terms of computational resources on HPC systems. In many cases, other like storage or network not allocatable shared across the entire system. By looking at particular, any workflow application should be able to select both its preferred data manager required capability capacity. To achieve such a goal, new mechanisms introduced. this work, we present tool that dynamically provisions management system top devices. We...

10.1109/ipdpsw50202.2020.00173 article EN 2022 IEEE International Parallel and Distributed Processing Symposium Workshops (IPDPSW) 2020-05-01

Summary We present a methodology to enable the complete software development life cycle on Cray XC systems within container that can hold any version of Programming Environment (CPE). The installation CPE inside facilitates many aspects typical HPC support and operation workloads managing such as testing new CPEs, comparing performances, or keeping built with an old running updated systems. procedure for creating consists three steps: creation holding targeted CPE, compilation desired...

10.1002/cpe.5543 article EN Concurrency and Computation Practice and Experience 2019-11-11

Complex applications and workflows needs are often exclusively expressed in terms of computational resources on HPC systems. In many cases, other like storage or network not allocatable shared across the entire system. By looking at resource particular, any workflow application should be able to select both its preferred data manager required capability capacity. To achieve such a goal, new mechanisms introduced. this work, we introduce mechanism for dynamically provision management system...

10.3929/ethz-b-000405606 article EN 2019-05-01

Complex applications and workflows needs are often exclusively expressed in terms of computational resources on HPC systems. In many cases, other like storage or network not allocatable shared across the entire system. By looking at resource particular, any workflow application should be able to select both its preferred data manager required capability capacity. To achieve such a goal, new mechanisms introduced. this work, we introduce mechanism for dynamically provision management system...

10.48550/arxiv.1911.12162 preprint EN cc-by-sa arXiv (Cornell University) 2019-01-01

As supercomputing systems gradually become an integral part of data driven workflows such as ML and AI or tightly-coupled pre- post-processing pipelines, users need programmable access to shared resources avoid moving large volume dedicated public cloud providers. Public clouds the private ones using technologies like OpenStack, multi-tenancy on hardware has been a commonplace over decade, offering privileged compute, network storage. Such is unavailable batch-scheduled, multi-Petascale...

10.1109/scws55283.2021.00021 article EN 2021-11-01
Coming Soon ...