ESiWACE2

The Centre of Excellence in Simulation of Weather and Climate in Europe (ESiWACE2) is an H2020 funded project and successor of the ESiWACE project.

Within the project, the research group is responsible for various contribution to work packages and particularly WP 4.

Please see the official web page of ESiWACE for further information.

Contact Dr. Julian Kunkel

  • Deutsches Klimarechenzentrum GmbH (coordinator)
  • Centre National de la Recherche Scientifique
  • European Centre for Medium-Range Weather Forecasts
  • Barcelona Supercomputing Center
  • Max-Planck-Gesellschaft zur Förderung der Wissenschaften e.V./ Max-Planck-Institut für Meteorologie
  • Sveriges meteorologiska och hydrologiska institut
  • Centre Européen de Recherche et de Formation Avancée en Calcul Scientifique
  • National University of Ireland Galway (Irish Centre for High End Computing)
  • Met Office
  • Fondazione Centro Euro-Mediterraneo sui Cambiamenti Climatici
  • The University of Reading
  • Science and Technology Facilities Council
  • BULL SAS
  • Seagate Systems UK Limited
  • ETH Zürich
  • The University of Manchester
  • Netherlands eScience Center
  • Federal Office of Meteorology and Climatology
  • DataDirect Networks
  • Mercator Océan

While we are involved in various work packages, the main focus lies is WP4 which will provide the necessary toolchain to handle data at pre-exa-scale and exa-scale, for single simulations and ensembles.

Specifically, we will

  1. Support data reduction in ensembles and avoid un-necessary subsequent data manipulations by providing tools to carry out ensemble statistics “in-flight” and compress ensemble members on the way to storage.
  2. Provide tools to: a) transparently hide complexity of multiple-storage tiers from applications at runtime by developing middleware that lies between the familiar NetCDF interface and storage, and prototype commercially credible storage appliances which can appear at the backend of such middleware, and; b) support manual migration of semantically important content between primary storage on disk, tape, and object stores, including appropriate user-space caching tools (thus allowing some portable data management within weather and climate workflows).

The work will build upon the generated prototypes build in ESiWACE-1.

architecture-d4.2.pdf

ESiWACE2 Architecture Milestone Document

ESDM builds upon a data model similar to NetCDF and utilizes a self-describing on-disk data format for storing structured data. We aim to deliver the NetCDF integrated version by the end of the ESiWACE1 project. This improvement can then be used as a drop-in replacement for typical use-cases without changing anything from the application perspective. While our current version utilises the manual configuration by data-center experts, the ultimate long-term goal is to employ machine learning to automatise the decision making and reduce the burden for users and experts.

Here are some results achieved from the ESiWACE 1 project. We run our ESDM prototype at Mistral with several larger number of processes. The results for running the benchmarks on 200 nodes with varying numbers of processes are shown in Figure 1. The figure shows the results for different processes per node (x-axis) considering ten timesteps of 300 GB data each.

As the baseline for exploring the efficiency, we run the IOR benchmark using optimal settings (i.e., large sequential I/O). The graphic shows two IOR results: storing file-per-process (fpp) on Lustre (ior-fpp), as this yielded better performance than the results for shared file access, and storing fpp on local storage (ior-fpptmp).

Mistral has two file systems (Lustre01 and Lustre02) and five configurations with ESDM were tested: storing data only in Lustre02, settings where data are stored on both Lustre file systems concurrently (both), and environments with in-memory storage (local tmpfs). We also explored if fragmenting data into 100MB files and 500MB files is beneficial (the large configurations). Note that the performance achieved on a single file system is slightly faster to the best-case performance achieved with optimal settings using the benchmarks. We conclude that the fragmentation into chunks accelerates the benchmark.

By utilizing the two file systems resembling a heterogeneous environment effectively, we can improve the performance from 150 GB/s to 200 GB/s (133% of a single file system). While this was just a benchmark testing, it shows that we are able to exploit the available performance and thus.

Publications

  • \myPub{2019}{Toward Understanding I/O Behavior in HPC Workflows}{Jakob Lüttgau, Shane Snyder, Philip Carns, Justin M. Wozniak, Julian Kunkel, Thomas Ludwig}{In IEEE/ACM 3rd International Workshop on Parallel Data Storage & Data Intensive Scalable Computing Systems (PDSW-DISCS), pp. 64–75, IEEE Computer Society, PDSW-DISCS, Dallas, Texas, ISBN: 978-1-7281-0192-7}
  • \myPub{}{Beating data bottlenecks in weather and climate science}{Bryan N. Lawrence, Julian Kunkel, Jonathan Churchill, Neil Massey, Philip Kershaw, Matt Pritchard}{In Extreme Data Workshop 2018, Schriften des Forschungszentrums Jülich IAS Series (40), pp. 31–36, Forschungszentrum Jülich, Extreme Data Workshop, Jülich, Germany, ISBN: 978-3-95806-392-1, ISSN: 1868-8489}
  • \myPub{}{Cost and Performance Modeling for Earth System Data Management and Beyond}{Jakob Lüttgau, Julian Kunkel}{In High Performance Computing: ISC High Performance 2018 International Workshops, Frankfurt/Main, Germany, June 28, 2018, Revised Selected Papers, Lecture Notes in Computer Science (11203), pp. 23–35, Springer, HPC-IODC workshop, ISC HPC, Frankfurt, Germany, ISBN: 978-3-030-02465-9, ISSN: 1611-3349}
  • \myPub{2018}{A Survey of Storage Systems for High-Performance Computing}{Jakob Lüttgau, Michael Kuhn, Kira Duwe, Yevhen Alforov, Eugen Betke, Julian Kunkel, Thomas Ludwig}{In Supercomputing Frontiers and Innovations, Series: Volume 5, Number 1, pp. 31–58, Publishing Center of South Ural State University}
  • \myPub{2017}{Adaptive Tier Selection for NetCDF and HDF5}{Jakob Lüttgau, Eugen Betke, Olga Perevalova, Julian Kunkel, Michael Kuhn}{Poster, SC17, Denver, CO, USA}
  • \myPub{}{Simulation of Hierarchical Storage Systems for TCO and QoS}{Jakob Lüttgau, Julian Kunkel}{In High Performance Computing: ISC High Performance 2017 International Workshops, DRBSD, ExaComm, HCPM, HPC-IODC, IWOPH, IXPUG, P^3MA, VHPC, Visualization at Scale, WOPSSS, Lecture Notes in Computer Science (10524), pp. 116–128, Springer, ISC High Performance, Frankfurt, Germany, ISBN: 978-3-319-67629-6}
  • \myPub{2016}{Modeling and Simulation of Tape Libraries for Hierarchical Storage Systems}{Jakob Lüttgau, Julian Kunkel}{Poster, SC16, Salt Lake City, Utah, USA}

Talks

  • \myPub{2020}{Data Systems at Scale in Climate and Weather: Activities in the ESiWACE Project}{HPC IODC Workshop}{Virtual}
  • \myPub{}{Data-Centric IO: Potential for Climate/Weather}{6th ENES HPC Workshop}{Virtual/Hamburg, Germany}
  • \myPub{}{Progress of WP4: Data at Scale}{ESiWACE General Assembly}{Virtual/Hamburg, Germany}
  • \myPub{}{Toward Next Generation Interfaces for Exploiting Workflows}{SIG IO UK}{University of Reading, Reading, UK}
  • \myPub{}{Potential of I/O-Aware Workflows in Climate and Weather}{Supercomputing Frontiers Europe}{Virtual/Warshaw Poland}
  • \myPub{}{Challenges and Approaches for Extreme Data Processing}{EPSRC Centre for Doctoral Training Mathematics of Planet Earth}{University of Reading, Reading, UK}
  • \myPub{2019}{Smarter Management using Metadata and Workflow Expertise}{BoF: Knowledge Is Power: Unleashing the Potential of Your Archives Through Metadata}{Supercomputing, Denver, USA}
  • \myPub{}{Exploiting Different Storage Types with the Earth-System Data Middleware}{Parallel Data Systems Workshop}{Supercomputing, Denver, USA}
  • \myPub{}{OpenSource Software}{Hacktoberfest}{University of Reading, Reading, UK}
  • \myPub{}{The Earth-System Data Middleware: An Approach for Heterogeneous Storage Infrastructure}{SPPEXA Final Symposium}{Dresden, Germany}
  • \myPub{}{Utilizing Heterogeneous Storage Infrastructures via the Earth-System Data Middleware}{NEXTGenIO Workshop on applications of NVRAM storage to exascale I/O}{ECMWF, Reading, Germany}
  • \myPub{}{Data-Centric I/O and Next Generation Interfaces}{HPC IODC Workshop}{ISC HPC, Frankfurt, Germany}
  • \myPub{}{The goldilocks node: getting the RAM just right}{BoF: Data-Centric I/O}{ISC HPC, Frankfurt, Germany}
  • \myPub{}{High-Level Workflows – Potential for Innovation? The NGI Initiative and More}{BoF: Data-Centric I/O}{ISC HPC, Frankfurt, Germany}
  • \myPub{}{Fighting the Data Deluge with Data-Centric Middleware}{PASC Minisymposium: The Exabyte Data Challenge}{Zürich, Switzerland}
  • \myPub{2018}{Opportunities for Integrating I/O Capabilities with Cylc}{Cylc Weather User Group; Supercomputing 2018}{Dallas, USA}
  • \myPub{}{Status of WP4: Exploitability}{ESiWACE review meeting}{DKRZ, Hamburg, Germany}
  • \myPub{}{Overcoming Storage Issues of Earth-System Data with Intelligent Storage Systems}{18th Workshop on high performance computing in meteorology}{ECMWF, Reading, UK}
  • \myPub{}{RD&E in Compression}{NCAS Meeting}{Reading, UK}
  • \myPub{}{Challenges and Opportunities for I/O}{Gung Ho Networking Meeting}{Reading, UK}
  • \myPub{}{Cost and Performance Modeling for Earth System Data Management and Beyond}{HPC IODC Workshop}{Frankfurt, Germany}
  • \myPub{}{Community Development of Next Generation Semantic Interfaces}{Joint session of the HPC IODC Workshop and WOPSSS}{Frankfurt, Germany}
  • \myPub{}{The Need for Next Generation Semantic Interfaces to Process Climate/Weather Workflows}{SIG IO UK}{University of Reading, UK}
  • \myPub{}{Towards Intelligent Storage Systems}{Computer Science Workshop}{University of Reading, UK}
  • \myPub{}{Exploiting the Heterogeneous Storage Landscape in a Data Center}{Per3s: Performance and Scalability of Storage Systems Workshop}{Rennes, France}
  • \myPub{2017}{Exploiting Weather and Climate Data at Scale (WP4)}{ESiWACE General Assembly}{Berlin, Germany}

  • Impressum
  • Privacy
  • research/projects/reading/esiwace2/start.txt
  • Last modified: 2023-08-28 10:40
  • by 127.0.0.1