Author:

Gary Grider

Company : Los Alamos National Laboratory

Title : HPC Division Leader

 
 
author

Innovations, Challenges, and Lessons Learned in HPC Storage Yesterday, Today, and Tomorrow

Submitted by Anonymous (not verified) on

In this tutorial, we will introduce the audience to the lunatic fringe of extreme high-performance computing and its storage systems. The most difficult challenge in HPC storage is caused by millions (soon to be billions) of simultaneously writing threads. Although cloud providers handle workloads of comparable, or larger, aggregate scale, the HPC challenge is unique because the concurrent writers are modifying shared data. We will begin with a brief history of HPC computing covering the previous few decades, bringing us into the petaflop era which started in 2009.

MarFS: Near-POSIX Access to Object-Storage

Submitted by Anonymous (not verified) on

Many computing sites need long-term retention of mostly-cold data, often referred to as “data lakes”. The main function of this storage-tier is capacity, but non-trivial bandwidth/access requirements may also exist. For many years, tape was the most economical solution. However, data sets have grown larger more quickly than tape bandwidth has improved, such that disk is now becoming more economically feasible for this storage-tier.

MarFS: A Scalable Near-POSIX File System over Cloud Objects for HPC Cool Storage - SDC 2016

Submitted by Anonymous (not verified) on

Many computing sites need long-term retention of mostly cold data often “data lakes”. The main function of this storage tier is capacity but non trivial bandwidth/access requirements exist. For many years, tape was the best economic solution. Data sets have grown larger more quickly than tape bandwidth improvements and access demands have increased in the HPC environment. Disk can be more economically for this storage tier. The Cloud Community has moved towards erasure based object stores to gain scalability and durability using commodity hardware.

MarFS: A Scalable Near-POSIX File System over Cloud Objects for HPC Cool Storage - DSI 2016

Submitted by Anonymous (not verified) on

Many computing sites need long-term retention of mostly cold data often “data lakes”. The main function of this storage tier is capacity but non trivial bandwidth/access requirements exist. For many years, tape was the best economic solution. Data sets have grown larger more quickly than tape bandwidth improvements and access demands have increased. Disk can be more economically for this storage tier. The Cloud Community has moved towards erasure based object stores to gain scalability and durability using commodity hardware.

HPC For Science Based Motivations for Computation Near Storage

Submitted by Anonymous (not verified) on

Scientific data is mostly stored in linear bytes in files but it almost always has hidden structure that resembles records with keys and  values, often times in multiple dimensions.  Further, the bandwidths required to service HPC simulation workloads will soon approach tens of terabytes/sec with single data files surpassing a petabyte and single sets of data from a campaign approaching 100 petabytes.  Multiple tasks from distributed analytical/indexing functions to data management tasks like compression, erasure, encoding, dedup, are all potentially more efficiently and econ

HPC Scientific Simulation Computational Storage Saga

Submitted by Anonymous (not verified) on

Los Alamos is working to revolutionize how scientific data management is done, moving from large Petabyte sized files generated periodically by extreme scale simulations to a record and column based approach. Along the journey, the NVME Computational Storage efforts became a strategic way to help accomplish this revolution. Los Alamos has been working on a series of proof of concepts with a set of data storage industry partners and these partnerships have proven to be the key to success.

Leveraging Computational Storage for Simulation Science Storage System Design

Submitted by Anonymous (not verified) on

High-performance computing data centers supporting large-scale simulation applications can routinely generate a large amount of data. To minimize time-to-result, it is crucial that this data be promptly absorbed, processed, and potentially even multidimensionally indexed so that it can be efficiently retrieved when the scientists need it for insights.

Versity Gateway, an Open Source High Performance Object to File Translation Tool

Submitted by Anonymous (not verified) on

Gary Grider will kick off the talk with a brief background of the challenges currently faced in mass storage systems, and why it is so difficult for modern S3 based workloads to utilize them. Ben McClelland, who architected the system and did the bulk of the development, will introduce the Versity Gateway, an open-source high-performance object-to-file translation tool. He will explain that the Versity Gateway was written from scratch in Go, highlighting the benefits of this choice.

Subscribe to Gary Grider