The group behind AnalyzeThis tested their system on complex, real-word workflows, including Montage, a program used to produce astronomical images like this, the Large Magellanic Cloud. This mosaic compiled from the NASA Spitzer Space Telescope was made as part of a project called SAGE (Surveying the Agents of a Galaxy’s Evolution). (Image: NASA/JPL-Caltech/M. Meixner (Space Telescope Science Institute) and the SAGE legacy team.)

High-performance computing (HPC) systems are like a burgeoning restaurant that can’t widen its service window fast enough to keep pace with its expanding dining room. Computer clusters crunch numbers while ambitious simulations and experiments generate ever bigger and more complex data-analysis orders.

To save HPC resources for the computational heavy lifting of running simulations and managing major experiments, the resulting data are analyzed offline. They’re read from storage and transmitted through a narrow-bandwidth window to a cluster outside the high-performance machine. After the cluster completes the analysis, the final data must then be written back into storage through the same small portal.

Soon, data-analysis orders pile up, awaiting their turn for processing. HPC machines in this position can never reach their full potential, no matter how fast processing power increases. Much of their time is spent just moving data from one place to another, and the problem will only get worse as HPC rises toward exascale – a million trillion calculations per second.

Part of the solution may be akin to waiters preparing meals at the tables, avoiding the service window bottleneck altogether.

That’s the idea behind AnalyzeThis, a novel data storage and analysis system developed by Sudharshan Vazhkudai and colleagues. Vazhkudai leads the Technology Integration Group at the National Center for Computational Sciences at Oak Ridge National Laboratory. With colleagues at Virginia Tech and Lawrence Berkeley National Laboratory, the team is souping up the computational power of solid-state devices (SSDs) – an approach they call active flash, in which scientific data analysis is conducted where the data already resides.

“You have a bandwidth bottleneck exacerbated by the need to do data wrangling,” Vazhkudai says. “Why not do the analysis on the storage component, where the data already resides?”

The group has developed ways to do just that, which they will describe next week during a session on scalable storage systems at the SC15 supercomputing conference in Austin, Texas.

AnalyzeThis grew out of the group’s work exploring ways to use SSDs in input/output (I/O) and in memory extension. At once, a third line of inquiry naturally emerged: How might SSDs be used for in situ data analysis – as the simulation runs?

Each SSD is already equipped with a controller – that is, a multicore processor that manages storage and retrieval of data. The research team realized that, when the controllers are not busy managing I/O, they might also execute data analysis kernels. That opened the possibility for active flash devices.

An early challenge was that active flash devices are not commercially available, “so we have to rely on emulations and build a test bed,” Vazhkudai says. Once an array of active flash devices is in place, how can users direct them to do the necessary analyses?

The answer is AnalyzeThis. In this system, active flash devices within a high-performance machine crunch numbers while the machine itself is busy computing and not using the flash devices for I/O. AnalyzeThis directs data to leapfrog from one active flash system to another, allowing each device to automatically execute predetermined analytics kernels on any data it receives. A file-system layer gives users the ability to dictate which analytics are done and to track the progress of data through the system. In the end, the devices complete much of the analysis without ever transferring the data from the HPC system. In fact, the data are moved little among the active flash devices.

In Vazhkudai’s words, this approach makes data analyses “first-class citizens in the storage system.” The data and analyses are blended into the storage system so that the location reflects the analytics performed and the analytics indicate the location (provenance). The team calls their system an “analysis workflow-aware storage system.”

The idea of in situ data analysis isn’t new, Vazhkudai explains. In the 1990s, others had worked to develop active disks, but the technology wasn’t ready for the leap. Application needs and other factors frustrated development. Active flash devices now have more powerful, self-contained controllers that make the concept viable.

First author on the SC15 paper is Hyogi Sim, a post-master’s research assistant at Oak Ridge and a Ph.D. student at Virginia Tech, advised by Ali Butt. Sim did the implementation, which is about 10,000 lines of code, over the course of a year.

The group used their emulation and test bed to put the new storage system to work on four “real-world, complex” workflows: Montage (astronomical imaging), Brain Atlas (brain-mapping), Sipros (DNA database searching), and Grep (keyword searching). In each case, using a variety of scheduling techniques, one or more iterations of AnalyzeThis outperformed traditional offline analysis.

The team’s next step is to develop AnalyzeThis for distributed storage using the GlusterFS file system.

Vazhkudai and his co-workers believe AnalyzeThis can help break the tradition of developing storage systems and workflow systems independently of one another – a “disconnect,” he says, that makes data-movement expensive. “AnalyzeThis helps you take a data-centric view, and computation is surrounding the data.”

Andy Boyles

Share
Published by
Andy Boyles

Recent Posts

Predicting chaos

A Caltech fellowship recipient works on the physics underlying turbulence, or the chaotic gain of… Read More

November, 2024

Scale-tamer

A recent program alum interweaves large and small scales in wind-energy and ocean models. Read More

October, 2024

Exploring electrons

At UC Berkeley, a fellow applies machine learning to sharpen microscopy. Read More

September, 2024

Subduing software surprises

A Cornell University fellowship recipient works on methods for ensuring software functions as expected. Read More

August, 2024

‘Crazy ideas’

A UCSD engineering professor and former DOE CSGF recipient combines curiosity and diverse research experiences… Read More

July, 2024

Star treatment

A UT Austin-based fellow blends physics and advanced computing to reveal cosmic rays’ role in… Read More

July, 2024