ExaFEL and CoPA: Rapid Imaging of Molecular Systems

May 3, 2024

May 3, 2024 — The Linac Coherent Light Source (LCLS) at the SLAC National Accelerator Laboratory is the world’s first hard X-ray free electron laser (XFEL) facility, using X-rays to take snapshots of atoms and molecules at a specific moment in time.

A schematic of ExaFEL’s workflow, demonstrating how it reconstructs the structure of a target from its diffraction pattern. Credit: ECP.

When XFEL beams scatter off a target, they produce a diffraction pattern—a pattern of light of various intensities, which can be used to study the target—before the radiation damages the molecular structure of the target. This makes XFELs uniquely suitable for studying biological molecules, since an image of its diffraction pattern can be taken before the molecule is destroyed. LCLS can therefore hep scientists understand how atoms interact and move in everything from photosynthesis to the formation of chemical bonds.

A recent upgrade to LCLS—dubbed LCLS-II—enables the brightest X-ray images in the world, using a million X-ray pulses per second, each lasting quadrillionths of a second. This is an 8,000-fold increase from LCLS’s 120 pulses per second, and the ultrahigh repetition rate and brightness provide scientists with adequate resolving power to study the structure and natural variations between individual molecules. With more rapid shots, more pixels per detector, and improvements in X-ray quality, the LCLS-II upgrade collects significantly more data than its predecessor, pointing to a need for accelerating the analysis process.

“All of this means that in the future, what used to be weeks [for data turnaround] could be years, unless we invest into innovative approaches to data analysis,” says Johannes Blaschke, an application performance scientist at the National Energy Research Scientific Computing Center (NERSC) at Lawrence Berkeley National Laboratory. “It’s becoming necessary for certain experiments to require real-time data processing, because it will be impossible to make any decisions on a small subset of data, or to save all this data for post-processing.” This rapid feedback capability could inform on-the-fly adjustments to experiments, better utilizing LCLS resources.

High-performance computing (HPC) can help. Exascale workflows are able to process bursts of data in short periods of time, enabling efficient data processing at granular atomic resolution. The Exascale Computing Project’s (ECP’s) ExaFEL utilizes exascale computers to rapidly analyze X-ray diffraction data at SLAC, cutting analysis down from weeks to minutes, while keeping up with future data collection rates.

Collecting Molecular Images

To prevent computational data throughput from becoming the bottleneck in achieving useful experimental data, ExaFEL must deliver results as quickly as possible.

“We’re kind of used to finding workarounds to data transfer problems,” says Blaschke, the principal NERSC liaison to ExaFEL. A prime example is the Event Horizon Telescope’s 2017 delivery of the first image of the supermassive black hole at the center of the Milky Way, Sagittarius A*. Astronomers used commercial airplanes to transport hard drives to combine terabytes of data collected by various telescopes into the image. “These workarounds will not be tenable in the future as detectors and accelerators are getting more powerful.”

In stark contrast to cargo transport, ExaFEL—which provides scientists with data processing code for X-ray nanocrystallography and single-particle imaging—has a data turnaround time of mere minutes. These near-real-time results are important in assuring researchers that their experiment is not going awry, as LCLS operating time and the targets used are both costly.

Because the data rate of LCLS is so large, only HPC systems can handle the vast amount of image processing and complex algorithms required. ExaFEL can also help researchers make stop-motion molecular “movies,” stitching together snapshots of different stages of a molecular interaction at LCLS—a computationally challenging and extremely data intensive process. ExaFEL’s enhanced workflow makes this possible in as little as ten minutes.

“I like to imagine this challenge has a certain opportunity as well,” says Blaschke. “A lot of analysis techniques were not considered in the past because they would require a stupendous amount of data, or large amount of compute. Having this available now means we can finally also start to ask scientific questions we might not have asked in the past, because it would not have been practical.”

All of this would be unimaginable in a pre-HPC environment. Blaschke describes ExaFEL as a particularly intersectional challenge, and ECP brought together diverse scientists—beamline scientists, X-ray scientists, crystallographers, computing experts, and more—to come up with solutions to complicated scientific problems for everyone’s benefit.

“In addition to computing scientists and domain scientists, we need mathematicians. If you’re going to develop a data analysis tool, it’s good to have a mathematician,” Blaschke says. “That’s why I really love working on problems like this—it is taking the state-of-the-art from one area, and helping to make the state-of-the-art in another area better.”

One avenue in which mathematicians were instrumental to ExaFEL was for fast Fourier transforms (FFTs).

A Computing Lens for Diffraction

Fourier transforms are a mathematical tool that breaks a function down into its constituent frequencies, similar to breaking a musical chord down into its individual notes and their intensities. In signal processing, Fourier transforms are a step in reconstructing the structure of a target from its X-ray diffraction pattern, like simulating a lens that brings the diffracted light back together to produce the original image. To achieve this key step, ExaFEL depends on another ECP product, FFTX.

FFTX is one of three main libraries and applications included in ECP’s Co-design Center for Particle Applications (CoPA). Along with FFTX, CoPA—whose products include the Cabana Pariticle Library and PROGRESS/BML Libraries for electronic structure solvers and quantum molecular dynamics algorithm —offers a suite of particle application capabilities for molecular dynamics, fusion simulations, and more.

Standard FFT libraries cannot be easily scaled up to exascale hardware systems. FFTX not only meets this challenge, but also offers additional optimizations. While traditional vendor FFT libraries are often a “black box”—that is, they complete their task without any transparency of their inner workings—FFTX allows users to optimize the entire process that calls FFTs. Rather than forcing scientists to sequentially apply and optimize a set of operations, FFTX uses a code generation system to combine the FFT, any necessary linear operations, and an inverse FFT all into one process.

“The big advantage [of FFTX] is the integrated applications, where you’re combining FFTs with linear operators and optimizing that as a whole, reducing the communication with the computer and speeding it up,” says Peter McCorquodale, a computational scientist at Lawrence Berkeley National Laboratory, who leads the FFTX project. According to McCorquodale, this merging of capabilities speeds up the process by about a factor of four.

For scientists that only need to calculate an FFT, the FFTX system is plug-and-play. But achieving the full process of an optimized FFT along with linear operations requires some integration support from the FFTX team, which the ExaFEL project took advantage of.

In particular, the FFTX team did specific work to make the library useable for ExaFEL. Though the standard FFTX is based on a C++ interface, ExaFEL uses Python, so the team developed a custom Python interface for ExaFEL. Other FFTX integrations include the ECP projects WarpX and NWChemEx, which address plasma accelerator and biofuel problems, respectively.

“These libraries are good not just for the ECP project, but they’re good for other projects—ECP or not—for the future,” says Susan Mniszewski, the principal investigator for CoPA.

Looking Ahead at Exascale Processing

ExaFEL is designed with scalability and portability in mind, making it adaptable to future HPC systems beyond ECP and beyond the exascale era of supercomputing.

With the ongoing increase in scientific data and imaging resolution, resources like ExaFEL and FFTX will continue to extend the types of systems scientists can image and study.


Source: Anashe Bandari, Lawrence Livermore National Laboratory; ECP

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industry updates delivered to you every week!

Google Announces Sixth-generation AI Chip, a TPU Called Trillium

May 17, 2024

On Tuesday May 14th, Google announced its sixth-generation TPU (tensor processing unit) called Trillium.  The chip, essentially a TPU v6, is the company's latest weapon in the AI battle with GPU maker Nvidia and clou Read more…

ISC 2024 Student Cluster Competition

May 16, 2024

The 2024 ISC 2024 competition welcomed 19 virtual (remote) and eight in-person teams. The in-person teams participated in the conference venue and, while the virtual teams competed using the Bridges-2 supercomputers at t Read more…

Grace Hopper Gets Busy with Science 

May 16, 2024

Nvidia’s new Grace Hopper Superchip (GH200) processor has landed in nine new worldwide systems. The GH200 is a recently announced chip from Nvidia that eliminates the PCI bus from the CPU/GPU communications pathway.  Read more…

Europe’s Race towards Quantum-HPC Integration and Quantum Advantage

May 16, 2024

What an interesting panel, Quantum Advantage — Where are We and What is Needed? While the panelists looked slightly weary — their’s was, after all, one of the last panels at ISC 2024 — the discussion was fascinat Read more…

The Future of AI in Science

May 15, 2024

AI is one of the most transformative and valuable scientific tools ever developed. By harnessing vast amounts of data and computational power, AI systems can uncover patterns, generate insights, and make predictions that Read more…

Some Reasons Why Aurora Didn’t Take First Place in the Top500 List

May 15, 2024

The makers of the Aurora supercomputer, which is housed at the Argonne National Laboratory, gave some reasons why the system didn't make the top spot on the Top500 list of the fastest supercomputers in the world. At s Read more…

Google Announces Sixth-generation AI Chip, a TPU Called Trillium

May 17, 2024

On Tuesday May 14th, Google announced its sixth-generation TPU (tensor processing unit) called Trillium.  The chip, essentially a TPU v6, is the company's l Read more…

Europe’s Race towards Quantum-HPC Integration and Quantum Advantage

May 16, 2024

What an interesting panel, Quantum Advantage — Where are We and What is Needed? While the panelists looked slightly weary — their’s was, after all, one of Read more…

The Future of AI in Science

May 15, 2024

AI is one of the most transformative and valuable scientific tools ever developed. By harnessing vast amounts of data and computational power, AI systems can un Read more…

Some Reasons Why Aurora Didn’t Take First Place in the Top500 List

May 15, 2024

The makers of the Aurora supercomputer, which is housed at the Argonne National Laboratory, gave some reasons why the system didn't make the top spot on the Top Read more…

ISC 2024 Keynote: High-precision Computing Will Be a Foundation for AI Models

May 15, 2024

Some scientific computing applications cannot sacrifice accuracy and will always require high-precision computing. Therefore, conventional high-performance c Read more…

Shutterstock 493860193

Linux Foundation Announces the Launch of the High-Performance Software Foundation

May 14, 2024

The Linux Foundation, the nonprofit organization enabling mass innovation through open source, is excited to announce the launch of the High-Performance Softw Read more…

ISC 2024: Hyperion Research Predicts HPC Market Rebound after Flat 2023

May 13, 2024

First, the top line: the overall HPC market was flat in 2023 at roughly $37 billion, bogged down by supply chain issues and slowed acceptance of some larger sys Read more…

Top 500: Aurora Breaks into Exascale, but Can’t Get to the Frontier of HPC

May 13, 2024

The 63rd installment of the TOP500 list is available today in coordination with the kickoff of ISC 2024 in Hamburg, Germany. Once again, the Frontier system at Read more…

Synopsys Eats Ansys: Does HPC Get Indigestion?

February 8, 2024

Recently, it was announced that Synopsys is buying HPC tool developer Ansys. Started in Pittsburgh, Pa., in 1970 as Swanson Analysis Systems, Inc. (SASI) by John Swanson (and eventually renamed), Ansys serves the CAE (Computer Aided Engineering)/multiphysics engineering simulation market. Read more…

Nvidia H100: Are 550,000 GPUs Enough for This Year?

August 17, 2023

The GPU Squeeze continues to place a premium on Nvidia H100 GPUs. In a recent Financial Times article, Nvidia reports that it expects to ship 550,000 of its lat Read more…

Comparing NVIDIA A100 and NVIDIA L40S: Which GPU is Ideal for AI and Graphics-Intensive Workloads?

October 30, 2023

With long lead times for the NVIDIA H100 and A100 GPUs, many organizations are looking at the new NVIDIA L40S GPU, which it’s a new GPU optimized for AI and g Read more…

Choosing the Right GPU for LLM Inference and Training

December 11, 2023

Accelerating the training and inference processes of deep learning models is crucial for unleashing their true potential and NVIDIA GPUs have emerged as a game- Read more…

Shutterstock 1606064203

Meta’s Zuckerberg Puts Its AI Future in the Hands of 600,000 GPUs

January 25, 2024

In under two minutes, Meta's CEO, Mark Zuckerberg, laid out the company's AI plans, which included a plan to build an artificial intelligence system with the eq Read more…

AMD MI3000A

How AMD May Get Across the CUDA Moat

October 5, 2023

When discussing GenAI, the term "GPU" almost always enters the conversation and the topic often moves toward performance and access. Interestingly, the word "GPU" is assumed to mean "Nvidia" products. (As an aside, the popular Nvidia hardware used in GenAI are not technically... Read more…

Nvidia’s New Blackwell GPU Can Train AI Models with Trillions of Parameters

March 18, 2024

Nvidia's latest and fastest GPU, codenamed Blackwell, is here and will underpin the company's AI plans this year. The chip offers performance improvements from Read more…

Shutterstock 1285747942

AMD’s Horsepower-packed MI300X GPU Beats Nvidia’s Upcoming H200

December 7, 2023

AMD and Nvidia are locked in an AI performance battle – much like the gaming GPU performance clash the companies have waged for decades. AMD has claimed it Read more…

Leading Solution Providers

Contributors

Eyes on the Quantum Prize – D-Wave Says its Time is Now

January 30, 2024

Early quantum computing pioneer D-Wave again asserted – that at least for D-Wave – the commercial quantum era has begun. Speaking at its first in-person Ana Read more…

The GenAI Datacenter Squeeze Is Here

February 1, 2024

The immediate effect of the GenAI GPU Squeeze was to reduce availability, either direct purchase or cloud access, increase cost, and push demand through the roof. A secondary issue has been developing over the last several years. Even though your organization secured several racks... Read more…

Intel Plans Falcon Shores 2 GPU Supercomputing Chip for 2026  

August 8, 2023

Intel is planning to onboard a new version of the Falcon Shores chip in 2026, which is code-named Falcon Shores 2. The new product was announced by CEO Pat Gel Read more…

The NASA Black Hole Plunge

May 7, 2024

We have all thought about it. No one has done it, but now, thanks to HPC, we see what it looks like. Hold on to your feet because NASA has released videos of wh Read more…

Some Reasons Why Aurora Didn’t Take First Place in the Top500 List

May 15, 2024

The makers of the Aurora supercomputer, which is housed at the Argonne National Laboratory, gave some reasons why the system didn't make the top spot on the Top Read more…

GenAI Having Major Impact on Data Culture, Survey Says

February 21, 2024

While 2023 was the year of GenAI, the adoption rates for GenAI did not match expectations. Most organizations are continuing to invest in GenAI but are yet to Read more…

How the Chip Industry is Helping a Battery Company

May 8, 2024

Chip companies, once seen as engineering pure plays, are now at the center of geopolitical intrigue. Chip manufacturing firms, especially TSMC and Intel, have b Read more…

Q&A with Nvidia’s Chief of DGX Systems on the DGX-GB200 Rack-scale System

March 27, 2024

Pictures of Nvidia's new flagship mega-server, the DGX GB200, on the GTC show floor got favorable reactions on social media for the sheer amount of computing po Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire