Leveraging Machine Learning in Dark Matter Research for the Aurora Exascale System 

By Linda Barney 

September 25, 2023

Scientists have unlocked many secrets about particle interactions at atomic and subatomic levels. However, one mystery that has eluded researchers is dark matter. Current supercomputers don’t have the capability to run custom machine learning (ML) architectures that can tackle calculating the properties and interactions of large atomic nuclei needed to help in solving this mystery. 

A team of scientists, using supercomputing resources at the US Department of Energy’s (DOE) Argonne National Laboratory are conducting research seeking to find answers to this puzzle with simulations using lattice quantum chromodynamics (LQCD). The team is developing novel ML algorithms to determine possible interactions between nuclei and a large class of dark matter candidate particles. The goal of the research is to enable calculations in nuclear physics on the Aurora exascale high performance computer (HPC) system that are not computationally possible with traditional approaches on existing petascale supercomputers.  

Introducing the dark matter research team  

The research team conducting the dark matter research includes Dr. William Detmoldand Dr. Phiala Shanahanas co-principal investigators from MIT. The team also includes researchers at New York University, and their collaborators in Europe and at Argonne National Laboratory. The researchers are members of the U.S. Lattice Quantum Chromodynamics (USQCD) collaboration, a national infrastructure for LQCD hardware and software. The ESP research is supported by the US Department of Energy and National Science Foundation. Software development is funded by a grant from the Department of Energy Scientific Discovery through Advanced Computing (SciDAC)program. The project is an awardee in the Argonne Leadership Computing Facility’s (ALCF) Early Science Program (ESP) for Aurora. 

Dark matter research based on the standard model of particle physics 

The standard model of particle physics is a theory that seeks to explain the strong force and weak force as well as electromagnetism, but doesn’t include gravity. Based on past standard model research, scientists currently understand that protons and atomic nuclei are made up of quarks and gluons that are the fundamental building blocks of the universe. Dark matter is the name applied to unknown matter in the universe that has not been detected by current scientific instruments but is inferred based on its gravitational effects.  

Detmold states, “When we talk about the standard model, we focus on things that at a very small scale– smaller than the atom basically. Our ESP team’s research is based on the theory of quantum chromodynamics (QCD), which explains the way quarks interact with one another inside the nucleus of an atom. We use LQCD simulations related to contemporary physics experiments to try to understand how those interactions work to determine the atomic constituents and their potential interactions with dark matter.”  

Machine learning software developed for dark matter research 

The team developed their own ML software for dark matter research to solve some of the challenging computational tasks. Detmold states, “There are big computational bottlenecks in certain parts of the LQCD calculation. Our ML software is designed to speed up HPC algorithms in parts of the LQCD calculation such as matrix inversions and big linear algebra calculations.”  

The team’s ML algorithm is optimized to take advantage of other software tools such as USQCD libraries, TensorFlow, HDF5, and PyTorch. The ML software uses a self-training method where the model generates samples of typical configurations of quarks and gluons, and the program learns from the samples to more accurately generate new samples.  

According to Dr. Shanahan, “Our team is developing novel machine-learning algorithms to enable next-generation lattice QCD calculations of nuclear physics on Aurora.” Parts of the LQCD calculations can only run on large-scale supercomputers. On Aurora, the team will calculate convolutions that work in four dimensions. Detmold indicates that dealing with a four-dimensional structure makes calculations a more challenging numerical problem and requires new software development. 

The numerical calculations use a spacetime lattice (grid) to determine the properties and interactions of the nuclei, including their potential interactions with dark matter. The researchers initially examine a small volume and increase the region to larger volumes and extrapolate the results to the infinite box size limit.   

Figure 1. Artist’s impression of a nucleus in a lattice QCD calculation. Courtesy of Argonne National Laboratory.

Preparing for work on Aurora  

In preparation for work on the future Aurora supercomputer, the team has historically worked on petascale supercomputers including ALCF Mira and Theta, Summit at ORNL, and Marconi at CINECA in Italy.  

The future Aurora supercomputer architecture is designed to optimize deep learning and the ML software stack will run at scale. Aurora will incorporate new Intel compute engines, including the Intel Xeon CPU Max series, and the Intel Data Center GPU Max series, as well as DAOS storage. Aurora will take advantage of the Intel-led cross-industry oneAPI initiative designed to unify and simplify application development across diverse computing architectures. Detmold indicates that HPC researchers need tools such as oneAPI to save time. 

Summary  

Dark matter research is computationally challenging and there are many unanswered questions. A team of researchers is doing dark matter research as part of the Argonne ESP program. They are developing novel machine-learning (ML) algorithms to determine possible interactions between nuclei and a large class of dark matter candidate particles. 

Being able to move to an exascale HPC system will allow the team to perform research that is not currently possible on petascale supercomputers. Detmold indicates that having access to an exascale supercomputer will help the team compare their numerical LQCD calculations against physical dark matter experiments as well as predictions from the standard model of particle physics to learn more about interactions at a subatomic particle level.  

“Aurora will enable us to scale-up and deploy custom machine learning architectures developed for physics to the exascale for the first time. The hope is that this will enable calculations in nuclear physics that are not computationally tractable with traditional approaches, but it will also represent the first at-scale application of machine learning at all in this context,” states Shanahan.  

The ALCF is a DOE Office of Science User Facility. 

References 

Linda Barney is the founder and owner of Barney and Associates, a technical/marketing writing, training, and web design firm in Beaverton, OR. 


This article was produced as part of Intel’s editorial program, with the goal of highlighting cutting-edge science, research and innovation driven by the HPC and AI communities through advanced technology. The publisher of the content has final editing rights and determines what articles are published. 

 

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industry updates delivered to you every week!

IonQ Plots Path to Commercial (Quantum) Advantage

July 2, 2024

IonQ, the trapped ion quantum computing specialist, delivered a progress report last week firming up 2024/25 product goals and reviewing its technology roadmap. Next up on the product roadmap is Forte Enterprise, intende Read more…

Best Networking Experience on the Planet: Join the 2024 SCinet CommUNITY Program

July 1, 2024

Join the SC24 SCinet team in Atlanta, GA, and learn high-performance networking while you network with high-performance people! Applications close July 15. Apply Now The CommUNITY@SC24 Professional Development program Read more…

Nvidia Economics: Make $5-$7 for Every $1 Spent on GPUs

June 30, 2024

Nvidia is saying that companies could make $5 to $7 for every $1 invested in GPUs over a four-year period. Customers are investing billions in new Nvidia hardware to keep up with newer AI models to drive revenue and prod Read more…

Four Steps to Ensure GenAI Safety and Ethics

June 27, 2024

With the deployment of generative artificial intelligence (GenAI) happening at a rapid pace, organizations of all sizes are tasked with navigating the challenges around implementation, especially regarding ethics and Read more…

AI-augmented HPC and the Inflation of Science and Technology

June 27, 2024

Everyone is aware of the inflationary model of the early universe in which the volume of space expands exponentially then slows down. AI-augmented HPC (AHPC for short) has started to expand creating new space in the scie Read more…

Top Three Pitfalls to Avoid When Processing Data with LLMs

June 26, 2024

It’s a truism of data analytics: when it comes to data, more is generally better. But the explosion of AI-powered large language models (LLMs) like ChatGPT and Google Gemini (formerly Bard) challenges this conventional Read more…

IonQ Plots Path to Commercial (Quantum) Advantage

July 2, 2024

IonQ, the trapped ion quantum computing specialist, delivered a progress report last week firming up 2024/25 product goals and reviewing its technology roadmap. Read more…

Shutterstock_1687123447

Nvidia Economics: Make $5-$7 for Every $1 Spent on GPUs

June 30, 2024

Nvidia is saying that companies could make $5 to $7 for every $1 invested in GPUs over a four-year period. Customers are investing billions in new Nvidia hardwa Read more…

Shutterstock 2338659951

AI-augmented HPC and the Inflation of Science and Technology

June 27, 2024

Everyone is aware of the inflationary model of the early universe in which the volume of space expands exponentially then slows down. AI-augmented HPC (AHPC for Read more…

Summer Reading: DARPA Showcases Quantum Benchmarking Progress

June 25, 2024

Last week, the Defense Advanced Research Projects Agency (DARPA) issued an interim progress update from the second phase of its Quantum Benchmark (QB) program. Read more…

Spelunking the HPC and AI GPU Software Stacks

June 21, 2024

As AI continues to reach into every domain of life, the question remains as to what kind of software these tools will run on. The choice in software stacks – Read more…

HPE and NVIDIA Join Forces and Plan Conquest of Enterprise AI Frontier

June 20, 2024

The HPE Discover 2024 conference is currently in full swing, and the keynote address from Hewlett-Packard Enterprise (HPE) CEO Antonio Neri on Tuesday, June 18, Read more…

Slide Shows Samsung May be Developing a RISC-V CPU for In-memory AI Chip

June 19, 2024

Samsung may have unintentionally revealed its intent to develop a RISC-V CPU, which a presentation slide showed may be used in an AI chip. The company plans to Read more…

Qubits 2024: D-Wave’s Steady March to Quantum Success

June 18, 2024

In his opening keynote at D-Wave’s annual Qubits 2024 user meeting, being held in Boston, yesterday and today, CEO Alan Baratz again made the compelling pitch Read more…

Atos Outlines Plans to Get Acquired, and a Path Forward

May 21, 2024

Atos – via its subsidiary Eviden – is the second major supercomputer maker outside of HPE, while others have largely dropped out. The lack of integrators and Atos' financial turmoil have the HPC market worried. If Atos goes under, HPE will be the only major option for building large-scale systems. Read more…

Comparing NVIDIA A100 and NVIDIA L40S: Which GPU is Ideal for AI and Graphics-Intensive Workloads?

October 30, 2023

With long lead times for the NVIDIA H100 and A100 GPUs, many organizations are looking at the new NVIDIA L40S GPU, which it’s a new GPU optimized for AI and g Read more…

Everyone Except Nvidia Forms Ultra Accelerator Link (UALink) Consortium

May 30, 2024

Consider the GPU. An island of SIMD greatness that makes light work of matrix math. Originally designed to rapidly paint dots on a computer monitor, it was then Read more…

Nvidia H100: Are 550,000 GPUs Enough for This Year?

August 17, 2023

The GPU Squeeze continues to place a premium on Nvidia H100 GPUs. In a recent Financial Times article, Nvidia reports that it expects to ship 550,000 of its lat Read more…

Nvidia’s New Blackwell GPU Can Train AI Models with Trillions of Parameters

March 18, 2024

Nvidia's latest and fastest GPU, codenamed Blackwell, is here and will underpin the company's AI plans this year. The chip offers performance improvements from Read more…

Some Reasons Why Aurora Didn’t Take First Place in the Top500 List

May 15, 2024

The makers of the Aurora supercomputer, which is housed at the Argonne National Laboratory, gave some reasons why the system didn't make the top spot on the Top Read more…

Choosing the Right GPU for LLM Inference and Training

December 11, 2023

Accelerating the training and inference processes of deep learning models is crucial for unleashing their true potential and NVIDIA GPUs have emerged as a game- Read more…

Nvidia Shipped 3.76 Million Data-center GPUs in 2023, According to Study

June 10, 2024

Nvidia had an explosive 2023 in data-center GPU shipments, which totaled roughly 3.76 million units, according to a study conducted by semiconductor analyst fir Read more…

Leading Solution Providers

Contributors

Synopsys Eats Ansys: Does HPC Get Indigestion?

February 8, 2024

Recently, it was announced that Synopsys is buying HPC tool developer Ansys. Started in Pittsburgh, Pa., in 1970 as Swanson Analysis Systems, Inc. (SASI) by John Swanson (and eventually renamed), Ansys serves the CAE (Computer Aided Engineering)/multiphysics engineering simulation market. Read more…

Intel’s Next-gen Falcon Shores Coming Out in Late 2025 

April 30, 2024

It's a long wait for customers hanging on for Intel's next-generation GPU, Falcon Shores, which will be released in late 2025.  "Then we have a rich, a very Read more…

Google Announces Sixth-generation AI Chip, a TPU Called Trillium

May 17, 2024

On Tuesday May 14th, Google announced its sixth-generation TPU (tensor processing unit) called Trillium.  The chip, essentially a TPU v6, is the company's l Read more…

AMD Clears Up Messy GPU Roadmap, Upgrades Chips Annually

June 3, 2024

In the world of AI, there's a desperate search for an alternative to Nvidia's GPUs, and AMD is stepping up to the plate. AMD detailed its updated GPU roadmap, w Read more…

The NASA Black Hole Plunge

May 7, 2024

We have all thought about it. No one has done it, but now, thanks to HPC, we see what it looks like. Hold on to your feet because NASA has released videos of wh Read more…

Shutterstock_1687123447

Nvidia Economics: Make $5-$7 for Every $1 Spent on GPUs

June 30, 2024

Nvidia is saying that companies could make $5 to $7 for every $1 invested in GPUs over a four-year period. Customers are investing billions in new Nvidia hardwa Read more…

Q&A with Nvidia’s Chief of DGX Systems on the DGX-GB200 Rack-scale System

March 27, 2024

Pictures of Nvidia's new flagship mega-server, the DGX GB200, on the GTC show floor got favorable reactions on social media for the sheer amount of computing po Read more…

AMD MI3000A

How AMD May Get Across the CUDA Moat

October 5, 2023

When discussing GenAI, the term "GPU" almost always enters the conversation and the topic often moves toward performance and access. Interestingly, the word "GPU" is assumed to mean "Nvidia" products. (As an aside, the popular Nvidia hardware used in GenAI are not technically... Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire