Purdue’s Anvil Helps Undergraduates Gain Real-World HPC Experience

February 14, 2024

Feb. 14, 2024 — Five students from the University of Illinois Urbana-Champaign (UIUC) took part in a semester-long research experience program to gain practical knowledge on data analytics and statistical research using common computing HPC resources.

From left: Austin Shwatal, Georges Durand, Spencer Bauer, Mallory Klostermann, and Sarah Yang.

Daniel Ries, a Principal Data Scientist at Sandia National Laboratories, led the project, teaching undergraduates in statistics and computer science the skills they will need once entering the workforce.

According to Ries, the intent behind this research experience was multifaceted—he wanted to utilize a real-world research problem to teach students how to scope, execute and refine, and draw conclusions, apply knowledge gained from coursework to actual research, and—importantly—create reproducible research on a common computing platform. By obtaining access to the Anvil supercomputer, the students were able to accomplish all of this and more.

“Getting the students on Anvil was not only a benefit in terms of reproducibility, but in terms of what these students will be doing when they either go to grad school or get a full-time job in the data science world,” said Ries. “Most of the work that’s done at a company, at a research lab, in academia—the computing is done on servers. You don’t do computing on your own laptop or your own computer anymore. Just given the scale of models, the scale of data, it’s very common to have to get used to working in a server environment, a Linux environment, things like that, and I don’t think actually any of the students had experience with that. So it actually turned out to be a very good experience for them.”

Learning how to work on a common computing platform such as Anvil may have been initially daunting, but the students quickly proved they were up to the task.

“For the first couple weeks, it might have been a headache for the students,” said Ries, “But it’s the price they pay to actually learn how to do this and something they’re probably going to be doing at a real job later. They’ll be able to say, ‘Hey, I’m maybe not an expert in Linux scripting, but I can work my way around a server.’”

Not only did the students get hands-on HPC experience in an actual research application, but the research itself had practical implications. The group set their sights on a mode of predictive modeling known as “nowcasting.” With nowcasting, a research team is looking to predict weather conditions in the near future based on conditions in the very recent past.

For part of this research project, Ries wanted the undergraduate students to build a predictive model that could determine where lightning would strike in the next 15 to 60 minutes—a type of nowcasting that is immensely useful across multiple sectors. Lightning is dangerous. Being able to forecast where it will strike in real-time has safety implications (recreational, construction, power line workers, etc.) as well as economic implications—where forest fires are likely to start, crop damage, house/building damage, etc. This is precisely the type of work that Ries does in his role at Sandia National Laboratories, which highlights just how important an opportunity like this can be for the students.

“I thought that problem was very interesting, and so I was able to kind of tie it to a problem that the undergraduates could actually work on,” said Ries. “And, so, what they were doing—some of those results—I’ll very much be able to actually use them to further some of the work that I’m doing. So it wasn’t just work, you know, to create a flashy poster.”

The five students worked on three different models, focusing their efforts on the upper Midwest region. Using data collected from multiple sources—lightning information from the National Lightning Detection Network (NLDN) and remote sensing data from the GOES-16 Advanced Baseline Imager—the students were able to develop two traditional statistical models and a third, U-Net deep learning model.

The two traditional models, while typically not memory or computationally expensive, benefitted from the use of Anvil due to the sheer size of the data sets. And the U-Net model was trained on the Anvil GPUs, saving the team an enormous amount of time (30-60 minutes per training run versus a day or more without). By the end of the semester, the students successfully developed all three models.

“[The students] definitely created models that had the ability to predict,” said Ries. “I was definitely impressed with what they were able to do. In particular, both of the statistical models that they kind of approached ended up being actually more sophisticated than I anticipated at the beginning of the semester.”

Ries continued: “In terms of the U-Net deep learning model, that was actually very much a long shot. It was something that I had read about—these are being implemented by other research institutes and even commercial entities to try to nowcast lightning and other weather phenomenon. And I had never tried them, so I said, ‘Well, let’s see what these undergraduates can do.’ And they largely took it upon themselves. Other than me explaining it at a high level, they kind of did everything from there, so I was really impressed that they were able to get it working. They were able to transfer U-Net to use in different applications—the code and the data formats; everything—and bring it over to the application we were doing with lightning and the types of data we were working with. They did all of that, and I was really impressed they were able to do that in such a short amount of time.”

Throughout the semester, the students learned the importance of consistency and reproducibility in research and how to take advantage of high-performance computing. They also gained experience presenting research—the five undergraduates participated in a research symposium at U of I, and showcased their work in a smaller presentation to the U of I Statistics Department.

Overall, the research experience was a tremendous success, greatly benefiting the students, and Ries was thrilled with the result. He hopes to continue this type of research experience in the future, and make it a two-semester program instead of just one.

For more information on the types of research conducted at Sandia, please visit the Sandia National Laboratories website.

To learn more about HPC and how it can help you, please visit the “Why HPC?” page.


Source: Purdue University Rosen Center for Advanced Computing

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industry updates delivered to you every week!

Ayar Labs CEO: Optical Chiplets Coming to SOCs Soon

October 22, 2024

In AI, time is money. Top AI players are spending billions to create computing infrastructures to satisfy that need for speed. However, these companies are bottlenecked by computing constraints at the chip, memory, and I Read more…

Quantum Nuggets: Riverlane’s 2024 QEC Study, IBM’s V-score, Twisted Semiconductors

October 22, 2024

Quantum error correction specialist Riverlane today released a fascinating report — The Quantum Error Correction Report 2024 — that’s worth scanning; IBM and 28 collaborators last week released V-score, a new metri Read more…

Celebrating Intel-AMD Unity: Looking Back at x86 Flubs

October 21, 2024

Both AMD and Intel were founded in the Disco era, and it took them decades to establish a brotherhood to protect the long-term interests of x86. But what took them so long? There's a history of bad blood between the riva Read more…

HPC Debrief: Matthew Shaxted, CEO of Parallel Works

October 20, 2024

In this installment of The HPC Debrief, we will discuss a big topic in HPC -- cluster provisioning. Getting hardware on-prem or in the cloud is often the easy part of standing up an HPC or AI cluster. Indeed, cloud techn Read more…

Nvidia, Google to Speak About RISC-V Use at Annual Summit

October 19, 2024

Nvidia will discuss how it uses the RISC-V architecture at the RISC-V Summit from October 22 to 24. The GPU maker has used the RISC-V CPU architecture in its GPU microcontrollers for nine years. A 20-minute keynote from Read more…

In This Club, You Must ‘Earn the Exa’

October 17, 2024

There have been some recent press releases and headlines with the phrase "AI Exascale" in them. Other than flaunting the word exascale or even zettascale, these stories do not provide enough information to justify using Read more…

Ayar Labs CEO: Optical Chiplets Coming to SOCs Soon

October 22, 2024

In AI, time is money. Top AI players are spending billions to create computing infrastructures to satisfy that need for speed. However, these companies are bott Read more…

Celebrating Intel-AMD Unity: Looking Back at x86 Flubs

October 21, 2024

Both AMD and Intel were founded in the Disco era, and it took them decades to establish a brotherhood to protect the long-term interests of x86. But what took t Read more…

HPC Debrief: Matthew Shaxted, CEO of Parallel Works

October 20, 2024

In this installment of The HPC Debrief, we will discuss a big topic in HPC -- cluster provisioning. Getting hardware on-prem or in the cloud is often the easy p Read more…

In This Club, You Must ‘Earn the Exa’

October 17, 2024

There have been some recent press releases and headlines with the phrase "AI Exascale" in them. Other than flaunting the word exascale or even zettascale, these Read more…

Research Insights, HPC Expertise, Meaningful Collaborations Abound at TACCSTER 2024

October 17, 2024

It's a wrap! The Texas Advanced Computing Center (TACC) at UT Austin welcomed more than 100 participants for the 7th annual TACC Symposium for Texas Researchers Read more…

Nvidia’s Blackwell Platform Powers AI Progress in Open Compute Project

October 16, 2024

Nvidia announced it has contributed foundational elements of its Blackwell accelerated computing platform design to the Open Compute Project (OCP). Shared at th Read more…

On Paper, AMD’s New MI355X Makes MI325X Look Pedestrian

October 15, 2024

Advanced Micro Devices has detailed two new GPUs that unambiguously reinforce it as the only legitimate GPU alternative to Nvidia. AMD shared new facts on its n Read more…

Nvidia Is Increasingly the Secret Sauce in AI Deployments, But You Still Need Experience

October 14, 2024

I’ve been through a number of briefings from different vendors from IBM to HP, and there is one constant: they are all leaning heavily on Nvidia for their AI Read more…

Shutterstock_2176157037

Intel’s Falcon Shores Future Looks Bleak as It Concedes AI Training to GPU Rivals

September 17, 2024

Intel's Falcon Shores future looks bleak as it concedes AI training to GPU rivals On Monday, Intel sent a letter to employees detailing its comeback plan after Read more…

Granite Rapids HPC Benchmarks: I’m Thinking Intel Is Back (Updated)

September 25, 2024

Waiting is the hardest part. In the fall of 2023, HPCwire wrote about the new diverging Xeon processor strategy from Intel. Instead of a on-size-fits all approa Read more…

Ansys Fluent® Adds AMD Instinct™ MI200 and MI300 Acceleration to Power CFD Simulations

September 23, 2024

Ansys Fluent® is well-known in the commercial computational fluid dynamics (CFD) space and is praised for its versatility as a general-purpose solver. Its impr Read more…

Shutterstock_1687123447

Nvidia Economics: Make $5-$7 for Every $1 Spent on GPUs

June 30, 2024

Nvidia is saying that companies could make $5 to $7 for every $1 invested in GPUs over a four-year period. Customers are investing billions in new Nvidia hardwa Read more…

Shutterstock 1024337068

Researchers Benchmark Nvidia’s GH200 Supercomputing Chips

September 4, 2024

Nvidia is putting its GH200 chips in European supercomputers, and researchers are getting their hands on those systems and releasing research papers with perfor Read more…

AMD Clears Up Messy GPU Roadmap, Upgrades Chips Annually

June 3, 2024

In the world of AI, there's a desperate search for an alternative to Nvidia's GPUs, and AMD is stepping up to the plate. AMD detailed its updated GPU roadmap, w Read more…

Nvidia Shipped 3.76 Million Data-center GPUs in 2023, According to Study

June 10, 2024

Nvidia had an explosive 2023 in data-center GPU shipments, which totaled roughly 3.76 million units, according to a study conducted by semiconductor analyst fir Read more…

Comparing NVIDIA A100 and NVIDIA L40S: Which GPU is Ideal for AI and Graphics-Intensive Workloads?

October 30, 2023

With long lead times for the NVIDIA H100 and A100 GPUs, many organizations are looking at the new NVIDIA L40S GPU, which it’s a new GPU optimized for AI and g Read more…

Leading Solution Providers

Contributors

xAI Colossus: The Elon Project

September 5, 2024

Elon Musk's xAI cluster, named Colossus (possibly after the 1970 movie about a massive computer that does not end well), has been brought online. Musk recently Read more…

IBM Develops New Quantum Benchmarking Tool — Benchpress

September 26, 2024

Benchmarking is an important topic in quantum computing. There’s consensus it’s needed but opinions vary widely on how to go about it. Last week, IBM introd Read more…

Intel Customizing Granite Rapids Server Chips for Nvidia GPUs

September 25, 2024

Intel is now customizing its latest Xeon 6 server chips for use with Nvidia's GPUs that dominate the AI landscape. The chipmaker's new Xeon 6 chips, also called Read more…

Quantum and AI: Navigating the Resource Challenge

September 18, 2024

Rapid advancements in quantum computing are bringing a new era of technological possibilities. However, as quantum technology progresses, there are growing conc Read more…

IonQ Plots Path to Commercial (Quantum) Advantage

July 2, 2024

IonQ, the trapped ion quantum computing specialist, delivered a progress report last week firming up 2024/25 product goals and reviewing its technology roadmap. Read more…

US Implements Controls on Quantum Computing and other Technologies

September 27, 2024

Yesterday the Commerce Department announced export controls on quantum computing technologies as well as new controls for advanced semiconductors and additive Read more…

Google’s DataGemma Tackles AI Hallucination

September 18, 2024

The rapid evolution of large language models (LLMs) has fueled significant advancement in AI, enabling these systems to analyze text, generate summaries, sugges Read more…

Microsoft, Quantinuum Use Hybrid Workflow to Simulate Catalyst

September 13, 2024

Microsoft and Quantinuum reported the ability to create 12 logical qubits on Quantinuum's H2 trapped ion system this week and also reported using two logical qu Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire