U Illinois Urbana-Champaign Named CUDA Center of Excellence

The University of Illinois at Urbana-Champaign (UIUC) has been named Nvidia's first CUDA "Center of Excellence" for adopting parallel computing strategies and applying them to research. The university also said it has received $500,000 from Nvidia to support ongoing research and to help develop parallel computing facilities.

CUDA is a suite of tools, including a C-compiler and SDK, for developing multi-core and parallel processing applications on Nvidia's GPUs. Forty-one universities are currently using CUDA for multi-core and parallel processing programming, which is about double the number of universities that were using it seven months ago.

UIUC is using parallel processing for research in computational biophysics and has "successfully accelerated NAMD/VMD--a popular parallel molecular dynamics application that analyzes large biomolecular systems," according to the university.

"We're very excited to partner with Nvidia and anticipate that together we will achieve breakthroughs in biomedicine, leading to a better understanding of disease and more effective treatments," said Klaus Schulten, Swanlund Professor of Physics and director of the Theoretical and Computational Biophysics Group at UIUC, in a statement released Monday. "This generous gift will be a great stimulus for Illinois' team of outstanding young programmers. It will help to extend their ranks and equip them with the necessary tools to advance computing in decades to come."

Universities that wish to become CUDA Centers of Excellence include CUDA technologies in their research efforts and teach at least one course in CUDA.

Around this time last year, UIUC partnered with Nvidia to offer a course in parallel computing such tools as Nvidia's CUDA C-programming environment taught by both the chair of the department of Electrical and Computer Engineering and the chief scientist at Nvidia, David Kirk.

Those institutions that are accepted as CUDA Centers of Excellence receive funding, equipment donations, and support for setting up a GPU computing cluster. In addition to the $500,000 donation it received this week, UIUC has received about $800,000 in equipment in the form of a 16-node CUDA technology cluster administered by UIUC's National Center for Supercomputing Applications (NCSA).

Further information on CUDA can be found here.

About the Author

David Nagel is the former editorial director of 1105 Media's Education Group and editor-in-chief of THE Journal, STEAM Universe, and Spaces4Learning. A 30-year publishing veteran, Nagel has led or contributed to dozens of technology, art, marketing, media, and business publications.

He can be reached at [email protected]. You can also connect with him on LinkedIn at https://www.linkedin.com/in/davidrnagel/ .


Featured

  • A panel discussion from SXSW EDU 2025

    12 Ways to Dive into AI at SXSW EDU

    This March 9-12, the SXSW EDU Conference & Festival returns to Austin, TX, to celebrate innovation, experimentation, and learning across every stage of education.

  • glowing crystal ball with network connections

    Call for Opinions: 2026 Predictions for Higher Ed IT

    How will the technology landscape in higher education change in the coming year? We're inviting our readership to weigh in with their predictions, wishes, or worries for 2026.

  • glowing brain above stacked coins

    The Higher Ed Playbook for AI Affordability

    Fulfilling the promise of AI in higher education does not require massive budgets or radical reinvention. By leveraging existing infrastructure, embracing edge and localized AI, collaborating across institutions, and embedding AI thoughtfully across the enterprise, universities can move from experimentation to impact.

  • AI word on microchip and colorful light spread

    Microsoft Unveils Maia 200 Inference Chip to Cut AI Serving Costs

    Microsoft recently introduced Maia 200, a custom-built accelerator aimed at lowering the cost of running artificial intelligence workloads at cloud scale, as major providers look to curb soaring inference expenses and lessen dependence on Nvidia graphics processors.