NSF Funds Indiana University Cybersecurity Research

The National Science Foundation is expanding its partnership with Indiana University's Trusted CI to translate cybersecurity research into practice.

digital padlock in concentric circles

Indiana University's Trusted CI center was started as a National Science Foundation Cybersecurity Center of Excellence (CCoE) in 2016. Now, the program is getting a $2.5 million grant to expand its activities into 2019 and 2020.

The funding will position Trusted CI to assist in transitioning cybersecurity research into practice for the NSF community, creating a Cybersecurity Fellows program to broaden the impact of the CCoE and the creation of an Open Science Cybersecurity Framework.

The Indiana University Center for Applied Cybersecurity Research is the lead organization for the NSF CCoE and it was founded in collaboration with the National Center for Supercomputing Applications, the Pittsburgh Supercomputing Center and the University of Wisconsin-Madison. Oklahoma State and the Lawrence Berkeley National Laboratory are joining the NSF CCoE to support Trusted CI's expanded research activities.

"Trusted CI has been leading the NSF community in building a set of technical, policy and cultural best practices necessary to ensure the security of that infrastructure and ensure the trustworthy nature of the science it produces," said Von Welch, director and principal investigator at Trusted CI. "We're very pleased to play an expanded role with our new collaborators in this very important work of protecting NSF-funded research and we thank the NSF for their support."

The Open Science Cybersecurity Framework is a project to help maintain and establish an open science cybersecurity program at any project scale and stage in a project's lifecycle. The framework is built on four pillars: mission alignment, governance, resources and controls. The first version of the framework is expected to be included in NSF's March 2019 version of the Large Facilities Manual to guide infrastructure projects that receive NSF funding.

At Indiana University, NSF is also investing in a new virtual Research Security Operations Center with a $4.9 million NSF grant. The ResearchSOC is distributed across cybersecurity entities from Indiana University, Duke University, the Pittsburgh Supercomputing Center and the University of California San Diego. It will help provide the research and educational community with cybersecurity services, training and information sharing in order to make scientific computing resilient to cyberattacks.

More information about Trusted CI can be found here.

About the Author

Sara Friedman is a reporter/producer for Campus Technology, THE Journal and STEAM Universe covering education policy and a wide range of other public-sector IT topics.

Friedman is a graduate of Ithaca College, where she studied journalism, politics and international communications.

Friedman can be contacted at [email protected] or follow her on Twitter @SaraEFriedman.

Click here for previous articles by Friedman.


Featured

  • glowing digital brain-shaped neural network surrounded by charts, graphs, and data visualizations

    Google Releases Advanced AI Model for Complex Reasoning Tasks

    Google has released Gemini 2.5 Deep Think, an advanced artificial intelligence model designed for complex reasoning tasks.

  • abstract pattern of cybersecurity, ai and cloud imagery

    OpenAI Report Identifies Malicious Use of AI in Cloud-Based Cyber Threats

    A report from OpenAI identifies the misuse of artificial intelligence in cybercrime, social engineering, and influence operations, particularly those targeting or operating through cloud infrastructure. In "Disrupting Malicious Uses of AI: June 2025," the company outlines how threat actors are weaponizing large language models for malicious ends — and how OpenAI is pushing back.

  • cybersecurity book with a shield and padlock

    NIST Proposes New Cybersecurity Guidelines for AI Systems

    The National Institute of Standards and Technology has unveiled plans to issue a new set of cybersecurity guidelines aimed at safeguarding artificial intelligence systems, citing rising concerns over risks tied to generative models, predictive analytics, and autonomous agents.

  • magnifying glass highlighting a human profile silhouette, set over a collage of framed icons including landscapes, charts, and education symbols

    AWS, DeepBrain AI Launch AI-Generated Multimedia Content Detector

    Amazon Web Services (AWS) and DeepBrain AI have introduced AI Detector, an enterprise-grade solution designed to identify and manage AI-generated content across multiple media types. The collaboration targets organizations in government, finance, media, law, and education sectors that need to validate content authenticity at scale.