<p>Get Ready to Quantify Your Computing and Network Capacity</p>

Every so often the National Science Foundation (NSF) conducts a survey of science and engineering research infrastructure at research-performing colleges and universities. This tends to be a biennial project and is a critically important and, to some, forbidding task. Those responsible for managing space on campuses are now getting ready for the call this coming October, when the NSF will once again unveil its redesigned survey and collect data.

But this time there’s a surprise! In recognition of the fact that “infrastructure” means more, now, than mortar and bricks, there’s a new Part 2 to the survey. It will call for institutions to report on their computing and networking capacity. So, CIOs and other IT staff can expect to hear in late October – from whomever their president appoints as “institutional coordinator” for collection and provision – of some pretty interesting statistics. If you’re not already expecting it, this NSF request will be using up some as-yet-unbudgeted-for staff time.

There’s not a lot on the NSF website about the details of this year’s infrastructure survey yet. If you’re inclined, you can see the 2001 survey’s methodology and the report based on the survey from 2001. Even though that survey collected only physical space data, you might find it useful to take a look. And for some early thoughts on the addition of computer and networking capacity to the survey, you can read this article, “Science Board Mulls Cyber Infrastructure,” from 2000.

We were privileged to attend a conference session recently which presented the methodology for the creation of the new survey and described the data which will be collected and the process for that data collection. We can’t share specific survey questions yet, but you’ll be asked to provide information such as the number of networks, the nodes on the networks, wireless hot spots, speed of connections, and the like. Items such as “advanced computing resources; digital libraries; shared data and information bases; research and education networks; distributed user facilities; and standards and protocols” are the focus of the new Part 2 of this survey.

For some, the collection of these data will present quite a challenge. And it is likely that IT staff will have to work more closely with other campus departments in ways they perhaps do not work now. Although we can’t tell you the details of what you’ll be asked for, we can describe a bit of the process.

* Institutional presidents will receive a letter, probably in October, from the NSF providing detailed survey information and asking them to designate an “institutional coordinator.”

* Campuses will be expected to designate that person and to register with NSF within two weeks. Then a familiarization program will begin.

* Coordinators will designate someone to respond to the computing and networking capacity section, likely a request that will come through the CIO’s office. However, the appointed coordinator, not the IT staff, will be the person who actually holds the Web-based survey “keys” and submit the data to NSF.

* It is expected that reported data will be available through the NSF by as early as February 2004.

In the past, there was significant confidentiality that kept most data from being identified as from specific institutions when it was reported. The intent for the 2003 survey is to make much more of it available in an institution-specific manner – to make it more useful for inter-institutional peer comparisons and the like.

So . . . we don’t know the details of the data which will be requested, just some generalities. But it is clear that some fairly senior IT staff on your campus will find a “new” project on their plates by November.

It might be a good idea to query your president’s office or your space management or institutional research staff to see who the institutional coordinator was in 2001, since that person, or someone in their office, may well be designated again in 2003. If you get those lines of communication open now, you’ll likely save some time and stress later.

Featured

  • AI robot with cybersecurity symbol on its chest

    Microsoft Adds New Agentic AI Tools to Security Copilot

    Microsoft has announced a major expansion of its AI-powered cybersecurity platform, introducing a suite of autonomous agents to help organizations counter rising threats and manage the growing complexity of cloud and AI security.

  • modern college building with circuit and brain motifs

    Anthropic Launches Claude for Education

    Anthropic has announced a version of its Claude AI assistant tailored for higher education institutions. Claude for Education "gives academic institutions secure, reliable AI access for their entire community," the company said, to enable colleges and universities to develop and implement AI-enabled approaches across teaching, learning, and administration.

  • central cloud platform connected to various AI icons—including a brain, robot, and network nodes

    Linux Foundation to Host Protocol for AI Agent Interoperability

    The Linux Foundation has announced it will host the Agent2Agent (A2A) protocol project, an open standard originally developed by Google to support secure communication and interoperability among AI agents.

  • open laptop in a college classroom with holographic AI icons like a brain and data charts rising from the screen

    4 Ways Universities Are Using Google AI Tools for Learning and Administration

    In a recent blog post, Google shared an array of education customer stories, showcasing ways institutions are using AI tools like Gemini and NotebookLM to transform both learning and administrative tasks.