Community College Researching Use of Online Discussion Tool to Engage Underrepresented Students

Through a partnership with online discussion platform Packback, Ivy Tech Community College is exploring the potential of discussion activities to help address equity issues. The goal: "identifying the most effective pedagogical approaches and interactions to improve engagement for students from historically underrepresented backgrounds," according to a news announcement.

Packback uses artificial intelligence to provide students with a real-time assessment of their discussion contributions. It measures factors such as curiosity, credibility, communication and convention to quantify the effectiveness of a discussion post, and allows students to revise and improve their work before final submission.

Initial pilots of the technology have suggested that classes who use Packback are more likely to be engaged in discussion, and are also more likely to report higher levels of satisfaction and higher overall course grades, the company said. For instance, results from the first cohort of research at Ivy Tech found that the use of the technology correlated with a 13.39 percent increase in the number of students earning A, B, or C grades.

"The lasting impact of the COVID-19 pandemic has made it more important than ever to ensure that the online learning experience is as effective — and fulfilling — as in-person programs," commented Nicole Duttlinger, adjunct faculty member at Ivy Tech, in a statement. "Packback's technology, including its automatic feedback system, is rooted in sound pedagogy and a commitment to fostering students' intrinsic motivation. Their platform has boosted participation rates in four sections, and enabled me to focus less on discussion logistics and more on course content and mastery."

"We are really excited about the research aspect of this partnership — and especially its potential to help close persistent equity gaps," said Matthew Pittman, assistant vice president for Distance Learning & Educational Technology at Ivy Tech. "Our initial pilot has already provided critical insights into engagement in discussion activities, particularly among different demographic groups. That data can help provide direction in our course design as we look to address inequalities as part of our strategic plan. As we continue this work with Packback, we are looking forward to leveraging this sort of data and insight to begin addressing the disparities we see among different populations in our online courses and beyond."

About the Author

Rhea Kelly is editor in chief for Campus Technology, THE Journal, and Spaces4Learning. She can be reached at [email protected].

Featured

  • pattern featuring interconnected lines, nodes, lock icons, and cogwheels

    Red Hat Enterprise Linux 9.5 Expands Automation, Security

    Open source solution provider Red Hat has introduced Red Hat Enterprise Linux (RHEL) 9.5, the latest version of its flagship Linux platform.

  • glowing lines connecting colorful nodes on a deep blue and black gradient background

    Juniper Launches AI-Native Networking and Security Management Platform

    Juniper Networks has introduced a new solution that integrates security and networking management under a unified cloud and artificial intelligence engine.

  • a digital lock symbol is cracked and breaking apart into dollar signs

    Ransomware Costs Schools Nearly $550,000 per Day of Downtime

    New data from cybersecurity research firm Comparitech quantifies the damage caused by ransomware attacks on educational institutions.

  • landscape photo with an AI rubber stamp on top

    California AI Watermarking Bill Garners OpenAI Support

    ChatGPT creator OpenAI is backing a California bill that would require tech companies to label AI-generated content in the form of a digital "watermark." The proposed legislation, known as the "California Digital Content Provenance Standards" (AB 3211), aims to ensure transparency in digital media by identifying content created through artificial intelligence. This requirement would apply to a broad range of AI-generated material, from harmless memes to deepfakes that could be used to spread misinformation about political candidates.