Online Discussion Platform Adds AI-Powered Feedback Tool

Packback, an online platform that uses artificial intelligence to facilitate class discussions, has introduced an Instant Feedback feature that provides students with a real-time assessment of their contributions and allows them to revise and improve their work before final submission.

When students create a discussion question or response, the Instant Feedback tool uses AI algorithms to provide personalized guidance. The tool measures such factors as curiosity (taking into account the "open-endedness" of the question, the depth of a student's description, the repetitiveness of text and more), credibility (whether or not sources are cited), communication and convention. The resulting "Curiosity Score" is meant to quantify the effectiveness of the discussion post.  

"Supplementing instructor-provided feedback with the consistent and instant feedback made possible by AI can give all students the best of both," said Jessica Tenuta, co-founder of Packback, in a statement. Students can apply and learn from that feedback before submitting their work, without any penalty, helping them "develop their skills while gaining a healthier relationship to feedback," she added.

"Meaningful discussion is a critical part of our work to ensure a high-quality online learning experience – and has become even more important as we shift to an online-ready model in the wake of the COVID-19 pandemic," commented John Locke, coordinator of technology-enhanced learning and distance education at SUNY Plattsburgh. "This resource will provide students with the kind of immediate feedback that technology can offer at scale — and enable faculty to dive deeper into substantive, engaging discussions that cultivate both curiosity and confidence."

For more information, visit the Packback site.

About the Author

Rhea Kelly is editor in chief for Campus Technology, THE Journal, and Spaces4Learning. She can be reached at [email protected].

Featured

  • student reading a book with a brain, a protective hand, a computer monitor showing education icons, gears, and leaves

    4 Steps to Responsible AI Implementation

    Researchers at the University of Kansas Center for Innovation, Design & Digital Learning (CIDDL) have published a new framework for the responsible implementation of artificial intelligence at all levels of education.

  • three glowing stacks of tech-themed icons

    Research: LLMs Need a Translation Layer to Launch Complex Cyber Attacks

    While large language models have been touted for their potential in cybersecurity, they are still far from executing real-world cyber attacks — unless given help from a new kind of abstraction layer, according to researchers at Carnegie Mellon University and Anthropic.

  • Hand holding a stylus over a tablet with futuristic risk management icons

    Why Universities Are Ransomware's Easy Target: Lessons from the 23% Surge

    Academic environments face heightened risk because their collaboration-driven environments are inherently open, making them more susceptible to attack, while the high-value research data they hold makes them an especially attractive target. The question is not if this data will be targeted, but whether universities can defend it swiftly enough against increasingly AI-powered threats.

  • magnifying glass revealing the letters AI

    New Tool Tracks Unauthorized AI Usage Across Organizations

    DevOps platform provider JFrog is taking aim at a growing challenge for enterprises: users deploying AI tools without IT approval.