Professor Uses ChatGPT for Lectures, Students Demand Tuition Refund
- May 19, 2025
- 0
In a surprising turn of events, a professor at Northeastern University in Boston has found himself at the center of a growing debate on the ethical use of
In a surprising turn of events, a professor at Northeastern University in Boston has found himself at the center of a growing debate on the ethical use of
In a surprising turn of events, a professor at Northeastern University in Boston has found himself at the center of a growing debate on the ethical use of artificial intelligence in education. Rick Arrowood, a business professor, reportedly used AI tools like ChatGPT, Perplexity, and Gamma to prepare his class materials. This decision caused an uproar after students discovered inconsistencies in the content and realized AI had been used — despite Arrowood explicitly banning students from using AI for their own assignments. The incident has ignited serious conversations around transparency, fairness, and the double standards that may be emerging in academic institutions when it comes to AI.
Students Spot AI Errors and Demand Tuition Refund
The controversy gained traction when business student Ella Stapleton noticed spelling errors, oddly generated images, and even direct references to ChatGPT searches in the lecture materials. Feeling betrayed, Stapleton filed a complaint with the university administration and formally requested a refund of her $8,000 tuition fee, citing misleading conduct by the professor. Her request, however, was ultimately denied after several meetings with university officials. The incident has sparked mixed reactions across campus, with some students supporting Stapleton’s stance, arguing that the professor’s actions undermined the academic integrity of the course.
Professor Defends AI Use, Acknowledges Oversight
Professor Rick Arrowood admitted to using AI tools but stated his intention was to make his lectures more engaging and efficient. He acknowledged the presence of errors and claimed that he had reviewed the content before presenting it but failed to catch all the AI-related mistakes. While Arrowood has defended the use of AI as a modern teaching aid, he also conceded that he should have been more transparent with his students. He now believes that faculty members must be upfront about their use of such tools and hopes that his experience can serve as a learning moment for others navigating AI in academia.
Rising Tensions Over AI Ethics in Higher Education
The Northeastern case is not an isolated incident but part of a broader conversation about how AI is transforming education. As tools like ChatGPT become more mainstream, the lack of clear guidelines is leading to confusion and frustration on both sides of the classroom. Students are increasingly questioning why they are penalized for using AI while professors quietly adopt the same tools behind the scenes. Many argue that universities must urgently set consistent standards and ensure that both faculty and students are held to the same ethical expectations when it comes to AI usage.
The Road Ahead: Transparency, Trust, and AI Literacy
This incident has laid bare the growing gap between technological adoption and academic policy. As higher education institutions grapple with the rapid integration of AI, there is a pressing need for clear communication, transparency, and inclusive policies that apply to everyone. Students deserve to know when AI is used in their education, and professors must model the same academic honesty they expect from their pupils. Whether Arrowood’s case will lead to policy reform remains to be seen, but it has undoubtedly opened the floor for much-needed dialogue. The future of AI in classrooms now hinges on how openly and fairly it is used by both educators and learners alike.
Get the latest in business, markets, startups, and policy—visit businessnewsindia.in for in-depth updates and follow us on Instagram @businessnewsindia.in for daily bites of what matters most.
Source : indiatoday.in

©2024 | businessnewsindia.in | All Rights Reserved.