Latest Posts

AI and the future of education. Disruptions, dilemmas and directions


Source

UNESCO

Summary

This UNESCO report provides policy guidance on integrating artificial intelligence (AI) into education systems worldwide. It stresses both the opportunities—such as personalised learning, enhanced efficiency, and expanded access—and the risks, including bias, privacy concerns, and the erosion of teacher and learner agency. The document frames AI as a powerful tool that can help address inequalities and support sustainable development, but only if implemented responsibly and inclusively.

Central to the report is the principle that AI in education must remain human-centred, promoting equity, transparency, and accountability. It highlights the importance of teacher empowerment, digital literacy, and robust governance frameworks. The guidance calls for capacity building at all levels, from policy to classroom practice, and for international cooperation to ensure that AI use aligns with ethical standards and local contexts. Ultimately, the report argues that AI should augment—not replace—human intelligence in education.

Key Points

  • AI offers opportunities for personalised learning and system efficiency.
  • Risks include bias, inequity, and privacy breaches if left unchecked.
  • AI in education must be guided by human-centred, ethical frameworks.
  • Teachers remain central; AI should support rather than replace them.
  • Digital literacy for learners and educators is essential.
  • Governance frameworks must ensure transparency and accountability.
  • Capacity building and training are critical for sustainable adoption.
  • AI should contribute to equity and inclusion, not exacerbate divides.
  • International collaboration is vital for responsible AI use in education.
  • AI’s role is to augment human intelligence, not supplant it.

Conclusion

UNESCO concludes that AI has the potential to transform education systems for the better, but only if adoption is deliberate, ethical, and values-driven. Policymakers must prioritise equity, inclusivity, and transparency while ensuring that human agency and the role of teachers remain central to education in the age of AI.

Keywords

URL

https://www.unesco.org/en/articles/ai-and-future-education-disruptions-dilemmas-and-directions

Summary generated by ChatGPT 5


AI Detectors in Education


Source

Associate Professor Mark A. Bassett

Summary

This report critically examines the use of AI text detectors in higher education, questioning their accuracy, fairness, and ethical implications. While institutions often adopt detectors as a visible response to concerns about generative AI in student work, the paper highlights that their statistical metrics (e.g., false positive/negative rates) are largely meaningless in real-world educational contexts. Human- and AI-written text cannot be reliably distinguished, making detector outputs unreliable as evidence. Moreover, reliance on detectors risks reinforcing inequities: students with access to premium AI tools are less likely to be flagged, while others face disproportionate scrutiny.

Bassett argues that AI detectors compromise fairness and transparency in academic integrity processes. Comparisons to metal detectors, smoke alarms, or door locks are dismissed as misleading, since those tools measure objective, physical phenomena with regulated standards, unlike the probabilistic guesswork of AI detectors. The report stresses that detector outputs shift the burden of proof unfairly onto students, often pressuring them into confessions or penalising them based on arbitrary markers like writing style or speed. Instead of doubling down on flawed tools, the focus should be on redesigning assessments, clarifying expectations, and upholding procedural fairness.

Key Points

  • AI detectors appear effective but offer no reliable standard of evidence.
  • Accuracy metrics (TPR, FPR, etc.) are meaningless in practice outside controlled tests.
  • Detectors unfairly target students without addressing systemic integrity issues.
  • Reliance risks inequity: affluent or tech-savvy students can evade detection more easily.
  • Using multiple detectors or comparing student work to AI outputs reinforces bias, not evidence.
  • Analogies to locks, smoke alarms, or metal detectors are misleading and invalid.
  • Procedural fairness demands that institutions—not students—carry the burden of proof.
  • False positives have serious consequences for students, unlike benign fire alarm errors.
  • Deterrence through fear undermines trust and shifts education toward surveillance.
  • Real solutions lie in redesigning assessment practices, not deploying flawed detection tools.

Conclusion

AI detectors are unreliable, unregulated, and ethically problematic as tools for ensuring academic integrity. Rather than treating detector outputs as evidence, institutions should prioritise fairness, transparency, and assessment redesign. Ensuring that students learn and are evaluated equitably requires moving beyond technological quick fixes toward principled, values-based approaches.

Keywords

URL

https://drmarkbassett.com/assets/AI_Detectors_in_education.pdf

Summary generated by ChatGPT 5


How AI Is Changing—Not ‘Killing’—College


A diverse group of college students is gathered in a modern university library or common area, with some holding tablets or looking at laptops. Above them, a large, glowing word cloud hovers, filled with terms related to artificial intelligence and its impact. Prominent words include "HELPFUL," "FUTURE," "ETHICS," "CHEATING," "BIAS," and "CONCERNING," reflecting a range of student opinions. The overall impression is one of active discussion and varied perspectives on AI. Image (and typos) generated by Nano Banana.
What do the next generation of leaders and innovators think about artificial intelligence? This visual captures the dynamic and often contrasting views of college students on AI’s role in their education, future careers, and daily lives. Image (and typos) generated by Nano Banana.

Source

Inside Higher Ed

Summary

A new Student Voice survey by Inside Higher Ed and Generation Lab captures how U.S. college students are adapting to generative AI in their studies and what they expect from institutions. Of the 1,047 students surveyed, 85 per cent had used AI tools in the past year—mainly for brainstorming, tutoring, and studying—while only a quarter admitted to using them for completing assignments. Most respondents called for universities to provide education on ethical AI use and clearer, standardised policies, rather than policing or banning the technology. Although students are divided about AI’s impact on critical thinking, most agree it can enhance learning if used responsibly. The majority do not view AI as diminishing the value of college; some even see it as increasing it.

Key Points

  • 85 per cent of students have used AI tools for coursework, mainly for brainstorming and study support.
  • 97 per cent want universities to respond to AI’s impact on academic integrity through education, not restriction.
  • Over half say AI has mixed effects on critical thinking; 27 per cent find it enhances learning.
  • Students want institutions to offer professional and ethical AI training, not leave it to individual faculty.
  • Only 18 per cent believe AI reduces the value of college; 23 per cent say it increases it.

Keywords

URL

https://www.insidehighered.com/news/students/academics/2025/08/29/survey-college-students-views-ai

Summary generated by ChatGPT 5


QQI Generative Artificial Intelligence Survey Report 2025


Source

Quality and Qualifications Ireland (QQI), August 2025

Summary

This national survey captures the views of 1,229 staff and 1,005 learners across Ireland’s further, higher, and English language education sectors on their knowledge, use, and perceptions of generative AI (GenAI). The report reveals growing engagement with GenAI but also wide disparities in understanding, policy, and preparedness. Most respondents recognise AI’s transformative impact but remain uncertain about its role in assessment, academic integrity, and employability.

While over 80% of staff and learners believe GenAI will significantly change education and work over the next five years, few feel equipped to respond. Only 20% of staff and 14% of learners report access to GenAI training. Policies are inconsistent or absent, with most institutions leaving decisions on use to individual educators. Both staff and learners support transparent, declared use of GenAI but express concerns about bias, overreliance, loss of essential skills, and declining trust in qualifications. Respondents call for coherent national and institutional policies, professional development, and curriculum reform that balances innovation with integrity.

Key Points

  • 82% of respondents expect GenAI to transform learning and work within five years.
  • 63% of staff and 36% of learners believe GenAI literacy should be explicitly taught.
  • Fewer than one in five institutions currently provide structured GenAI training.
  • Policies on GenAI use are inconsistent, unclear, or absent in most institutions.
  • Over half of respondents fear skill erosion and reduced academic trust from AI use.
  • 70% of staff say assessment rules for GenAI lack clarity or consistency.
  • 83% of learners believe GenAI will change how they are assessed.
  • Staff and learners call for transparent declaration of GenAI use in assignments.
  • 61% of staff feel learners are unprepared to use GenAI responsibly in the workplace.
  • Respondents emphasise ethical governance, inclusion, and sustainable AI adoption.

Conclusion

The survey highlights a critical moment for Irish education: generative AI is already influencing learning and work, yet systems for policy, training, and ethics are lagging behind. To maintain public trust and educational relevance, QQI recommends a coordinated national response centred on transparency, AI literacy, and values-led governance that equips both learners and educators for an AI-driven future.

Keywords

URL

https://www.qqi.ie/sites/default/files/2025-08/generative-artificial-intelligence-survey-report-2025.pdf

Summary generated by ChatGPT 5


Will AI Make You Stupid?


A digital representation of a human brain with glowing teal data streams and circuit-like patterns flowing out from its right side, against a dark, technical background with a subtle digital frame. Image (and typos) generated by Nano Banana.
Exploring the cognitive impact of artificial intelligence: Will reliance on AI enhance our intellect or diminish our critical thinking abilities? Image (and typos) generated by Nano Banana.

Source

The Economist

Summary

A Massachusetts Institute of Technology study has found that students using ChatGPT during essay-writing tasks showed reduced brain activity in areas linked to creativity and attention. Similar research from Microsoft and the SBS Swiss Business School supports the claim that frequent AI use may diminish critical thinking, fostering “cognitive miserliness,” or the tendency to offload mental effort. While experts caution that the evidence is not yet conclusive, they warn that excessive reliance on AI could erode problem-solving and creative skills over time. Historical parallels—such as Socrates’ scepticism about writing—suggest technological tools often reshape, but do not destroy, cognitive abilities. The article concludes that using AI thoughtfully—prompting step by step and reflecting critically—can help preserve intellectual engagement even as automation advances.

Key Points

  • MIT researchers observed reduced creative and attentional brain activity in AI-assisted students.
  • Frequent AI users performed worse on critical-thinking tests in a Swiss study.
  • Over-reliance on AI can create “cognitive offloading” and feedback loops of dependence.
  • Experts urge reflective, guided use—AI as assistant, not replacement.
  • Strategies such as incremental prompting and “cognitive forcing” can sustain mental effort.
  • Evidence remains mixed: AI may change, but not necessarily weaken, human intelligence.

Keywords

URL

https://www.economist.com/science-and-technology/2025/07/16/will-ai-make-you-stupid

Summary generated by ChatGPT 5