Universities: GenAI – There’s No Stopping, Start Shaping!

By Frances O’Donnell, Instructional Designer, ATU
Estimated reading time: 8 minutes
A group of diverse higher education professionals and students standing in front of a modern university building, engaged in a collaborative discussion around a glowing digital interface displaying the Gen.S.A.R. framework. In the foreground, traditional open books and a graduation cap are intertwined with glowing neural network nodes, symbolizing the integration of generative AI with traditional academic foundations. Image (and typos) generated by Nano Banana.
Moving from debate to action: Implementing a cross-departmental strategy to shape the future of GenAI in higher education. Image (and typos) generated by Nano Banana.

Debate continues to swing between those pushing rapid adoption and those advocating caution of GenAI, for example, panic about “AI taking over the classroom” and outrage at Big Tech’s labour practices. Both are important, but are these and other concerns causing inaction? In many cases, we are quietly watching students hand their data and their critical thinking over to the very Big Tech companies we are arguing against (while we still fly on holidays, stream on smart TVs and buy the same devices from the same companies). Pretending that GenAI in education is the one place we finally draw an ethical line, while doing nothing to make its use safer or more equitable, is not helpful. By all means, keep debating, but not at the cost of another three or four cohorts.

This opinion post suggests three things universities should address now: a minimal set of GenAI functions that should be available to staff and students, and a four-step teaching process to help lecturers rethink their role with GenAI.

Three things universities need to address now

1. Tell students and staff clearly what they can use (Déjà vu?)

Students and staff deserve clarity on which GenAI tools they have access to, what they can use them for and which ones are institutionally supported. Has your university provided this? No more grey areas or “ask your lecturer”. If people do not know this, it pushes GenAI use into secrecy. That secrecy hands more power to Big Tech to extract data and embed bias, while also quietly taking away their cognitive ability.

2. Untangle GenAI from “academic integrity”

Tightly linking GenAI to academic integrity was a mistake! It has created an endless debate about whether to permit or prohibit GenAI, which pushes use further underground. At this point, there is no real equity and no real academic integrity. Use of GenAI cannot simply be stopped, proved or disproved, so pretending otherwise, while holding endless anti‑AI discussions, will not lead to a solution. There is no putting GenAI back in the bottle!

3. Treat GenAI as a shared responsibility

GenAI affects curriculum design, assessment, student support, digital literacy, employability, libraries, disability support, IT, policy and everywhere in between. It cannot sit on the shoulders of one department or lead. Every university needs a cross‑departmental AI strategy that includes the student union, academic leads, IT, the data protection office, careers, student support, administration and teaching and learning personnel. Until leadership treats GenAI as systemic, lecturers will keep firefighting contradictions and marking assignments they know were AI-generated. Bring everyone to the table, and don’t adjourn until decisions have been made on students and staff clarity (even if this clarity is dynamic in nature – do not continue to leave them navigating this alone for another three years).

What GenAI functions should be provided

At a minimum, institutions should give safe, equitable access to:

  • A campus-licensed GenAI model
    One model for all staff and students to ask questions, draft, summarise, explain and translate text, including support for multilingual learners.
  • Multimodal creation tools
    Tools to create images, audio, video (including avatars), diagrams, code, etc., with clear ethical and legal guidance.
  • Research support tools
    Tools to support research, transcribing, coding, summaries, theme mapping, citations, etc., that reinforce critical exploration
  • Assessment and teaching design tools
    Tools to draft examples, case variations, rubrics, flashcards, questions, etc., are stored inside institutional systems.
  • Custom agents
    Staff create and share custom AI agents configured for specific purposes: subject-specific scaffolding for students, or workflow agents for planning, resource creation and content adaptation. Keep interactions within institutional systems.
  • Accessibility focused GenAI
    Tools that deliver captions, plain language rewrites, alt text and personalised study materials. Many institutions already have these in place.

Safer GenAI tools for exploration, collaboration and reflection. Now what do they do with them? This is where something like Gen.S.A.R comes in. A potential approach where staff and students explore together with GenAI, and one that is adaptable to different contexts/disciplines.

Gen.S.A.R.

Gen.S.A.R. is simply a suggested starting point; there is no magic wand, but this may help to ignite practical ideas from others. It suggests a shift from passive content delivery to constructivist and experiential learning.

  • GenAI exploration and collaborative knowledge construction
  • Scrutinise and share
  • Apply in real-world contexts with a low or no-tech approach
  • Reflect and evaluate

It keeps critical thinking, collaboration and real-world application at the centre, with GenAI as a set of tools rather than a replacement for learning. Note: GenAI is a set of tools, not a human!

Phase 1: GenAI, constructing, not copy-pasting

Students use GenAI, the lecturer, and reputable sources to explore a concept or problem linked to the learning outcomes. Lecturers guide this exploration as students work individually or in groups. With ongoing lecturer input, students may choose whether to use GenAI or other sources, but all develop an understanding of GenAI’s role in learning.

Phase 2: Scrutinise and Share

The second phase focuses on scrutinising and sharing ideas with others, not just presenting them as finished facts. Students bring GenAI outputs, reputable sources and their own thinking into dialogue. They interrogate evidence, assumptions and perspectives in groups or class discussion (social constructivism, dialogic teaching). The lecturer – the content expert – oversees this process and identifies the errors, draws attention to the errors and helps students clarify GenAI outputs.

Phase 3: Apply, low-tech, real-world

Screens step back. Students apply what they have discovered in low or no-tech ways: diagrams, mind maps, zines, prototypes, role plays, scenarios. They connect what they discovered to real contexts and show understanding through doing, making, explaining and practical application.

Phase 4: Reflect, evaluate and look forward

Students then evaluate and reflect on both their learning process and the role of GenAI. Using written, audio, video or visual reflections, they consider what they learned, how GenAI supported or distorted that learning and how this connects to their future. This reflective work, combined with artefacts from earlier phases, supports peer, self and lecturer assessment and moves us towards competency and readiness-based judgements.

Resourcing Gen.S.A.R. Yes, smaller class sizes and support would be required, but aspects of this can be implemented now (and are being implemented by some already). Time shifts to facilitation, co-learning, process feedback, and authentic evaluation (less three-thousand-word essays). This approach is not perfect but at least it’s an approach and one that draws on long‑standing learning theories, including constructivism, social constructivism, experiential learning, and traditions in inclusive and competency‑based education.

There’s No Stopping It, Time to Shape It

GenAI is not going away. Exploitative labour practices, data abuse and profit motives are real (and not exclusive to AI), and naming these harms is essential, but continuing to let these debates dominate any movement is not helpful. Universities can choose to lead (and I commend, not condemn, those who already are) with clear guidance, equitable access to safe GenAI tools and learning design. The alternative is all the risks associated with students and staff relying on personal accounts and workarounds.

For the integrity of education itself, it is time to translate debates into action. The genie is not going back in the bottle, and our profit-driven society is not only shaped by Big Tech but also by the everyday choices of those of us living privileged lives in westernised societies. It is time to be honest about our own complicity, to step out of the ivory tower and work with higher education students to navigate the impact GenAI is having on their lives right now.

Note: My views on GenAI for younger learners is very different; the suggestions here focus specifically on higher education.

Frances O’Donnell

Instructional Designer
ATU

Exploring the pros and cons of AI & GenAI in education, and indeed in society. Currently completing a Doctorate in Education with a focus on AI & Emerging Technologies.

Passionate about the potential education has to develop one’s self-confidence and self-worth, but frustrated by the fact it often does the opposite. AI has magnified our tendency to overassess and our inability to truly move away from rote learning.

Whether I’m carrying out the role of an instructional designer, or delivering workshops or researching, I think we should work together to make education a catalyst of change where learners are empowered to become confident as well as socially and environmentally conscious members of society. With or without AI, let’s change the perception of what success looks like for young people.

Keywords


How AI Adoption May Erode Key Skills US Students Need in an Automated World


A highly conceptual visual of a digital circuit board with key areas representing skills like "Critical Thinking," "Problem Solving," and "Originality" fading and becoming obscured by an overwhelming cloud of generic, high-speed AI data. Image (and typos) generated by Nano Banana.
The automation paradox: Experts warn that while AI drives efficiency, its widespread adoption in education may inadvertently erode the crucial cognitive and creative skills US students need to thrive in a future dominated by technology. Image (and typos) generated by Nano Banana.

Source

Times of India (Education International Desk)

Summary

This article explores concerns that widespread adoption of AI tools in education may undermine essential skills that students require for long-term success in an increasingly automated world. Educators and analysts interviewed argue that easy access to generative AI for writing, problem solving and research may weaken students’ capacity for critical thinking, creativity and independent judgement. They note that while AI can accelerate tasks, it may also reduce opportunities for deep learning and cognitive struggle, both of which are crucial for intellectual development. The article raises concerns that students who rely heavily on AI may experience diminished confidence in producing original work and solving complex problems without technological support. Experts recommend curriculum renewal that blends responsible AI literacy with explicit instruction in foundational skills, ensuring that students can use AI effectively without sacrificing their broader intellectual growth. The discussion reflects a recurring theme in the global AI-in-education debate: the need to preserve human expertise and cognitive resilience in an era of pervasive automation. The article calls for educators, policymakers and institutions to strike a balance between embracing AI and safeguarding human capabilities.

Key Points

  • Widespread AI use may weaken foundational cognitive skills
  • Risks include reduced independent thinking and reduced confidence
  • Educators call for curriculum redesign with balanced AI integration
  • Highlights need for responsible AI literacy
  • Addresses long-term workforce preparation concerns

Keywords

URL

https://timesofindia.indiatimes.com/education/news/how-ai-adoption-may-erode-key-skills-us-students-need-in-an-automated-world/articleshow/125672541.cms

Summary generated by ChatGPT 5.1


Students using ChatGPT beware: Real learning takes legwork, study finds


split image illustrating two contrasting study methods. On the left, a student in a blue-lit setting uses a laptop for "SHORT-CUT LEARNING" with "EASY ANSWERS" floating around. On the right, a student in a warm, orange-lit setting is engaged in "REAL LEGWORK LEARNING," writing in a notebook with open books and calculations. A large question mark divides the two scenes. Image (and typos) generated by Nano Banana.
The learning divide: A visual comparison highlights the potential pitfalls of relying on AI for “easy answers” versus the proven benefits of diligent study and engagement, as a new study suggests. Image (and typos) generated by Nano Banana.

Source

The Register

Summary

A new study published in PNAS Nexus finds that people who rely on ChatGPT or similar AI tools for research develop shallower understanding compared with those who gather information manually. Conducted by researchers from the University of Pennsylvania’s Wharton School and New Mexico State University, the study involved over 10,000 participants. Those using AI-generated summaries retained fewer facts, demonstrated less engagement, and produced advice that was shorter, less original, and less trustworthy. The findings reinforce concerns that overreliance on AI can “deskill” learners by replacing active effort with passive consumption. The researchers conclude that AI should support—not replace—critical thinking and independent study.

Key Points

  • Study of 10,000 participants compared AI-assisted and traditional research.
  • AI users showed shallower understanding and less factual recall.
  • AI summaries led to homogenised, less trustworthy responses.
  • Overreliance on AI risks reducing active learning and cognitive engagement.
  • Researchers recommend using AI as a support tool, not a substitute.

Keywords

URL

https://www.theregister.com/2025/11/03/chatgpt_real_understanding/

Summary generated by ChatGPT 5


Dr. Strange-Syllabus or: How My Students Learned to Mistrust AI and Trust Themselves

by Tadhg Blommerde – Assistant Professor, Northumbria University
Estimated reading time: 5 minutes
A stylized image featuring a character resembling Doctor Strange, dressed in his iconic attire, standing in a magical classroom setting. He holds up a glowing scroll labeled "SYLLABUS." In the foreground, two students (one Hispanic, one Black) are seated at a table, working on laptops that display a red 'X' over an AI-like interface, symbolizing mistrust of AI. Above Doctor Strange, a glowing, menacing AI entity with red eyes and outstretched arms hovers, presenting a digital screen, representing the seductive but potentially harmful nature of AI. Magical, glowing runes, symbols, and light effects fill the air around the students and the central figure, illustrating complex learning. Image (and typos) generated by Nano Banana.
In an era dominated by AI, educators are finding innovative ways to guide students. This image, inspired by a “Dr. Strange-Syllabus,” represents a pedagogical approach focused on fostering self-reliance and critical thinking, helping students to navigate the complexities of AI and ultimately trust their own capabilities. Image (and typos) generated by Nano Banana.

There is a scene I have witnessed many times in my classroom over the last couple of years. A question is posed, and before the silence has a chance to settle and spark a thought, a hand shoots up. The student confidently provides an answer, not from their own reasoning, but read directly from a glowing phone or laptop screen. Sometimes the answer is wrong and other times it is plausible but subtly wrong, lacking the specific context of our course materials. Almost always the reasoning behind the answer cannot be satisfactorily explained. This is the modern classroom reality. Students arrive with generative AI already deeply embedded in their personal lives and academic processes, viewing it not as a tool, but as a magic machine, an infallible oracle. Their initial relationship with it is one of unquestioning trust.

The Illusion of the All-Knowing Machine

Attempting to ban this technology would be a futile gesture. Instead, the purpose of my teaching became to deliberately make students more critical and reflective users of it. At the start of my module, their overreliance is palpable. They view AI as an all-knowing friend, a collaborator that can replace the hard work of thinking and writing. In the early weeks, this manifests as a flurry of incorrect answers shouted out in class, the product of poorly constructed prompts fed into (exclusively) ChatGPT, and a complete faith in the response it generated. It was clear there was a dual deficit: a lack of foundational knowledge on the topic, and a complete absence of critical engagement with the AI’s output.

Remedying this begins not with a single ‘aha!’ moment, but through a cumulative, twelve-week process of structured exploration. I introduce a prompt engineering and critical analysis framework that guides students through writing more effective prompts and critically engaging with AI output. We move beyond simple questions and answers. I task them with having AI produce complex academic work, such as literature reviews and research proposals, which they would then systematically interrogate. Their task is to question everything. Does the output actually adhere to the instructions in the prompt? Can every claim and statement be verified with a credible, existing source? Are there hidden biases or a leading tone that misrepresents the topic or their own perspective?

Pulling Back the Curtain on AI

As they began this work, the curtain was pulled back on the ‘magic’ machine. Students quickly discovered the emperor had no clothes. They found AI-generated literature reviews cited non-existent sources or completely misrepresented the findings of real academic papers. They critiqued research proposals that suggested baffling methodologies, like using long-form interviews in a positivist study. This process forced them to rely on their own developing knowledge of module materials to spot the flaws. They also began to critique the writing itself, noting that the prose was often excessively long-winded, failed to make points succinctly, and felt bland. A common refrain was that it simply ‘didn’t sound like them’. They came to realise that AI, being sycophantic by design, could not provide the truly critical feedback necessary for their intellectual or personal growth.

This practical work was paired with broader conversations about the ethics of AI, from its significant environmental impact to the copyrighted material used in its training. Many students began to recognise their own over-dependence, reporting a loss of skills when starting assignments and a profound lack of satisfaction in their work when they felt they had overused this technology. Their use of the technology began to shift. Instead of a replacement for their own intellect, it became a device to enhance it. For many, this new-found scepticism extended beyond the classroom. Some students mentioned they were now more critical of content they encountered on social media, understanding how easily inaccurate or misleading information could be generated and spread. The module was fostering not just AI literacy, but a broader media literacy.

From Blind Trust to Critical Confidence

What this experience has taught me is that student overreliance on AI is often driven by a lack of confidence in their own abilities. By bringing the technology into the open and teaching them to expose its limitations, we do more than just create responsible users. We empower them to believe in their own knowledge and their own voice. They now see AI for what it is: not an oracle, but a tool with serious shortcomings. It has no common sense and cannot replace their thinking. In an educational landscape where AI is not going anywhere, our greatest task is not to fear it, but to use it as a powerful instrument for teaching the very skills it threatens to erode: critical inquiry, intellectual self-reliance, and academic integrity.

Tadhg Blommerde

Assistant Professor
Northumbria University

Tadhg is a lecturer (programme and module leader) and researcher that is proficient in quantitative and qualitative social science techniques and methods. His research to date has been published in Journal of Business Research, The Service Industries Journal, and European Journal of Business and Management Research. Presently, he holds dual roles and is an Assistant Professor (Senior Lecturer) in Entrepreneurship at Northumbria University and an MSc dissertation supervisor at Oxford Brookes University.

His interests include innovation management; the impact of new technologies on learning, teaching, and assessment in higher education; service development and design; business process modelling; statistics and structural equation modelling; and the practical application and dissemination of research.


Keywords


Teachers Worry AI Will Impede Students’ Critical Thinking Skills. Many Teens Aren’t So Sure


A split image contrasting teachers' concerns about AI with teenagers' perspectives. On the left, a worried female teacher stands in a traditional classroom, gesturing with open hands towards a laptop on a desk. A glowing red 'X' mark covers the words "CRITICAL THINKING" and gears/data on the laptop screen, symbolizing the perceived threat to cognitive skills. On the right, three engaged teenagers (two boys, one girl) are working collaboratively on laptops in a bright, modern setting. Glowing keywords like "PROBLEM-SOLVING," "INNOVATION," and "CREATIVITY" emanate from their screens, representing AI's perceived benefits. A large question mark is placed in the middle top of the image. Image (and typos) generated by Nano Banana.
A clear divide emerges in the debate over AI’s impact on critical thinking: while many teachers express concern that AI will hinder students’ cognitive development, a significant number of teenagers remain unconvinced, often viewing AI as a tool that can enhance their problem-solving abilities. This image visualises the contrasting viewpoints on this crucial educational challenge. Image (and typos) generated by Nano Banana.

Source

Education Week

Summary

Alyson Klein reports on the growing divide between teachers and students over how artificial intelligence is affecting critical thinking. While educators fear that AI tools like ChatGPT are eroding students’ ability to reason independently, many teens argue that AI can actually enhance their thinking when used responsibly. Teachers cite declining originality and over-reliance on AI-generated answers, expressing concern that students are losing confidence in forming their own arguments. Students, however, describe AI as a useful study companion—helping clarify concepts, model strong writing, and guide brainstorming. Experts suggest that the key issue is not whether AI harms or helps, but how schools teach students to engage with it critically. Educators who integrate AI into lessons rather than banning it outright are finding that students can strengthen, rather than surrender, their analytical skills.

Key Points

  • Teachers fear AI use is diminishing critical thinking and originality in student work.
  • Many students view AI as a learning aid that supports understanding and creativity.
  • The divide reflects differing expectations around what “thinking critically” means.
  • Experts recommend structured AI literacy education over prohibition or punishment.
  • Responsible AI use depends on reflection, questioning, and teacher guidance.

Keywords

URL

https://www.edweek.org/technology/teachers-worry-ai-will-impede-students-critical-thinking-skills-many-teens-arent-so-sure/2025/10

Summary generated by ChatGPT 5