2025 Review – A Shared Conversation, Built Over Time

Estimated reading time: 9 minutes

From Individual Questions to Collective Practice

Since September, the GenAI:N3 blog has hosted a weekly series of reflections exploring what generative AI means for higher education: for teaching, learning, assessment, academic identity, and institutional responsibility. Early contributions captured a sector grappling with disruption, uncertainty, and unease, asking difficult questions about trust, integrity, creativity, and control at a moment when generative AI arrived faster than policy, pedagogy, or professional development could respond.

As the series unfolded, a clear shift began to emerge. Posts moved from individual reactions and early experimentation towards more structured sense-making, discipline-specific redesign, and crucially shared learning. The introduction of communities of practice as a deliberate strategy for AI upskilling marked a turning point in the conversation: from “How do I deal with this?” to “How do we learn, adapt, and shape this together?” Taken as a whole, the series traces that journey from disruption to agency, and from isolated responses to collective practice.

What makes this series distinctive is not simply its focus on generative AI, but the diversity of voices it brings together. Contributors include academic staff, professional staff, educational developers, students, and sector partners, each writing from their own context while engaging with a set of common challenges. The result is not a single narrative, but a constellation of perspectives that reflect the complexity of teaching and learning in an AI-shaped world.

29th September 2025 – Jim O’Mahony

Something Wicked This Way Comes

The GenAI:N3 blog series opens with a deliberately unsettling provocation, asking higher education to confront the unease, disruption, and uncertainty that generative AI has introduced into teaching, assessment, and academic identity. Rather than framing AI as either saviour or villain, this piece invites a more honest reckoning with fear, denial, and institutional inertia. It sets the tone for the series by arguing that ignoring GenAI is no longer an option; what matters now is how educators respond, individually and collectively, to a technology that has already crossed the threshold into everyday academic practice.

6th October 2025 – Dr Yannis

3 Things AI Can Do for You: The No-Nonsense Guide

Building directly on that initial unease, this post grounds the conversation in pragmatism. Stripping away hype and alarmism, it focuses on concrete, immediately useful ways AI can support academic work, from sense-making to productivity. The emphasis is not on replacement but augmentation, encouraging educators to experiment cautiously, critically, and with intent. In the arc of the series, this contribution marks a shift from fear to agency, demonstrating that engagement with AI can be practical, purposeful, and aligned with professional judgement.

13th October 2025 – Sue Beckingham & Peter Hartley

New Elephants in the Generative AI Room? Acknowledging the Costs of GenAI to Develop ‘Critical AI Literacy’

As confidence in experimentation grows, this post re-introduces necessary friction by surfacing the hidden costs of generative AI. Environmental impact, labour implications, equity, and ethical responsibility are brought into sharp focus, challenging overly simplistic narratives of efficiency and innovation. The authors remind readers that responsible adoption requires confronting uncomfortable trade-offs. Within the wider series, this piece deepens the discussion, insisting that values, sustainability, and social responsibility must sit alongside pedagogical opportunity.

20th October 2025 Jonathan Sansom

Making Sense of GenAI in Education: From Force Analysis to Pedagogical Copilot Agents

Here the conversation turns toward structured sense-making. Drawing on strategic and pedagogical frameworks, this post explores how educators and institutions can move beyond reactive responses to more deliberate design choices. The idea of AI as a “copilot” rather than an autonomous actor reframes the relationship between teacher, learner, and technology. In the narrative of the series, this contribution offers conceptual tools for navigating complexity, helping readers connect experimentation with strategy.

27th October 2025 – Patrick Shields

AI Adoption & Education for SMEs

Widening the lens beyond universities, this post examines AI adoption through the perspective of small and medium-sized enterprises, highlighting the skills, mindsets, and educational approaches needed to support workforce readiness. The crossover between higher education, lifelong learning, and industry becomes explicit. This piece situates GenAI not just as an academic concern, but as a societal one, reinforcing the importance of education systems that are responsive, connected, and outward-looking.

3rd November 2025 – Tadhg Blommerde

Dr Strange-Syllabus or: How My Students Learned to Mistrust AI and Trust Themselves

Returning firmly to the classroom, this reflective account explores what happens when students are encouraged to engage critically with AI rather than rely on it unthinkingly. Through curriculum design and assessment choices, learners begin to question outputs, assert their own judgement, and reclaim intellectual agency. This post is a turning point in the series, showing how thoughtful pedagogy can transform AI from a threat to academic integrity into a catalyst for deeper learning.

10th November 2025 – Brian Mulligan

AI Could Revolutionise Higher Education in a Way We Did Not Expect

This contribution steps back to consider second-order effects, arguing that the most significant impact of AI may not be efficiency or automation, but a reconfiguration of how learning, expertise, and value are understood. It challenges institutions to think beyond surface-level policy responses and to anticipate longer-term cultural shifts. Positioned mid-series, the post broadens the horizon, encouraging readers to think systemically rather than tactically.

17th November 2025 – Kerith George-Briant & Jack Hogan

This Is Not the End but a Beginning: Responding to “Something Wicked This Way Comes”

Explicitly dialoguing with the opening post, this response reframes the initial sense of threat as a starting point rather than a conclusion. The authors emphasise community, dialogue, and shared responsibility, arguing that collective reflection is essential if higher education is to navigate GenAI well. This piece reinforces one of the central through-lines of the series: that no single institution or individual has all the answers, but progress is possible through collaboration.

24th November 2025 – Bernie Goldbach

The Transformative Power of Communities of Practice in AI Upskilling for Educators

This post makes the case that the most sustainable way to build AI capability in education is not through one-off training sessions, but through communities of practice that support ongoing learning, experimentation, and shared problem-solving. It highlights how peer-to-peer dialogue helps educators move from cautious curiosity to confident, critical use of tools, while also creating space to discuss ethics, assessment, and evolving norms without judgement. Positioned within the blog series, it serves as a bridge between individual experimentation and institutional change: a reminder that AI upskilling is fundamentally social, and that collective learning structures are one of the best defences against both hype and paralysis.

1st December 2025 – Hazel Farrell et al.

Teaching the Future: How Tomorrow’s Music Educators Are Reimagining Pedagogy

Offering a discipline-specific lens, this post explores how music education is being rethought in light of AI, creativity, and emerging professional realities. Rather than diluting artistic practice, AI becomes a catalyst for re-examining what it means to teach, learn, and create. Within the series, this contribution demonstrates how GenAI conversations translate into authentic curriculum redesign, grounded in disciplinary values rather than generic solutions.

8th December 2025 – Ken McCarthy

Building the Manifesto: How We Got Here and What Comes Next

This reflective piece pulls together many of the threads running through the series, documenting the collaborative process behind the Manifesto for Generative AI in Higher Education. It positions the Manifesto not as a prescriptive policy document but as a living statement shaped by diverse voices, shared concerns, and collective aspiration. In the narrative arc, it represents a moment of synthesis, turning discussion into a shared point of reference.

15th December 2025 – Leigh Graves Wolf

Rebuilding Thought Networks in the Age of AI

Moving from frameworks to cognition, this post explores how AI is reshaping thinking itself. Rather than outsourcing thought, the author argues for intentionally rebuilding intellectual networks so that AI becomes part of, not a replacement for, human sense-making. This contribution deepens the series philosophically, reminding readers that the stakes of GenAI are as much cognitive and epistemic as they are technical.

22nd December 2025 – Frances O’Donnell

Universities: GenAI – There’s No Stopping, Start Shaping!

The series culminates with a clear call to action. Acknowledging both inevitability and responsibility, this post urges universities to move decisively from reaction to leadership. The emphasis is on shaping futures rather than resisting change, grounded in values, purpose, and public good. As a closing note, it captures the spirit of the entire series: GenAI is already here, but how it reshapes higher education remains a choice.

With Thanks – and an Invitation

This series exists because of the generosity, openness, and intellectual courage of its contributors. Each author took the time to reflect publicly, to question assumptions, to share practice, and to contribute thoughtfully to a conversation that is still very much in motion. Collectively, these posts embody the spirit of GenAI:N3 – collaborative, reflective, and committed to shaping the future of higher education with care rather than fear.

We would like to extend our sincere thanks to all who have contributed to the blog to date, and to those who have engaged with the posts through reading, sharing, and discussion. The conversation does not end here. If you are experimenting with generative AI in your teaching, supporting others to do so, grappling with its implications, or working with students as partners in this space, we warmly invite you to write a blog post of your own. Your perspective matters, and your experience can help others navigate this rapidly evolving landscape.

If you would like to contribute, please get in touch (blog@genain3.ie) we would love to hear from you.


DateTitleAuthorLink
29 September 2025Something Wicked This Way ComesJim O’Mahonyhttps://genain3.ie/something-wicked-this-way-comes/
6 October 20253 Things AI Can Do for You: The No-Nonsense GuideDr Yannishttps://genain3.ie/3-things-ai-can-do-for-you-the-no-nonsense-guide/
13 October 2025New Elephants in the Generative AI Room? Acknowledging the Costs of GenAI to Develop ‘Critical AI Literacy’Sue Beckingham & Peter Hartleyhttps://genain3.ie/new-elephants-in-the-generative-ai-room-acknowledging-the-costs-of-genai-to-develop-critical-ai-literacy/
20 October 2025Making Sense of GenAI in Education: From Force Analysis to Pedagogical Copilot AgentsJonathan Sansomhttps://genain3.ie/making-sense-of-genai-in-education-from-force-analysis-to-pedagogical-copilot-agents/
27 October 2025AI Adoption & Education for SMEsPatrick Shieldshttps://genain3.ie/ai-adoption-education-for-smes/
3 November 2025Dr Strange-Syllabus or: How My Students Learned to Mistrust AI and Trust ThemselvesTadhg Blommerdehttps://genain3.ie/dr-strange-syllabus-or-how-my-students-learned-to-mistrust-ai-and-trust-themselves/
10 November 2025AI Could Revolutionise Higher Education in a Way We Did Not ExpectBrian Mulliganhttps://genain3.ie/ai-could-revolutionise-higher-education-in-a-way-we-did-not-expect/
17 November 2025This Is Not the End but a Beginning: Responding to “Something Wicked This Way Comes”Kerith George-Briant & Jack Hoganhttps://genain3.ie/this-is-not-the-end-but-a-beginning-responding-to-something-wicked-this-way-comes/
24 November 2025The Transformative Power of Communities of Practice in AI Upskilling for EducatorsBernie Goldbachhttps://genain3.ie/the-transformative-power-of-communities-of-practice-in-ai-upskilling-for-educators/
1 December 2025Teaching the Future: How Tomorrow’s Music Educators Are Reimagining PedagogyHazel Farrell et al.https://genain3.ie/teaching-the-future-how-tomorrows-music-educators-are-reimagining-pedagogy/
8 December 2025Building the Manifesto: How We Got Here and What Comes NextKen McCarthyhttps://genain3.ie/building-the-manifesto-how-we-got-here-and-what-comes-next/
15 December 2025Rebuilding Thought Networks in the Age of AILeigh Graves Wolfhttps://genain3.ie/rebuilding-thought-networks-in-the-age-of-ai/
22 December 2025Universities: GenAI – There’s No Stopping, Start Shaping!Frances O’Donnellhttps://genain3.ie/universities-genai-theres-no-stopping-start-shaping/

Keywords


Universities: GenAI – There’s No Stopping, Start Shaping!

By Frances O’Donnell, Instructional Designer, ATU
Estimated reading time: 8 minutes
A group of diverse higher education professionals and students standing in front of a modern university building, engaged in a collaborative discussion around a glowing digital interface displaying the Gen.S.A.R. framework. In the foreground, traditional open books and a graduation cap are intertwined with glowing neural network nodes, symbolizing the integration of generative AI with traditional academic foundations. Image (and typos) generated by Nano Banana.
Moving from debate to action: Implementing a cross-departmental strategy to shape the future of GenAI in higher education. Image (and typos) generated by Nano Banana.

Debate continues to swing between those pushing rapid adoption and those advocating caution of GenAI, for example, panic about “AI taking over the classroom” and outrage at Big Tech’s labour practices. Both are important, but are these and other concerns causing inaction? In many cases, we are quietly watching students hand their data and their critical thinking over to the very Big Tech companies we are arguing against (while we still fly on holidays, stream on smart TVs and buy the same devices from the same companies). Pretending that GenAI in education is the one place we finally draw an ethical line, while doing nothing to make its use safer or more equitable, is not helpful. By all means, keep debating, but not at the cost of another three or four cohorts.

This opinion post suggests three things universities should address now: a minimal set of GenAI functions that should be available to staff and students, and a four-step teaching process to help lecturers rethink their role with GenAI.

Three things universities need to address now

1. Tell students and staff clearly what they can use (Déjà vu?)

Students and staff deserve clarity on which GenAI tools they have access to, what they can use them for and which ones are institutionally supported. Has your university provided this? No more grey areas or “ask your lecturer”. If people do not know this, it pushes GenAI use into secrecy. That secrecy hands more power to Big Tech to extract data and embed bias, while also quietly taking away their cognitive ability.

2. Untangle GenAI from “academic integrity”

Tightly linking GenAI to academic integrity was a mistake! It has created an endless debate about whether to permit or prohibit GenAI, which pushes use further underground. At this point, there is no real equity and no real academic integrity. Use of GenAI cannot simply be stopped, proved or disproved, so pretending otherwise, while holding endless anti‑AI discussions, will not lead to a solution. There is no putting GenAI back in the bottle!

3. Treat GenAI as a shared responsibility

GenAI affects curriculum design, assessment, student support, digital literacy, employability, libraries, disability support, IT, policy and everywhere in between. It cannot sit on the shoulders of one department or lead. Every university needs a cross‑departmental AI strategy that includes the student union, academic leads, IT, the data protection office, careers, student support, administration and teaching and learning personnel. Until leadership treats GenAI as systemic, lecturers will keep firefighting contradictions and marking assignments they know were AI-generated. Bring everyone to the table, and don’t adjourn until decisions have been made on students and staff clarity (even if this clarity is dynamic in nature – do not continue to leave them navigating this alone for another three years).

What GenAI functions should be provided

At a minimum, institutions should give safe, equitable access to:

  • A campus-licensed GenAI model
    One model for all staff and students to ask questions, draft, summarise, explain and translate text, including support for multilingual learners.
  • Multimodal creation tools
    Tools to create images, audio, video (including avatars), diagrams, code, etc., with clear ethical and legal guidance.
  • Research support tools
    Tools to support research, transcribing, coding, summaries, theme mapping, citations, etc., that reinforce critical exploration
  • Assessment and teaching design tools
    Tools to draft examples, case variations, rubrics, flashcards, questions, etc., are stored inside institutional systems.
  • Custom agents
    Staff create and share custom AI agents configured for specific purposes: subject-specific scaffolding for students, or workflow agents for planning, resource creation and content adaptation. Keep interactions within institutional systems.
  • Accessibility focused GenAI
    Tools that deliver captions, plain language rewrites, alt text and personalised study materials. Many institutions already have these in place.

Safer GenAI tools for exploration, collaboration and reflection. Now what do they do with them? This is where something like Gen.S.A.R comes in. A potential approach where staff and students explore together with GenAI, and one that is adaptable to different contexts/disciplines.

Gen.S.A.R.

Gen.S.A.R. is simply a suggested starting point; there is no magic wand, but this may help to ignite practical ideas from others. It suggests a shift from passive content delivery to constructivist and experiential learning.

  • GenAI exploration and collaborative knowledge construction
  • Scrutinise and share
  • Apply in real-world contexts with a low or no-tech approach
  • Reflect and evaluate

It keeps critical thinking, collaboration and real-world application at the centre, with GenAI as a set of tools rather than a replacement for learning. Note: GenAI is a set of tools, not a human!

Phase 1: GenAI, constructing, not copy-pasting

Students use GenAI, the lecturer, and reputable sources to explore a concept or problem linked to the learning outcomes. Lecturers guide this exploration as students work individually or in groups. With ongoing lecturer input, students may choose whether to use GenAI or other sources, but all develop an understanding of GenAI’s role in learning.

Phase 2: Scrutinise and Share

The second phase focuses on scrutinising and sharing ideas with others, not just presenting them as finished facts. Students bring GenAI outputs, reputable sources and their own thinking into dialogue. They interrogate evidence, assumptions and perspectives in groups or class discussion (social constructivism, dialogic teaching). The lecturer – the content expert – oversees this process and identifies the errors, draws attention to the errors and helps students clarify GenAI outputs.

Phase 3: Apply, low-tech, real-world

Screens step back. Students apply what they have discovered in low or no-tech ways: diagrams, mind maps, zines, prototypes, role plays, scenarios. They connect what they discovered to real contexts and show understanding through doing, making, explaining and practical application.

Phase 4: Reflect, evaluate and look forward

Students then evaluate and reflect on both their learning process and the role of GenAI. Using written, audio, video or visual reflections, they consider what they learned, how GenAI supported or distorted that learning and how this connects to their future. This reflective work, combined with artefacts from earlier phases, supports peer, self and lecturer assessment and moves us towards competency and readiness-based judgements.

Resourcing Gen.S.A.R. Yes, smaller class sizes and support would be required, but aspects of this can be implemented now (and are being implemented by some already). Time shifts to facilitation, co-learning, process feedback, and authentic evaluation (less three-thousand-word essays). This approach is not perfect but at least it’s an approach and one that draws on long‑standing learning theories, including constructivism, social constructivism, experiential learning, and traditions in inclusive and competency‑based education.

There’s No Stopping It, Time to Shape It

GenAI is not going away. Exploitative labour practices, data abuse and profit motives are real (and not exclusive to AI), and naming these harms is essential, but continuing to let these debates dominate any movement is not helpful. Universities can choose to lead (and I commend, not condemn, those who already are) with clear guidance, equitable access to safe GenAI tools and learning design. The alternative is all the risks associated with students and staff relying on personal accounts and workarounds.

For the integrity of education itself, it is time to translate debates into action. The genie is not going back in the bottle, and our profit-driven society is not only shaped by Big Tech but also by the everyday choices of those of us living privileged lives in westernised societies. It is time to be honest about our own complicity, to step out of the ivory tower and work with higher education students to navigate the impact GenAI is having on their lives right now.

Note: My views on GenAI for younger learners is very different; the suggestions here focus specifically on higher education.

Frances O’Donnell

Instructional Designer
ATU

Exploring the pros and cons of AI & GenAI in education, and indeed in society. Currently completing a Doctorate in Education with a focus on AI & Emerging Technologies.

Passionate about the potential education has to develop one’s self-confidence and self-worth, but frustrated by the fact it often does the opposite. AI has magnified our tendency to overassess and our inability to truly move away from rote learning.

Whether I’m carrying out the role of an instructional designer, or delivering workshops or researching, I think we should work together to make education a catalyst of change where learners are empowered to become confident as well as socially and environmentally conscious members of society. With or without AI, let’s change the perception of what success looks like for young people.

Keywords


Rebuilding Thought Networks in the Age of AI

By Leigh Graves Wolf, University College Dublin Teaching & Learning
Estimated reading time: 5 minutes
A highly conceptual visual showing a partially fragmented human brain structure with new, glowing neural pathways being actively reconnected and rebuilt by a series of fine, digital threads, symbolizing the conscious effort to strengthen cognitive skills amidst reliance on AI. Image (and typos) generated by Nano Banana.
Strengthening the mind: Highlighting the crucial need and methodology for intentionally restructuring and reinforcing human cognitive and critical thinking skills in an environment increasingly dominated by artificial intelligence. Image (and typos) generated by Nano Banana.

Thinking is a social activity. This isn’t a new insight (scholars have studied this for ages) but it’s one I keep coming back to lately as I try to stay afloat in the “AI Era.”

For a long stretch as I developed as an academic, I thought with others through technology (i.e. del.icio.us, Typepad, and Twitter.) We would bounce ideas off each other, glean golden nuggets of information, share resources that sparked new connections in our minds. There was something magical about that era, the serendipitous discovery of a colleague’s bookmark that led you down an unexpected intellectual rabbit hole, or a Twitter thread that challenged your thinking in ways you hadn’t anticipated. These weren’t just tools; they were extensions of our collective scholarly brain.

Then, all of that broke. (And I’m still. not. over it.)

When Our Digital Commons Began to Fracture

Koutropoulos et. al (2024) speak more eloquently to this fragmentation. They capture something I’ve been feeling but fail to articulate as clearly, the way our digital spaces have become increasingly unstable, the way platforms that once felt like home can shift beneath our feet overnight. Their collaborative autoethnography explores the metaphors we use to describe this movement and ultimately concludes that no single term captures what’s happening. What resonates most is their observation that we were never truly in control of these spaces, we were building communities on “the fickle and shifting sands of capitalism.”

Commercial generative AI feels social. You’re chatting, prompting, getting responses that (by design) seem thoughtful and engaged. But fundamentally, it is not social. You’re talking with biased algorithms. There is no human in the loop, no colleague who might push back on your thinking from their own lived experience, no peer who might share a resource you’d never have found on your own, no friend who might simply say “I’ve been thinking about this too.”

I haven’t seen genuine sharing built into any commercial generative AI tools. NotebookLM will let you share content that others can interact with, other tools allow you to create bots – but again, you’re not linking with a human. You’re not building a web of ambient findability (Morville, 2005) that made those early social media days so generative. There’s no AI equivalent of stumbling upon a colleague’s carefully curated collection and thinking, “Oh, they’re interested in this too – I should reach out.”

So in this fragmented, overly connected yet profoundly disconnected world, how do we stay connected to each other and each other’s ideas? I need my thought network now more than ever. And I suspect you do too.

Choosing Human Connection in an Algorithmic Age

Here are a few tools that have helped me navigate this landscape:

Raindrop.io – it’s not as social as del.icio.us was (oh, how I miss those days!), but it is a bookmark management tool that helps me keep track of the deluge of AI articles (and all sorts of other things) coming my way. I’ve made my collection public because (surprise!) I believe in working out loud and sharing what I’m learning. You can find it here: https://raindrop.io/leigh-wolf/ai-62057797.

RSS is Awesome is an “in-progress passion project” by Tom Hazledine. It has now become my morning ritual to open up this lovely, lightweight, no-login-needed, browser-based tool to catch up on my feeds. There’s something deeply satisfying about returning to RSS – a technology that puts the reader in control rather than an algorithm. (And yes, you can add the GenAI:N3 Blog to your feed simply by adding this URL: https://genain3.ie/blog/)

We need each other more than ever to navigate this sea of (mis)information. The platforms are fragmenting, the algorithms are optimising for engagement rather than insight, and AI offers the feeling of conversation without its substance – but we still have each other. We still have the ability to share, to curate, to point each other toward ideas worth wrestling with.

As Koutropoulos et al. (2024) challenge us, the solution isn’t to find the perfect platform – it’s to “take charge of our own data” and to invest in relationships with the humans in our educational networks. The platforms will always ebb and flow. But the connections we build with each other can (and do!) persist across whatever digital landscape emerges next.

Hold on to each other. Hold on to the tools that are enabling rather than disabling us to do this work together. And maybe, just maybe, start (re)building those thought networks – one shared bookmark, one RSS subscription, one genuine human connection at a time.

What tools are helping you stay connected to others’ thinking? What spaces have you found that still feel like home? I would love to know – please reach out in the comments below!!

Reference

Koutropoulos, A., Stewart, B., Singh, L., Sinfield, S., Burns, T., Abegglen, S., Hamon, K., Honeychurch, S., & Bozkurt, A. (2024). Lines of flight: The digital fragmenting of educational networks. Journal of Interactive Media in Education, 2024(1), 11. https://doi.org/10.5334/jime.850

Morville, P. (2005). Ambient findability: What we find changes who we become. O’Reilly.

Leigh Graves Wolf

Assistant Professor
University College Dublin

Leigh Graves Wolf is teacher-scholar and an Assistant Professor in Educational Development with Teaching and Learning at UCD. Her work focuses on online education, critical digital pedagogy, educator professional development and relationships mediated by and with technology. She has worked across the educational spectrum from primary to higher to further and lifelong. She believes passionately in collaboration and community.

Keywords


Building the Manifesto: How We Got Here and What Comes Next

By Ken McCarthy
Estimated reading time: 6 minutes
A minimalist illustration featuring the silhouette of a person standing and gazing toward a horizon line formed by soft, glowing digital patterns and abstract light streams. The scene blends naturalistic contemplation with modern technology, symbolizing human agency in shaping the future of AI against a clean, neutral background. Image (and typos) generated by Nano Banana.
Looking ahead: As we navigate the complexities of generative AI in higher education, it is crucial to remember that technology does not dictate our path. Through ethical inquiry and reimagined learning, the horizon is still ours to shape. Image (and typos) generated by Nano Banana.

When Hazel and I started working with GenAI in higher education, we did not set out to write a manifesto. We were simply trying to make sense of a fast-moving landscape. GenAI arrived quickly, finding its way into classrooms and prompting new questions about academic integrity and AI integration long before we had time to work through what it all meant. Students were experimenting earlier than many staff felt prepared for. Policies were still forming.

What eventually became the Manifesto for Generative AI in Higher Education began as our attempt to capture our thoughts. Not a policy, not a fully fledged framework, not a strategy. Just a way to hold the questions, principles, and tensions that kept surfacing. It took shape through notes gathered in margins, comments shared after workshops, ideas exchanged in meetings, and moments in teaching sessions that stayed with us long after they ended. It was never a single project. It gathered itself slowly.

From the start, we wanted it to be a short read that opened the door to big ideas. The sector already has plenty of documents that run to seventy or eighty pages. Many of them are helpful, but they can be difficult to take into a team meeting or a coffee break. We wanted something different. Something that could be read in ten minutes, but still spark thought and conversation. A series of concise statements that felt recognisable to anyone grappling with the challenges and possibilities of GenAI. A document that holds principles without pretending to offer every answer. We took inspiration from the Edinburgh Manifesto for Teaching Online, which reminded us that a series of short, honest statements can travel further than a long policy ever will.

The manifesto is a living reflection. It recognises that we stand at a threshold between what learning has been and what it might become. GenAI brings possibility and uncertainty together, and our role is to respond with imagination and integrity to keep learning a deeply human act .

Three themes shaped the work

As the ideas settled, three themes emerged that helped give structure to the thirty statements.

Rethinking teaching and learning responds to an age of abundance. Information is everywhere. The task of teaching shifts toward helping students interpret, critique, and question rather than collect. Inquiry becomes central. Several statements address this shift, emphasising that GenAI does not replace thinking. It reveals the cost of not thinking. They point toward assessment design that rewards insight over detection and remind us that curiosity drives learning in ways that completion never can .

Responsibility, ethics, and power acknowledges that GenAI is shaped by datasets, values, and omissions. It is not neutral. This theme stresses transparency, ethical leadership, and the continuing importance of academic judgement. It challenges institutions to act with care, not just efficiency. It highlights that prompting is an academic skill, not a technical trick, and that GenAI looks different in every discipline, which means no single approach will fit all contexts.

Imagination, humanity, and the future encourages us to look beyond the disruption of the present moment and ask what we want higher education to become. It holds inclusion as a requirement rather than an aspiration. It names sustainability as a learning outcome. It insists that ethics belong at the beginning of design processes. It ends with the reminder that the horizon is still ours to shape and that the future classroom is a conversation where people and systems learn in dialogue without losing sight of human purpose

How it came together

The writing process was iterative. Some statements arrived whole. Others needed several attempts. We removed the ones that tried to do too much and kept the ones that stayed clear in the mind after a few days. We read them aloud to test the rhythm. The text only settled into its final shape once we noticed the three themes forming naturally.

The feedback from our reviewers, Tom Farrelly and Sue Beckingham, strengthened the final version. Their comments helped us tighten the language and balance the tone. The manifesto may have two named authors, but it is built from many voices.

Early responses from the sector

In the short time since the manifesto was released, the webpage has been visited by more than 750 people from 40 countries. For a document that began as a few lines in a notebook, this has been encouraging. It suggests the concerns and questions we tried to capture are widely shared. More importantly, it signals that there is an appetite for a conversation that is thoughtful, practical, and honest about the pace of change.

This early engagement reinforces something we felt from the start. The manifesto is only the beginning. It is not a destination. It is a point of departure for a shared journey.

Next steps: a book of voices across the sector

To continue that journey, we are developing a book of short essays and chapters that respond to the manifesto. Each contribution will explore a statement within the document. The chapters will be around 1,000 words. They can draw on practice, research, disciplinary experience, student partnership, leadership, policy, or critique. They can support, question, or challenge the manifesto. The aim is not agreement. The aim is insight.

We want to bring together educators, librarians, technologists, academic developers, researchers, students, and professional staff. The only requirement is that contributors have something to say about how GenAI is affecting their work, their discipline, or their students.

An invitation to join us

If you would like to contribute, we would welcome your expression of interest. You do not need specialist expertise in AI. You only need a perspective that might help the sector move forward with clarity and confidence.

Your chapter should reflect on a single statement. It could highlight emerging practice or ask questions that do not yet have answers. It could bring a disciplinary lens or a broader institutional one.

The manifesto was built from shared conversations. The next stage will be shaped by an even wider community. If this work is going to stay alive, it needs many hands.

The horizon is still ours to shape. If you would like to help shape it with us, please submit an expression of interest through the following link: https://forms.gle/fGTR9tkZrK1EeoLH8

Ken McCarthy

Head of Centre for Academic Practice
South East Technological University

As Head of the Centre for Academic Practice at SETU, I lead strategic initiatives to enhance teaching, learning, and assessment across the university. I work collaboratively with academic staff, professional teams, and students to promote inclusive, research-informed, and digitally enriched education.
I’m passionate about fostering academic excellence through professional development, curriculum design, and scholarship of teaching and learning. I also support and drive innovation in digital pedagogy and learning spaces.

Keywords


Teaching the Future: How Tomorrow’s Music Educators Are Reimagining Pedagogy

By James Hanley, Oliver Harris, Caitlin Walsh, Sam Blanch, Dakota Venn-Keane, Eve Whelan, Luke Kiely, Jake Power, and Alex Rockett Power in collaboration with ChatGPT and Dr Hazel Farrell
Estimated reading time: 7 minutes
A group of eight music students from the BA (Hons) Music program at SETU are pictured in a futuristic, neon-lit "futureville" setting. They are gathered around a piano, which glows with digital accents, against a backdrop of towering, illuminated cityscapes and flowing data streams.
The future is now! BA (Hons) Music students from SETU in a vibrant “futureville” setting, blending the timeless artistry of music with cutting-edge technological imagination.

In recognition of how deeply AI is becoming embedded in the educational landscape, a co-created assignment exploring possibilities for music educators was considered timely. As part of the Year 3 Music Pedagogy module at South East Technological University (SETU), students were tasked with designing a learning activity that meaningfully integrated AI into the process. They were asked not only to create a resource but to trial it, evaluate it, and critically reflect on how AI shaped the learning experience. A wide range of free AI tools were used, including ChatGPT, SUNO, Audacity, Napkin, Google Gemini, Notebook LM, and Eleven Labs, and each student focused on a teaching resource that resonated with them, such as interactive tools, infographics, lesson plans, and basic websites.

Across their written and audio reflections, a rich picture emerged: AI is powerful, fallible, inspiring, frustrating, and always dependent on thoughtful human oversight. This blog is based on their reflections which reveal a generation of educators learning not just how to use AI, but why it must be used with care.

Expanding Pedagogical Possibilities

Students consistently highlighted AI’s ability to accelerate creativity and resource development. Several noted that AI made it easier to create visually engaging materials, such as diagrams, colourful flashcards, or child‑friendly graphics. One student reflected, “With just a click of the mouse, anyone can generate their own diagrams and flash cards for learning,” emphasising how AI allowed them to design tools they would otherwise struggle to produce manually.

Others explored AI‑generated musical content. One student used a sight‑reading generator to trial melodic exercises, observing that while the exercises themselves were well‑structured, “the feedback was exceedingly generous.” Another used ChatGPT to build a lesson structure, describing the process as “seamless and streamlined,” though still requiring adjustments to ensure accuracy and alignment with Irish terminology. One reflection explained, “AI can create an instrumental track in a completely different style, but it still needs human balance through EQ, compression, and reverb to make it sound natural.” This demonstrated how AI and hands-on editing can work together to develop both musical and technical skills.

An interactive rhythm game for children was designed by another student who used ChatGPT to progressively refine layout, colour schemes, difficulty levels, and supportive messages such as “Nice timing!” and “Perfect rhythm!” They described an iterative process requiring over 30 versions as the model continuously adapted to new instructions. The result was a working single‑player prototype that demonstrated both the creative potential and technical limits of AI‑assisted design.

The Teacher’s Role Remains Central

Across all reflections, students expressed strong awareness that AI cannot replace fundamental aspects of music teaching. Human judgment, accuracy, musical nuance, and relational connection were seen as irreplaceable. One student wrote that although AI can generate ideas and frameworks, “the underlying educational thinking remained a human responsibility.” Another reflected on voice‑training tools, noting that constant pitch guidance from AI could become “a crutch,” misleading students into believing they were singing correctly even when not. Many recognised that while AI can speed up creative processes, the emotional control, balance, and overall musical feel must still come from human input. One reflection put it simply: “AI gives you the idea, but people give it life.”

There was also a deep recognition of the social dimension of teaching. As one student put it, the “teacher–student relationship bears too much of an impact” to be substituted by automated tools. Many emphasised that confidence‑building, emotional support, and adaptive feedback come from real educators, not algorithms.

Challenges, Risks, and Ethical Considerations

The assignment surfaced several important realisations, including the fact that technical inaccuracies were common. Students identified incorrect musical examples, inconsistent notation, malfunctioning website features, and audio‑mixing problems. One student documented how, over time, the “quality of the site got worse,” illustrating AI’s tendency to forget earlier instructions in long interactions. This reinforced the need for rigorous verification when creating learning materials.

Another reflection noted that not all AI websites perform equally; some produce excellent results, while others generate distorted or incomplete outputs, forcing teachers to try multiple tools before finding one that works. It also reminded educators that even free or simple programs, like basic versions of Audacity, can still teach valuable mixing and editing skills without needing expensive software. A parallel concern was over‑reliance. Students worried that teachers might outsource too much planning to AI or that learners might depend on automated feedback rather than developing critical listening skills. As one reflection warned, “AI can and will become a key tool… the crucial factor is that we as real people know where the line is between a ‘tool’ and a ‘free worker.’”

Equity of access also arose as a barrier. Subscription‑based AI tools required credits or payment, creating challenges for students and highlighting ethical tensions between commercial technologies and educational use. Students demonstrated strong awareness of academic integrity. They distinguished between using AI to support structure and clarity versus allowing AI to generate entire lessons or presentations. One student cautioned that presenting AI‑produced content as one’s own is “blatant plagiarism,” highlighting the need for transparent and ethical practice.

Learning About Pedagogy and Professional Identity

Many students described developing a clearer sense of themselves as educators. They reflected on the complexity of communicating clearly, engaging learners, and designing accessible content. Some discovered gaps in their teaching confidence; others found new enthusiasm for pedagogical design. One wrote, “Teaching and clearly communicating my views was more challenging than I assumed,” acknowledging the shift from student to teacher mindset. Another recognised that while AI could support efficiency, it made them more aware of their responsibility for accuracy and learner experience.

Imagining the Future of AI in Music Education

Students were divided between optimism and caution. Some saw AI becoming a standard part of educational resource creation, enabling personalised practice, interactive learning, and rapid content generation. Others expressed concern about the possibility of AI replacing human instruction if not critically managed. However, all students agreed on one point: AI works best when treated as a supportive tool rather than an autonomous teacher. As one reflection summarised, “It is clear to me that AI is by no means a replacement for musical knowledge or teaching expertise.” Another added, “AI can make the process faster and more creative, but it still needs the human touch to sound right.”

Dr Hazel Farrell

Academic Lead for GenAI, Programme Leader BA (Hons) Music
South East Technological University

Dr Hazel Farrell is the SETU Academic Lead for Generative AI, and lead for the N-TUTORR National Gen AI Network project GenAI:N3, which aims to draw on expertise across the higher education sector to create a network and develop resources to support staff and students. She has presented her research on integrating AI into the classroom in a multitude of national and international forums focusing on topics such as Gen AI and student engagement, music education, assessment re-design, and UDL.

Keywords