Skip to main content

Field notes

Teaching AI to Non-CS Students Without Dumbing It Down

Published Wed Apr 29 2026 00:00:00 GMT+0000 (Coordinated Universal Time)

How to teach genuinely useful AI to law, journalism, design, business, and humanities students — pedagogical patterns that work and the most common ways AI literacy courses fail.

AI curriculum · 2026-04-11 · ~10 min read

There is a common assumption that AI literacy for non-CS students has to be either a watered-down version of the engineering course or a tour of "ethics and society" with some demos. Both are wrong, and the difference between them and a course that actually works is mostly pedagogical, not technical.

Mixed-discipline classroom with students reviewing AI output on shared laptops

This piece is about the pedagogy. It draws on what we have learned building Track D — AI Literacy for All Disciplines, which has 10 discipline modules running across Journalism, Law, Healthcare, Education, Liberal Arts, Economics, Psychology, Performing Arts, Sports & Hospitality, and Languages.

The difference between a course that works and one that doesn't is rarely the syllabus. It is the pedagogy underneath it.

Diagnose the two failure modes

Most AI-literacy courses for non-CS students fail in one of two predictable ways. Both are visible in the first three weeks of the semester.

Recognise "AI for poets"

The course is conceptual, never goes near a model, talks about ethics and futures and impact. Students are left able to write a thoughtful essay about AI and unable to use it to do anything.

Recognise "B.Tech-lite"

The course tries to compress a CS curriculum into one semester, runs a Python primer, gestures at gradient descent, and assigns a homework involving sklearn. Students who cannot already code are bewildered and demoralised. Students who can already code are bored.

Both failure modes share an assumption that does not survive contact with the classroom: that you have to choose between real and accessible. You don't.

Dimension
AI for poets
B.Tech-lite
What it teaches
Conceptual only. Ethics-and-futures essays.
Compressed CS curriculum. Python primer + sklearn homework.
Where it lands
Students can describe AI but not use it.
Bewilders non-coders, bores those who already code.
Underlying assumption
Real means CS. So we strip the realness.
Accessible means dumbed down. So we keep the CS.

Reframe the goal as capability without translation

The reframe is simple to state. A non-CS undergraduate does not need to be turned into a junior engineer. They need to become someone who can:

  • Use AI to do their own discipline's work substantially better
  • Exercise judgment about where it works and where it doesn't
  • Hold enough mechanical understanding to debug their own use
  • Recognise the failure modes specific to their domain

That is a different goal than "be a CS student." It is also a different goal than "have an opinion about AI." And it is achievable in a semester or two, if the course is built for it.

Map the discipline-specific outcomes

The capability you are aiming for is not generic. A journalism graduate needs different AI fluency than a law graduate. The course design should make those outcomes legible from week one.

  • Journalism

    Detect deepfakes and synthetic media, verify claims at scale, draft with AI while preserving editorial voice.

  • Law

    Stress-test arguments, accelerate doctrinal research, audit AI-generated drafts for hallucinated citations.

  • Design

    Use generative tools as ideation partners, critique outputs against brand systems, brief models like collaborators.

  • Healthcare

    Read AI-summarised case notes critically, recognise where models flatten clinical nuance, document safely.

  • Management

    Run AI-augmented analysis on memos and decks, evaluate vendor claims, design workflows that survive model drift.

  • Liberal arts & languages

    Interrogate translations and summaries for cultural loss, use AI for archival reading, defend interpretive judgment.

Discipline-anchoring is not a design flourish. It is the load-bearing variable.

Apply six pedagogical patterns that work

Six patterns, all of which we have stress-tested across disciplines. Each one trades a familiar pedagogical shortcut for something harder to operationalise — and substantially more effective.

Faculty member coaching a small group through an AI evaluation rubric

Anchor every concept to a discipline-native problem

A psychology student does not learn what an LLM is via a generic chatbot demo. She learns it via a comparative analysis of how three different LLMs respond to the same therapy intake transcript — which one preserves clinical nuance, which one flattens it, why. The discipline carries the conceptual weight.

This pattern needs real discipline expertise on the teaching side. A CS instructor cannot do it alone. This is why our Track E couples discipline faculty with AI methodology training — neither half alone is sufficient.

Teach evaluation before generation

The most common failure pattern in non-CS AI use is that students learn to make AI produce things before they learn to judge whether the output is any good. This produces overconfident graduates. Our discipline modules invert the order: evaluation rubrics first, generation second.

A journalism student first learns to evaluate AI-summarised articles for factual drift. Only after that does she use AI to draft her own.

Use code-aware interfaces, not raw notebooks

A B.A. student does not need to learn Jupyter as a precondition for AI work. She needs to learn to think clearly about what she wants the system to do, how to verify that it did it, and where its limits are.

The interface for that is a code-aware chat interface (Claude, ChatGPT, Cursor, etc.) where she can read and review code without having to write it from scratch. Programming-track students who want depth go to Track A. Track D students should not be gatekept by syntax.

Build mixed-discipline teams deliberately

A capstone group of three psychology students, two journalism students, and a CSE student produces visibly better output than three groups of six same-discipline students. The structure forces translation, surfaces assumptions, and gives the CSE student practice explaining their work to non-engineers — which is, separately, a skill they will need.

This is administratively annoying for universities to operationalise. It is one of the things our Track D deployment includes by default.

Assess on artefacts, in panels, with externals

The grading scheme that produces actual capability is one where:

  • The assessment is a real artefact — a portfolio piece, a final project, a published-grade output
  • The review panel includes at least one practitioner from the discipline
  • Externals are named in the syllabus so students know who they are writing for

The MCQ-graded version of this course produces nothing.

Refresh on a 9-month cycle, not a 3-year one

AI moves faster than most academic refresh cycles. A Track D module that worked in autumn 2024 had to be substantially redone for autumn 2025 because the toolchain changed. Universities need to commit operationally to refresh cycles that do not align with traditional curriculum review timelines.

  1. Faculty cohort

    Pair a discipline expert with an AI-trained co-instructor.

  2. Evaluation rubrics

    Write the grading sheet before the lecture deck.

  3. Discipline-native problems

    Source live artefacts from the field, not generic demos.

  4. Mixed teams

    Composition set centrally, not student-selected.

  5. Panel assessment

    At least one external practitioner per panel.

  6. 9-month refresh

    Calendar review baked into the operational plan.

Avoid the patterns that look reasonable but fail

Three patterns that look reasonable and reliably underperform:

  1. Self-paced online modules with proctored end-of-term exams. The discipline-anchoring (Pattern 1) is the high-leverage variable, and self-paced modules cannot do it.
  2. "Bring your own AI" project structures with no scaffolding. Students need pattern libraries before they can produce. Open-ended projects from Day 1 produce shallow output.
  3. Guest-lecture-heavy structures. Coherence beats novelty. A course taught by twelve different industry experts feels exciting and produces less learning than one taught by two faculty across the semester.
Empty lecture hall with a guest-speaker poster on a stand

Confront the faculty problem, again

Every pattern above pushes more weight onto faculty. The teaching skill required to run a Track D module well is genuinely higher than the skill required to run a traditional discipline course, and the field is too new for there to be a deep teaching pool.

Without an aligned faculty cohort on the partner campus, even a perfectly designed Track D module collapses in execution.

This is the structural reason we built Track E before scaling Track D. The constraint is human, not curricular.

Faculty members in a workshop annotating a syllabus together

Take the operational order of operations to your campus

If you are building this course on your own campus, the order of operations that we recommend:

  1. Get the faculty right, even if it slows down the launch
  2. Build evaluation rubrics before content
  3. Anchor every concept to discipline-native problems
  4. Use mixed-discipline teams
  5. Assess in panels, with externals
  6. Build refresh into the operational calendar

Use the open materials

Track D structures and rubrics are public.

Module structures, evaluation rubrics, and per-discipline artefact briefs are in the curriculum library. Adapt them, push back on them, send us what breaks.

Open the curriculum library

The full Track D structure and the open materials are in our curriculum library. Use them, adapt them, push back on them. We update them based on what we learn from people doing this in the wild.

Related reading