Skip to content
The Moving Finish Line
7 min read 8 February 2026

The rush to summarise: what enterprise L&D loses when AI reads for us

Microsoft Research found that professionals put less effort into critical thinking when working with AI. If the goal of enterprise L&D is behaviour change, learners need to engage with material - not just receive it.

James Pierechod

Founder, Visual Content Consultancy

TL;DR

  • AI summaries save time but cost comprehension depth
  • Enterprise L&D is outsourcing the thinking, not just the reading
  • Friction in learning isn't a bug - it's where understanding happens

The rush to summarise — when AI reads for us, what do we lose?

There’s a pattern emerging in enterprise L&D conversations that I’ve been sitting with for a while now. It goes something like this: we’ve got too much content, learners don’t have time, so let’s use AI to summarise everything and surface the right information at the right moment.

It sounds like progress. It might be a problem.

In the previous piece, I wrote about calculators - how educators spent twenty years figuring out what students should memorise versus what machines should handle. They got it right because they asked the question consciously. GPS is the counter-example: we offloaded spatial navigation by default, and research now shows measurable cognitive decline in habitual users.

The question for enterprise L&D is which path we’re on. Are we making conscious decisions about what AI should handle? Or are we sleepwalking into cognitive offloading?

The efficiency trap

Advait Sarkar, a senior researcher at Microsoft leading their Tools for Thought initiative, put it sharply in a recent TED talk: we’re at risk of becoming “professional validators of robots’ opinions.” Knowledge workers who no longer engage with the materials of their craft. Intellectual tourists who visit ideas but don’t inhabit them.

His team surveyed 319 professionals about their AI use and found something that should concern anyone designing learning experiences. Workers reported putting less effort into critical thinking when working with AI than when working manually. And this effect was stronger when they had higher confidence in AI and lower confidence in themselves.

The research gets more uncomfortable from there. On a collective level, knowledge workers using AI assistance produce a smaller range of ideas than groups working manually. People remember less of what they wrote when AI helped them write it. And when they read AI-generated summaries - which is increasingly how learning content gets consumed - they remember less than if they’d read the original document.

The summary tradeoff — what we gain in speed we lose in comprehension depth

The metacognition gap

What’s being lost isn’t just information. It’s what Sarkar calls “metacognition” - the ability to think about your own thinking process.

Working with AI introduces new cognitive demands: reasoning about task goals, decomposing the task, evaluating whether AI is even applicable, assessing the quality of output. These are things that are built into the process of working directly with material. They become problematic when that material engagement becomes intermediated.

To put it bluntly: we’ve become middle managers for our own thoughts.

This has direct implications for how we design enterprise learning. If the goal is behaviour change - and it usually is - then learners need to engage with material, not just receive it. They need to form their own judgments, struggle with ambiguity, and build the pattern recognition that comes from wrestling with complexity.

When AI summarises that struggle away, what exactly are we training?

The measurement problem

The measurement problem — why current L&D metrics may be actively misleading us

Here’s where this gets practically difficult. Most enterprise L&D measurement is still built around the old model: completion rates, assessment scores, time-to-competency. These metrics assume that consuming content equals learning.

But if the research is right, the relationship between consumption and capability is being fundamentally disrupted. A learner who reads an AI summary might score well on an immediate assessment while retaining almost nothing a week later. A learner who struggled through the original material might score worse initially but build durable capability.

We don’t have good frameworks for measuring this yet. I’ve been exploring what evaluation and retention scoring might look like in AI-augmented delivery structures - how you’d predict and benchmark retention aligned to capability sprints and learning frameworks. It’s early-stage thinking, but it feels like the right question.

The current metrics aren’t just inadequate. They might be actively misleading us about what’s working.

Kirkpatrick evolved — rethinking L&D evaluation for AI-augmented delivery

What this means for L&D builds

When I’m working on learning strategy with enterprise clients, I’m increasingly thinking about where human engagement is non-negotiable versus where AI orchestration adds genuine value.

The calculator precedent is useful here. Educators didn’t ban calculators - they redesigned curricula to focus on conceptual understanding, problem formulation, and mathematical reasoning. The arithmetic got offloaded; the thinking didn’t.

It’s not a blanket answer for L&D either. Some content genuinely benefits from AI summarisation and point-of-need delivery. Procedural knowledge, reference material, compliance updates - these can often be surfaced contextually without requiring deep cognitive engagement.

But capability building is different. Strategic thinking, judgment, pattern recognition, complex decision-making - these require what the researchers call “material engagement.” The learner needs to read, to struggle, to form opinions, to be wrong sometimes.

The design challenge is knowing where to place that friction intentionally.

The scaffold versus substitute question

Recent cognitive research draws a useful distinction between AI as scaffold and AI as substitute.

Scaffolding is temporary, adaptive, and empowering. The goal is to strengthen internal capacities so the technology becomes progressively less necessary. A learning system that gradually reduces support as the learner builds capability exemplifies scaffolding.

Substitution is permanent and dependency-creating. The technology assumes responsibility in ways that diminish intrinsic skills. An AI system that learners rely on continuously, without building transferable capability, exemplifies substitution.

The difference depends less on technical sophistication than on design philosophy, integration context, and patterns of use.

Most AI-augmented L&D I see defaults toward substitution without intending to. The efficiency gains are immediate and measurable. The capability costs are delayed and invisible - until they’re not.

Scaffold versus substitute — contrasting temporary support that builds capability with permanent dependency that diminishes skills

The opportunity in the problem

None of this is an argument against AI in enterprise learning. The tools are too powerful and the efficiency gains too real to ignore.

But it is an argument for intentionality. For understanding what we’re optimising for. For recognising that faster delivery and deeper capability might sometimes be in tension, and designing accordingly.

Sarkar’s research suggests that with the right principles of design, you can build tools that are the best of both worlds - applying the speed and flexibility of AI while protecting and enhancing human thought. His team has demonstrated that you can reintroduce critical thinking into AI-assisted workflows, reverse the loss of creativity, and build tools for memory that enable people to work at speed while retaining what they learn.

The principles aren’t complicated: preserve material engagement at strategic points, offer productive resistance rather than just frictionless delivery, and scaffold metacognition so learners are always thinking about their thinking.

The execution is harder. But that’s the work.


This is the second in a series on AI and enterprise learning. Next: a framework for designing learning experiences that protect and enhance thinking.

Common questions

Quick answers

Got another question?

What is cognitive offloading in enterprise learning?

Cognitive offloading is when learners rely on AI to handle thinking tasks - like reading, summarising, and forming judgments - rather than doing the cognitive work themselves. Research shows this leads to reduced retention and weaker critical thinking skills.

Does AI-assisted learning reduce critical thinking?

Microsoft Research found that professionals put less effort into critical thinking when working with AI, and the effect was stronger in those with higher confidence in AI. People also remember less of what they wrote when AI helped, and retain less from AI summaries than from reading originals.

What's the difference between AI as scaffold and AI as substitute in learning?

Scaffolding is temporary support that strengthens internal capacities so technology becomes progressively less necessary. Substitution is permanent and creates dependency. Most AI-augmented L&D defaults toward substitution without intending to.

Want to discuss this?

If this resonates with a challenge you're facing, let's talk.

Book a conversation