L&B Blog – Education & Teacher Conferences Skip to main content
Interleaving Meets Prequestioning: A Powerful Combination for Learning?
Andrew Watson
Andrew Watson

A few weeks ago, I wrote about an intriguing research question. We know that retrieval practice helps students learn. And we know that generative drawing helps students learn. So: what happens if we combine those two research-informed strategies?

As I wrote back then, I think we have three answers to that question:

  • Answer #1: combining strategies didn’t help. Students remembered the same amount whether or not they retrieved as they drew.
  • Answer #2: Wait a minute: I’m actually not so sure about answer #1. The “retrieval” in this study didn’t really match the typical definition of “retrieval practice.” The students recalled something they had just read, not something they had learned a few days before. In other words: they were “retrieving” from working memory, not from long-term memory. That’s not really retrieval practice.
  • Answer #3: although we haven’t learned the answer to the question–at least in my opinion–we have learned that research into combining strategies gets tricky.

I’ve just come across another study asking a similar “is two better than one?” question. Here goes:

  • We know that jumbling practice on related topics together helps students learn. That is: rather than practice “adjectives” and “adverbs” separately, students benefit from doing practice questions on both topics together. (We call that approach “interleaving.”)
  • And we know that asking students to answer questions before they have studied the answer helps students learn. (We call that technique “prequestioning” or “pretesting.”)
  • Dr. Steven Pan and his colleagues wanted to know: does interleaving + prequestioning result in more learning than either strategy on its own? Or, the same amount of learning? Less learning?

Let’s explore…

“What Is Bipolar Disorder?”

Using an online research platform, Pan and his colleagues asked adults to learn about six different mental conditions: ADHD, bipolar disorder, etc.. To learn this information, participants read short “case study” passages about three people with each condition.

  • Half of the participants read three consecutive passages about any one condition. That is: they read 3 passages about a person with schizophrena; then 3 about someone with autism; and so forth. (This structure is the opposite of “interleaving”; we call it “blocking.”)
  • The other half of the participants read those same passages–but all jumbled together. So, they might read a passage about someone with ADHD, then another about someone with schizophrenia, then another about someone with bipolar disorder. (Yup: this is “interleaving.”)

So: all the participants read the same 18 passages in groups of 3 passages. But half of the participants read them in “blocked” groups; the other half read them jumbled up in an “interleaving” structure.

So far, this research explores the relative benefits of blocking or interleaving. But what about the “prequestions“?

  • Within each of those groups, half of the participants saw the list of technical names for the six conditions that they’re learning about: “cyclothymic affect disorder,” “resonance development disorder,” “schismic cognition disorder,” and so forth. (The researchers deliberately chose obscure technical names so that participants couldn’t rely on prior knowledge—they had to learn the associations from scratch.) Participants had to guess: which one of these six is the passage about?
    • They had to answer these questions without having yet learned about the conditions. That’s a “prequestion.”
  • The other half saw the technical name at the top of each passage they were reading. So: a case study about bipolar disorder would be labeled “cyclothymic affect disorder.”
    • No prequestion here.

With these groups and subgroups, Pan and Co. can now discover which combination helps students learn the most:

  • It could be BLOCKING, with our without PREQUESTIONS
  • It could be INTERLEAVING, with our without PREQUESTIONS

To be extra careful, the researchers checked to see how much the participants learned five minutes later. And then — with a completely different group of participants — they checked to see how much they learned 48 hours later.

We’ve Got Answers (I’ve Got Questions)

Team Pan wanted to know: does combining interleaving (a structure) with prequestioning (a technique) increase learning?

In a word: YES.

  • Both right away and 48 hours later, these adults learned more about identifying cognitive disorders if they read jumbled case studies (“interleaving) and had to guess what each one was (“prequestioning”).
  • The group that read the case studies all clumped together (“blocking”) without prequestions learned the least.
  • And the groups that either studied with interleaved case studies or answered prequestions scored in the middle — roughly the same as each other.

On the one hand, this helpfully direct answer gives teachers useful guidance. Sure enough, using an interleaving structure with a prequestioning technique helps students learn.

I do, however, have two questions.

First: A quiz after two days gives us more useful information than a quiz right away. But “two days” isn’t yet learning. In fact, we’ve got all sorts of reasons to worry that “short term performance does not accurately predict long-term learning.” (Check out this research review.) I will feel MUCH more confident offering this guidance if we have studies confirming these results after noticeably longer periods of time.

While I’m at it, I should say: adult learners might benefit from different combinations than K-12 learners. For this reason, I’d also like to see replications with different age groups.

To be clear, Pan and his colleagues haven’t done anything wrong by limiting their study to two days, and working with adults. No one study answers all questions. They’ve given us a helpful set of data points. I myself think we need several more collections of data points before we offer strong advice. They themselves raise this point in their study.

Second: The procedure described above uses a version of “prequestioning”; but it’s not exactly “prequestioning.” With typical prequestions, students almost certainly don’t know the answer because they haven’t yet studied the material. In this case, the students (probably) don’t know the answer at the beginning of the study. However, over time, they have more and more reason to know the answer.

To my way of thinking, this study STARTS by combining interleaving with prequestioning, and ENDS UP combining interleaving with a version of retrieval practice. After all: by the final round or two, the students can start recalling the correct answers that they’ve been studying up to now.

Here again, I’m not describing a flaw in the study. At the same time, I do think it’s helpful to have our terminology as clear as possible.

To Combine Or Not

Based on this research, we have a tentative reason to believe that starting with questions that students can’t yet answer (“prequestioning”) and jumbling practice problems together (“interleaving”) helps students learn.

As teachers, we can start trying out this combination of research-supported strategies.

And: we can keep an eager eye out for more research into strategy combinations. A whole new fruitful field awaits…


Pan, S. C., Selvarajan, G., & Murphy, C. S. (2024). Interleaved pretesting enhances category learning and classification skills. Journal of Applied Research in Memory and Cognition13(3), 393.

From Facts to Application: The Surprising Power of Repeated Retrieval
Andrew Watson
Andrew Watson

The more time you’ve spent in Learning and the Brain world, the more you’ve heard about retrieval practice.

  • If students practice by looking back over work they have done, we call that “simple review.”
    • “Simple review” doesn’t benefit learning very much, if at all.
  • If students practice by trying to recall information and concepts, we call that “retrieval practice.”
    • Compared to “simple review,” “retrieval practice” enhances learning A LOT.

So: rather than reviewing their notes or rereading the textbook, students should ask themselves and each other questions. The additional mental struggle of trying to answer those questions will — in the long run — help them learn the material.

This basic idea has been around for decades, and has gotten lots of research love in the last 20 years. If you’d like to brush up on previous research, you can:

Of course, researchers constantly explore new questions within this well-established field. One perfectly reasonable question sounds like this:

“Okay, we know that retrieval practice helps students recall facts and concepts. But does it help them APPLY those facts and concepts? We want students to USE information, not just HAVE information.”

Good point!

A recent study by Corral and Carpenter explores this question across three experiments, and offers a few striking conclusions.

The Setup

Corral and Carpenter ran three experiments with several hundred college students. These students studied slides that defined terminology relevant for a research methodology course: e.g., “reverse causation.”

After looking at the slides:

  • Some students simply reviewed the information; they looked at the slides again.
  • Some took a short-answer quiz about it (that’s retrieval practice).
  • Some studied the quiz questions and answers, but didn’t have to answer them. (A different kind of review.)
  • And some did none of those things; that is, they saw the information only once.

As a final step, students took a test. Some test questions checked to see if they recalled the information. Students could get these answers right if they simply memorized the slides. Crucially, other test questions checked to see if they could apply the information in a novel situation. For these questions, simple memorization wouldn’t get the job done — students had to transfer their knowledge to a new situation.

For instance:

  • A RECALL question about reverse causation:
    • “When it is unclear whether variable X causes variable Y or whether variable Y causes variable X, what kind of causal inference problem do we have?”
  • An APPLICATION question about reverse causation:
    • Lisa just received her master’s for her work in astronomy. Lisa found that planets that are closer to their moons rotate faster than planets that are farther away from their moons. However, Lisa is not sure how to interpret her data because it is unclear whether a planet that rotates faster draws its moons in closer to the planet or whether moons that are closer to a planet cause the planet to rotate faster. Ignoring the possibility that other factors may be involved, what kind of problem does Lisa have?

Headline Number One: A Big Nuthin’

In the short term, the review strategy didn’t matter.

That is: when there was only an eight minute gap between the review and the final test, then (on average) none of the review strategies helped students score better than their peers who didn’t review. Not retrieval practice; not quiz review; not slide study. Nothing.

And, by the way, they found this same result for both the recall questions and the application questions.

This finding reminds us of an important truth: we’re looking for a study strategy that helps students learn more over the long term. Of course, EIGHT MINUTES isn’t the long term. You can be in and out of Starbucks in under eight minutes. Heck, you can probably read this blog post in under eight minutes.

I’m emphasizing this point because it answers an essential question: “if retrieval practice is so great, why didn’t teachers and students figure it out on our own? If it’s obviously better than review, why didn’t we stumble across this strategy centuries ago?”

We didn’t figure it out because its benefits take a while to appear. Retrieval practice doesn’t help right away. It helps long-term learning.

Headline Number Two: More Reps + More Time = Better Application

Corral and Carpenter repeated their experiment twice: this time with two important changes.

  • First: students used their review strategy — either “simple review” or “retrieval practice” or “quiz review” — THREE TIMES, not just once.
  • Second: they waited a week to take that final test. (You have already noticed: one week is a lot more time than eight minutes.)

Sure enough, once a meaningful amount of time had passed, students who took three retrieval practice quizes scored higher on the recall questions. AND, they scored higher on the application questions.

In other words: repeated retrieval practice helps students recall facts and definitions. But, it doesn’t only do that. It also helps students transfer that knowledge and apply it in new ways.

This finding merits attention because it pushes back against one of the most common concerns about retrieval practice: “it’s just for memorizing disconnected facts! It doesn’t help students use their learning!”

Well, if Corral and Carpenter are correct, retrieval practice does help students use their learning. Of course, retrieval practice might not help right away — after 8 minutes. But repeated retrieval practice does help students over the long run.

Teaching Implications

First, in language proposed by Agarwal and Bain:

  • Rather than ask students to put information back into their brains (that’s “simple review”),
  • Ask them to try to take information out of their brains (that’s retrieval practice).

Second: the exact form of retrieval practice isn’t terribly important. According to an ENORMOUS meta-analysis I describe here,

  • Multiple choice and short-answer questions work just fine. (Contrary to popular advice, “brain dumps” are less effective.)
  • Combining retrieval practice with correct-answer feedback is good, but retrieval practice helps even without feedback.
  • Graded or ungraded: doesn’t matter. (Me: I prefer ungraded.)
  • Repeated retrieval practice helps students learn and transfer more than one-time retrieval practice. (No doubt there’s a limit here, but I suspect most students benefit from “more than they are currently doing.”)

Third: don’t be misled by the critics. Repeated retrieval practice helps remember facts, ideas, and procedures; and it helps students use those facts, ideas, and procedures.

This research offers real encouragement: when we build repeated retrieval practice into our teaching, we’re equipping students with knowledge they can actually use.


Corral, D., & Carpenter, S. K. (2025). Effects of retrieval practice on retention and application of complex educational concepts. Learning and Instruction100, 102219.

From Lab to Life: Testing Study Strategies with 2,500+ Real Students
Andrew Watson
Andrew Watson

Research-based conclusions often appeal to us because of their tidy clarity. For example, research shows that…

  • Retrieval practice helps students learn, compared to simple review;
  • Spreading practice out (“spacing”) help students learn better than doing all the practice at once (“massing”);
  • A quiet study environment fosters concentration and learning;
  • Study strategies like rereading and highlighting are almost certainly a waste of time.

These suggestions — and many, many others — have LOTS of research behind them. We’ve got numbers and graphs and equations and peer review and replication. All that science-y good stuff.

Of course, the actual gritty world of classroom learning rarely achieves this level of regulated clarity. Students want to play with their friends. A spider will appear, seemingly with fangs. The copying machine got glitchy, and we have only half as many handouts as we need.

For all these reasons, we need to do research BOTH in psychology labs AND in actual classrooms. This second kind of research lets us know what happens when scientific rigor meets day-to-day school muddle. (A classic of this genre: Powerful Teaching by Dr. Pooja Agarwal and Patrice Bain.)

Alas, even classroom-based research doesn’t tell the whole story. Once students start doing academic work at home, then their own beliefs and practices add an additional layer of reesarch complexity. If classrooms are “messy” compared to psychology labs, imagine how much “MESSIER” an individual student’s academic life could be.

A recent study surveys Spanish high school students — from 7th to 10th grade — to understand exactly what they do to study at home, and exactly what they believe about learning. It then correlates those practices and beliefs with students’ academic success. If everything works out just right, we can discover:

  • Are students DOING all those research-based study practices, like retrieval practice and spacing?
  • Are they AVOIDING all the useless practices, like highlighting?
  • Are their grades confirming the researchers’ predictions? That is: do students who study in quiet rooms learn more than those who don’t?

In other words: do students (in their cluttered lives) get the same results from study practices that researchers (in their pristine psychology labs) do?

Good News/Bad News, Part I

I read A LOT of studies for my work, and this one stands out to me for its combination of research chops and common sense. The easiest point to notice: the authors report data on almost 2500 students! Given this very large sample size, we can have some confidence that flukey outliers are not skewing the results.

Although I won’t go into the methodological calculations, I also admire the care with which the authors developed their question list.

All this procedural good news does come with some caveats.

First: This study asks students what do and think. Alas, self-report data don’t inspire lots of confidence. If I ask you what you ate for lunch on August 15, you might remember. But your self report of that menu isn’t as reliable as most scientific variables. So too: when we ask students how they study, they might not remember — or might not admit that they study with music playing.

Second: the researchers use grades to measure academic success — although they admit that grades don’t measure learning very well. We have many reasons to doubt the relationship between grades and learning. For example:

  • Imagine that a student uses ineffective study strategies at home, but her teacher uses effective teaching practices in the classroom. The effective teaching might mask the effects of the ineffective studying.
  • Although cramming is bad for long-term learning, it’s good for short-term performance. If I cram for a test, I might get an A on that test…even though I won’t remember those ideas over the long term.

To make the broader point a different way:

  • Research in a psychology lab is useful, but not perfect;
  • Research in classrooms is useful, but not perfect;
  • Reasearch into students’ study practices is useful, but not perfect.

This study can give us helpful data, but it doesn’t provide some kind of comprehensive “final answer” to question about study practices. (No one research method can provide such a comprehensive answer.)

Good News/Bad News, Part II

If you ask 2500 students lots of questions, you’re going to get LOTS of data. For that reason, I will touch on a few highlights, but I’m not aiming for a comprehensive survey of this study’s findings. (If you want to know more, you can read the study itself here. It’s open access!)

The good news:

  1. Sure enough, research-supported practices (largely) support learning. So:
    • students who study in a quiet atmosphere get higher grades than those who listen to music while they study.
    • (by the way: most students report that they don’t listen to music while they study!)
    • students who use retrieval practice score higher than those who don’t.
    • those who “elaborate” on their learning — by creating concept maps, for example, or by generating their own examples — learn more than those who don’t.
  2. Sure enough, research-discouraged practices don’t help learning. So:
    • highlighting, copying, and rereading didn’t really do much to enhance learning.
  3. Sure enough: high “self-efficacy” belief correlates with more learning (although we don’t know if it enhances learning, or results from learning, or both).

The bad news:

  1. Compared to college students, high-school students use much less retrieval practice. (This result might not be surprising. Students who don’t use retrieval practice get lower grades…and are therefore less likely to go to college.)
  2. More puzzling: students who report spacing their practice don’t see benefits in higher grades. In other words, this research-supported practice doesn’t produce the touted benefits.

So, what’s going on with spacing?

As noted above, the authors have a potential explanation here. Cramming doesn’t help long-term learning, but it does improve short-term performance. Students who cram for a test might get the same high score as those who spaced out their practice. For this reason, the long-term learning benefits of spacing might not show up in research that uses grades as a proxy for learning.

Two students studying together sitting on a bench outdoors.

I myself have a second hypothesis. I think that students CAN control some study behaviors: retrieval practice, studying in a quiet space, not wasting time highlighting, and so forth. But some study practices are really OUTSIDE student control.

For spacing to work as researchers intended, a particular topic should be re-opened and re-practiced at unpredictable intervals over long periods of time. Spacing, in other words, requires syllabus-level planning. I just don’t think it’s reasonable to ask students — especially young high-school students — to do that. In brief: I suspect that some students believe that they’re “spacing,” and say so on those surveys. Alas, their “spacing” might not match researchers’ definitions of spacing. If students are not really spacing, they don’t get the benefit of spacing.

TL;DR

We have good research — both in psychology labs and in classrooms — to guide students’ study habits. In all that research, retrieval practice and elaboration help; highlighting and underlining don’t. Quiet study spaces foster learning; so do “self-efficacy” beliefs.

This study from Spain — with 2500 high-school students! — shows that those strategies probably work in real life as well as in more tightly controlled research settings.

Although teachers should probably be responsible for spacing, students can — and SHOULD — rely on those other study practices. When teachers give this advice to students, our guidance has the backing of multiple layers of research: in labs, in classrooms, and in students’ real lives.


Ruiz-Martín, H., Blanco, F., & Ferrero, M. (2024). Which learning techniques supported by cognitive research do students use at secondary school? Prevalence and associations with students’ beliefs and achievement. Cognitive Research: Principles and Implications9(1), 44.

Research Reality Check: ‘More Is Better’ Doesn’t Always Apply to Learning Strategies
Andrew Watson
Andrew Watson

We’ve got lots of strong research suggesting that drawing can improve memory for specific words.

And

We’ve got LOTS of strong research showing that retrieval practice enhances learning.

So

Simple logic suggests that combining drawing with retrieval practice will super-duper strengthen memory.

To capture this specific suggestion, let’s propose a general equation:

  • [Research-supported technique A] + [research-supported technique B] = even MORE benefit

That equation just sounds like common sense. Of course, in the world of science, we always have to stop and check our common-sense assumptions. (It just makes sense to believe that the earth is flat. But because we stopped to check that assumption, we now know it to be untrue — knowledge that has proven HIGHLY useful.)

I’ve found a study that explores the specific question raised above: “do students learn more when they combine drawing with retrieval practice?”

Of course, it also invites us to consider that general equation: [technique A] + [technique B] = MORE benefit.

Let’s dig in.

Drawing Biology; Drawing Conclusions

The researchers here wanted to know if students who combined drawing with retrieval practice learned more than those who simply drew. To do so, they had some students [retrieve + draw], while others [drew without retrieval].

Specifically: 8th and 9th grade students in Germany read a 670 word passage for biology class. This five-paragraph passage discussed the human immune response to the flu.

The [retrieve + draw] students followed a multi-step procedure. (If you want to skim these details, that’s ok. I’ve skipped a few steps to keep things simple.)

  1. They read the five-paragraph passage.
  2. After each paragraph, they drew a picture to capture the key ideas that they just read. Note: they could NOT look back at the paragraph while they did so. In other words: they had to retrieve the important ideas as they drew.
  3. Once they had completed the full passage and generated five drawings, they then looked at the passage while they revised and corrected those earlier drawings.

The [draw without retrieval] group did all that, but they could always see the passages while they drew. So: lots of drawing, but no retrieving.

To measure the potential benefits of [retrieving + drawing], the researchers measured three outcomes:

  • How well did students remember the information — as measured by a written quiz?
  • How well did students remember the information — as measured by drawings?
  • Did the students understand other ideas — for instance, HIV infection — better based on learning these related concepts?

The research team asked these questions TWICE: both 2 weeks later and 8 weeks later. This point merits attention, because research only rarely looks for results so long after an experiment. Because we have data gathered two months after the students combined [retrieving + drawing], we can have a plausible basis to start talking about “learning.”

So: how much better did the [retrieve + draw] group do than the [draw only] group?

When a Bust is Not a Bust

We can summarize the findings in three words:

  • “Nope,”
  • “Nope,” and
  • “Nope.”

That is:

  • [Retrieving + drawing] didn’t improve scores on a written quiz after 2 weeks. Or 8 weeks.
  • It didn’t improve scores on drawings.
  • And, it didn’t improve scores on learning related new information.

These results might feel like a bust. The researchers did ALL THAT WORK, and they found nothing. But I think this “what a bust” response misses several key points.

First: the researchers DID find something. They found that combining [retrieving + drawing] DIDN’T produce extra learning. That information gives us useful guidance — because it tells us not to devote our scarce resources to this extra work.

Second: These results help us rethink our general equation: “[research-supported technique A] + [research-supported technique B] = even MORE benefit.”

While that equation initially sounded like common sense, we can see that it’s not necessarily true. (Just as: it’s common sense to think that the earth is flat, but…)

As we enter the world of research-based teaching advice, we should be careful how far we extrapolate research findings.

Intriguing Possibilities

This research suggests that [reading followed immediately by drawing] doesn’t improve learning.

BUT…

I myself don’t think that IMMEDIATE retrieval is, in fact, “retrieval practice.” As it’s typically understood, retrieval practice involves attempts to recall from long-term memory. But in this case, the information the students just read hasn’t gotten anywhere near long-term memory. It’s probably awash in some working-memory-ish zone.

So, drawing right after reading might not help learning. But

  • Maybe [drawing the next day as retrieval practice] might help. Or,
  • Maybe [reading] then [immediate generative drawing] then [retrieving the drawing the next day] might help. Or,
  • Perhaps [an alternating schedule of drawing and retrieving], or…

In other words: I don’t think this study rules out the benefits of [drawing + retrieval practice]. I think it shows that IMMEDIATE drawing doesn’t produce the benefits of retrieval practice — because it isn’t retrieval practice. (By the way: I have checked with my friend Dr. Cindy Nebel — one of the Learning Scientists — to confirm my suspicion that immediate recall isn’t technically retrieval practice.)

In Sum

This study provides a specific “no”: drawing as retrieval practice doesn’t improve learning (if the drawing happens immediately after reading).

It also provides a broader “no”: we shouldn’t assume that two research-informed teaching strategies interact positively.

Finally, it offers an intriguing possibility: we might get the benefits of both drawing and retrieval practice if that retrieval practice happens after some time. We don’t know yet…but it’s a question worth asking.

After all: research-based teaching strategies don’t always enhance each other, but they might!


Kim, S., Abel, R., Leutner, D., Schmiemann, P., & Roelle, J. (2025). Can the effects of generative drawing on lasting learning be optimized through integration of retrieval practice?. Learning and Instruction98, 102143.

Are Science Videos Really the Answer for Struggling Readers?
Andrew Watson
Andrew Watson

Obviously, we want our students to learn science. But: what if they struggle to learn the science by reading?

One solution: we could use VIDEOS to teach science. Videos can transform the written words into an audio track. Voila — no reading required!

A recent study explores this very question. To paraphrase its conclusion: “YUP — students learn more from science videos, so we can/should use them in science classes.”

While the study itself strikes me as well done in a number of ways, I want to push back against this conclusion. Ultimately, this disagreement might reveal the tension between a researcher’s perspective and a teacher’s perspective.

The outline for this post:

  • I’ll start by summarizing the study, and noting its strengths
  • I’ll then try to explain why I’m not persuaded that this study should make science teachers favor videos — even for weaker-than-average readers.

All The Right Steps

Doing education research well requires enormous care, because it can go wrong in so many different ways. Consider all the expectations that we have — and should have! — for research.

  1. We want to have enough participants in the study for the results to be meaningful.
  2. We want researchers to compare plausible alternatives — not just show that “doing something” is better than “doing nothing.”
  3. We’d like reports on meaningful results — not just “did the students have fun” or something like that.
  4. We’d be REALLY happy if the research took place in typical school conditions — not in some hermetically sealed zone that bears little resemblance to actual classrooms.
  5. We expect lots of statsy/mathy results (even if we don’t promise to read them carefully).

And so on. If researchers skip any of these steps, we can complain from the sidelines: “this study doesn’t include an ‘active control group,’ so we shouldn’t rely on its results.” (Honestly, I’m one of the people who object to studies without active control groups.)

Because I’m so aware of these expectations — and lots more — I felt more and more impressed as I made my way through this study. This team has clearly thought through many of the possible objections and found ways to anticipate and mollify them.

  • Enough participants? More than 100.
  • Active controls? Students learned by watching videos, or by reading an illustrated text (with identical words!). In my view, that’s an entirely plausible comparison.
  • Meaningful results? Researchers measured how well the students remembered and transfered their learning…up to a week later!
  • Within a school? Yup. In fact, students had RECESS in the middle of the study, because that’s how they roll in Finland.
  • All the stats? Yes. (More on this point in a minute.)

Thus, I was inclined to be persuaded that, as the abstract says:

The results indicate that videos are beneficial to most children across reading skill levels, especially those with weaker reading skills. This suggests that incorporating videos into primary school science intruction supports diverse learning needs associated with weaker reading skills.

By the way, in this case, “primary school” includes 5th and 6th grade.

A Teacher Has Doubts

Despite this study’s strengths — and I’m being quite serious when I compliment them — I was struck by the actual statistical findings.

The research team focused on three results:

  • how much cognitive load did the students experience while watching videos or reading illustrated texts?
  • how much difference did video vs. illustrated text make for remembering the information?
  • how much difference it make for using information in another context — that is, for transfer?

To answer these questions, they focused on a statistical measure called “R²ₘₐᵣ”. As is always true with stats, it’s tricky to explain what they mean. But here’s a rough-n-ready explanation.

Smiling girl wearing a pink shirt and blue headphones taking notes while looking at a laptop; a library-like setting in the background.

Let’s say that when I study for a quiz using Method A I score 0 points, and when I study using Method B I get 100 points. The R²ₘₐᵣ tells me how much of that difference comes from the two different methods.

So, if R²ₘₐᵣ = .25, that means 25% of the difference between the two scores came from the difference in the two study methods. The other 75 points came from other stuff.

Typically, according to this measure:

  • any R²ₘₐᵣ bigger than 0.25 is “large,”
  • a value between .09 and .25 is “medium,” and
  • a value between .01 and .09 is “small.”

Now that we have a very introductory understanding to this measurement, how meaningful were the results in this study?

  • The “cognitive load” R²ₘₐᵣ was .046: right in the middle of “small.”
  • The R²ₘₐᵣ for remembering information was .016: barely above the bottom of the scale. And
  • R²ₘₐᵣ for transferring information was .003. That’s too small even to register as small.

In brief: did the researchers find STATISTICALLY significant results? It seems they did. Did they find MEANINGFUL differences between videos and illutrated texts? I’m not so sure.

Opportunity Costs

My objection at this point might reflect this difference between a researcher’s perspective and a teacher’s perspective.

The researchers can — entirely reasonably — say: “we ran a scrupulous experiment, and came up with statistically significant results. The data show that videos helped students learn science content better than illustrated texts. Teachers should at least think about using videos to teach science — especially for weak readers.”

As a teacher, my perspective has this additional variable: one of school’s core functions is to teach students to read. And: they get better at reading by — among other strategies — practicing reading.

In other words: according to this study, the benefits of video are so small as to be statistically almost invisible. The benefits of reading practice are — over time — likely to be quite important. I hesitate to give up on one of school’s essential functions (reading) to get such a small benefit (marginal increase in science knowledge) in return.

TL;DR

Someone might say to you — as a friend said to me — “this study shows that we should use videos to teach science content.”

If you hear that claim, be aware that this well executed study found only minute differences between videos and illustrated texts. We should consider this finding alongside a clear understanding of our broader educational mission: teach ALL students to read.


Haavisto, M., Lepola, J., & Jaakkola, T. (2025). The “simple” view of learning from illustrated texts and videos. Learning and Instruction100, 102200.

The Shelf-Life of Research: When Education Theories Evolve…
Andrew Watson
Andrew Watson

To most of us, research conclusions have an air of finality about them. If research shows that … say … prequestions help students learn, well then: case closed. Teachers MUST start class with prequestions!

But wait…we have many reasons to push back against this kind of blunt finality.

  • We could have other strategies that help more: retrieval practice might provide more benefits than prequestions.
  • Perhaps prequestions help these students in this discipline: say — 6th graders learning science concepts. But they might not help 2nd graders practicing phonics.
  • In some cases, initial research findings don’t hold up over time. I used to cite research about “expressive writing,” but I’ve lost confidence in that suggestion because later studies don’t consistently support it.

But there’s another crucial limitation we often overlook: the theories themselves change over time.

Here’s the story.

Once Upon a Time

Scientific researchers rarely arrive at theories in a single great leap. Instead, those theories build up incrementally: study after study, year after year, decade after decade.

In one well-known example: Dr. Carol Dweck started her research in the early 1970s. For most of the subsequent decades, she and her fellow researchers focused on “entity theorists” and “incremental theorists.” Only in the early 2000s did she rebrand to “Mindset.” That is:

  • An “entity theorist” thinks that intelligence is a specific entity, a completed mental capacity. It can’t be changed. These days, we call that a “fixed mindset.”
  • An “incremental theorist” thinks that intelligence can change incrementally. It’s not complete. Today, we talk about a “growth mindset.”

Dweck took, quite literally, more than 30 years to develop the theory to the point that we recognize it today. And since the publication of her book — almost 20 years ago now!! — she has developed it still further.

In other words: mindset theory circa 2005 differs meaningfully from mindset theory circa 2025.

Crucially, this evolution does not mean that Dweck has done anything wrong. Instead, she and her colleagues have been doing what they should be doing. They have a hypothesis; they test it; they adjust the theory based on the results. That updated theory in turn prompts a new hypothesis, so Team Dweck repeats the cycle.

The evolution of these theories isn’t a bug in the system; it’s a feature of the system.

Buggy Features

Although I’ve described this evolution as a feature (not a bug), this feature does allow for some bugs to creep into advice based on theories. A recent example comes quickly to mind.

Cognitive load theory helps many thinkers organize their ideas about working memory in the classroom. (In my view, few topics could be more important than working memory in the classroom.)

For all the normal reasons, cognitive load theory has evolved over time. For instance, it used to distinguish among “intrinsic load,” “extrinsic load,” and “germane load.”

However, in recent years, the scholars who have championed the theory have removed “germane load” from the theoretical framework. It now focuses on “intrinsic” and “extrinsic” only.

The reasons for the change don’t really matter here. What does matter is: cognitive load theory in 2025 doesn’t say exactly what it did in 2015 — and this kind of change is normal.

What to Do?

This insight — “theories do change over time” — offers at least two practical suggestion to teachers and school leaders.

FIRST: Evaluate your own understanding.

That is: if you’re thinking about going all in on a particular theory, be sure you’re planning to enact its most up-to-date version.

In the case of growth mindset, you might start by reading Dweck’s book to understand this foundational text. At the same time, you will no doubt notice that the book was written almost 20 years ago. Next step: take time to see what’s happened to the theory since then.

In the case of mindset, this investigation should be fairly simple. LOTS of people have written books on growth mindset, and research in this field regularly gets attention.

In other cases, double-checking the currency of your information might require some creative sleuthing.

For instance, I asked Claude (the LLM) how Mindset Theory and Cognitive Load Theory have changed over the years; in both cases, Claude’s answer hit the revelant highlights. No doubt other strategies can help you navigate this kind of research history.

Features and Filters

A SECOND practical suggestion also arises from this insight: use it to evaluate the “research-based” advice you hear from someone else.

That is: if a PD speaker says “cognitive load theory tells you to do XYZ to manage your students’ germane load,” you’ve gained an important data point. Although you might find cognitive load theory itself persuasive, you probably shouldn’t take this speaker’s advice about it — because his/her understanding of the theory is more than five years out of date.

In other words: our understanding that theories evolve helps us filter out advice from self-styled experts whose understanding stops with a long-ago version of the theory.

To be fair, I could become overly obsessed with having the very most recent update. Few people can be fully current on all the nuances.

But if a PD speaker says that “a growth mindset is all about effort,” that’s a (mis)interpretation that Dweck has corrected many times. (The thoughtful quality of the effort, and focus on the process, matter more that sheer determination.)

TL;DR

Because educatation is complicated, research-based theories about education are complicated — and they change over time.

To ensure that our students benefit from the work we do, we need to make sure that our own mental models haven’t gone beyond their expiration dates.

Enjoyment or Skill? The Case of Reading [Repost]
Andrew Watson
Andrew Watson

Student motivation has always been a concern; never more so than since the pandemic. What guidance do we get from recent research into reading?


Do we want our students to ENJOY math, or to BE SKILLED AT math?

At first, this question sounds like a false choice. Obviously, we want BOTH.

As an English teacher, I want my students to have fun analyzing the books we read…and I want their analyses to have heft, merit, and substance.

I suspect that most teachers, no matter the subject  — Math, English, Chemistry, Religion, Pickleball — want our students to revel in core ideas and arrive at correct answers.

At some times, alas, we probably need to prioritize one or the other. Especially at the beginning of a unit, should I focus on …

… ensuring that my students like this stuff (even if they don’t immediately understand it), or on

… ensuring they understand the stuff (even if they don’t immediately like it)?

In teaching as in life: if I try to accomplish both goals simultaneously, I’m likely to accomplish neither.

Reading Research

I’m not surprised to discover in a recent study that students’ enjoyment of reading correlates with their skill at reading.

That is: students who get high scores on various reading tests report enjoying reading more than their low-test-scoring peers.

Of course, correlation (say it with me) isn’t causation.

Does the enjoyment lead to the skill? The skill lead to the enjoyment?

Both?

Neither?

To answer these questions, Elsje van Bergen’s research team looked at twins in Finland — more than 3500 of them.

In theory, if we ask all the right questions, gather the right data, and run the right calculations, we should glean insight into the correlation/causation question.

So: what did Team van Bergen find?

But First…

Before you read the answers to that question, you might pause to make a committment. Try to decide NOW if you’re inclined to trust this methodology.

That is:

a) you think well-done twin studies are likely to be a good way to answer this question. For that reason, you will be inclined to accept this answer even if you initially disagree with it.

or

b) you think twin studies can’t answer questions about skill and enjoyment. Thus, you will not cite this study to support your beliefs even if it aligns with those beliefs.

If we’re going to use research to make decisions about education, we should be scrupulous about doing so even when research contradicts the conclusions we had initially held.

Answers, and Questions

Now, back to this post’s main narrative…

Unlike many studies, this one can be summarized in a few pithy sentences.

A young student looks at a book open on her desk and scratches her head in confusion

Based on the twin data they analyzed, van Bergen’s team concludes that:

  • reading skill increases reading enjoyment,
  • reading enjoyment has no effect on reading skill,
  • genetics influences both positively.

Unsurprisingly, the stats get all stats-y. But the above-the-fold headlines are that simple.

Because I don’t teach reading, I’ve always hesitated to be too opinionated on the topic. Now that this study is in the wild, I do think it adds a useful perspective while the reading wars rage on.

For instance: teachers whom I like and respect have told me that older methods might not have science behind them, but they’re excellent at “making students feel like readers.”

This claim has always puzzled me. How can a student feel like a reader if s/he can’t read?

Van Bergen’s study, I think, gives me permission to address that point directly: “this study suggests that skill at reading will be the more important place to start in reading instruction.”

Zooming the Camera Back

While this study and this post have focused on reading instruction, I do think there’s a broader message here as well.

We frequently hear about the importance of intrinsic motivation; that is, a motivation that springs from students’ natural interests, not from external encouragement (or pressure).

This study, to the contrary, finds that the work teachers do to improve students’ skill simultaneously enhances their motivation. That motivation might be — in effect — extrinsic; but, it’s working. (Working = students read better, and want to read more.)

Overall, I believe we need a substantial rethink of the (false) intrinsic/extrinsic dichotomy, and the (unhelpful) criticism of motivational strategies that many teachers currently find themselves using.

If you want to join me for just such a rethink, I’m giving a webinar for Learning and the Brain on April 5th. We’ll be talking about several research-informed approaches to intrinsic motivation, and brainstorming strategies to make those ideas fit in our classrooms.

I hope I’ll persuade you that we have better ways to talk about motivation than “intrinsic/extrinsic,” and those better ways give us useful teacherly guidance.

I hope you’ll join us!


van Bergen, E., Hart, S. A., Latvala, A., Vuoksimaa, E., Tolvanen, A., & Torppa, M. (2023). Literacy skills seem to fuel literacy enjoyment, rather than vice versa. Developmental Science26(3), e13325.

Book Review: 10 to 25, by David Yeager
Andrew Watson
Andrew Watson

As long as humans have lived into their teens, adults have complained about their behavior. Aristotle famously described this age cohort as “fickle, passionate, irascible, and apt to be carried away by their impulses.” If your experience aligns with mine, you’ve seen your share of grouchy selfishness and demotivation in your classroom — especially before 10 am.

Although complaints about adolescents have been around as long as adolescence, those gripes have grown sharper in recent months. PANDEMIC this and CELL PHONE that and AI CHEATING to boot – all these additional concerns have ramped up adult exhaustion with teenage too-muchness.

Given this bleak outlook – both historic and contemporary – what’s a middle-school or high-school teacher to do?

Happily, noted researcher Dr. David Yeager has wise thoughts – and LOTS of research – to give us hope. His recent book 10 to 25: The Science of Motivating Young People will guide and inspire teachers and school leaders.

START HERE

Before getting to specific suggestions, Yeager wants us to rethink our starting point. Whereas Plato and others start with a “deficit mindset” – seeing only bad behavior and gloomy prognosis – Yeager wants us to see the strengths and potentials in humans between the ages of 10 and 25.

So, for instance: you’ll often hear that “the human pre-frontal cortex isn’t fully wired until people reach their mid to late 20s!” The implication of this claim: without this “self-control center” fully developed, young ‘uns are doomed to erratic and immature behavior.

Yeager, however, has a different perspective. This additional time for important brain regions to develop gives growing youth the chance to adapt to the ever-changing world. As the computer people say: the drawn-out schedule of brain development “isn’t a bug, it’s a feature!”

In Yeager’s analysis, most adults respond to this time period with some blend of expectations and support. But most of us get that blend wrong.

Specifically:

  • Some of us default to HIGH expectations – but don’t offer much by way of support. It’s my way or the highway – and teens often find themselves out on that road. Yeager calls this blend the “enforcer” mindset.
  • Others offer HIGH support – and don’t worry too much about expectations. Not wanting to make a difficult time even harder, we soften standards when they seem too onerous. Yeager thinks of these folks as having a “protector” mindset.

Perhaps you can anticipate Yeager’s next move.

  • Adults with a “mentor” mindset balance HIGH expectations with HIGH support. This combination, in Yeager’s view, offers the best path to help adolescents navigate the 10-to-25 stage with optimal growth and development.

One of Yeager’s many strengths in this book: he doesn’t claim to be the only person to advocate for this “mentor mindset” balance. In fact, he carefully and respectfully charts the terminology that other scholars – Lewin, Baumrind, Scott – have used to describe these perspectives. I myself am likelier to take advice from scholars who conspicuously acknowledge their debts to others in the field; Yeager does just that.

WHAT TO DO

To help us translate this mentor’s mindset into specific action, Yeager outlines five broad approaches to help teens through these years: transparency, questioning, stress, purpose, and belonging. He devotes a chapter to each, offering both uplifting stories and scrupulous research to specify and support his case.

For instance, looks at the problem of stress – familiar to anyone dealing with emerging adults. We might, by way of shorthand, define stress as a situation where the DEMANDS of the moment exceed a student’s RESOURCES.

How might we respond?

  • Those with an enforcer mindset say: “Deal with it. The demands are the demands, so get the job done – with whatever resources you have.”
  • Those with a protector mindset say: “Well, that looks stressful. Let’s reduce the demands until the align with the resources you’ve got.”
  • A mentor’s mindset would encourage a different approach: “Let me help you increase your resources so that they match the demands that you face.” The mentor doesn’t reduce the requirements of the moment, but helps develop the skills or knowledge necessary to face it.

Yeager then explores a body of research (by Jameison and others) showing how to boost resources.

Specifically, students who understand the evolutionary history of stress responses recognize that all those seemingly unpleasant symptoms – the sweaty palms, the stomach butterflies – signal helpful physiological responses.

Once students have this additional resource – the knowledge to reframe their physiological responses to stress – they can face the demands of the current situation.

Voila: a mentor’s mindset, combined with research, helps us coach and support a student through an age-appropriate challenge.

HESITATIONS?

I always think a book review should include a caveat or two, but 10 to 25 makes this belief something of a challenge. Yeager writes conversationally, keeps the jargon to a minimum, and modestly acknowledges the difficulties of applying research-based ideas to real-life situations. He even includes a chapter of guided practice to ensure the book’s suggestions take root.

My only consistent concern: as noted above, the book draws on both research and “uplifting” stories. More than most readers, perhaps, I find myself reacting to uplift with suspicion. An uplifting story is – from another perspective – simply an anecdote. I don’t think we should rely on anecdotes to make teaching decisions. And I especially resist stories about life far outside of school – say, in Microsoft training programs.

(I should say: I’m probably an outlier here. Many people find anecdotes a helpful way to translate research-y principles into real-life situations.)

This modest critique aside, Yeager’s book both explains and organizes lots of research. In this way, it helps teachers think differently about real challenges in our work, and gives us clear guidance about what to do.

If you wish you had a clearer set of principles to help motivate and support an emerging adult, Yeager’s book is for you.

Returning to the Classroom in the Age of AI: Why I’m Cautiously Optimistic
Guest Post
Guest Post

This week’s guest post in by Rod Naquin.


This fall, I’m heading back to the high school English classroom after some time away, and I couldn’t be more excited about the possibilities ahead. The landscape has changed dramatically since I last taught full-time, particularly with the emergence of large language models like ChatGPT becoming commonplace in education.

The Research That’s Got Everyone Talking

Recent headlines about AI in education have been decidedly mixed. A new MIT Media Lab study titled “Your Brain on ChatGPT” has garnered attention for suggesting that AI use leads to “weaker neural connectivity” and reduced cognitive engagement during writing tasks. The research, led by Nataliya Kosmyna, used EEG to monitor brain activity as participants wrote essays with different tools: ChatGPT, search engines, or no assistance at all.

The findings are striking. Students using ChatGPT showed the weakest overall neural coupling, while those writing without any assistance exhibited the strongest, most wide-ranging brain networks. The LLM group also reported lower ownership of their essays and struggled to quote from work they had completed just minutes earlier.

But here’s where I think we need to pump the brakes on the panic. The MIT researchers themselves emphasize crucial limitations that often get lost in sensational headlines. Key study limitations include:

  • Small sample size of 54 participants from a specific geographic area
  • Single context focus on essay writing only
  • Not yet peer-reviewed research
  • Testing limited to ChatGPT rather than other AI tools
  • No breakdown of writing process into specific subtasks

The authors explicitly warn against using alarmist language like “brain damage” or claims that AI is “making us dumber.” They’re calling for more research, not educational moratoriums. This kind of preliminary work should inform our thinking, not dictate our policies. Dr. Cindy Nebel and Andrew Watson wrote a more detailed analysis of the MIT study’s methodology and implications that you can read here.

What Students Are Actually Saying

What I find more illuminating are the student voices that often get overlooked in these discussions. Jason Gulya recently compiled reflections from three high schoolers about their experiences with AI, and their insights are far more nuanced than much of the adult commentary.

William Liang offers a particularly sharp observation: many students see assignments as “logistical hurdles” rather than learning opportunities. This perspective helps explain why students might turn to AI tools – not necessarily out of laziness, but because our current educational system often prioritizes compliance over genuine engagement.

Olivia Han’s journey is even more revealing. She initially embraced AI for efficiency, telling herself she was “working smarter.” But over time, she noticed something troubling: AI’s voice was gradually replacing her own. What’s remarkable is her conscious decision to step back and reclaim her authentic expression. She chose the harder path to maintain ownership of her thinking and writing.

Mary Ruskell raises perhaps the most profound question of all: how do we build meaningful relationships and understanding when we can’t always distinguish between authentic and artificial content? Her concerns go beyond academic performance to fundamental questions about truth and trust in an AI-mediated world.

My Plan for Thoughtful Integration

These student perspectives highlight why I’m approaching AI integration thoughtfully rather than wholesale adoption or rejection. When I return to teaching, I plan to use AI for what it genuinely does well. AI tools can support:

  • Text analysis and pattern recognition for quick assessment
  • Example generation across multiple difficulty levels
  • Assignment differentiation for diverse learners
  • Feedback compilation and organization
  • Time-saving on routine administrative tasks

For instance, AI can help me rapidly identify patterns in student writing, suggest targeted feedback, or create assignments for diverse learners. These applications leverage AI’s strengths in pattern recognition and text manipulation while keeping human judgment and relationships at the center of learning.

What I won’t do is use AI as a substitute for the core work of education: helping students develop their own voices, critical thinking skills, and authentic expression. The MIT study’s findings about reduced ownership and cognitive engagement serve as important warnings about what happens when AI assistance crosses the line from support to replacement.

Beyond the Binary Thinking

This distinction matters enormously. Educational researcher Daniel Willingham has long argued that good teaching focuses on student thinking processes rather than just the artifacts students produce. AI tools, used wisely, can actually support this shift by handling routine tasks and freeing up time for deeper engagement with student ideas.

The rhetorical science around AI in education often misses these nuances.

  • Preliminary research gets transformed into definitive conclusions
  • Complex findings get reduced to binary choices: AI good or AI bad
  • Student voices get drowned out by adult anxieties and commercial interests

We need better frameworks for thinking about AI integration – ones that acknowledge both potential benefits and legitimate concerns while keeping student development at the center. This means being intentional about when and how we use these tools, monitoring their effects on learning, and remaining flexible as we learn more.

Essential questions for AI integration include:

  • How can we measure real impact on learning rather than just efficiency?
  • How do tools support specific learning objectives rather than replace them?
  • When does AI enhance versus replace authentic student thinking?
  • What metacognitive skills do students need to use AI effectively?
  • How do we maintain and strengthen student voice in an AI-mediated environment?
  • What boundaries help preserve the human elements of education?

Looking Forward

I’m optimistic about returning to teaching in this moment because I believe educators are uniquely positioned to shape how AI gets integrated into learning environments. We can model thoughtful usage, help students develop metacognitive awareness about these tools, and ensure that technology serves pedagogical goals rather than driving them.

The conversation about AI in education is just beginning, and it’s crucial that classroom teachers have a voice in it. We’re the ones who see daily how students learn, struggle, and grow. Our insights about what works and what doesn’t will be essential as we navigate this new landscape.

The goal isn’t to eliminate AI from education or to embrace it uncritically. It’s to use these powerful tools in service of timeless educational goals: helping students think clearly, express themselves authentically, and engage meaningfully with ideas and each other.

That’s work worth returning to the classroom for.


References

Gulya, J. (2025, July 1). 3 student reflections on gen-AI and what we can learn from them. The AI Edventure. https://higherai.substack.com/p/3-student-reflections-on-gen-ai-and?triedRedirect=true

Kosmyna, N. (2025). Your brain on ChatGPT [Research project]. MIT Media Lab. https://www.media.mit.edu/projects/your-brain-on-chatgpt/overview/Willingham, D. (2025). How ChatGPT might inadvertently improve student thinking. Substack. https://substack.com/inbox/post/167822624?triedRedirect=true

Willingham, D. (2025). How ChatGPT might inadvertently improve student thinking. Substack. https://substack.com/inbox/post/167822624?triedRedirect=true


Rod J. Naquin is an educational leadership and professional development specialist dedicated to practicing and teaching the science of dialogue. Through 15 years of transforming educational practices, he explores how meaningful conversations and emerging technologies like AI can reshape teaching, learning, and leadership. With degrees from UNO and Southeastern Louisiana University (where he’s now working on his doctorate), Rod runs “The Science of Dialogue,” examining how thoughtful discussion and tech tools can revolutionize educational leadership. Through his blog, podcast, and consulting work, he helps educators master the intersection of strategic dialogue and innovative technology – catch him at @rodjnaquin across socials, on Substack, and tune into “The Science of Dialogue” on your favorite podcast platform.

The Best Way to Teach: When Clarity Leads to Muddle [Repost]
Andrew Watson
Andrew Watson

Our blogger read this claim: “research shows that active learning is better than direct instruction.” Back in November, he responded to that research, and that claim.


Most teachers want to be better teachers. You’re probably reading this blog for research-based guidance on doing so.

A young student wearing plastic goggles carefully pours something into a beaker slightly filled with green liquid

I recently read a study that offers emphatic — and paradoxical — guidance. Exploring this research — as well as its paradoxes — might be helpful as we think about being better teachers.

Here’s the story.

A research team, led by Louis Deslauriers, worked with students in an introductory physics class at Harvard. This class was taught by an experienced professor who mostly lectured; he also supplemented the class with “demonstrations, … occasional interactive quizzes or conceptual questions.”

Let’s call this approach “interactive lecture.”

In Deslauriers’s study, students also attended two additional classes. One was taught with Method A and the other with Method B.

In Method A, an experienced professor:

  • presented slides
  • gave explanations
  • solved sample problems
  • strove for fluency of presentation

What abotu Method B? Another experienced teacher:

  • used principles of deliberate practice
  • instructed students to solve sample problems together in small groups
  • circulated through the room to answer questions
  • ultimately provided a full and correct answer to the problems

The researchers strove, as much as possible, to make the class content identical; only the pedagogy differed.

What did the researchers learn about the relative benefits of Methods A and B?

Paradox #1: Effective and Unloved

First off, the students learned more from Method B.

That is: when they solved problems in small groups, wrestled with the content, and ultimately heard the right answer, students scored relatively higher on an end-of-class multiple choice test. When they experienced Method A (the prof explained all the info and solved all the problems), they scored relatively lower.

But — paradoxically — the students preferred Method A, and believed that they learned more from it. They even suggested that all their classes be taught according to Method A — the method that resulted in less learning.

The researchers offer several explanations for this paradox. The headlines sound like this:

  • When students hear straightforward explanations and see clear/succesful demonstrations of solutions strategies (Method A), the content seems easy and clear. Students think they understand.
  • But, when they have to do the cognitive heavy lifting (Method B), class feels more difficult. Result: students worry they didn’t understand.
  • Because the students are — relatively speaking — novices, they don’t know enough to know when they understand.

Team Deslauriers, sensibly enough, suggests that we can help students appreciate and accept the more challenging methods — like Method B — if we explain the reseasoning behind them.

I, by the way, take this suggestion myself. For instance: I explain the benefits of retrieval practice to my students. They don’t always love RP exercises, because retrieval practice feels harder than simple review. But they understand the logic behind my approach.

Paradox #2: Clarity vs. Muddle

Up to this point, Deslauriers and Co. pursue a sensible path.

They know that MOST college profs use Method A (the bad one), so they want those profs to change. To encourage that change, the undertake a study showing a better option: Method B!

Given these research results, Deslauriers and Co. offer two clear and emphatic suggestions:

First: teachers should use Method B teaching strategies, not Method A strategies.

Second: to counteract students’ skepticism about Method B, we should explain the logic behind it.

What could be more helpful?

Alas, these clear suggestions can lead to another muddle. This muddle results from the freighted NAMES that this study gives to Methods A and B.

Method B — the good one — is called “active.”

Method A — the bad one — is called (inevitably) “passive.”

So, this study summarizes its findings by saying that “active” teaching is better than “passive” teaching.

These labels create real problems with the research conclusions.

Because these labels lack precision, I can apply them quite loosely to any teaching approach that I believe to be good or bad.

For instance: recall the experienced professor who regularly teaches this physics course. He mostly lectures; he also supplements the class with “demonstrations, … occasional interactive quizzes or conceptual questions.”

If I disapprove of that combination, I can call it “passive” — he mostly lectures!

If I approve, I can call it “active” — consider all those demonstractions, interactions, and conceptual questions!!

These labels, in other words are both loaded and vague — a perilous combination.

The peril arises here: literally no one in the world of cognitive science champions Method A.

EVERYONE who draws on cognitive science research — from the most ardent “constructivist” to the most passionate advocate for direct instruction — believes that students should actively participate in learning by problem solving, discussion, creation, and so forth.

Advocates for those two groups have different names for this mental activity: “desirable difficulties,” “productive struggle.” They think quite differently about the best way to achieve all that active thinking. But they all agree that students must struggle with some degree of difficulty.

Slippery Logic

This naming muddle creates unfortunate logical slips.

The study certainly suggests that Method B benefits students more than Method A. But, it doesn’t suggest that Method B is better than other methods that might reasonably be called by the open-ended named “active.”

For instance: it doesn’t necessarily mean that “constructivism” is better than direct instruction. And yet — because of those highly flexible labels — the study can be misinterpreted to support that claim.

My concern isn’t hypothetical. Someone sent me this study precisely to support the argument that inquiry learning promotes more learning than direct instruction.

But: “Method B” isn’t inquiry learning. And direct instruction isn’t Method A.

The Big Picture

I said at the beginning of this post that teachers might draw on research to be better teachers.

I worry that readers will draw this inaccurate conclusion based on this study:

“Research proves that ‘active learning’ (like projects and inquiry) is better than ‘passive learning’ (like direct instruction).”

Instead, this study suggests that asking students to do additional, productive mental work results in more learning than reducing their mental work.

Champions of both projects/inquiry and direct instruction want students to do additional, productive mental work.

Those schools of though have sharply different ideas of the best ways to accomplish that goal. But dismissing one of them as “passive” — and therefore obviously bad — obscures the important insights of that approach.


Deslauriers, L., McCarty, L. S., Miller, K., Callaghan, K., & Kestin, G. (2019). Measuring actual learning versus feeling of learning in response to being actively engaged in the classroom. Proceedings of the National Academy of Sciences116(39), 19251-19257.