How EXACTLY will your competent? doctor use this to get you fully recovered? Completely upend their rehab protocols! Repetition has always been the mantra for stroke rehab.
Less Experience Leads to Faster Neural Adaptation
Summary: For over a century, the cornerstone of psychology has been the Pavlovian idea that we learn through repetition—the more a bell rings before food, the stronger the association. However, a groundbreaking study is upending this 100-year-old assumption.
Researchers discovered that the brain actually learns more efficiently when rewards are rare and spaced far apart. Rather than “practice makes perfect,” the brain’s dopamine system prioritizes the timing between events. This discovery suggests that our neural circuitry is designed to extract maximum information from infrequent experiences, providing a new biological explanation for why “cramming” for exams fails while spaced-out learning succeeds.
Key Facts
- The Timing Rule: The brain determines how much to learn based on the time between cue-reward pairings, rather than the total number of repetitions.
- Dopamine Acceleration: When rewards are spaced further apart, the brain requires significantly fewer repetitions before it begins releasing dopamine in anticipation of the reward.
- Sparse Learning Efficiency: Mice that received rewards only 10% of the time learned at the same rate—or faster—than those who received rewards 20 times more frequently.
- The “Cramming” Effect: When experiences happen too close together, the brain “downregulates” its learning, explaining why frequent, repetitive exposure can lead to diminishing returns in memory.
- AI Implications: This discovery could lead to faster artificial intelligence. Current AI requires billions of data points to learn, but a model based on this “sparse learning” theory could learn more quickly from fewer experiences.
Source: UCSF
More than a century ago, Pavlov trained his dog to associate the sound of a bell with food. Ever since, scientists assumed the dog learned this through repetition: The more times the dog heard the bell and then got fed, the better it learned that the sound meant food would soon follow.
Now, scientists at UC San Francisco are upending this 100-year-old assumption about associative learning. The new theory asserts that it depends less on how many times something happens and more on how much time passes between rewards.

“It turns out that the time between these cue-reward pairings helps the brain determine how much to learn from that experience,” said Vijay Mohan K. Namboobidiri, PhD, an associate professor of Neurology and senior author of the study, published Feb. 12 in Nature Neuroscience.
When the experiences happen closer together, the brain learns less from each instance, Namboodiri said, adding that this could explain why students who cram for exams don’t do as well as those who studied throughout the semester.
Learning the cues
Scientists have traditionally thought of associative learning as a process of trial and error. Once the brain has detected that certain cues might lead to rewards, it begins to predict them. Scientists have postulated that at first the brain only releases dopamine when a reward like tasty food arrives.
But if the reward arrives often enough, the brain begins to anticipate it with a release of dopamine as soon as it gets the cue. The dopamine hit refines the brain’s prediction, the theory goes, strengthening the link with the cue if the reward arrives — or weakening it if the reward fails to appear.
Namboodiri and postdoctoral scholar Dennis Burke, PhD, trained mice to associate a brief sound with getting sugar-sweetened water, varying the time between trials. They spaced the trials 30 to 60 seconds apart for some of the mice, and five to 10 minutes apart, or more, for others. The result was that the mice whose trials were closer together received many more rewards than those who trials were spaced farther apart in the same amount of time.
If associative learning depended only on repetition, the mice with more trials should have learned faster. Instead, the mice that got very few rewards learned the same amount as those that got 20 times more trials over the same amount of time.
“What this tells us is that associative learning is less ‘practice makes perfect’ and more ‘timing is everything,’” said Burke, the first author of the study.
Namboodiri and Burke then looked at what dopamine was doing in the mouse brain.
When the rewards were spaced further apart, the mice needed fewer repetitions before their brains began to respond to the sound with dopamine.
Then, the researchers tried a different variation. They repeatedly played the sound — spacing the cues 60 seconds apart — but only gave the mice sugar water 10% of the time. These mice needed far fewer rewards before they began releasing dopamine after the cue, regardless of whether it was followed by a reward.
More rapid learning
The findings could shift the way we look at learning and addiction. Smoking, for example, is intermittent and can involve cues — like the sight or smell of cigarettes — that increase the urge to smoke. Because a nicotine patch delivers nicotine constantly, it may disrupt the brain’s association between nicotine and the resulting dopamine reward, blunting the urge to smoke and making it easier to quit.
Next, Namboodiri plans to investigate how his new theory could speed up artificial intelligence. Current AI systems learn quite slowly, because they are based on the prevailing model of associative learning, making small refinements after every interaction between billions of data points.
“A model that borrows from what we’ve discovered could potentially learn more quickly from fewer experiences,” Namboodiri said. “For the moment, though, our brains can learn a lot faster than our machines and this study helps explain why.”
Authors: Additional authors on the study include Annie Taylor, Huijeong Jeong, SeulAh Lee, Leo Zsembik, Brenda Wu, Joseph Floeder, Gautam Naik, and Ritchie Chan, all of UCSF.
Funding: This work was supported by the National Institutes of Health (grants R00MH118422, R01MH129582, F32DA060044). the National Science Foundation, the Klingenstein-Simons Fellowship, the David and Lucile Packard Foundation, and Shurl and Kay Curci Foundation.
Key Questions Answered:
A: Not necessarily, but it means “spacing” is more important than “grinding.” If you’re trying to learn a new language or instrument, your brain will actually absorb more from three 20-minute sessions spread throughout the day than one solid hour of repetition.
A: From an evolutionary standpoint, rare rewards (like finding a hidden fruit tree) are more “informative” than common ones. If something happens all the time, the brain treats it as background noise. If it’s rare, the brain pays extra attention to the timing to make sure it doesn’t miss the next opportunity.
A: It explains why intermittent rewards (like gambling or social media notifications) are so addictive. Because the rewards are unpredictable and spaced out, the brain’s dopamine system remains highly sensitive and “learns” the habit much more deeply than if the reward was constant.
Editorial Notes:
- This article was edited by a Neuroscience News editor.
- Journal paper reviewed in full.
- Additional context added by our staff.
About this learning and neuroscience research
Author: Laura Kurtzman
Source: UCSF
Contact: Laura Kurtzman – UCSF
Image: The image is credited to Neuroscience News
Original Research: Open access.
“Duration between rewards controls the rate of behavioral and dopaminergic learning” by Dennis A. Burke, Annie Taylor, Huijeong Jeong, SeulAh Lee, Leo Zsembik, Brenda Wu, Joseph R. Floeder, Gautam A. Naik, Ritchie Chen & Vijay Mohan K Namboodiri. Nature Neuroscience
DOI:10.1038/s41593-026-02206-2


