Download:
Abstract:
Generalization from past experience is an important feature of intelligent systems. When faced with a new task, one efficient computational approach is to evaluate solutions to earlier tasks as candidates for reuse. Consistent with this idea, we found that human participants (n = 38) learned optimal solutions to a set of training tasks and generalized them to novel test tasks in a reward-selective manner. This behavior was consistent with a computational process based on the successor representation known as successor features and generalized policy improvement (SF&GPI). Neither model-free perseveration or model-based control using a complete model of the environment could explain choice behavior. Decoding from functional magnetic resonance imaging data revealed that solutions from the SF&GPI algorithm were activated on test tasks in visual and prefrontal cortex. This activation had a functional connection to behavior in that stronger activation of SF&GPI solutions in visual areas was associated with increased behavioral reuse. These findings point to a possible neural implementation of an adaptive algorithm for generalization across tasks.
Neural activation of previously reinforced policy on novel task
Citation
Hall-McMaster, S., Tomov, M. S., Gershman, S. J., Schuck, N. W. (2025). “Neural evidence that humans reuse strategies to solve new tasks.” PLOS Biology, 23(6), e3003174. https://doi.org/10.1371/journal.pbio.3003174.
@article{hall2025neural,
doi = {10.1371/journal.pbio.3003174},
author = {Hall-McMaster, Sam AND Tomov, Momchil S. AND Gershman, Samuel J. AND Schuck, Nicolas W.},
journal = {PLOS Biology},
publisher = {Public Library of Science},
title = {Neural evidence that humans reuse strategies to solve new tasks},
year = {2025},
month = {06},
volume = {23},
url = {https://doi.org/10.1371/journal.pbio.3003174},
pages = {1-31},
abstract = {Generalization from past experience is an important feature of intelligent systems. When faced with a new task, one efficient computational approach is to evaluate solutions to earlier tasks as candidates for reuse. Consistent with this idea, we found that human participants (n = 38) learned optimal solutions to a set of training tasks and generalized them to novel test tasks in a reward-selective manner. This behavior was consistent with a computational process based on the successor representation known as successor features and generalized policy improvement (SF&GPI). Neither model-free perseveration or model-based control using a complete model of the environment could explain choice behavior. Decoding from functional magnetic resonance imaging data revealed that solutions from the SF&GPI algorithm were activated on test tasks in visual and prefrontal cortex. This activation had a functional connection to behavior in that stronger activation of SF&GPI solutions in visual areas was associated with increased behavioral reuse. These findings point to a possible neural implementation of an adaptive algorithm for generalization across tasks.},
number = {6},
}