Optimal structure of metaplasticity for adaptive learning

Autor: Alireza Soltani, Peyman Khorsand
Jazyk: angličtina
Rok vydání: 2017
Předmět:
0301 basic medicine
Statistical methods
Physiology
Computer science
Distributed computing
Social Sciences
computer.software_genre
Nervous System
Synaptic Transmission
Synapse
Learning and Memory
0302 clinical medicine
Medicine and Health Sciences
Psychology
Probability Estimation
skin and connective tissue diseases
lcsh:QH301-705.5
media_common
Neuronal Plasticity
Ecology
Noise (signal processing)
Simulation and Modeling
Synaptic efficacy
Brain
Adaptation
Physiological

Electrophysiology
Monte Carlo method
Physical sciences
Computational Theory and Mathematics
Modeling and Simulation
Engineering and Technology
Adaptive learning
Anatomy
Research Article
Optimization
media_common.quotation_subject
Models
Neurological

Neurophysiology
Statistics (mathematics)
Machine learning
Adaptability
Bottleneck
03 medical and health sciences
Cellular and Molecular Neuroscience
Developmental Neuroscience
Reward
Control theory
Metaplasticity
Genetics
Learning
Humans
Computer Simulation
Adaptation (computer science)
Molecular Biology
Ecology
Evolution
Behavior and Systematics

Structure (mathematical logic)
Models
Statistical

business.industry
Mechanism (biology)
fungi
Cognitive Psychology
Biology and Life Sciences
Noise Reduction
Research and analysis methods
Range (mathematics)
030104 developmental biology
lcsh:Biology (General)
Cellular Neuroscience
Synapses
Signal Processing
Synaptic plasticity
Mathematical and statistical techniques
Cognitive Science
Artificial intelligence
sense organs
Nerve Net
business
computer
Mathematics
030217 neurology & neurosurgery
Neuroscience
Synaptic Plasticity
Zdroj: PLoS Computational Biology, Vol 13, Iss 6, p e1005630 (2017)
PLoS Computational Biology
ISSN: 1553-7358
Popis: Learning from reward feedback in a changing environment requires a high degree of adaptability, yet the precise estimation of reward information demands slow updates. In the framework of estimating reward probability, here we investigated how this tradeoff between adaptability and precision can be mitigated via metaplasticity, i.e. synaptic changes that do not always alter synaptic efficacy. Using the mean-field and Monte Carlo simulations we identified ‘superior’ metaplastic models that can substantially overcome the adaptability-precision tradeoff. These models can achieve both adaptability and precision by forming two separate sets of meta-states: reservoirs and buffers. Synapses in reservoir meta-states do not change their efficacy upon reward feedback, whereas those in buffer meta-states can change their efficacy. Rapid changes in efficacy are limited to synapses occupying buffers, creating a bottleneck that reduces noise without significantly decreasing adaptability. In contrast, more-populated reservoirs can generate a strong signal without manifesting any observable plasticity. By comparing the behavior of our model and a few competing models during a dynamic probability estimation task, we found that superior metaplastic models perform close to optimally for a wider range of model parameters. Finally, we found that metaplastic models are robust to changes in model parameters and that metaplastic transitions are crucial for adaptive learning since replacing them with graded plastic transitions (transitions that change synaptic efficacy) reduces the ability to overcome the adaptability-precision tradeoff. Overall, our results suggest that ubiquitous unreliability of synaptic changes evinces metaplasticity that can provide a robust mechanism for mitigating the tradeoff between adaptability and precision and thus adaptive learning.
Author summary Successful learning from our experience and feedback from the environment requires that the reward value assigned to a given option or action to be updated by a precise amount after each feedback. In the standard model for reward-based learning known as reinforcement learning, the learning rates determine the strength of such update. A large learning rate allows fast update of values (large adaptability) but introduces noise (small precision), whereas a small learning rate does the opposite. Thus, learning seems to be bounded by a tradeoff between adaptability and precision. Here, we asked whether there are synaptic mechanisms that are capable of adjusting the brain’s level of plasticity according to reward statistics, and, therefore, allow the learning process to be adaptive. We showed that metaplasticity, changes in the synaptic state that shape future synaptic modifications without any observable changes in the strength of synapses, could provide such a mechanism and furthermore, identified the optimal structure of such metaplasticity. We propose that metaplasticity, which sometimes causes no observable changes in behavior and thus could be perceived as a lack of learning, can provide a robust mechanism for adaptive learning.
Databáze: OpenAIRE