Login / Signup

Inferring neural activity before plasticity as a foundation for learning beyond backpropagation.

Yuhang SongBeren MillidgeTommaso SalvatoriThomas LukasiewiczZhenghua XuRafal Bogacz
Published in: Nature neuroscience (2024)
For both humans and machines, the essence of learning is to pinpoint which components in its information processing pipeline are responsible for an error in its output, a challenge that is known as 'credit assignment'. It has long been assumed that credit assignment is best solved by backpropagation, which is also the foundation of modern machine learning. Here, we set out a fundamentally different principle on credit assignment called 'prospective configuration'. In prospective configuration, the network first infers the pattern of neural activity that should result from learning, and then the synaptic weights are modified to consolidate the change in neural activity. We demonstrate that this distinct mechanism, in contrast to backpropagation, (1) underlies learning in a well-established family of models of cortical circuits, (2) enables learning that is more efficient and effective in many contexts faced by biological organisms and (3) reproduces surprising patterns of neural activity and behavior observed in diverse human and rat learning experiments.
Keyphrases
  • machine learning
  • magnetic resonance
  • healthcare
  • magnetic resonance imaging
  • artificial intelligence