<div class="csl-bib-body">
<div class="csl-entry">Millidge, B., Song, Y., Salvatori, T., Lukasiewicz, T., & Bogacz, R. (2023). A Theoretical Framework for Inference and Learning in Predictive Coding Networks. In <i>The Eleventh International Conference on Learning Representations, ICLR 2023</i> (pp. 1–24). http://hdl.handle.net/20.500.12708/192478</div>
</div>
-
dc.identifier.uri
http://hdl.handle.net/20.500.12708/192478
-
dc.description.abstract
Predictive coding (PC) is an influential theory in computational neuroscience, which argues that the cortex forms unsupervised world models by implementing a hierarchical process of prediction error minimization. PC networks (PCNs) are trained in two phases. First, neural activities are updated to optimize the network’s response to external stimuli. Second, synaptic weights are updated to consolidate this change in activity — an algorithm called prospective configuration. While previous work has shown how in various limits, PCNs can be found to approximate backpropagation (BP), recent work has demonstrated that PCNs operating in this standard regime, which does not approximate BP, nevertheless obtain competitive training and generalization performance to BP-trained networks while outperforming them on various tasks. However, little is understood theoretically about the properties and dynamics of PCNs in this regime. In this paper, we provide a comprehensive theoretical analysis of the properties of PCNs trained with prospective configuration. We first derive analytical results concerning the inference equilibrium for PCNs and a previously unknown close connection relationship to target propagation (TP). Secondly, we provide a theoretical analysis of learning in PCNs as a variant of generalized expectation-maximization and use that to prove the convergence of PCNs to critical points of the BP loss function, thus showing that deep PCNs can, in theory, achieve the same generalization performance as BP, while maintaining their unique advantages.
en
dc.language.iso
en
-
dc.subject
predictive coding
en
dc.subject
backpropagation
en
dc.subject
generalized expectation-maximization
en
dc.subject
target propagation
en
dc.title
A Theoretical Framework for Inference and Learning in Predictive Coding Networks
en
dc.type
Inproceedings
en
dc.type
Konferenzbeitrag
de
dc.contributor.affiliation
University of Oxford, United Kingdom of Great Britain and Northern Ireland (the)
-
dc.contributor.affiliation
University of Oxford, United Kingdom of Great Britain and Northern Ireland (the)
-
dc.contributor.affiliation
University of Oxford, United Kingdom of Great Britain and Northern Ireland (the)
-
dc.description.startpage
1
-
dc.description.endpage
24
-
dc.type.category
Full-Paper Contribution
-
tuw.booktitle
The Eleventh International Conference on Learning Representations, ICLR 2023