Free-Energy Reinforcement Learning
This is a framework proposed by Sallans and Hinton in 2004 (Sallans and Hinton, n.d.). The key insight is that a product of experts allows for model parameters to be learnt efficiently, because values and derivatives for the product of experts can be efficiently computed.
The weights of the RBM are tweaked such that the free energy of a network configuration equals to the reward signal for the given state-action pair.
An action is selected by performing Gibbs sampling, holding the state variables fixed. The action with the lowest free energy is produced, corresponding to the highest expected reward for the given state.
Spiking neural networks can be used to implement RBMs, hence used for FERL (Nakano and Otsuka, n.d.).
Nakano, Takashi, and Makoto Otsuka. n.d. “Spiking Neural Network Model of Free-Energy-Based Reinforcement Learning” 12 (S1):P244. https://doi.org/10.1186/1471-2202-12-s1-p244.