Associative Learning via Inhibitory Search
暂无分享,去创建一个
ALVIS is a reinforcement-based connectionist architecture that learns associative maps in continuous multidimensional environments. The discovered locations of positive and negative reinforcements are recorded in "do be" and "don't be" subnetworks, respectively. The outputs of the subnetworks relevant to the current goal are combined and compared with the current location to produce an error vector. This vector is backpropagated through a motor-perceptual mapping network to produce an action vector that leads the system towards do-be locations and away from don't-be locations. ALVIS is demonstrated with a simulated robot posed a target-seeking task.
[1] Geoffrey E. Hinton,et al. Learning representations by back-propagation errors, nature , 1986 .
[2] Geoffrey E. Hinton,et al. Learning representations by back-propagating errors , 1986, Nature.
[3] Yoshiro Miyata,et al. The learning and planning of actions , 1988 .