Learning where to look for a hidden target

Survival depends on successfully foraging for food, for which evolution has selected diverse behaviors in different species. Humans forage not only for food, but also for information. We decide where to look over 170,000 times per day, approximately three times per wakeful second. The frequency of these saccadic eye movements belies the complexity underlying each individual choice. Experience factors into the choice of where to look and can be invoked to rapidly redirect gaze in a context and task-appropriate manner. However, remarkably little is known about how individuals learn to direct their gaze given the current context and task. We designed a task in which participants search a novel scene for a target whose location was drawn stochastically on each trial from a fixed prior distribution. The target was invisible on a blank screen, and the participants were rewarded when they fixated the hidden target location. In just a few trials, participants rapidly found the hidden targets by looking near previously rewarded locations and avoiding previously unrewarded locations. Learning trajectories were well characterized by a simple reinforcement-learning (RL) model that maintained and continually updated a reward map of locations. The RL model made further predictions concerning sensitivity to recent experience that were confirmed by the data. The asymptotic performance of both the participants and the RL model approached optimal performance characterized by an ideal-observer theory. These two complementary levels of explanation show how experience in a novel environment drives visual search in humans and may extend to other forms of search such as animal foraging.

[1]  R. C. Langford How People Look at Pictures, A Study of the Psychology of Perception in Art. , 1936 .

[2]  R J HERRNSTEIN,et al.  Relative and absolute strength of response as a function of frequency of reinforcement. , 1961, Journal of the experimental analysis of behavior.

[3]  A. L. I︠A︡rbus Eye Movements and Vision , 1967 .

[4]  M. Potter Meaning in visual search. , 1975, Science.

[5]  E. Charnov Optimal foraging, the marginal value theorem. , 1976, Theoretical population biology.

[6]  D. Sparks,et al.  Size and distribution of movement fields in the monkey superior colliculus , 1976, Brain Research.

[7]  C. Bradshaw,et al.  Behavior of humans in variable-interval schedules of reinforcement. , 1976, Journal of the experimental analysis of behavior.

[8]  W M Baum,et al.  Choice, changeover, and travel. , 1982, Journal of the experimental analysis of behavior.

[9]  R A Abrams,et al.  Speed and accuracy of saccadic eye movements: characteristics of impulse variability in the oculomotor system. , 1989, Journal of experimental psychology. Human perception and performance.

[10]  T. Sejnowski,et al.  The predictive brain: temporal coincidence and temporal order in synaptic learning mechanisms. , 1994, Learning & memory.

[11]  T. A. Mark,et al.  Kinetics of matching. , 1994, Journal of experimental psychology. Animal behavior processes.

[12]  Peter Dayan,et al.  A Neural Substrate of Prediction and Reward , 1997, Science.

[13]  U. Greggers,et al.  Matching behavior of honeybees in a multiple-choice situation: The differential effect of environmental stimuli on the choice process , 1997 .

[14]  D H Brainard,et al.  The Psychophysics Toolbox. , 1997, Spatial vision.

[15]  Todd S. Horowitz,et al.  Visual search has no memory , 1998, Nature.

[16]  K. Rayner Eye movements in reading and information processing: 20 years of research. , 1998, Psychological bulletin.

[17]  M. Chun,et al.  Contextual Cueing: Implicit Learning and Memory of Visual Context Guides Spatial Attention , 1998, Cognitive Psychology.

[18]  H. Stanley,et al.  Optimizing the success of random searches , 1999, Nature.

[19]  P Reinagel,et al.  Natural scene statistics at the centre of gaze. , 1999, Network.

[20]  Michael L. Platt,et al.  Neural correlates of decision variables in parietal cortex , 1999, Nature.

[21]  C. Koch,et al.  A saliency-based search mechanism for overt and covert shifts of visual attention , 2000, Vision Research.

[22]  J M Wolfe,et al.  Search for multiple targets: Remember the targets, forget the search , 2001, Perception & psychophysics.

[23]  Eileen Kowler,et al.  Eye movements during visual search: the costs of choosing the optimal path , 2001, Vision Research.

[24]  B. Hood,et al.  Is Visual Search Really like Foraging? , 2001, Perception.

[25]  Isaac Meilijson,et al.  Evolution of Reinforcement Learning in Uncertain Environments: A Simple Explanation for Complex Foraging Behaviors , 2002, Adapt. Behav..

[26]  Jillian H. Fecteau,et al.  Exploring the consequences of the previous trial , 2003, Nature Reviews Neuroscience.

[27]  Derrick J. Parkhurst,et al.  Scene content selected by active vision. , 2003, Spatial vision.

[28]  Christopher D. Carello,et al.  Manipulating Intent Evidence for a Causal Role of the Superior Colliculus in Target Selection , 2004, Neuron.

[29]  Michael C. Mozer,et al.  Top-Down Control of Visual Attention: A Rational Account , 2005, NIPS.

[30]  B. Hood,et al.  Children's Search Behaviour in Large-Scale Space: Developmental Components of Exploration , 2005, Perception.

[31]  D. Ballard,et al.  Eye movements in natural behavior , 2005, Trends in Cognitive Sciences.

[32]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[33]  Richard S. Sutton,et al.  Learning to predict by the methods of temporal differences , 1988, Machine Learning.

[34]  O Bénichou,et al.  Optimal search strategies for hidden targets. , 2005, Physical review letters.

[35]  Reinhold Kliegl,et al.  SWIFT: a dynamical model of saccade generation during reading. , 2005, Psychological review.

[36]  P. Glimcher,et al.  JOURNAL OF THE EXPERIMENTAL ANALYSIS OF BEHAVIOR 2005, 84, 555–579 NUMBER 3(NOVEMBER) DYNAMIC RESPONSE-BY-RESPONSE MODELS OF MATCHING BEHAVIOR IN RHESUS MONKEYS , 2022 .

[37]  M. Behrmann,et al.  Spatial probability as an attentional cue in visual search , 2005, Perception & psychophysics.

[38]  Jillian H. Fecteau,et al.  Salience, relevance, and firing: a priority map for target selection , 2006, Trends in Cognitive Sciences.

[39]  R. Baddeley,et al.  The long and the short of it: Spatial statistics at fixation vary with saccade amplitude and task , 2006, Vision Research.

[40]  Liming Xiang,et al.  Kernel-Based Reinforcement Learning , 2006, ICIC.

[41]  Antonio Torralba,et al.  Building the gist of a scene: the role of global image features in recognition. , 2006, Progress in brain research.

[42]  Antonio Torralba,et al.  Contextual guidance of eye movements and attention in real-world scenes: the role of global features in object search. , 2006, Psychological review.

[43]  Melissa R. Beck,et al.  Memory for where, but not what, is used during visual search. , 2006, Journal of experimental psychology. Human perception and performance.

[44]  Gregory J. Zelinsky,et al.  Scene context guides eye movements during visual search , 2006, Vision Research.

[45]  C. Gallistel,et al.  Is matching innate? , 2007, Journal of the experimental analysis of behavior.

[46]  David M Milstein,et al.  The Influence of Expected Value on Saccadic Preparation , 2007, The Journal of Neuroscience.

[47]  Julia Trommershäuser,et al.  Eye movements during rapid pointing under risk , 2007, Vision Research.

[48]  M. Platt,et al.  Neural Correlates of Social Target Value in Macaque Parietal Cortex , 2008, Current Biology.

[49]  Jonathan D. Cohen,et al.  Sequential effects: Superstition or rational behavior? , 2008, NIPS.

[50]  B. Tatler,et al.  The prominence of behavioural biases in eye guidance , 2009 .

[51]  F. Scharnowski,et al.  Long-lasting modulation of feature integration by transcranial magnetic stimulation. , 2009, Journal of vision.

[52]  J. Bisley,et al.  Been there, seen that: a neural mechanism for performing efficient visual search. , 2009, Journal of neurophysiology.

[53]  R. Shadmehr,et al.  The intrinsic value of visual information affects saccade velocities , 2009, Experimental Brain Research.

[54]  Barbara Anne Dosher,et al.  Task precision at transfer determines specificity of perceptual learning. , 2009, Journal of vision.

[55]  K. Rayner,et al.  Eye movements when looking at unusual/weird scenes: are there cultural differences? , 2009, Journal of experimental psychology. Learning, memory, and cognition.

[56]  Matt Jones,et al.  Sequential Effects Reflect Parallel Learning of Multiple Environmental Regularities , 2022 .

[57]  Scott D. Brown,et al.  Detecting and predicting changes , 2009, Cognitive Psychology.

[58]  Michael L. Mack,et al.  Viewing task influences eye movement control during active scene perception. , 2009, Journal of vision.

[59]  R. Shadmehr,et al.  Temporal Discounting of Reward and the Cost of Time in Motor Control , 2010, The Journal of Neuroscience.

[60]  John M Henderson,et al.  The time course of initial scene processing for eye movement guidance in natural scene search. , 2010, Journal of vision.

[61]  M. Goldberg,et al.  Attention, intention, and priority in the parietal lobe. , 2010, Annual review of neuroscience.

[62]  M. Castelhano,et al.  The relative contribution of scene context and target features to visual search in scenes , 2010, Attention, perception & psychophysics.

[63]  Todd S Horowitz,et al.  Distinguishing between parallel and serial accounts of multiple object tracking. , 2010, Journal of vision.

[64]  P. König,et al.  Developmental Changes in Natural Viewing Behavior: Bottom-Up and Top-Down Differences between Children, Young Adults and Older Adults , 2010, Front. Psychology.

[65]  R. Shadmehr Control of movements and temporal discounting of reward , 2010, Current Opinion in Neurobiology.

[66]  Nicolas E. Humphries,et al.  Environmental context explains Lévy and Brownian movement patterns of marine predators , 2010, Nature.

[67]  J. Bisley,et al.  Microstimulation of posterior parietal cortex biases the selection of eye movement goals during search. , 2010, Journal of neurophysiology.

[68]  Michael L. Mack,et al.  Modeling categorization of scenes containing consistent versus inconsistent objects. , 2010, Journal of vision.

[69]  A. M. Edwards,et al.  Assessing Lévy walks as models of animal foraging , 2011, Journal of The Royal Society Interface.

[70]  M. Castelhano,et al.  Scene context influences without scene gist: Eye movements guided by spatial associations in visual search , 2011, Psychonomic bulletin & review.

[71]  Michelle R. Greene,et al.  Visual search in scenes involves selective and nonselective pathways , 2011, Trends in Cognitive Sciences.

[72]  John M. Pearson,et al.  Neuronal basis of sequential foraging decisions in a patchy environment , 2011, Nature Neuroscience.

[73]  J. Snider Optimal random search for a single hidden target. , 2010, Physical review. E, Statistical, nonlinear, and soft matter physics.

[74]  Matthew H. Wilder,et al.  An integrative, experience-based theory of attentional control. , 2011, Journal of vision.

[75]  G. T. Buswell How People Look At Pictures: A Study Of The Psychology Of Perception In Art , 2012 .

[76]  O. Hikosaka,et al.  Robust Representation of Stable Object Values in the Oculomotor Basal Ganglia , 2012, The Journal of Neuroscience.

[77]  Alexander C. Schütz,et al.  Dynamic integration of information about salience and value for saccadic eye movements , 2012, Proceedings of the National Academy of Sciences.

[78]  Michelle R. Greene,et al.  Reconsidering Yarbus: A failure to predict observers’ task from eye movement patterns , 2012, Vision Research.

[79]  H. Seo,et al.  Neural basis of reinforcement learning and decision making. , 2012, Annual review of neuroscience.

[80]  O. Hikosaka,et al.  Learning to represent reward structure: A key to adapting to complex environments , 2012, Neuroscience Research.

[81]  John M. Pearson,et al.  Neuroethology of decision-making , 2012, Current Opinion in Neurobiology.