fbpx
Wikipedia

Visual search

Visual search is a type of perceptual task requiring attention that typically involves an active scan of the visual environment for a particular object or feature (the target) among other objects or features (the distractors).[1] Visual search can take place with or without eye movements. The ability to consciously locate an object or target amongst a complex array of stimuli has been extensively studied over the past 40 years. Practical examples of using visual search can be seen in everyday life, such as when one is picking out a product on a supermarket shelf, when animals are searching for food among piles of leaves, when trying to find a friend in a large crowd of people, or simply when playing visual search games such as Where's Wally?

Much previous literature on visual search used reaction time in order to measure the time it takes to detect the target amongst its distractors. An example of this could be a green square (the target) amongst a set of red circles (the distractors). However, reaction time measurements do not always distinguish between the role of attention and other factors: a long reaction time might be the result of difficulty directing attention to the target, or slowed decision-making processes or slowed motor responses after attention is already directed to the target and the target has already been detected. Many visual search paradigms have therefore used eye movement as a means to measure the degree of attention given to stimuli.[2][3] However, eyes can move independently of attention, and therefore eye movement measures do not completely capture the role of attention.[4][5]

Search types edit

Feature search edit

 
feature based search task

Feature search (also known as "disjunctive" or "efficient" search)[6] is a visual search process that focuses on identifying a previously requested target amongst distractors that differ from the target by a unique visual feature such as color, shape, orientation, or size.[7] An example of a feature search task is asking a participant to identify a white square (target) surrounded by black squares (distractors).[6] In this type of visual search, the distractors are characterized by the same visual features.[7] The efficiency of feature search in regards to reaction time (RT) and accuracy depends on the "pop out" effect,[8] bottom-up processing,[8] and parallel processing.[7] However, the efficiency of feature search is unaffected by the number of distractors present.[7]

The "pop out" effect is an element of feature search that characterizes the target's ability to stand out from surrounding distractors due to its unique feature.[8] Bottom-up processing, which is the processing of information that depends on input from the environment,[8] explains how one utilizes feature detectors to process characteristics of the stimuli and differentiate a target from its distractors.[7] This draw of visual attention towards the target due to bottom-up processes is known as "saliency."[9] Lastly, parallel processing is the mechanism that then allows one's feature detectors to work simultaneously in identifying the target.[7]

Conjunction search edit

 
Conjunctive based search task.

Conjunction search (also known as inefficient or serial search)[6] is a visual search process that focuses on identifying a previously requested target surrounded by distractors possessing no distinct features from the target itself.[10] An example of a conjunction search task is having a person identify a red X (target) amongst distractors composed of black Xs (same shape) and red Os (same color).[10] Unlike feature search, conjunction search involves distractors (or groups of distractors) that may differ from each other but exhibit at least one common feature with the target.[10] The efficiency of conjunction search in regards to reaction time (RT) and accuracy is dependent on the distractor-ratio[10] and the number of distractors present.[7] As the distractors represent the differing individual features of the target more equally amongst themselves (distractor-ratio effect), reaction time(RT) increases and accuracy decreases.[10] As the number of distractors present increases, the reaction time (RT) increases and the accuracy decreases.[6] However, with practice the original reaction time (RT) restraints of conjunction search tend to show improvement.[11] In the early stages of processing, conjunction search utilizes bottom-up processes to identify pre-specified features amongst the stimuli.[7] These processes are then overtaken by a more serial process of consciously evaluating the indicated features of the stimuli[7] in order to properly allocate one's focal spatial attention towards the stimulus that most accurately represents the target.[12]

In many cases, top-down processing affects conjunction search by eliminating stimuli that are incongruent with one's previous knowledge of the target-description, which in the end allows for more efficient identification of the target.[8][9] An example of the effect of top-down processes on a conjunction search task is when searching for a red 'K' among red 'Cs' and black 'Ks', individuals ignore the black letters and focus on the remaining red letters in order to decrease the set size of possible targets and, therefore, more efficiently identify their target.[13]

Real world visual search edit

In everyday situations, people are most commonly searching their visual fields for targets that are familiar to them. When it comes to searching for familiar stimuli, top-down processing allows one to more efficiently identify targets with greater complexity than can be represented in a feature or conjunction search task.[8] In a study done to analyze the reverse-letter effect, which is the idea that identifying the asymmetric letter among symmetric letters is more efficient than its reciprocal, researchers concluded that individuals more efficiently recognize an asymmetric letter among symmetric letters due to top-down processes.[9] Top-down processes allowed study participants to access prior knowledge regarding shape recognition of the letter N and quickly eliminate the stimuli that matched their knowledge.[9] In the real world, one must use prior knowledge everyday in order to accurately and efficiently locate objects such as phones, keys, etc. among a much more complex array of distractors.[8] Despite this complexity, visual search with complex objects (and search for categories of objects, such as "phone", based on prior knowledge) appears to rely on the same active scanning processes as conjunction search with less complex, contrived laboratory stimuli,[14][15] although global statistical information available in real-world scenes can also help people locate target objects.[16][17][18] While bottom-up processes may come into play when identifying objects that are not as familiar to a person, overall top-down processing highly influences visual searches that occur in everyday life.[8][19][20] Familiarity can play especially critical roles when parts of objects are not visible (as when objects are partly hidden from view because they are behind other objects). Visual information from hidden parts can be recalled from long-term memory and used to facilitate search for familiar objects.[21][22]

Reaction time slope edit

It is also possible to measure the role of attention within visual search experiments by calculating the slope of reaction time over the number of distractors present.[23] Generally, when high levels of attention are required when looking at a complex array of stimuli (conjunction search), the slope increases as reaction times increase. For simple visual search tasks (feature search), the slope decreases due to reaction times being fast and requiring less attention.[24] However, the use of a reaction time slope to measure attention is controversial because non-attentional factors can also affect reaction time slope.[25][26][27]

Visual orienting and attention edit

 
A photograph that simulates foveation

One obvious way to select visual information is to turn towards it, also known as visual orienting. This may be a movement of the head and/or eyes towards the visual stimulus, called a saccade. Through a process called foveation, the eyes fixate on the object of interest, making the image of the visual stimulus fall on the fovea of the eye, the central part of the retina with the sharpest visual acuity.

There are two types of orienting:

  • Exogenous orienting is the involuntary and automatic movement that occurs to direct one's visual attention toward a sudden disruption in his peripheral vision field.[28] Attention is therefore externally guided by a stimulus, resulting in a reflexive saccade.
  • Endogenous orienting is the voluntary movement that occurs in order for one to focus visual attention on a goal-driven stimulus.[28] Thus, the focus of attention of the perceiver can be manipulated by the demands of a task. A scanning saccade is triggered endogenously for the purpose of exploring the visual environment.
 
A plot of the saccades made while reading text. The plot shows the path of eye movements and the size of the circles represents the time spent at any one location.

Visual search relies primarily on endogenous orienting because participants have the goal to detect the presence or absence of a specific target object in an array of other distracting objects.

Early research suggested that attention could be covertly (without eye movement) shifted to peripheral stimuli,[29] but later studies found that small saccades (microsaccades) occur during these tasks, and that these eye movements are frequently directed towards the attended locations (whether or not there are visible stimuli).[30][31][32] These findings indicate that attention plays a critical role in understanding visual search.

Subsequently, competing theories of attention have come to dominate visual search discourse.[33] The environment contains a vast amount of information. We are limited in the amount of information we are able to process at any one time, so it is therefore necessary that we have mechanisms by which extraneous stimuli can be filtered and only relevant information attended to. In the study of attention, psychologists distinguish between pre-attentive and attentional processes.[34] Pre-attentive processes are evenly distributed across all input signals, forming a kind of "low-level" attention. Attentional processes are more selective and can only be applied to specific preattentive input. A large part of the current debate in visual search theory centres on selective attention and what the visual system is capable of achieving without focal attention.[33]

Theory edit

Feature integration theory (FIT) edit

A popular explanation for the different reaction times of feature and conjunction searches is the feature integration theory (FIT), introduced by Treisman and Gelade in 1980. This theory proposes that certain visual features are registered early, automatically, and are coded rapidly in parallel across the visual field using pre-attentive processes.[35] Experiments show that these features include luminance, colour, orientation, motion direction, and velocity, as well as some simple aspects of form.[36] For example, a red X can be quickly found among any number of black Xs and Os because the red X has the discriminative feature of colour and will "pop out." In contrast, this theory also suggests that in order to integrate two or more visual features belonging to the same object, a later process involving integration of information from different brain areas is needed and is coded serially using focal attention. For example, when locating an orange square among blue squares and orange triangles, neither the colour feature "orange" nor the shape feature "square" is sufficient to locate the search target. Instead, one must integrate information of both colour and shape to locate the target.

Evidence that attention and thus later visual processing is needed to integrate two or more features of the same object is shown by the occurrence of illusory conjunctions, or when features do not combine correctly For example, if a display of a green X and a red O are flashed on a screen so briefly that the later visual process of a serial search with focal attention cannot occur, the observer may report seeing a red X and a green O.

The FIT is a dichotomy because of the distinction between its two stages: the preattentive and attentive stages.[37] Preattentive processes are those performed in the first stage of the FIT model, in which the simplest features of the object are being analyzed, such as color, size, and arrangement. The second attentive stage of the model incorporates cross-dimensional processing,[38] and the actual identification of an object is done and information about the target object is put together. This theory has not always been what it is today; there have been disagreements and problems with its proposals that have allowed the theory to be amended and altered over time, and this criticism and revision has allowed it to become more accurate in its description of visual search.[38] There have been disagreements over whether or not there is a clear distinction between feature detection and other searches that use a master map accounting for multiple dimensions in order to search for an object. Some psychologists support the idea that feature integration is completely separate from this type of master map search, whereas many others have decided that feature integration incorporates this use of a master map in order to locate an object in multiple dimensions.[37]

The FIT also explains that there is a distinction between the brain's processes that are being used in a parallel versus a focal attention task. Chan and Hayward[37] have conducted multiple experiments supporting this idea by demonstrating the role of dimensions in visual search. While exploring whether or not focal attention can reduce the costs caused by dimension-switching in visual search, they explained that the results collected supported the mechanisms of the feature integration theory in comparison to other search-based approaches. They discovered that single dimensions allow for a much more efficient search regardless of the size of the area being searched, but once more dimensions are added it is much more difficult to efficiently search, and the bigger the area being searched the longer it takes for one to find the target.[37]

Guided search model edit

A second main function of preattentive processes is to direct focal attention to the most "promising" information in the visual field.[33] There are two ways in which these processes can be used to direct attention: bottom-up activation (which is stimulus-driven) and top-down activation (which is user-driven). In the guided search model by Jeremy Wolfe,[39] information from top-down and bottom-up processing of the stimulus is used to create a ranking of items in order of their attentional priority. In a visual search, attention will be directed to the item with the highest priority. If that item is rejected, then attention will move on to the next item and the next, and so forth. The guided search theory follows that of parallel search processing.

An activation map is a representation of visual space in which the level of activation at a location reflects the likelihood that the location contains a target. This likelihood is based on preattentive, featural information of the perceiver. According to the guided search model, the initial processing of basic features produces an activation map, with every item in the visual display having its own level of activation. Attention is demanded based on peaks of activation in the activation map in a search for the target.[39] Visual search can proceed efficiently or inefficiently. During efficient search, performance is unaffected by the number of distractor items. The reaction time functions are flat, and the search is assumed to be a parallel search. Thus, in the guided search model, a search is efficient if the target generates the highest, or one of the highest activation peaks. For example, suppose someone is searching for red, horizontal targets. Feature processing would activate all red objects and all horizontal objects. Attention is then directed to items depending on their level of activation, starting with those most activated. This explains why search times are longer when distractors share one or more features with the target stimuli. In contrast, during inefficient search, the reaction time to identify the target increases linearly with the number of distractor items present. According to the guided search model, this is because the peak generated by the target is not one of the highest.[39]

Biological basis edit

 
A pseudo-color image showing activation of the primary visual cortex during a perceptual task using functional magnetic resonance imaging (fMRI)

During visual search experiments the posterior parietal cortex has elicited much activation during functional magnetic resonance imaging (fMRI) and electroencephalography (EEG) experiments for inefficient conjunction search, which has also been confirmed through lesion studies. Patients with lesions to the posterior parietal cortex show low accuracy and very slow reaction times during a conjunction search task but have intact feature search remaining to the ipsilesional (the same side of the body as the lesion) side of space.[40][41][42][43] Ashbridge, Walsh, and Cowey in (1997)[44] demonstrated that during the application of transcranial magnetic stimulation (TMS) to the right parietal cortex, conjunction search was impaired by 100 milliseconds after stimulus onset. This was not found during feature search. Nobre, Coull, Walsh and Frith (2003)[45] identified using functional magnetic resonance imaging (fMRI) that the intraparietal sulcus located in the superior parietal cortex was activated specifically to feature search and the binding of individual perceptual features as opposed to conjunction search. Conversely, the authors further identify that for conjunction search, the superior parietal lobe and the right angular gyrus elicit bilaterally during fMRI experiments.

 
Visual search primarily activates areas of the parietal lobe.

In contrast, Leonards, Sunaert, Vam Hecke and Orban (2000)[46] identified that significant activation is seen during fMRI experiments in the superior frontal sulcus primarily for conjunction search. This research hypothesises that activation in this region may in fact reflect working memory for holding and maintaining stimulus information in mind in order to identify the target. Furthermore, significant frontal activation including the ventrolateral prefrontal cortex bilaterally and the right dorsolateral prefrontal cortex were seen during positron emission tomography for attentional spatial representations during visual search.[47] The same regions associated with spatial attention in the parietal cortex coincide with the regions associated with feature search. Furthermore, the frontal eye field (FEF) located bilaterally in the prefrontal cortex, plays a critical role in saccadic eye movements and the control of visual attention.[48][49][50]

Moreover, research into monkeys and single cell recording found that the superior colliculus is involved in the selection of the target during visual search as well as the initiation of movements.[51] Conversely, it also suggested that activation in the superior colliculus results from disengaging attention, ensuring that the next stimulus can be internally represented. The ability to directly attend to a particular stimuli during visual search experiments has been linked to the pulvinar nucleus (located in the midbrain) while inhibiting attention to unattended stimuli.[52] Conversely, Bender and Butter (1987)[53] found that during testing on monkeys, no involvement of the pulvinar nucleus was identified during visual search tasks.

There is evidence for the V1 Saliency Hypothesis that the primary visual cortex (V1) creates a bottom-up saliency map to guide attention exogenously,[54][55] and this V1 saliency map is read out by the superior colliculus which receives monosynaptic inputs from V1.

Evolution edit

There is a variety of speculation about the origin and evolution of visual search in humans. It has been shown that during visual exploration of complex natural scenes, both humans and nonhuman primates make highly stereotyped eye movements.[56] Furthermore, chimpanzees have demonstrated improved performance in visual searches for upright human or dog faces,[57] suggesting that visual search (particularly where the target is a face) is not peculiar to humans and that it may be a primal trait. Research has suggested that effective visual search may have developed as a necessary skill for survival, where being adept at detecting threats and identifying food was essential.[58][59]

 
Henri Rousseau, Jungle with Lion

The importance of evolutionarily relevant threat stimuli was demonstrated in a study by LoBue and DeLoache (2008) in which children (and adults) were able to detect snakes more rapidly than other targets amongst distractor stimuli.[60] However, some researchers question whether evolutionarily relevant threat stimuli are detected automatically.[61]

Face recognition edit

Over the past few decades there have been vast amounts of research into face recognition, specifying that faces endure specialized processing within a region called the fusiform face area (FFA) located in the mid fusiform gyrus in the temporal lobe.[62] Debates are ongoing whether both faces and objects are detected and processed in different systems and whether both have category specific regions for recognition and identification.[63][64] Much research to date focuses on the accuracy of the detection and the time taken to detect the face in a complex visual search array. When faces are displayed in isolation, upright faces are processed faster and more accurately than inverted faces,[65][66][67][68] but this effect was observed in non-face objects as well.[69] When faces are to be detected among inverted or jumbled faces, reaction times for intact and upright faces increase as the number of distractors within the array is increased.[70][71][72] Hence, it is argued that the 'pop out' theory defined in feature search is not applicable in the recognition of faces in such visual search paradigm. Conversely, the opposite effect has been argued and within a natural environmental scene, the 'pop out' effect of the face is significantly shown.[73] This could be due to evolutionary developments as the need to be able to identify faces that appear threatening to the individual or group is deemed critical in the survival of the fittest.[74] More recently, it was found that faces can be efficiently detected in a visual search paradigm, if the distracters are non-face objects,[75][76][77] however it is debated whether this apparent 'pop out' effect is driven by a high-level mechanism or by low-level confounding features.[78][79] Furthermore, patients with developmental prosopagnosia, who have impaired face identification, generally detect faces normally, suggesting that visual search for faces is facilitated by mechanisms other than the face-identification circuits of the fusiform face area.[80]

Patients with forms of dementia can also have deficits in facial recognition and the ability to recognize human emotions in the face. In a meta-analysis of nineteen different studies comparing normal adults with dementia patients in their abilities to recognize facial emotions,[81] the patients with frontotemporal dementia were seen to have a lower ability to recognize many different emotions. These patients were much less accurate than the control participants (and even in comparison with Alzheimer's patients) in recognizing negative emotions, but were not significantly impaired in recognizing happiness. Anger and disgust in particular were the most difficult for the dementia patients to recognize.[81]

Face recognition is a complex process that is affected by many factors, both environmental and individually internal. Other aspects to be considered include race and culture and their effects on one's ability to recognize faces.[82] Some factors such as the cross-race effect can influence one's ability to recognize and remember faces.

Considerations edit

Ageing edit

Research indicates that performance in conjunctive visual search tasks significantly improves during childhood and declines in later life.[83] More specifically, young adults have been shown to have faster reaction times on conjunctive visual search tasks than both children and older adults, but their reaction times were similar for feature visual search tasks.[52] This suggests that there is something about the process of integrating visual features or serial searching that is difficult for children and older adults, but not for young adults. Studies have suggested numerous mechanisms involved in this difficulty in children, including peripheral visual acuity,[84] eye movement ability,[85] ability of attentional focal movement,[86] and the ability to divide visual attention among multiple objects.[87]

Studies have suggested similar mechanisms in the difficulty for older adults, such as age related optical changes that influence peripheral acuity,[88] the ability to move attention over the visual field,[89] the ability to disengage attention,[90] and the ability to ignore distractors.[91]

A study by Lorenzo-López et al. (2008) provides neurological evidence for the fact that older adults have slower reaction times during conjunctive searches compared to young adults. Event-related potentials (ERPs) showed longer latencies and lower amplitudes in older subjects than young adults at the P3 component, which is related to activity of the parietal lobes. This suggests the involvement of the parietal lobe function with an age-related decline in the speed of visual search tasks. Results also showed that older adults, when compared to young adults, had significantly less activity in the anterior cingulate cortex and many limbic and occipitotemporal regions that are involved in performing visual search tasks.[92]

Alzheimer's disease edit

Research has found that people with Alzheimer's disease (AD) are significantly impaired overall in visual search tasks.[93] People with AD manifest enhanced spatial cueing, but this benefit is only obtained for cues with high spatial precision.[94] Abnormal visual attention may underlie certain visuospatial difficulties in patients with (AD). People with AD have hypometabolism and neuropathology in the parietal cortex, and given the role of parietal function for visual attention, patients with AD may have hemispatial neglect, which may result in difficulty with disengaging attention in visual search.[95]

An experiment conducted by Tales et al. (2000)[93] investigated the ability of patients with AD to perform various types of visual search tasks. Their results showed that search rates on "pop-out" tasks were similar for both AD and control groups, however, people with AD searched significantly slower compared to the control group on a conjunction task. One interpretation of these results is that the visual system of AD patients has a problem with feature binding, such that it is unable to communicate the different feature descriptions for the stimulus efficiently.[93] Binding of features is thought to be mediated by areas in the temporal and parietal cortex, and these areas are known to be affected by AD-related pathology.

Another possibility for the impairment of people with AD on conjunction searches is that there may be some damage to general attentional mechanisms in AD, and therefore any attention-related task will be affected, including visual search.[93]

Tales et al. (2000) detected a double dissociation with their experimental results on AD and visual search. Earlier work was carried out on patients with Parkinson's disease (PD) concerning the impairment patients with PD have on visual search tasks.[96][97] In those studies, evidence was found of impairment in PD patients on the "pop-out" task, but no evidence was found on the impairment of the conjunction task. As discussed, AD patients show the exact opposite of these results: normal performance was seen on the "pop-out" task, but impairment was found on the conjunction task. This double dissociation provides evidence that PD and AD affect the visual pathway in different ways, and that the pop-out task and the conjunction task are differentially processed within that pathway.

Autism edit

Studies have consistently shown that autistic individuals performed better and with lower reaction times in feature and conjunctive visual search tasks than matched controls without autism.[98][99] Several explanations for these observations have been suggested. One possibility is that people with autism have enhanced perceptual capacity.[99] This means that autistic individuals are able to process larger amounts of perceptual information, allowing for superior parallel processing and hence faster target location.[100] Second, autistic individuals show superior performance in discrimination tasks between similar stimuli and therefore may have an enhanced ability to differentiate between items in the visual search display.[101] A third suggestion is that autistic individuals may have stronger top-down target excitation processing and stronger distractor inhibition processing than controls.[98] Keehn et al. (2008) used an event-related functional magnetic resonance imaging design to study the neurofunctional correlates of visual search in autistic children and matched controls of typically developing children.[102] Autistic children showed superior search efficiency and increased neural activation patterns in the frontal, parietal, and occipital lobes when compared to the typically developing children. Thus, autistic individuals' superior performance on visual search tasks may be due to enhanced discrimination of items on the display, which is associated with occipital activity, and increased top-down shifts of visual attention, which is associated with the frontal and parietal areas.

Consumer psychology edit

In the past decade, there has been extensive research into how companies can maximise sales using psychological techniques derived from visual search to determine how products should be positioned on shelves. Pieters and Warlop (1999)[103] used eye tracking devices to assess saccades and fixations of consumers while they visually scanned/searched an array of products on a supermarket shelf. Their research suggests that consumers specifically direct their attention to products with eye-catching properties such as shape, colour or brand name. This effect is due to a pressured visual search where eye movements accelerate and saccades minimise, thus resulting in the consumer's quickly choosing a product with a 'pop out' effect. This study suggests that efficient search is primarily used, concluding that consumers do not focus on items that share very similar features. The more distinct or maximally visually different a product is from surrounding products, the more likely the consumer is to notice it. Janiszewski (1998)[104] discussed two types of consumer search. One search type is goal directed search taking place when somebody uses stored knowledge of the product in order to make a purchase choice. The second is exploratory search. This occurs when the consumer has minimal previous knowledge about how to choose a product. It was found that for exploratory search, individuals would pay less attention to products that were placed in visually competitive areas such as the middle of the shelf at an optimal viewing height. This was primarily due to the competition in attention meaning that less information was maintained in visual working memory for these products.

References edit

  1. ^ Treisman, AM; Gelade, G (January 1980). "A feature-integration theory of attention". Cogn Psychol. 12 (1): 97–136. doi:10.1016/0010-0285(80)90005-5. PMID 7351125. S2CID 353246.
  2. ^ Shelga, B. M.; Riggio, L.; Rizzolatti, G. (1994). "Orienting of attention and eye movements". Experimental Brain Research. 98 (3): 507–522. doi:10.1007/bf00233988. PMID 8056071. S2CID 2280537.
  3. ^ Hoffman, J. E.; B. Subramaniam (1995). "The role of visual attention in saccadic eye movements". Perception and Psychophysics. 57 (6): 787–795. doi:10.3758/bf03206794. PMID 7651803.
  4. ^ Klein, R; Farrell, M (1989). "Search performance without eye movements". Percept Psychophys. 46 (5): 476–82. doi:10.3758/BF03210863. PMID 2813033.
  5. ^ Murthy, A; Thompson, KG; Schall, JD (2001). "Dynamic dissociation of visual selection from saccade programming in frontal eye field". J Neurophysiol. 86 (5): 2634–7. doi:10.1152/jn.2001.86.5.2634. PMID 11698551. S2CID 653798.
  6. ^ a b c d Treisman, A. M.; Gelade, G (1980). "A feature-integration theory of attention". Cognitive Psychology. 12 (1): 97–136. doi:10.1016/0010-0285(80)90005-5. PMID 7351125. S2CID 353246.
  7. ^ a b c d e f g h i McElree, B; Carrasco, M (December 1999). "The temporal dynamics of visual search: evidence for parallel processing in feature and conjunction searches". Journal of Experimental Psychology: Human Perception and Performance. 25 (6): 1517–39. doi:10.1037/0096-1523.25.6.1517. PMC 3313830. PMID 10641310.
  8. ^ a b c d e f g h Radvansky, Gabriel, A.; Ashcraft, Mark, H. (2016). Cognition (6 ed.). Pearson Education, Inc. – via online.{{cite book}}: CS1 maint: multiple names: authors list (link)
  9. ^ a b c d Zhaoping, L; Frith, U (August 2011). "A clash of bottom-up and top-down processes in visual search: the reversed letter effect revisited". Journal of Experimental Psychology: Human Perception and Performance. 37 (4): 997–1006. doi:10.1037/a0023099. PMID 21574744. S2CID 12986008.
  10. ^ a b c d e Shen, J; Reingold, EM; Pomplun, M (June 2003). "Guidance of eye movements during conjunctive visual search: the distractor-ratio effect". Canadian Journal of Experimental Psychology. 57 (2): 76–96. CiteSeerX 10.1.1.59.251. doi:10.1037/h0087415. PMID 12822838.
  11. ^ Reavis, EA; Frank, SM; Greenlee, MW; Tse, PU (June 2016). "Neural correlates of context-dependent feature conjunction learning in visual search tasks". Human Brain Mapping. 37 (6): 2319–30. doi:10.1002/hbm.23176. PMC 6867346. PMID 26970441. S2CID 205849752.
  12. ^ Eimer, M; Grubert, A (October 2014). "The gradual emergence of spatially selective target processing in visual search: From feature-specific to object-based attentional control" (PDF). Journal of Experimental Psychology: Human Perception and Performance. 40 (5): 1819–31. doi:10.1037/a0037387. PMID 24999612.
  13. ^ Wolfe, J.M. (2014). Approaches to visual search: feature integration theory and guided search. The Oxford handbook of attention. Oxford: Oxford University Press. pp. 11–50. ISBN 9780199675111.
  14. ^ Alexander, Robert G.; Zelinsky, Gregory J. (2012). "Effects of part-based similarity on visual search: The Frankenbear experiment". Vision Research. 54: 20–30. doi:10.1016/j.visres.2011.12.004. PMC 3345177. PMID 22227607.
  15. ^ Alexander, Robert G.; Zelinsky, Gregory J. (2011). "Visual Similarity Effects in Categorical Search". Journal of Vision. 11 (8): 9. doi:10.1167/11.8.9. PMC 8409006. PMID 21757505.
  16. ^ Rosenholtz, Ruth; Huang, Jie; Raj, A.; Balas, Benjamin J.; Ilie, Livia (2012). "A summary statistic representation in peripheral vision explains visual search". Journal of Vision. 12 (4): 14. doi:10.1167/12.4.14. PMC 4032502. PMID 22523401.
  17. ^ Alexander, Robert G.; Schmidt, Joseph; Zelinsky, Gregory J. (2014). "Are summary statistics enough? Evidence for the importance of shape in guiding visual search". Visual Cognition. 22 (3–4): 595–609. doi:10.1080/13506285.2014.890989. PMC 4500174. PMID 26180505.
  18. ^ Wolfe, Jeremy M.; Võ, Melissa L.-H.; Evans, Karla K.; Greene, Michelle R. (2011). "Visual search in scenes involves selective and nonselective pathways". Trends in Cognitive Sciences. 15 (2): 77–84. doi:10.1016/j.tics.2010.12.001. PMC 3035167. PMID 21227734.
  19. ^ Siebold, Alisha; Van Zoest, Wieske; Donk, Mieke (2011). "Oculomotor evidence for top-down control following the initial saccade". PLOS ONE. 6 (9): e23552. Bibcode:2011PLoSO...623552S. doi:10.1371/journal.pone.0023552. PMC 3169564. PMID 21931603.
  20. ^ Malcolm, G. L.; Henderson, J. M. (2010). "Combining top-down processes to guide eye movements during real-world scene search". Journal of Vision. 10 (2): 4.1–11. doi:10.1167/10.2.4. PMID 20462305.
  21. ^ Alexander, Robert G.; Zelinsky, Gregory J. (2018). "Occluded information is restored at preview but not during visual search". J Vis. 18 (11): 4. doi:10.1167/18.11.4. PMC 6181188. PMID 30347091.
  22. ^ Plomp, G; Nakatani, C; Bonnardel, V; Leeuwen, C. v. (2004). "Amodal completion as reflected by gaze durations". Perception. 33 (10): 1185–1200. doi:10.1068/p5342x. PMID 15693664. S2CID 35581448.
  23. ^ Trick, Lana M.; Enns, James T. (1998-07-01). "Lifespan changes in attention: The visual search task". Cognitive Development. 13 (3): 369–386. CiteSeerX 10.1.1.522.1907. doi:10.1016/S0885-2014(98)90016-8.
  24. ^ Alvarez, G. A.; Cavanagh, P. (2004-02-01). "The capacity of visual short-term memory is set both by visual information load and by number of objects". Psychological Science. 15 (2): 106–111. doi:10.1111/j.0963-7214.2004.01502006.x. ISSN 0956-7976. PMID 14738517. S2CID 2286443.
  25. ^ Palmer, J. (1995). "Attention in visual search: Distinguishing four causes of a set-size effect". Current Directions in Psychological Science. 4 (4): 118–123. doi:10.1111/1467-8721.ep10772534. S2CID 145161732.
  26. ^ Eckstein, M. P. (2011). "Visual search: A retrospective". Journal of Vision. 11 (5): 14. doi:10.1167/11.5.14. PMID 22209816.
  27. ^ Algom, D; Eidels, A; Hawkins, R.X.D; Jefferson, B; Townsend, J. T. (2015). "Features of response times: Identification of cognitive mechanisms through mathematical modeling.". In Busemeyer, J; Wang, Z; Townsend, J. T.; Eidels, A (eds.). The Oxford handbook of computational and mathematical psychology. Oxford University Press.
  28. ^ a b Berger, A; Henik, A; Rafal, R (May 2005). "Competition between endogenous and exogenous orienting of visual attention". Journal of Experimental Psychology: General. 134 (2): 207–21. doi:10.1037/0096-3445.134.2.207. PMID 15869346. S2CID 19087912.
  29. ^ Wurtz, Robert H.; Michael E. Goldberg; David Lee Robinson (June 1982). "Brain Mechanisms of Visual Attention". Scientific American. 246 (6): 124–135. Bibcode:1982SciAm.246f.124W. doi:10.1038/scientificamerican0682-124. ISSN 0036-8733. PMID 7100892.
  30. ^ Laubrock, J; Kliegl, R; Rolfs, M; Engbert, R (2010). "When do microsaccades follow spatial attention?". Attention, Perception, & Psychophysics. 72 (3): 683–694. doi:10.3758/APP.72.3.683. PMID 20348575.
  31. ^ Martinez-Conde, S; Alexander, R. G. (2019). "A gaze bias in the mind's eye". Nature Human Behaviour. 3 (5): 424–425. doi:10.1038/s41562-019-0546-1. PMID 31089295. S2CID 71148025.
  32. ^ Laubrock; Engbert; Kliegl (2005). "Microsaccade dynamics during covert attention". Vision Research. 45 (6): 721–730. doi:10.1016/j.visres.2004.09.029. PMID 15639499.
  33. ^ a b c Müller, Hermann J.; Krummenacher, Joseph (2006). "Visual search and selective attention". Visual Cognition. 14 (4–8): 389–410. doi:10.1080/13506280500527676. ISSN 1350-6285. S2CID 671170.
  34. ^ Neisser, Ulric (1967). Cognitive Psychology. Meredith Publishing Company. ISBN 9781317566182. PsycNET 1967-35031-000. Retrieved 2012-11-17.
  35. ^ Treisman, A. M.; G. Gelade (1980). "A feature-integration theory of attention". Cognitive Psychology. 12 (1): 97–136. doi:10.1016/0010-0285(80)90005-5. PMID 7351125. S2CID 353246.
  36. ^ Wolfe, J. M. (1998). "What can 1 million trials tell us about visual search?". Psychological Science. 9 (1): 33–39. CiteSeerX 10.1.1.148.6975. doi:10.1111/1467-9280.00006. S2CID 11042813.
  37. ^ a b c d Chan, Louis K. H.; Hayward, William G. (2009). "Feature integration theory revisited: Dissociating feature detection and attentional guidance in visual search". Journal of Experimental Psychology: Human Perception and Performance. 35 (1): 119–132. doi:10.1037/0096-1523.35.1.119. PMID 19170475. S2CID 22704624.
  38. ^ a b Quinlan, Philip T. (September 2003). "Visual feature integration theory: Past, present, and future". Psychological Bulletin. 129 (5): 643–673. doi:10.1037/0033-2909.129.5.643. PMID 12956538. S2CID 25206656.
  39. ^ a b c Wolfe, J. M. (1994). "Guided search 2.0 A revised model of visual search". Psychonomic Bulletin & Review. 1 (2): 202–238. doi:10.3758/bf03200774. PMID 24203471.
  40. ^ Aglioti, S.; Smania, N.; Barbieri, C.; Corbetta, M. (1997). "Influence of stimulus salience and attentional demands on visual search patterns in hemispatial neglect". Brain and Cognition. 34 (3): 388–403. doi:10.1006/brcg.1997.0915. PMID 9292188. S2CID 13138212.
  41. ^ Eglin, M.; Robertson, L. C.; Knight, R. T. (1991). "Cortical substrates supporting visual search in humans". Cerebral Cortex. 1 (3): 262–272. doi:10.1093/cercor/1.3.262. PMID 1822736. S2CID 12364128.
  42. ^ Friedman-Hill, S. R.; Robertson, L. C.; Treisman, A. (1995). "Parietal contributions to visual feature binding: Evidence from a patient with bilateral lesions". Science. 269 (5225): 853–855. Bibcode:1995Sci...269..853F. doi:10.1126/science.7638604. PMID 7638604. S2CID 42706447.
  43. ^ Ellison, A.; Schindler, I.; Pattison, L. L.; Milner, A. D (2004). "An exploration of the role of the superior temporal gyrus in visualsearch and spatial perception using TMS.v". Brain. 127 (10): 2307–2315. doi:10.1093/brain/awh244. PMID 15292055.
  44. ^ Ashbridge, V.; Walsh, A.; Cowey, D (1997). "Temporal aspects of visual search studied by transcranial magnetic stimulation". Neuropsychologia. 35 (8): 1121–1131. doi:10.1016/s0028-3932(97)00003-1. PMID 9256377. S2CID 7305220.
  45. ^ Nobre, A. C.; J. T. Coull; V. Walsh; C. D. Frith (2003). "Brain activations during visual search: contributions of search efficiency versus feature binding". NeuroImage. 18 (1): 91–103. doi:10.1006/nimg.2002.1329. PMID 12507447. S2CID 32757191.
  46. ^ Leonards, U.; Suneart, S.; Van Hecke, P.; Orban, G. (2000). "Attention mechanisms in visual search—An fMRI study". Journal of Cognitive Neuroscience. 12: 61–75. doi:10.1162/089892900564073. PMID 11506648. S2CID 35061939.
  47. ^ Nobre, A.C,.; Sebestyen, G. N.; Gitelman, D. R.; Frith, C. D.; Mesulam, M. M. (2002). "Filtering of distractors during visual search studied by positron emission tomography". NeuroImage. 16 (4): 968–976. doi:10.1006/nimg.2002.1137. PMID 12202084. S2CID 1702722.{{cite journal}}: CS1 maint: multiple names: authors list (link)
  48. ^ Schall JD. (2004). "On the role of frontal eye field in guiding attention and saccades". Vision Research. 44 (12): 1453–1467. doi:10.1016/j.visres.2003.10.025. PMID 15066404.
  49. ^ . Archived from the original on 2011-11-09.
  50. ^ Mustari MJ, Ono S, Das VE (May 2009). "Signal processing and distribution in cortical-brainstem pathways for smooth pursuit eye movements". Ann. N. Y. Acad. Sci. 1164 (1): 147–54. Bibcode:2009NYASA1164..147M. doi:10.1111/j.1749-6632.2009.03859.x. PMC 3057571. PMID 19645893.
  51. ^ McPeek, R.M,.; Keller, E. L. (2002). "Saccade target selection in the superior colliculus during a visual search task". Journal of Neurophysiology. 18 (4): 2019–2034. doi:10.1152/jn.2002.88.4.2019. PMID 12364525. S2CID 16885851.{{cite journal}}: CS1 maint: multiple names: authors list (link)
  52. ^ a b Trick, L. M.; Enns, J. T. (1998). "Life-span changes in attention: The visual search task". Cognitive Development. 13 (3): 369–386. CiteSeerX 10.1.1.522.1907. doi:10.1016/s0885-2014(98)90016-8.
  53. ^ Bender, D.B,.; Butter, C. M. (1987). "Comparison of the effects of superior colliculus and pulvinar lesions on visual search and tachistoscopic pattern discrimination in monkeys" (PDF). Experimental Brain Research. 69 (1): 140–154. doi:10.1007/bf00247037. hdl:2027.42/46559. PMID 3436384. S2CID 2333254.{{cite journal}}: CS1 maint: multiple names: authors list (link)
  54. ^ Li, Zhaoping (2002-01-01). "A saliency map in primary visual cortex". Trends in Cognitive Sciences. 6 (1): 9–16. doi:10.1016/S1364-6613(00)01817-9. ISSN 1364-6613. PMID 11849610. S2CID 13411369.
  55. ^ Yan, Yin; Zhaoping, Li; Li, Wu (2018-10-09). "Bottom-up saliency and top-down learning in the primary visual cortex of monkeys". Proceedings of the National Academy of Sciences. 115 (41): 10499–10504. Bibcode:2018PNAS..11510499Y. doi:10.1073/pnas.1803854115. ISSN 0027-8424. PMC 6187116. PMID 30254154.
  56. ^ Mazer, James A; Jack L Gallant (2003-12-18). "Goal-Related Activity in V4 during Free Viewing Visual Search: Evidence for a Ventral Stream Visual Salience Map". Neuron. 40 (6): 1241–1250. doi:10.1016/S0896-6273(03)00764-5. ISSN 0896-6273. PMID 14687556.
  57. ^ Tomonaga, Masaki (2007-01-01). "Visual search for orientation of faces by a chimpanzee (Pan troglodytes): face-specific upright superiority and the role of facial configural properties". Primates. 48 (1): 1–12. doi:10.1007/s10329-006-0011-4. ISSN 0032-8332. PMID 16969584. S2CID 7313319.
  58. ^ Öhman, A; Mineka, S (2001). "Fears, phobias, and preparedness: Toward an evolved module of fear and fear learning". Psychological Review. 108 (3): 483–522. doi:10.1037/0033-295X.108.3.483. PMID 11488376. S2CID 7920871.
  59. ^ Öhman, A (1999). "Distinguishing unconscious from conscious emotional processes: Methodological considerations and theoretical implications.". In Dalgleish, T.; Powers, M. J. (eds.). Handbook of cognition and emotion. Chichester, England: Wiley. pp. 321–352.
  60. ^ LoBue, Vanessa; Judy S. DeLoache (2008-03-01). "Detecting the Snake in the Grass Attention to Fear-Relevant Stimuli by Adults and Young Children". Psychological Science. 19 (3): 284–289. doi:10.1111/j.1467-9280.2008.02081.x. ISSN 0956-7976. PMID 18315802. S2CID 12776572.
  61. ^ Quinlan, Philip T. (2013). "The visual detection of threat: A cautionary tale". Psychonomic Bulletin & Review. 20 (6): 1080–1101. doi:10.3758/s13423-013-0421-4. PMID 23504916.
  62. ^ Kanwisher, Nancy; McDermott, Josh; Chun, Marvin M. (1997). "The fusiform face area: a module in human extrastriate cortex specialized for face perception". The Journal of Neuroscience. 17 (11): 4302–4311. doi:10.1523/JNEUROSCI.17-11-04302.1997. PMC 6573547. PMID 9151747.
  63. ^ Tarr, M. J.; Gauthier, I. (2000). "FFA: a flexible fusiform area for subordinate-level visual processing automatized by expertise". Nature Neuroscience. 3 (8): 764–770. doi:10.1038/77666. PMID 10903568. S2CID 8355344.
  64. ^ Grill-Spector, K.; Knouf, N.; Kanwisher, N. (2004). "The fusiform face area subserves face perception, not generic within-category identification". Nature Neuroscience. 7 (5): 555–562. doi:10.1038/nn1224. PMID 15077112. S2CID 2204107.
  65. ^ Valentine, T; Bruce, V (1986). "The effects of distinctiveness in recognizing and classifying faces". Perception. 15 (5): 525–533. doi:10.1068/p150525. PMID 3588212. S2CID 9641249.
  66. ^ Purcell, D G; Stewart, A L (1986). "The face-detection effect". Bulletin of the Psychonomic Society. 24 (2): 118–120. doi:10.3758/bf03330521.
  67. ^ Purcell, D G; Stewart, A L (1988). "The face-detection effect: Configuration enhances perception". Perception & Psychophysics. 43 (4): 355–366. doi:10.3758/bf03208806. PMID 3362664.
  68. ^ Yovel, G.; Kanwisher, N. (2005). "The neural basis of the behavioural face-inversion effect". Current Biology. 15 (24): 2256–2262. doi:10.1016/j.cub.2005.10.072. PMID 16360687.
  69. ^ Purcell, D G; Stewart, A L (1991). "The object-detection effect: Configuration enhances perception". Perception & Psychophysics. 50 (3): 215–224. doi:10.3758/bf03206744. PMID 1754362.
  70. ^ Nothdurft, H. C. (1993). "Faces and facial expressions do not pop out". Perception. 22 (11): 1287–98. doi:10.1068/p221287. PMID 8047415. S2CID 33911653.
  71. ^ Kuehn, S. M.; Jolicoeur, P. (1994). "Impact of quality of the image, orientation, and similarity of the stimuli on visual search for faces". Perception. 23 (1): 95–122. doi:10.1068/p230095. PMID 7936979. S2CID 20262065.
  72. ^ Brown, V.; Huey, D.; Findlay, J. M. (1997). "Face detection in peripheral vision: do faces pop out?". Perception. 26 (12): 1555–1570. doi:10.1068/p261555. PMID 9616483. S2CID 39634780.
  73. ^ Lewis, Michael; Edmonds, Andrew (2005). "Searching for faces in scrambled scenes". Visual Cognition. 12 (7): 1309–1336. doi:10.1080/13506280444000535. S2CID 144115983.
  74. ^ Nelson, C. A. (2001). "The development and neural bases of face recognition". Infant and Child Development. 10 (1–2): 3–18. CiteSeerX 10.1.1.130.8912. doi:10.1002/icd.239.
  75. ^ Hershler, O.; Hochstein, S. (2005). "At first sight: A high-level pop out effect for faces". Vision Research. 45 (13): 1707–1724. doi:10.1016/j.visres.2004.12.021. PMID 15792845.
  76. ^ Hershler, O.; Golan, T.; Bentin, S.; Hochstein, S. (2010). "The wide window of face detection". Journal of Vision. 10 (10): 21. doi:10.1167/10.10.21. PMC 2981506. PMID 20884486.
  77. ^ Simpson, E. A., Husband, H. L., Yee, K., Fullerton, A., & Jakobsen, K. V. (2014). Visual Search Efficiency Is Greater for Human Faces Compared to Animal Faces.
  78. ^ VanRullen, R (2006). "On second glance: Still no high-level pop-out effect for faces" (PDF). Vision Research. 46 (18): 3017–3027. doi:10.1016/j.visres.2005.07.009. PMID 16125749. S2CID 1180752.
  79. ^ Hershler, O.; Hochstein, S. (2006). "With a careful look: Still no low-level confound to face pop-out". Vision Research. 46 (18): 3028–3035. doi:10.1016/j.visres.2006.03.023. PMID 16698058.
  80. ^ Golan, T.; Bentin, S.; DeGutis, J. M.; Robertson, L. C.; Harel, A. (2014). "Association and dissociation between detection and discrimination of objects of expertise: evidence from visual search". Attention, Perception, & Psychophysics. 76 (2): 391–406. doi:10.3758/s13414-013-0562-6. PMID 24338355. S2CID 1639650.
  81. ^ a b Bora, Emre; Velakoulis, Dennis; Walterfang, Mark (2016-07-01). "Meta-Analysis of Facial Emotion Recognition in Behavioral Variant Frontotemporal Dementia Comparison With Alzheimer Disease and Healthy Controls". Journal of Geriatric Psychiatry and Neurology. 29 (4): 205–211. doi:10.1177/0891988716640375. ISSN 0891-9887. PMID 27056068. S2CID 36274369.
  82. ^ Kaspar, K. (2016). Culture, group membership, and face recognition. Commentary: Will you remember me? Cultural differences in own-group face recognition biases. Frontiers in Psychology, 7.
  83. ^ Plude, D. J.; J. A. Doussard-Roosevelt (1989). "Aging, selective attention, and feature integration". Psychology and Aging. 4 (1): 98–105. doi:10.1037/0882-7974.4.1.98. PMID 2803617.
  84. ^ Akhtar, N. (1990). "Peripheral vision in young children: Implications for the study of visual attention". The development of attention: Research and theory. Elsevier. pp. 245–262. ISBN 9780080867236. Retrieved 2012-11-19.
  85. ^ Miller, L. K. (1973). "Developmental differences in the field of view during covert and overt search". Child Development. 44 (2): 247–252. doi:10.1111/j.1467-8624.1973.tb02147.x. JSTOR 1128043. PMID 4705552.
  86. ^ Enns, J. T.; D. A. Brodeur (1989). "A developmental study of covert orienting to peripheral visual cues". Journal of Experimental Child Psychology. 48 (2): 171–189. doi:10.1016/0022-0965(89)90001-5. PMID 2794852.
  87. ^ Day, M. C. (1978). "Visual search by children: The effect of background variation and the use of visual cues". Journal of Experimental Child Psychology. 25 (1): 1–16. doi:10.1016/0022-0965(78)90034-6. PMID 641439.
  88. ^ Harpur, L. L.; C. T. Scialfa; D. M. Thomas (1995). "Age differences in feature search as a function of exposure duration". Experimental Aging Research. 21 (1): 1–15. doi:10.1080/03610739508254264. PMID 7744167.
  89. ^ Hartley, A. A.; J. M. Kieley; E. H. Slabach (1990). "Age differences and similarities in the effects of cues and prompts". Journal of Experimental Psychology: Human Perception and Performance. 16 (3): 523–537. doi:10.1037/0096-1523.16.3.523. PMID 2144568.
  90. ^ Connelly, S. L.; L. Hasher (1993). "Aging and the inhibition of spatial location". Journal of Experimental Psychology: Human Perception and Performance. 19 (6): 1238–1250. doi:10.1037/0096-1523.19.6.1238. PMID 8294889.
  91. ^ Rabbitt, P. (1965). "An age-decrement in the ability to ignore irrelevant information". Journal of Gerontology. 20 (2): 233–238. doi:10.1093/geronj/20.2.233. PMID 14284802.
  92. ^ Lorenzo-López, L.; E. Amenedo; R. D. Pascual-Marqui; F. Cadaveira (2008). "Neural correlates of age-related visual search decline: a combined ERP and sLORETA study". NeuroImage. 41 (2): 511–524. doi:10.1016/j.neuroimage.2008.02.041. PMID 18395470. S2CID 10314546. Retrieved 2012-11-19.
  93. ^ a b c d Tales, A.; S. R. Butler; J. Fossey; I. D. Gilchrist; R. W. Jones; T. Troscianko (2002). "Visual search in Alzheimer's disease: a deficiency in processing conjunctions of features". Neuropsychologia. 40 (12): 1849–1857. CiteSeerX 10.1.1.538.4618. doi:10.1016/S0028-3932(02)00073-8. PMID 12207983. S2CID 16310213.
  94. ^ Parasuraman, R.; P. M. Greenwood; G. E. Alexander (2000). "Alzheimer disease constricts the dynamic range of spatial attention in visual search" (PDF). Neuropsychologia. 38 (8): 1126–1135. doi:10.1016/s0028-3932(00)00024-5. PMID 10838147. S2CID 28425852. Retrieved 2012-11-19.
  95. ^ Mendez, M. F.; M. M. Cherrier; J. S. Cymerman (1997). "Hemispatial neglect on visual search tasks in Alzheimer's disease". Neuropsychiatry, Neuropsychology & Behavioral Neurology. 10 (3): 203–8. PMID 9297714.
  96. ^ Troscianko, T.; J. Calvert (1993). "Impaired parallel visual search mechanisms in Parkinson's disease: implications for the role of dopamine in visual attention". Clinical Vision Sciences. 8 (3): 281–287.
  97. ^ Weinstein, A.; T. Troscianko; J. Calvert (1997). "Impaired visual search mechanisms in Parkinson's disease (PD): a psychophysical and event-related potentials study". Journal of Psychophysiology. 11: 33–47.
  98. ^ a b O'Riordan, Michelle A.; Kate C. Plaisted; Jon Driver; Simon Baron-Cohen (2001). "Superior visual search in autism". Journal of Experimental Psychology: Human Perception and Performance. 27 (3): 719–730. doi:10.1037/0096-1523.27.3.719. ISSN 1939-1277. PMID 11424657.
  99. ^ a b Remington, Anna M; John G Swettenham; Nilli Lavie (May 2012). "Lightening the load: perceptual load impairs visual detection in typical adults but not in autism". Journal of Abnormal Psychology. 121 (2): 544–551. doi:10.1037/a0027670. ISSN 1939-1846. PMC 3357114. PMID 22428792.
  100. ^ Remington, Anna; John Swettenham; Ruth Campbell; Mike Coleman (2009-11-01). "Selective Attention and Perceptual Load in Autism Spectrum Disorder" (PDF). Psychological Science. 20 (11): 1388–1393. doi:10.1111/j.1467-9280.2009.02454.x. ISSN 0956-7976. PMID 19843262. S2CID 17119998.
  101. ^ Plaisted, Kate; Michelle O'Riordan; Simon Baron-Cohen (1998). "Enhanced Visual Search for a Conjunctive Target in Autism: A Research Note". Journal of Child Psychology and Psychiatry. 39 (5): 777–783. CiteSeerX 10.1.1.464.6677. doi:10.1111/1469-7610.00376. ISSN 1469-7610. PMID 9690940.
  102. ^ Keehn, Brandon; Laurie Brenner; Erica Palmer; Alan J. Lincoln; Ralph-Axel Müller (2008). "Functional brain organization for visual search in ASD". Journal of the International Neuropsychological Society. 14 (6): 990–1003. doi:10.1017/S1355617708081356. PMID 18954479.
  103. ^ Pieters, R.; Warlop, L. (1999). "Visual attention during brand choice: the impact of time pressure and task motivation" (PDF). International Journal of Research in Marketing. 16: 1–16. doi:10.1016/s0167-8116(98)00022-6.
  104. ^ Janiszewski, C. (1998). "The Influence of Display Characteristics on Visual Exploratory Search Behavior". Journal of Consumer Research. 25 (3): 290–301. doi:10.1086/209540. S2CID 33778354.

visual, search, this, article, about, vision, biology, computer, based, information, retrieval, visual, search, engine, content, based, image, retrieval, type, perceptual, task, requiring, attention, that, typically, involves, active, scan, visual, environment. This article is about vision in biology For computer based information retrieval see visual search engine and content based image retrieval Visual search is a type of perceptual task requiring attention that typically involves an active scan of the visual environment for a particular object or feature the target among other objects or features the distractors 1 Visual search can take place with or without eye movements The ability to consciously locate an object or target amongst a complex array of stimuli has been extensively studied over the past 40 years Practical examples of using visual search can be seen in everyday life such as when one is picking out a product on a supermarket shelf when animals are searching for food among piles of leaves when trying to find a friend in a large crowd of people or simply when playing visual search games such as Where s Wally Much previous literature on visual search used reaction time in order to measure the time it takes to detect the target amongst its distractors An example of this could be a green square the target amongst a set of red circles the distractors However reaction time measurements do not always distinguish between the role of attention and other factors a long reaction time might be the result of difficulty directing attention to the target or slowed decision making processes or slowed motor responses after attention is already directed to the target and the target has already been detected Many visual search paradigms have therefore used eye movement as a means to measure the degree of attention given to stimuli 2 3 However eyes can move independently of attention and therefore eye movement measures do not completely capture the role of attention 4 5 Contents 1 Search types 1 1 Feature search 1 2 Conjunction search 1 3 Real world visual search 2 Reaction time slope 3 Visual orienting and attention 4 Theory 4 1 Feature integration theory FIT 4 2 Guided search model 5 Biological basis 6 Evolution 7 Face recognition 8 Considerations 8 1 Ageing 8 2 Alzheimer s disease 8 3 Autism 9 Consumer psychology 10 ReferencesSearch types editFeature search edit nbsp feature based search taskFeature search also known as disjunctive or efficient search 6 is a visual search process that focuses on identifying a previously requested target amongst distractors that differ from the target by a unique visual feature such as color shape orientation or size 7 An example of a feature search task is asking a participant to identify a white square target surrounded by black squares distractors 6 In this type of visual search the distractors are characterized by the same visual features 7 The efficiency of feature search in regards to reaction time RT and accuracy depends on the pop out effect 8 bottom up processing 8 and parallel processing 7 However the efficiency of feature search is unaffected by the number of distractors present 7 The pop out effect is an element of feature search that characterizes the target s ability to stand out from surrounding distractors due to its unique feature 8 Bottom up processing which is the processing of information that depends on input from the environment 8 explains how one utilizes feature detectors to process characteristics of the stimuli and differentiate a target from its distractors 7 This draw of visual attention towards the target due to bottom up processes is known as saliency 9 Lastly parallel processing is the mechanism that then allows one s feature detectors to work simultaneously in identifying the target 7 Conjunction search edit nbsp Conjunctive based search task Conjunction search also known as inefficient or serial search 6 is a visual search process that focuses on identifying a previously requested target surrounded by distractors possessing no distinct features from the target itself 10 An example of a conjunction search task is having a person identify a red X target amongst distractors composed of black Xs same shape and red Os same color 10 Unlike feature search conjunction search involves distractors or groups of distractors that may differ from each other but exhibit at least one common feature with the target 10 The efficiency of conjunction search in regards to reaction time RT and accuracy is dependent on the distractor ratio 10 and the number of distractors present 7 As the distractors represent the differing individual features of the target more equally amongst themselves distractor ratio effect reaction time RT increases and accuracy decreases 10 As the number of distractors present increases the reaction time RT increases and the accuracy decreases 6 However with practice the original reaction time RT restraints of conjunction search tend to show improvement 11 In the early stages of processing conjunction search utilizes bottom up processes to identify pre specified features amongst the stimuli 7 These processes are then overtaken by a more serial process of consciously evaluating the indicated features of the stimuli 7 in order to properly allocate one s focal spatial attention towards the stimulus that most accurately represents the target 12 In many cases top down processing affects conjunction search by eliminating stimuli that are incongruent with one s previous knowledge of the target description which in the end allows for more efficient identification of the target 8 9 An example of the effect of top down processes on a conjunction search task is when searching for a red K among red Cs and black Ks individuals ignore the black letters and focus on the remaining red letters in order to decrease the set size of possible targets and therefore more efficiently identify their target 13 Real world visual search edit In everyday situations people are most commonly searching their visual fields for targets that are familiar to them When it comes to searching for familiar stimuli top down processing allows one to more efficiently identify targets with greater complexity than can be represented in a feature or conjunction search task 8 In a study done to analyze the reverse letter effect which is the idea that identifying the asymmetric letter among symmetric letters is more efficient than its reciprocal researchers concluded that individuals more efficiently recognize an asymmetric letter among symmetric letters due to top down processes 9 Top down processes allowed study participants to access prior knowledge regarding shape recognition of the letter N and quickly eliminate the stimuli that matched their knowledge 9 In the real world one must use prior knowledge everyday in order to accurately and efficiently locate objects such as phones keys etc among a much more complex array of distractors 8 Despite this complexity visual search with complex objects and search for categories of objects such as phone based on prior knowledge appears to rely on the same active scanning processes as conjunction search with less complex contrived laboratory stimuli 14 15 although global statistical information available in real world scenes can also help people locate target objects 16 17 18 While bottom up processes may come into play when identifying objects that are not as familiar to a person overall top down processing highly influences visual searches that occur in everyday life 8 19 20 Familiarity can play especially critical roles when parts of objects are not visible as when objects are partly hidden from view because they are behind other objects Visual information from hidden parts can be recalled from long term memory and used to facilitate search for familiar objects 21 22 Reaction time slope editIt is also possible to measure the role of attention within visual search experiments by calculating the slope of reaction time over the number of distractors present 23 Generally when high levels of attention are required when looking at a complex array of stimuli conjunction search the slope increases as reaction times increase For simple visual search tasks feature search the slope decreases due to reaction times being fast and requiring less attention 24 However the use of a reaction time slope to measure attention is controversial because non attentional factors can also affect reaction time slope 25 26 27 Visual orienting and attention edit nbsp A photograph that simulates foveationOne obvious way to select visual information is to turn towards it also known as visual orienting This may be a movement of the head and or eyes towards the visual stimulus called a saccade Through a process called foveation the eyes fixate on the object of interest making the image of the visual stimulus fall on the fovea of the eye the central part of the retina with the sharpest visual acuity There are two types of orienting Exogenous orienting is the involuntary and automatic movement that occurs to direct one s visual attention toward a sudden disruption in his peripheral vision field 28 Attention is therefore externally guided by a stimulus resulting in a reflexive saccade Endogenous orienting is the voluntary movement that occurs in order for one to focus visual attention on a goal driven stimulus 28 Thus the focus of attention of the perceiver can be manipulated by the demands of a task A scanning saccade is triggered endogenously for the purpose of exploring the visual environment nbsp A plot of the saccades made while reading text The plot shows the path of eye movements and the size of the circles represents the time spent at any one location Visual search relies primarily on endogenous orienting because participants have the goal to detect the presence or absence of a specific target object in an array of other distracting objects Early research suggested that attention could be covertly without eye movement shifted to peripheral stimuli 29 but later studies found that small saccades microsaccades occur during these tasks and that these eye movements are frequently directed towards the attended locations whether or not there are visible stimuli 30 31 32 These findings indicate that attention plays a critical role in understanding visual search Subsequently competing theories of attention have come to dominate visual search discourse 33 The environment contains a vast amount of information We are limited in the amount of information we are able to process at any one time so it is therefore necessary that we have mechanisms by which extraneous stimuli can be filtered and only relevant information attended to In the study of attention psychologists distinguish between pre attentive and attentional processes 34 Pre attentive processes are evenly distributed across all input signals forming a kind of low level attention Attentional processes are more selective and can only be applied to specific preattentive input A large part of the current debate in visual search theory centres on selective attention and what the visual system is capable of achieving without focal attention 33 Theory editFeature integration theory FIT edit Main article Feature integration theory A popular explanation for the different reaction times of feature and conjunction searches is the feature integration theory FIT introduced by Treisman and Gelade in 1980 This theory proposes that certain visual features are registered early automatically and are coded rapidly in parallel across the visual field using pre attentive processes 35 Experiments show that these features include luminance colour orientation motion direction and velocity as well as some simple aspects of form 36 For example a red X can be quickly found among any number of black Xs and Os because the red X has the discriminative feature of colour and will pop out In contrast this theory also suggests that in order to integrate two or more visual features belonging to the same object a later process involving integration of information from different brain areas is needed and is coded serially using focal attention For example when locating an orange square among blue squares and orange triangles neither the colour feature orange nor the shape feature square is sufficient to locate the search target Instead one must integrate information of both colour and shape to locate the target Evidence that attention and thus later visual processing is needed to integrate two or more features of the same object is shown by the occurrence of illusory conjunctions or when features do not combine correctly For example if a display of a green X and a red O are flashed on a screen so briefly that the later visual process of a serial search with focal attention cannot occur the observer may report seeing a red X and a green O The FIT is a dichotomy because of the distinction between its two stages the preattentive and attentive stages 37 Preattentive processes are those performed in the first stage of the FIT model in which the simplest features of the object are being analyzed such as color size and arrangement The second attentive stage of the model incorporates cross dimensional processing 38 and the actual identification of an object is done and information about the target object is put together This theory has not always been what it is today there have been disagreements and problems with its proposals that have allowed the theory to be amended and altered over time and this criticism and revision has allowed it to become more accurate in its description of visual search 38 There have been disagreements over whether or not there is a clear distinction between feature detection and other searches that use a master map accounting for multiple dimensions in order to search for an object Some psychologists support the idea that feature integration is completely separate from this type of master map search whereas many others have decided that feature integration incorporates this use of a master map in order to locate an object in multiple dimensions 37 The FIT also explains that there is a distinction between the brain s processes that are being used in a parallel versus a focal attention task Chan and Hayward 37 have conducted multiple experiments supporting this idea by demonstrating the role of dimensions in visual search While exploring whether or not focal attention can reduce the costs caused by dimension switching in visual search they explained that the results collected supported the mechanisms of the feature integration theory in comparison to other search based approaches They discovered that single dimensions allow for a much more efficient search regardless of the size of the area being searched but once more dimensions are added it is much more difficult to efficiently search and the bigger the area being searched the longer it takes for one to find the target 37 Guided search model edit A second main function of preattentive processes is to direct focal attention to the most promising information in the visual field 33 There are two ways in which these processes can be used to direct attention bottom up activation which is stimulus driven and top down activation which is user driven In the guided search model by Jeremy Wolfe 39 information from top down and bottom up processing of the stimulus is used to create a ranking of items in order of their attentional priority In a visual search attention will be directed to the item with the highest priority If that item is rejected then attention will move on to the next item and the next and so forth The guided search theory follows that of parallel search processing An activation map is a representation of visual space in which the level of activation at a location reflects the likelihood that the location contains a target This likelihood is based on preattentive featural information of the perceiver According to the guided search model the initial processing of basic features produces an activation map with every item in the visual display having its own level of activation Attention is demanded based on peaks of activation in the activation map in a search for the target 39 Visual search can proceed efficiently or inefficiently During efficient search performance is unaffected by the number of distractor items The reaction time functions are flat and the search is assumed to be a parallel search Thus in the guided search model a search is efficient if the target generates the highest or one of the highest activation peaks For example suppose someone is searching for red horizontal targets Feature processing would activate all red objects and all horizontal objects Attention is then directed to items depending on their level of activation starting with those most activated This explains why search times are longer when distractors share one or more features with the target stimuli In contrast during inefficient search the reaction time to identify the target increases linearly with the number of distractor items present According to the guided search model this is because the peak generated by the target is not one of the highest 39 Biological basis edit nbsp A pseudo color image showing activation of the primary visual cortex during a perceptual task using functional magnetic resonance imaging fMRI During visual search experiments the posterior parietal cortex has elicited much activation during functional magnetic resonance imaging fMRI and electroencephalography EEG experiments for inefficient conjunction search which has also been confirmed through lesion studies Patients with lesions to the posterior parietal cortex show low accuracy and very slow reaction times during a conjunction search task but have intact feature search remaining to the ipsilesional the same side of the body as the lesion side of space 40 41 42 43 Ashbridge Walsh and Cowey in 1997 44 demonstrated that during the application of transcranial magnetic stimulation TMS to the right parietal cortex conjunction search was impaired by 100 milliseconds after stimulus onset This was not found during feature search Nobre Coull Walsh and Frith 2003 45 identified using functional magnetic resonance imaging fMRI that the intraparietal sulcus located in the superior parietal cortex was activated specifically to feature search and the binding of individual perceptual features as opposed to conjunction search Conversely the authors further identify that for conjunction search the superior parietal lobe and the right angular gyrus elicit bilaterally during fMRI experiments nbsp Visual search primarily activates areas of the parietal lobe In contrast Leonards Sunaert Vam Hecke and Orban 2000 46 identified that significant activation is seen during fMRI experiments in the superior frontal sulcus primarily for conjunction search This research hypothesises that activation in this region may in fact reflect working memory for holding and maintaining stimulus information in mind in order to identify the target Furthermore significant frontal activation including the ventrolateral prefrontal cortex bilaterally and the right dorsolateral prefrontal cortex were seen during positron emission tomography for attentional spatial representations during visual search 47 The same regions associated with spatial attention in the parietal cortex coincide with the regions associated with feature search Furthermore the frontal eye field FEF located bilaterally in the prefrontal cortex plays a critical role in saccadic eye movements and the control of visual attention 48 49 50 Moreover research into monkeys and single cell recording found that the superior colliculus is involved in the selection of the target during visual search as well as the initiation of movements 51 Conversely it also suggested that activation in the superior colliculus results from disengaging attention ensuring that the next stimulus can be internally represented The ability to directly attend to a particular stimuli during visual search experiments has been linked to the pulvinar nucleus located in the midbrain while inhibiting attention to unattended stimuli 52 Conversely Bender and Butter 1987 53 found that during testing on monkeys no involvement of the pulvinar nucleus was identified during visual search tasks There is evidence for the V1 Saliency Hypothesis that the primary visual cortex V1 creates a bottom up saliency map to guide attention exogenously 54 55 and this V1 saliency map is read out by the superior colliculus which receives monosynaptic inputs from V1 Evolution editThere is a variety of speculation about the origin and evolution of visual search in humans It has been shown that during visual exploration of complex natural scenes both humans and nonhuman primates make highly stereotyped eye movements 56 Furthermore chimpanzees have demonstrated improved performance in visual searches for upright human or dog faces 57 suggesting that visual search particularly where the target is a face is not peculiar to humans and that it may be a primal trait Research has suggested that effective visual search may have developed as a necessary skill for survival where being adept at detecting threats and identifying food was essential 58 59 nbsp Henri Rousseau Jungle with LionThe importance of evolutionarily relevant threat stimuli was demonstrated in a study by LoBue and DeLoache 2008 in which children and adults were able to detect snakes more rapidly than other targets amongst distractor stimuli 60 However some researchers question whether evolutionarily relevant threat stimuli are detected automatically 61 Face recognition editOver the past few decades there have been vast amounts of research into face recognition specifying that faces endure specialized processing within a region called the fusiform face area FFA located in the mid fusiform gyrus in the temporal lobe 62 Debates are ongoing whether both faces and objects are detected and processed in different systems and whether both have category specific regions for recognition and identification 63 64 Much research to date focuses on the accuracy of the detection and the time taken to detect the face in a complex visual search array When faces are displayed in isolation upright faces are processed faster and more accurately than inverted faces 65 66 67 68 but this effect was observed in non face objects as well 69 When faces are to be detected among inverted or jumbled faces reaction times for intact and upright faces increase as the number of distractors within the array is increased 70 71 72 Hence it is argued that the pop out theory defined in feature search is not applicable in the recognition of faces in such visual search paradigm Conversely the opposite effect has been argued and within a natural environmental scene the pop out effect of the face is significantly shown 73 This could be due to evolutionary developments as the need to be able to identify faces that appear threatening to the individual or group is deemed critical in the survival of the fittest 74 More recently it was found that faces can be efficiently detected in a visual search paradigm if the distracters are non face objects 75 76 77 however it is debated whether this apparent pop out effect is driven by a high level mechanism or by low level confounding features 78 79 Furthermore patients with developmental prosopagnosia who have impaired face identification generally detect faces normally suggesting that visual search for faces is facilitated by mechanisms other than the face identification circuits of the fusiform face area 80 Patients with forms of dementia can also have deficits in facial recognition and the ability to recognize human emotions in the face In a meta analysis of nineteen different studies comparing normal adults with dementia patients in their abilities to recognize facial emotions 81 the patients with frontotemporal dementia were seen to have a lower ability to recognize many different emotions These patients were much less accurate than the control participants and even in comparison with Alzheimer s patients in recognizing negative emotions but were not significantly impaired in recognizing happiness Anger and disgust in particular were the most difficult for the dementia patients to recognize 81 Face recognition is a complex process that is affected by many factors both environmental and individually internal Other aspects to be considered include race and culture and their effects on one s ability to recognize faces 82 Some factors such as the cross race effect can influence one s ability to recognize and remember faces Considerations editAgeing edit Research indicates that performance in conjunctive visual search tasks significantly improves during childhood and declines in later life 83 More specifically young adults have been shown to have faster reaction times on conjunctive visual search tasks than both children and older adults but their reaction times were similar for feature visual search tasks 52 This suggests that there is something about the process of integrating visual features or serial searching that is difficult for children and older adults but not for young adults Studies have suggested numerous mechanisms involved in this difficulty in children including peripheral visual acuity 84 eye movement ability 85 ability of attentional focal movement 86 and the ability to divide visual attention among multiple objects 87 Studies have suggested similar mechanisms in the difficulty for older adults such as age related optical changes that influence peripheral acuity 88 the ability to move attention over the visual field 89 the ability to disengage attention 90 and the ability to ignore distractors 91 A study by Lorenzo Lopez et al 2008 provides neurological evidence for the fact that older adults have slower reaction times during conjunctive searches compared to young adults Event related potentials ERPs showed longer latencies and lower amplitudes in older subjects than young adults at the P3 component which is related to activity of the parietal lobes This suggests the involvement of the parietal lobe function with an age related decline in the speed of visual search tasks Results also showed that older adults when compared to young adults had significantly less activity in the anterior cingulate cortex and many limbic and occipitotemporal regions that are involved in performing visual search tasks 92 Alzheimer s disease edit Research has found that people with Alzheimer s disease AD are significantly impaired overall in visual search tasks 93 People with AD manifest enhanced spatial cueing but this benefit is only obtained for cues with high spatial precision 94 Abnormal visual attention may underlie certain visuospatial difficulties in patients with AD People with AD have hypometabolism and neuropathology in the parietal cortex and given the role of parietal function for visual attention patients with AD may have hemispatial neglect which may result in difficulty with disengaging attention in visual search 95 An experiment conducted by Tales et al 2000 93 investigated the ability of patients with AD to perform various types of visual search tasks Their results showed that search rates on pop out tasks were similar for both AD and control groups however people with AD searched significantly slower compared to the control group on a conjunction task One interpretation of these results is that the visual system of AD patients has a problem with feature binding such that it is unable to communicate the different feature descriptions for the stimulus efficiently 93 Binding of features is thought to be mediated by areas in the temporal and parietal cortex and these areas are known to be affected by AD related pathology Another possibility for the impairment of people with AD on conjunction searches is that there may be some damage to general attentional mechanisms in AD and therefore any attention related task will be affected including visual search 93 Tales et al 2000 detected a double dissociation with their experimental results on AD and visual search Earlier work was carried out on patients with Parkinson s disease PD concerning the impairment patients with PD have on visual search tasks 96 97 In those studies evidence was found of impairment in PD patients on the pop out task but no evidence was found on the impairment of the conjunction task As discussed AD patients show the exact opposite of these results normal performance was seen on the pop out task but impairment was found on the conjunction task This double dissociation provides evidence that PD and AD affect the visual pathway in different ways and that the pop out task and the conjunction task are differentially processed within that pathway Autism edit Studies have consistently shown that autistic individuals performed better and with lower reaction times in feature and conjunctive visual search tasks than matched controls without autism 98 99 Several explanations for these observations have been suggested One possibility is that people with autism have enhanced perceptual capacity 99 This means that autistic individuals are able to process larger amounts of perceptual information allowing for superior parallel processing and hence faster target location 100 Second autistic individuals show superior performance in discrimination tasks between similar stimuli and therefore may have an enhanced ability to differentiate between items in the visual search display 101 A third suggestion is that autistic individuals may have stronger top down target excitation processing and stronger distractor inhibition processing than controls 98 Keehn et al 2008 used an event related functional magnetic resonance imaging design to study the neurofunctional correlates of visual search in autistic children and matched controls of typically developing children 102 Autistic children showed superior search efficiency and increased neural activation patterns in the frontal parietal and occipital lobes when compared to the typically developing children Thus autistic individuals superior performance on visual search tasks may be due to enhanced discrimination of items on the display which is associated with occipital activity and increased top down shifts of visual attention which is associated with the frontal and parietal areas Consumer psychology editIn the past decade there has been extensive research into how companies can maximise sales using psychological techniques derived from visual search to determine how products should be positioned on shelves Pieters and Warlop 1999 103 used eye tracking devices to assess saccades and fixations of consumers while they visually scanned searched an array of products on a supermarket shelf Their research suggests that consumers specifically direct their attention to products with eye catching properties such as shape colour or brand name This effect is due to a pressured visual search where eye movements accelerate and saccades minimise thus resulting in the consumer s quickly choosing a product with a pop out effect This study suggests that efficient search is primarily used concluding that consumers do not focus on items that share very similar features The more distinct or maximally visually different a product is from surrounding products the more likely the consumer is to notice it Janiszewski 1998 104 discussed two types of consumer search One search type is goal directed search taking place when somebody uses stored knowledge of the product in order to make a purchase choice The second is exploratory search This occurs when the consumer has minimal previous knowledge about how to choose a product It was found that for exploratory search individuals would pay less attention to products that were placed in visually competitive areas such as the middle of the shelf at an optimal viewing height This was primarily due to the competition in attention meaning that less information was maintained in visual working memory for these products References edit Treisman AM Gelade G January 1980 A feature integration theory of attention Cogn Psychol 12 1 97 136 doi 10 1016 0010 0285 80 90005 5 PMID 7351125 S2CID 353246 Shelga B M Riggio L Rizzolatti G 1994 Orienting of attention and eye movements Experimental Brain Research 98 3 507 522 doi 10 1007 bf00233988 PMID 8056071 S2CID 2280537 Hoffman J E B Subramaniam 1995 The role of visual attention in saccadic eye movements Perception and Psychophysics 57 6 787 795 doi 10 3758 bf03206794 PMID 7651803 Klein R Farrell M 1989 Search performance without eye movements Percept Psychophys 46 5 476 82 doi 10 3758 BF03210863 PMID 2813033 Murthy A Thompson KG Schall JD 2001 Dynamic dissociation of visual selection from saccade programming in frontal eye field J Neurophysiol 86 5 2634 7 doi 10 1152 jn 2001 86 5 2634 PMID 11698551 S2CID 653798 a b c d Treisman A M Gelade G 1980 A feature integration theory of attention Cognitive Psychology 12 1 97 136 doi 10 1016 0010 0285 80 90005 5 PMID 7351125 S2CID 353246 a b c d e f g h i McElree B Carrasco M December 1999 The temporal dynamics of visual search evidence for parallel processing in feature and conjunction searches Journal of Experimental Psychology Human Perception and Performance 25 6 1517 39 doi 10 1037 0096 1523 25 6 1517 PMC 3313830 PMID 10641310 a b c d e f g h Radvansky Gabriel A Ashcraft Mark H 2016 Cognition 6 ed Pearson Education Inc via online a href Template Cite book html title Template Cite book cite book a CS1 maint multiple names authors list link a b c d Zhaoping L Frith U August 2011 A clash of bottom up and top down processes in visual search the reversed letter effect revisited Journal of Experimental Psychology Human Perception and Performance 37 4 997 1006 doi 10 1037 a0023099 PMID 21574744 S2CID 12986008 a b c d e Shen J Reingold EM Pomplun M June 2003 Guidance of eye movements during conjunctive visual search the distractor ratio effect Canadian Journal of Experimental Psychology 57 2 76 96 CiteSeerX 10 1 1 59 251 doi 10 1037 h0087415 PMID 12822838 Reavis EA Frank SM Greenlee MW Tse PU June 2016 Neural correlates of context dependent feature conjunction learning in visual search tasks Human Brain Mapping 37 6 2319 30 doi 10 1002 hbm 23176 PMC 6867346 PMID 26970441 S2CID 205849752 Eimer M Grubert A October 2014 The gradual emergence of spatially selective target processing in visual search From feature specific to object based attentional control PDF Journal of Experimental Psychology Human Perception and Performance 40 5 1819 31 doi 10 1037 a0037387 PMID 24999612 Wolfe J M 2014 Approaches to visual search feature integration theory and guided search The Oxford handbook of attention Oxford Oxford University Press pp 11 50 ISBN 9780199675111 Alexander Robert G Zelinsky Gregory J 2012 Effects of part based similarity on visual search The Frankenbear experiment Vision Research 54 20 30 doi 10 1016 j visres 2011 12 004 PMC 3345177 PMID 22227607 Alexander Robert G Zelinsky Gregory J 2011 Visual Similarity Effects in Categorical Search Journal of Vision 11 8 9 doi 10 1167 11 8 9 PMC 8409006 PMID 21757505 Rosenholtz Ruth Huang Jie Raj A Balas Benjamin J Ilie Livia 2012 A summary statistic representation in peripheral vision explains visual search Journal of Vision 12 4 14 doi 10 1167 12 4 14 PMC 4032502 PMID 22523401 Alexander Robert G Schmidt Joseph Zelinsky Gregory J 2014 Are summary statistics enough Evidence for the importance of shape in guiding visual search Visual Cognition 22 3 4 595 609 doi 10 1080 13506285 2014 890989 PMC 4500174 PMID 26180505 Wolfe Jeremy M Vo Melissa L H Evans Karla K Greene Michelle R 2011 Visual search in scenes involves selective and nonselective pathways Trends in Cognitive Sciences 15 2 77 84 doi 10 1016 j tics 2010 12 001 PMC 3035167 PMID 21227734 Siebold Alisha Van Zoest Wieske Donk Mieke 2011 Oculomotor evidence for top down control following the initial saccade PLOS ONE 6 9 e23552 Bibcode 2011PLoSO 623552S doi 10 1371 journal pone 0023552 PMC 3169564 PMID 21931603 Malcolm G L Henderson J M 2010 Combining top down processes to guide eye movements during real world scene search Journal of Vision 10 2 4 1 11 doi 10 1167 10 2 4 PMID 20462305 Alexander Robert G Zelinsky Gregory J 2018 Occluded information is restored at preview but not during visual search J Vis 18 11 4 doi 10 1167 18 11 4 PMC 6181188 PMID 30347091 Plomp G Nakatani C Bonnardel V Leeuwen C v 2004 Amodal completion as reflected by gaze durations Perception 33 10 1185 1200 doi 10 1068 p5342x PMID 15693664 S2CID 35581448 Trick Lana M Enns James T 1998 07 01 Lifespan changes in attention The visual search task Cognitive Development 13 3 369 386 CiteSeerX 10 1 1 522 1907 doi 10 1016 S0885 2014 98 90016 8 Alvarez G A Cavanagh P 2004 02 01 The capacity of visual short term memory is set both by visual information load and by number of objects Psychological Science 15 2 106 111 doi 10 1111 j 0963 7214 2004 01502006 x ISSN 0956 7976 PMID 14738517 S2CID 2286443 Palmer J 1995 Attention in visual search Distinguishing four causes of a set size effect Current Directions in Psychological Science 4 4 118 123 doi 10 1111 1467 8721 ep10772534 S2CID 145161732 Eckstein M P 2011 Visual search A retrospective Journal of Vision 11 5 14 doi 10 1167 11 5 14 PMID 22209816 Algom D Eidels A Hawkins R X D Jefferson B Townsend J T 2015 Features of response times Identification of cognitive mechanisms through mathematical modeling In Busemeyer J Wang Z Townsend J T Eidels A eds The Oxford handbook of computational and mathematical psychology Oxford University Press a b Berger A Henik A Rafal R May 2005 Competition between endogenous and exogenous orienting of visual attention Journal of Experimental Psychology General 134 2 207 21 doi 10 1037 0096 3445 134 2 207 PMID 15869346 S2CID 19087912 Wurtz Robert H Michael E Goldberg David Lee Robinson June 1982 Brain Mechanisms of Visual Attention Scientific American 246 6 124 135 Bibcode 1982SciAm 246f 124W doi 10 1038 scientificamerican0682 124 ISSN 0036 8733 PMID 7100892 Laubrock J Kliegl R Rolfs M Engbert R 2010 When do microsaccades follow spatial attention Attention Perception amp Psychophysics 72 3 683 694 doi 10 3758 APP 72 3 683 PMID 20348575 Martinez Conde S Alexander R G 2019 A gaze bias in the mind s eye Nature Human Behaviour 3 5 424 425 doi 10 1038 s41562 019 0546 1 PMID 31089295 S2CID 71148025 Laubrock Engbert Kliegl 2005 Microsaccade dynamics during covert attention Vision Research 45 6 721 730 doi 10 1016 j visres 2004 09 029 PMID 15639499 a b c Muller Hermann J Krummenacher Joseph 2006 Visual search and selective attention Visual Cognition 14 4 8 389 410 doi 10 1080 13506280500527676 ISSN 1350 6285 S2CID 671170 Neisser Ulric 1967 Cognitive Psychology Meredith Publishing Company ISBN 9781317566182 PsycNET 1967 35031 000 Retrieved 2012 11 17 Treisman A M G Gelade 1980 A feature integration theory of attention Cognitive Psychology 12 1 97 136 doi 10 1016 0010 0285 80 90005 5 PMID 7351125 S2CID 353246 Wolfe J M 1998 What can 1 million trials tell us about visual search Psychological Science 9 1 33 39 CiteSeerX 10 1 1 148 6975 doi 10 1111 1467 9280 00006 S2CID 11042813 a b c d Chan Louis K H Hayward William G 2009 Feature integration theory revisited Dissociating feature detection and attentional guidance in visual search Journal of Experimental Psychology Human Perception and Performance 35 1 119 132 doi 10 1037 0096 1523 35 1 119 PMID 19170475 S2CID 22704624 a b Quinlan Philip T September 2003 Visual feature integration theory Past present and future Psychological Bulletin 129 5 643 673 doi 10 1037 0033 2909 129 5 643 PMID 12956538 S2CID 25206656 a b c Wolfe J M 1994 Guided search 2 0 A revised model of visual search Psychonomic Bulletin amp Review 1 2 202 238 doi 10 3758 bf03200774 PMID 24203471 Aglioti S Smania N Barbieri C Corbetta M 1997 Influence of stimulus salience and attentional demands on visual search patterns in hemispatial neglect Brain and Cognition 34 3 388 403 doi 10 1006 brcg 1997 0915 PMID 9292188 S2CID 13138212 Eglin M Robertson L C Knight R T 1991 Cortical substrates supporting visual search in humans Cerebral Cortex 1 3 262 272 doi 10 1093 cercor 1 3 262 PMID 1822736 S2CID 12364128 Friedman Hill S R Robertson L C Treisman A 1995 Parietal contributions to visual feature binding Evidence from a patient with bilateral lesions Science 269 5225 853 855 Bibcode 1995Sci 269 853F doi 10 1126 science 7638604 PMID 7638604 S2CID 42706447 Ellison A Schindler I Pattison L L Milner A D 2004 An exploration of the role of the superior temporal gyrus in visualsearch and spatial perception using TMS v Brain 127 10 2307 2315 doi 10 1093 brain awh244 PMID 15292055 Ashbridge V Walsh A Cowey D 1997 Temporal aspects of visual search studied by transcranial magnetic stimulation Neuropsychologia 35 8 1121 1131 doi 10 1016 s0028 3932 97 00003 1 PMID 9256377 S2CID 7305220 Nobre A C J T Coull V Walsh C D Frith 2003 Brain activations during visual search contributions of search efficiency versus feature binding NeuroImage 18 1 91 103 doi 10 1006 nimg 2002 1329 PMID 12507447 S2CID 32757191 Leonards U Suneart S Van Hecke P Orban G 2000 Attention mechanisms in visual search An fMRI study Journal of Cognitive Neuroscience 12 61 75 doi 10 1162 089892900564073 PMID 11506648 S2CID 35061939 Nobre A C Sebestyen G N Gitelman D R Frith C D Mesulam M M 2002 Filtering of distractors during visual search studied by positron emission tomography NeuroImage 16 4 968 976 doi 10 1006 nimg 2002 1137 PMID 12202084 S2CID 1702722 a href Template Cite journal html title Template Cite journal cite journal a CS1 maint multiple names authors list link Schall JD 2004 On the role of frontal eye field in guiding attention and saccades Vision Research 44 12 1453 1467 doi 10 1016 j visres 2003 10 025 PMID 15066404 Medical Neurosciences Archived from the original on 2011 11 09 Mustari MJ Ono S Das VE May 2009 Signal processing and distribution in cortical brainstem pathways for smooth pursuit eye movements Ann N Y Acad Sci 1164 1 147 54 Bibcode 2009NYASA1164 147M doi 10 1111 j 1749 6632 2009 03859 x PMC 3057571 PMID 19645893 McPeek R M Keller E L 2002 Saccade target selection in the superior colliculus during a visual search task Journal of Neurophysiology 18 4 2019 2034 doi 10 1152 jn 2002 88 4 2019 PMID 12364525 S2CID 16885851 a href Template Cite journal html title Template Cite journal cite journal a CS1 maint multiple names authors list link a b Trick L M Enns J T 1998 Life span changes in attention The visual search task Cognitive Development 13 3 369 386 CiteSeerX 10 1 1 522 1907 doi 10 1016 s0885 2014 98 90016 8 Bender D B Butter C M 1987 Comparison of the effects of superior colliculus and pulvinar lesions on visual search and tachistoscopic pattern discrimination in monkeys PDF Experimental Brain Research 69 1 140 154 doi 10 1007 bf00247037 hdl 2027 42 46559 PMID 3436384 S2CID 2333254 a href Template Cite journal html title Template Cite journal cite journal a CS1 maint multiple names authors list link Li Zhaoping 2002 01 01 A saliency map in primary visual cortex Trends in Cognitive Sciences 6 1 9 16 doi 10 1016 S1364 6613 00 01817 9 ISSN 1364 6613 PMID 11849610 S2CID 13411369 Yan Yin Zhaoping Li Li Wu 2018 10 09 Bottom up saliency and top down learning in the primary visual cortex of monkeys Proceedings of the National Academy of Sciences 115 41 10499 10504 Bibcode 2018PNAS 11510499Y doi 10 1073 pnas 1803854115 ISSN 0027 8424 PMC 6187116 PMID 30254154 Mazer James A Jack L Gallant 2003 12 18 Goal Related Activity in V4 during Free Viewing Visual Search Evidence for a Ventral Stream Visual Salience Map Neuron 40 6 1241 1250 doi 10 1016 S0896 6273 03 00764 5 ISSN 0896 6273 PMID 14687556 Tomonaga Masaki 2007 01 01 Visual search for orientation of faces by a chimpanzee Pan troglodytes face specific upright superiority and the role of facial configural properties Primates 48 1 1 12 doi 10 1007 s10329 006 0011 4 ISSN 0032 8332 PMID 16969584 S2CID 7313319 Ohman A Mineka S 2001 Fears phobias and preparedness Toward an evolved module of fear and fear learning Psychological Review 108 3 483 522 doi 10 1037 0033 295X 108 3 483 PMID 11488376 S2CID 7920871 Ohman A 1999 Distinguishing unconscious from conscious emotional processes Methodological considerations and theoretical implications In Dalgleish T Powers M J eds Handbook of cognition and emotion Chichester England Wiley pp 321 352 LoBue Vanessa Judy S DeLoache 2008 03 01 Detecting the Snake in the Grass Attention to Fear Relevant Stimuli by Adults and Young Children Psychological Science 19 3 284 289 doi 10 1111 j 1467 9280 2008 02081 x ISSN 0956 7976 PMID 18315802 S2CID 12776572 Quinlan Philip T 2013 The visual detection of threat A cautionary tale Psychonomic Bulletin amp Review 20 6 1080 1101 doi 10 3758 s13423 013 0421 4 PMID 23504916 Kanwisher Nancy McDermott Josh Chun Marvin M 1997 The fusiform face area a module in human extrastriate cortex specialized for face perception The Journal of Neuroscience 17 11 4302 4311 doi 10 1523 JNEUROSCI 17 11 04302 1997 PMC 6573547 PMID 9151747 Tarr M J Gauthier I 2000 FFA a flexible fusiform area for subordinate level visual processing automatized by expertise Nature Neuroscience 3 8 764 770 doi 10 1038 77666 PMID 10903568 S2CID 8355344 Grill Spector K Knouf N Kanwisher N 2004 The fusiform face area subserves face perception not generic within category identification Nature Neuroscience 7 5 555 562 doi 10 1038 nn1224 PMID 15077112 S2CID 2204107 Valentine T Bruce V 1986 The effects of distinctiveness in recognizing and classifying faces Perception 15 5 525 533 doi 10 1068 p150525 PMID 3588212 S2CID 9641249 Purcell D G Stewart A L 1986 The face detection effect Bulletin of the Psychonomic Society 24 2 118 120 doi 10 3758 bf03330521 Purcell D G Stewart A L 1988 The face detection effect Configuration enhances perception Perception amp Psychophysics 43 4 355 366 doi 10 3758 bf03208806 PMID 3362664 Yovel G Kanwisher N 2005 The neural basis of the behavioural face inversion effect Current Biology 15 24 2256 2262 doi 10 1016 j cub 2005 10 072 PMID 16360687 Purcell D G Stewart A L 1991 The object detection effect Configuration enhances perception Perception amp Psychophysics 50 3 215 224 doi 10 3758 bf03206744 PMID 1754362 Nothdurft H C 1993 Faces and facial expressions do not pop out Perception 22 11 1287 98 doi 10 1068 p221287 PMID 8047415 S2CID 33911653 Kuehn S M Jolicoeur P 1994 Impact of quality of the image orientation and similarity of the stimuli on visual search for faces Perception 23 1 95 122 doi 10 1068 p230095 PMID 7936979 S2CID 20262065 Brown V Huey D Findlay J M 1997 Face detection in peripheral vision do faces pop out Perception 26 12 1555 1570 doi 10 1068 p261555 PMID 9616483 S2CID 39634780 Lewis Michael Edmonds Andrew 2005 Searching for faces in scrambled scenes Visual Cognition 12 7 1309 1336 doi 10 1080 13506280444000535 S2CID 144115983 Nelson C A 2001 The development and neural bases of face recognition Infant and Child Development 10 1 2 3 18 CiteSeerX 10 1 1 130 8912 doi 10 1002 icd 239 Hershler O Hochstein S 2005 At first sight A high level pop out effect for faces Vision Research 45 13 1707 1724 doi 10 1016 j visres 2004 12 021 PMID 15792845 Hershler O Golan T Bentin S Hochstein S 2010 The wide window of face detection Journal of Vision 10 10 21 doi 10 1167 10 10 21 PMC 2981506 PMID 20884486 Simpson E A Husband H L Yee K Fullerton A amp Jakobsen K V 2014 Visual Search Efficiency Is Greater for Human Faces Compared to Animal Faces VanRullen R 2006 On second glance Still no high level pop out effect for faces PDF Vision Research 46 18 3017 3027 doi 10 1016 j visres 2005 07 009 PMID 16125749 S2CID 1180752 Hershler O Hochstein S 2006 With a careful look Still no low level confound to face pop out Vision Research 46 18 3028 3035 doi 10 1016 j visres 2006 03 023 PMID 16698058 Golan T Bentin S DeGutis J M Robertson L C Harel A 2014 Association and dissociation between detection and discrimination of objects of expertise evidence from visual search Attention Perception amp Psychophysics 76 2 391 406 doi 10 3758 s13414 013 0562 6 PMID 24338355 S2CID 1639650 a b Bora Emre Velakoulis Dennis Walterfang Mark 2016 07 01 Meta Analysis of Facial Emotion Recognition in Behavioral Variant Frontotemporal Dementia Comparison With Alzheimer Disease and Healthy Controls Journal of Geriatric Psychiatry and Neurology 29 4 205 211 doi 10 1177 0891988716640375 ISSN 0891 9887 PMID 27056068 S2CID 36274369 Kaspar K 2016 Culture group membership and face recognition Commentary Will you remember me Cultural differences in own group face recognition biases Frontiers in Psychology 7 Plude D J J A Doussard Roosevelt 1989 Aging selective attention and feature integration Psychology and Aging 4 1 98 105 doi 10 1037 0882 7974 4 1 98 PMID 2803617 Akhtar N 1990 Peripheral vision in young children Implications for the study of visual attention The development of attention Research and theory Elsevier pp 245 262 ISBN 9780080867236 Retrieved 2012 11 19 Miller L K 1973 Developmental differences in the field of view during covert and overt search Child Development 44 2 247 252 doi 10 1111 j 1467 8624 1973 tb02147 x JSTOR 1128043 PMID 4705552 Enns J T D A Brodeur 1989 A developmental study of covert orienting to peripheral visual cues Journal of Experimental Child Psychology 48 2 171 189 doi 10 1016 0022 0965 89 90001 5 PMID 2794852 Day M C 1978 Visual search by children The effect of background variation and the use of visual cues Journal of Experimental Child Psychology 25 1 1 16 doi 10 1016 0022 0965 78 90034 6 PMID 641439 Harpur L L C T Scialfa D M Thomas 1995 Age differences in feature search as a function of exposure duration Experimental Aging Research 21 1 1 15 doi 10 1080 03610739508254264 PMID 7744167 Hartley A A J M Kieley E H Slabach 1990 Age differences and similarities in the effects of cues and prompts Journal of Experimental Psychology Human Perception and Performance 16 3 523 537 doi 10 1037 0096 1523 16 3 523 PMID 2144568 Connelly S L L Hasher 1993 Aging and the inhibition of spatial location Journal of Experimental Psychology Human Perception and Performance 19 6 1238 1250 doi 10 1037 0096 1523 19 6 1238 PMID 8294889 Rabbitt P 1965 An age decrement in the ability to ignore irrelevant information Journal of Gerontology 20 2 233 238 doi 10 1093 geronj 20 2 233 PMID 14284802 Lorenzo Lopez L E Amenedo R D Pascual Marqui F Cadaveira 2008 Neural correlates of age related visual search decline a combined ERP and sLORETA study NeuroImage 41 2 511 524 doi 10 1016 j neuroimage 2008 02 041 PMID 18395470 S2CID 10314546 Retrieved 2012 11 19 a b c d Tales A S R Butler J Fossey I D Gilchrist R W Jones T Troscianko 2002 Visual search in Alzheimer s disease a deficiency in processing conjunctions of features Neuropsychologia 40 12 1849 1857 CiteSeerX 10 1 1 538 4618 doi 10 1016 S0028 3932 02 00073 8 PMID 12207983 S2CID 16310213 Parasuraman R P M Greenwood G E Alexander 2000 Alzheimer disease constricts the dynamic range of spatial attention in visual search PDF Neuropsychologia 38 8 1126 1135 doi 10 1016 s0028 3932 00 00024 5 PMID 10838147 S2CID 28425852 Retrieved 2012 11 19 Mendez M F M M Cherrier J S Cymerman 1997 Hemispatial neglect on visual search tasks in Alzheimer s disease Neuropsychiatry Neuropsychology amp Behavioral Neurology 10 3 203 8 PMID 9297714 Troscianko T J Calvert 1993 Impaired parallel visual search mechanisms in Parkinson s disease implications for the role of dopamine in visual attention Clinical Vision Sciences 8 3 281 287 Weinstein A T Troscianko J Calvert 1997 Impaired visual search mechanisms in Parkinson s disease PD a psychophysical and event related potentials study Journal of Psychophysiology 11 33 47 a b O Riordan Michelle A Kate C Plaisted Jon Driver Simon Baron Cohen 2001 Superior visual search in autism Journal of Experimental Psychology Human Perception and Performance 27 3 719 730 doi 10 1037 0096 1523 27 3 719 ISSN 1939 1277 PMID 11424657 a b Remington Anna M John G Swettenham Nilli Lavie May 2012 Lightening the load perceptual load impairs visual detection in typical adults but not in autism Journal of Abnormal Psychology 121 2 544 551 doi 10 1037 a0027670 ISSN 1939 1846 PMC 3357114 PMID 22428792 Remington Anna John Swettenham Ruth Campbell Mike Coleman 2009 11 01 Selective Attention and Perceptual Load in Autism Spectrum Disorder PDF Psychological Science 20 11 1388 1393 doi 10 1111 j 1467 9280 2009 02454 x ISSN 0956 7976 PMID 19843262 S2CID 17119998 Plaisted Kate Michelle O Riordan Simon Baron Cohen 1998 Enhanced Visual Search for a Conjunctive Target in Autism A Research Note Journal of Child Psychology and Psychiatry 39 5 777 783 CiteSeerX 10 1 1 464 6677 doi 10 1111 1469 7610 00376 ISSN 1469 7610 PMID 9690940 Keehn Brandon Laurie Brenner Erica Palmer Alan J Lincoln Ralph Axel Muller 2008 Functional brain organization for visual search in ASD Journal of the International Neuropsychological Society 14 6 990 1003 doi 10 1017 S1355617708081356 PMID 18954479 Pieters R Warlop L 1999 Visual attention during brand choice the impact of time pressure and task motivation PDF International Journal of Research in Marketing 16 1 16 doi 10 1016 s0167 8116 98 00022 6 Janiszewski C 1998 The Influence of Display Characteristics on Visual Exploratory Search Behavior Journal of Consumer Research 25 3 290 301 doi 10 1086 209540 S2CID 33778354 Retrieved from https en wikipedia org w index php title Visual search amp oldid 1181939080, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.