Feature-Based Attention Samples Stimuli Rhythmically
Fecha
2019Autor
Re, Daniele
Inbar, Maya
Richter, Craig G.
Landau, Ayelet N.
Metadatos
Mostrar el registro completo del ítem
Daniele Re, Maya Inbar, Craig G. Richter, Ayelet N. Landau, Feature-Based Attention Samples Stimuli Rhythmically, Current Biology, Volume 29, Issue 4, 2019, Pages 693-699.e4, ISSN 0960-9822, https://doi.org/10.1016/j.cub.2019.01.010.
Resumen
Attention supports the allocation of resources to
relevant locations and objects in a scene. Under
most conditions, several stimuli compete for neural
representation. Attention biases neural representation
toward the response associated with the attended
object [1, 2]. Therefore, an attended stimulus
enjoys a neural response that resembles the
response to that stimulus in isolation. Factors that
determine and generate attentional bias have been
researched, ranging from endogenously controlled
processes to exogenous capture of attention [1–4].
Recent studies investigate the temporal structure
governing attention. When participants monitor a
single location, visual-target detection depends on
the phase of an 8-Hz brain rhythm [5, 6]. When
two locations are monitored, performance fluctuates
at 4 Hz for each location [7, 8]. The hypothesis
is that 4-Hz sampling for two locations may reflect
a common sampler that operates at 8 Hz globally,
which is divided between relevant locations
[5–7, 9]. The present study targets two properties
of this phenomenon, called rhythmic-attentional
sampling: first, sampling is typically described for
selection over different locations. We examined
whether rhythmic sampling is limited to selection
over space or whether it extends to feature-based
attention. Second, we examined whether sampling
at 4 Hz results from the division of an 8-Hz rhythm
over two objects.We found that two overlapping objects
defined by features are sampled at 4 Hz per
object. In addition, performance on a single object
fluctuated at 8 Hz. Rhythmic sampling of features
did not result from temporal structure in eye movements.