====== Publication List ====== Fox, K. J., Birman, D. and Gardner, J. L. (2023) Gain, not concomitant changes in spatial receptive field properties, improves task performance in a neural network attention model //eLife// [[ https://doi.org/10.7554/eLife.78392|DOI]] 12:e78392 ++Abstract| \\ \\ Attention allows us to focus sensory processing on behaviorally relevant aspects of the visual world. One potential mechanism of attention is a change in the gain of sensory responses. However, changing gain at early stages could have multiple downstream consequences for visual processing. Which, if any, of these effects can account for the benefits of attention for detection and discrimination? Using a model of primate visual cortex we document how a Gaussian-shaped gain modulation results in changes to spatial tuning properties. Forcing the model to use only these changes failed to produce any benefit in task performance. Instead, we found that gain alone was both necessary and sufficient to explain category detection and discrimination during attention. Our results show how gain can give rise to changes in receptive fields which are not necessary for enhancing task performance.++[[https://www.biorxiv.org/content/10.1101/2022.03.04.483026v2.full.pdf|pdf]] Himmelberg, M. M., Gardner, J. L. and Winawer, J. (2022) What has vision science taught us about functional MRI? //Neuroimage// 262:119536. [[https://doi.org/10.1016/j.neuroimage.2022.119536|DOI]] ++Abstract| \\ \\ In the domain of human neuroimaging, much attention has been paid to the question of whether and how the development of functional magnetic resonance imaging (fMRI) has advanced our scientific knowledge of the human brain. However, the opposite question is also important; how has our knowledge of the visual system advanced our understanding of fMRI? Here, we discuss how and why scientific knowledge about the human and animal visual system has been used to answer fundamental questions about fMRI as a brain measurement tool and how these answers have contributed to scientific discoveries beyond vision science.++{{:reprints:1-s2.0-s1053811922006516-main.pdf|pdf}} Jagadeesh, A. V., and Gardner, J. L. (2022) Texture-like representation of objects in human visual cortex //Proceedings of the National Academy of Sciences// 119 (17) e2115302119. [[https://doi.org/10.1073/pnas.2115302119|DOI]] ++Abstract| \\ \\ Task-optimized convolutional neural networks (CNNs) show striking similarities to the ventral visual stream. However, human-imperceptible image perturbations can cause a CNN to make incorrect predictions. Here we provide insight into this brittleness by investigating the representations of models that are either robust or not robust to image perturbations. Theory suggests that the robustness of a system to these perturbations could be related to the power law exponent of the eigenspectrum of its set of neural responses, where power law exponents closer to and larger than one would indicate a system that is less susceptible to input perturbations. We show that neural responses in mouse and macaque primary visual cortex (V1) obey the predictions of this theory, where their eigenspectra have power law exponents of at least one. We also find that the eigenspectra of model representations decay slowly relative to those observed in neurophysiology and that robust models have eigenspectra that decay slightly faster and have higher power law exponents than those of non-robust models. The slow decay of the eigenspectra suggests that substantial variance in the model responses is related to the encoding of fine stimulus features. We therefore investigated the spatial frequency tuning of artificial neurons and found that a large proportion of them preferred high spatial frequencies and that robust models had preferred spatial frequency distributions more aligned with the measured spatial frequency distribution of macaque V1 cells. Furthermore, robust models were quantitatively better models of V1 than non-robust models. Our results are consistent with other findings that there is a misalignment between human and machine perception. They also suggest that it may be useful to penalize slow-decaying eigenspectra or to bias models to extract features of lower spatial frequencies during task-optimization in order to improve robustness and V1 neural response predictivity.++{{:reprints:pnas.2115302119.pdf|pdf}} Kong, N. C. L., Margalit, E., Gardner, J. L., and Norcia, A. M. (2022) Increasing neural network robustness improves match to macaque V1 eigenspectrum, spatial frequency preference and predictivity //PLoS Computational Biology// 18:e1009739. [[https://doi.org/10.1371/journal.pcbi.10009739|DOI]] ++Abstract| \\ \\ Task-optimized convolutional neural networks (CNNs) show striking similarities to the ventral visual stream. However, human-imperceptible image perturbations can cause a CNN to make incorrect predictions. Here we provide insight into this brittleness by investigating the representations of models that are either robust or not robust to image perturbations. Theory suggests that the robustness of a system to these perturbations could be related to the power law exponent of the eigenspectrum of its set of neural responses, where power law exponents closer to and larger than one would indicate a system that is less susceptible to input perturbations. We show that neural responses in mouse and macaque primary visual cortex (V1) obey the predictions of this theory, where their eigenspectra have power law exponents of at least one. We also find that the eigenspectra of model representations decay slowly relative to those observed in neurophysiology and that robust models have eigenspectra that decay slightly faster and have higher power law exponents than those of non-robust models. The slow decay of the eigenspectra suggests that substantial variance in the model responses is related to the encoding of fine stimulus features. We therefore investigated the spatial frequency tuning of artificial neurons and found that a large proportion of them preferred high spatial frequencies and that robust models had preferred spatial frequency distributions more aligned with the measured spatial frequency distribution of macaque V1 cells. Furthermore, robust models were quantitatively better models of V1 than non-robust models. Our results are consistent with other findings that there is a misalignment between human and machine perception. They also suggest that it may be useful to penalize slow-decaying eigenspectra or to bias models to extract features of lower spatial frequencies during task-optimization in order to improve robustness and V1 neural response predictivity.++{{:reprints:journal.pcbi.1009739.pdf|pdf}} Gardner, J. L., and Merriam, E.M. (2021) Population models, not analyses, of human neuroscience measurements //Annual Review of Vision Science// 7:1-31. [[https://doi.org/10.1146/annurev-vision-093019-111124|DOI]] ++Abstract| \\ \\ Selectivity for many basic properties of visual stimuli, such as orientation, is thought to be organized at the scale of cortical columns, making it difficult or impossible to measure directly with noninvasive human neuroscience measurement. However, computational analyses of neuroimaging data have shown that selectivity for orientation can be recovered by considering the pattern of response across a region of cortex. This suggests that computational analyses can reveal representation encoded at a finer spatial scale than is implied by the spatial resolution limits of measurement techniques. This potentially opens up the possibility to study a much wider range of neural phenomena that are otherwise inaccessible through noninvasive measurement. However, as we review in this article, a large body of evidence suggests an alternative hypothesis to this superresolution account: that orientation information is available at the spatial scale of cortical maps and thus easily measurable at the spatial resolution of standard techniques. In fact, a population model shows that this orientation information need not even come from single-unit selectivity for orientation tuning, but instead can result from population selectivity for spatial frequency. Thus, a categorical error of interpretation can result whereby orientation selectivity can be confused with spatial frequency selectivity. This is similarly problematic for the interpretation of results from numerous studies of more complex representations and cognitive functions that have built upon the computational techniques used to reveal stimulus orientation. We suggest in this review that these interpretational ambiguities can be avoided by treating computational analyses as models of the neural processes that give rise to measurement. Building upon the modeling tradition in vision science using considerations of whether population models meet a set of core criteria is important for creating the foundation for a cumulative and replicable approach to making valid inferences from human neuroscience measurements.++{{:reprints:annurev-vision-093019-111124.pdf|pdf}} Lin, Y., Zhou, X., Naya, Y., Gardner, J. L., and Sun, P. (2021) Voxel-wise linearity analysis of increments and decrements in BOLD responses in human visual cortex using a contrast adaptation paradigm //Frontiers in Human Neuroscience// 15:541314.[[https://doi.org/10.3389/fnhum.2021.541314|DOI]] ++Abstract| \\ \\ The linearity of BOLD responses is a fundamental presumption in most analysis procedures for BOLD fMRI studies. Previous studies have examined the linearity of BOLD signal increments, but less is known about the linearity of BOLD signal decrements. The present study assessed the linearity of both BOLD signal increments and decrements in the human primary visual cortex using a contrast adaptation paradigm. Results showed that both BOLD signal increments and decrements kept linearity to long stimuli (e.g., 3 s, 6 s), yet, deviated from linearity to transient stimuli (e.g., 1 s). Furthermore, a voxel-wise analysis showed that the deviation patterns were different for BOLD signal increments and decrements: while the BOLD signal increments demonstrated a consistent overestimation pattern, the patterns for BOLD signal decrements varied from overestimation to underestimation. Our results suggested that corrections to deviations from linearity of transient responses should consider the different effects of BOLD signal increments and decrements.++{{:reprints:fnhum-15-541314.pdf|pdf}} Lin W-H., Gardner J. L., Wu S-W. (2020) Context effects on probability estimation. //PLoS Biology// 18:e3000634.[[https://doi.org/10.1371/journal.pbio.3000634|DOI]] ++ Abstract| \\ \\ Many decisions rely on how we evaluate potential outcomes and estimate their corresponding probabilities of occurrence. Outcome evaluation is subjective because it requires consulting internal preferences and is sensitive to context. In contrast, probability estimation requires extracting statistics from the environment and therefore imposes unique challenges to the decision maker. Here, we show that probability estimation, like outcome evaluation, is subject to context effects that bias probability estimates away from other events present in the same context. However, unlike valuation, these context effects appeared to be scaled by estimated uncertainty, which is largest at intermediate probabilities. Blood-oxygen-level-dependent (BOLD) imaging showed that patterns of multivoxel activity in the dorsal anterior cingulate cortex (dACC), ventromedial prefrontal cortex (VMPFC), and intraparietal sulcus (IPS) predicted individual differences in context effects on probability estimates. These results establish VMPFC as the neurocomputational substrate shared between valuation and probability estimation and highlight the additional involvement of dACC and IPS that can be uniquely attributed to probability estimation. Because probability estimation is a required component of computational accounts from sensory inference to higher cognition, the context effects found here may affect a wide array of cognitive computations.++{{:reprints:lin_gardner_wu_2020.pdf|pdf}} Riesen, G., Norcia, A. M. and Gardner, J. L. (2019) Humans perceive binocular rivalry and fusion in a tristable dynamic state. //The Journal of Neuroscience// 39(43):8527-8537 [[https://doi.org/10.1523/JNEUROSCI.0713-19.2019|DOI]] ++Abstract| \\ \\ Human vision combines inputs from the two eyes into one percept. Small differences ‘fuse’ together, while larger differences are seen ‘rivalrously’ from one eye at a time. These outcomes are typically treated as mutually exclusive processes, with paradigms targeting one or the other and fusion being unreported in most rivalry studies. Is fusion truly a default, stable state that only breaks into rivalry for non-fusible stimuli? Or are monocular and fused percepts three sub-states of one dynamical system? To determine whether fusion and rivalry are separate processes, we measured human perception of Gabor patches with a range of inter-ocular orientation disparities. Observers (10 female, 5 male) reported rivalrous, fused and uncertain percepts over time. We found a dynamic “tristable” zone spanning from ~25-35 degrees of orientation disparity where fused, left- or right-eye dominant percepts could all occur. The temporal characteristics of fusion and non-fusion periods during tristability matched other bistable processes. We tested statistical models with fusion as a higher-level bistable process alternating with rivalry against our findings. None of these fit our data, but a simple bistable model extended to have three states reproduced many of our observations. We conclude that rivalry and fusion are multistable sub-states capable of direct competition, rather than separate bistable processes.++{{:reprints:riesen_et_al_jn_2019.pdf|pdf}} Birman, D. and Gardner, J. L. (2019) A flexible readout mechanism of human sensory representations. //Nature Communications// 10:3500 [[https://doi.org/10.1038/s41467-019-11448-7|DOI]] ++Abstract| \\ \\ Attention can both enhance and suppress cortical sensory representations. However, changing sensory representations can also be detrimental to behavior. Behavioral consequences can be avoided by flexibly changing sensory readout, while leaving the representations unchanged. Here, we asked human observers to attend to and report about either one of two features which control the visibility of motion while making concurrent measurements of cortical activity with BOLD imaging (fMRI). Extending a well-established linking model to account for the relationship between these measurements, we found that changes in sensory representation during directed attention were insufficient to explain perceptual reports. A flexible downstream readout was also necessary to best explain our data. Such a model implies that observers should be able to recover information about ignored features, a prediction which we confirmed behaviorally. Thus, flexible readout is a critical component of the cortical implementation of human adaptive behavior.++{{:reprints:nat_commun_2019_birman.pdf|pdf}} Fukuda H., Ma N., Suzuki S., Harasawa N., Ueno K., Gardner J.L., Ichinohe N., Haruno M., Cheng K., Nakahara H. (2019) Computing social value conversion in the human brain. //The Journal of Neuroscience// 39(26):5153-72 [[http://doi.org/10.1523/JNEUROSCI.3117-18.2019|DOI]] ++Abstract| \\ \\ Social signals play powerful roles in shaping self-oriented reward valuation and decision making. These signals activate social and valuation/decision areas, but the core computation for their integration into the self-oriented decision machinery remains unclear. Here, we study how a fundamental social signal, social value (others' reward value), is converted into self-oriented decision making in the human brain. Using behavioral analysis, modeling, and neuroimaging, we show three-stage processing of social value conversion from the offer to the effective value and then to the final decision value. First, a value of others' bonus on offer, called offered value, was encoded uniquely in the right temporoparietal junction (rTPJ) and also in the left dorsolateral prefrontal cortex (ldlPFC), which is commonly activated by offered self-bonus value. The effective value, an intermediate value representing the effective influence of the offer on the decision, was represented in the right anterior insula (rAI), and the final decision value was encoded in the medial prefrontal cortex (mPFC). Second, using psychophysiological interaction and dynamic causal modeling analyses, we demonstrated three-stage feedforward processing from the rTPJ and ldPFC to the rAI and then from rAI to the mPFC. Further, we showed that these characteristics of social conversion underlie distinct sociobehavioral phenotypes. We demonstrate that the variability in the conversion underlies the difference between prosocial and selfish subjects, as seen from the differential strength of the rAI and ldlPFC coupling to the mPFC responses, respectively. Together, these findings identified fundamental neural computation processes for social value conversion underlying complex social decision making behaviors.++{{:reprints:fukuda_et_al_2019.pdf|pdf}} Gardner, J. L. and Liu, T. (2019) Inverted encoding models reconstruct an arbitrary model response, not the stimulus. //eNeuro// 6(2) e0363-18.2019 1–11 [[https://doi.org/10.1523/ENEURO.0363-18.2019|DOI]] ++Abstract| \\ \\ Probing how large populations of neurons represent stimuli is key to understanding sensory representations as many stimulus characteristics can only be discerned from population activity and not from individual single-units. Recently, inverted encoding models have been used to produce channel response functions from large spatial-scale measurements of human brain activity that are reminiscent of single-unit tuning functions and have been proposed to assay “population-level stimulus representations” (Sprague et al., 2018a). However, these channel response functions do not assay population tuning. We show by derivation that the channel response function is only determined up to an invertible linear transform. Thus, these channel response functions are arbitrary, one of an infinite family and therefore not a unique description of population representation. Indeed, simulations demonstrate that bimodal, even random, channel basis functions can account perfectly well for population responses without any underlying neural response units that are so tuned. However, the approach can be salvaged by extending it to reconstruct the stimulus, not the assumed model. We show that when this is done, even using bimodal and random channel basis functions, a unimodal function peaking at the appropriate value of the stimulus is recovered which can be interpreted as a measure of population selectivity. More precisely, the recovered function signifies how likely any value of the stimulus is, given the observed population response. Whether an analysis is recovering the hypothetical responses of an arbitrary model rather than assessing the selectivity of population representations is not an issue unique to the inverted encoding model and human neuroscience, but a general problem that must be confronted as more complex analyses intervene between measurement of population activity and presentation of data.++{{:reprints:enu002192895p.pdf|pdf}} Gardner, J. L. (2019) Optimality and heuristics in perceptual neuroscience. //Nature Neuroscience// 22:514-523 [[https://doi.org/10.1038/s41593-019-0340-4|DOI]] ++Abstract| \\ \\ The foundation for modern understanding of how we make perceptual decisions about what it is that we see or where to look comes from considering the optimal way to perform these behaviors. While statistical computation is useful for deriving the optimal solution to a perceptual problem, optimality requires perfect knowledge of priors and often complex computation. Accumulating evidence, however, suggests that optimal perceptual goals can be achieved or approximated more simply by human observers using heuristic approaches. Perceptual neuroscientists captivated by optimal explanations of sensory behaviors will fail in their search for the neural circuits and cortical processes that implement an optimal computation whenever that behavior is actually achieved through heuristics. This article provides a cross-disciplinary review of decision-making with the aim of building perceptual theory that uses optimality to set the computational goals for perceptual behavior, but through consideration of ecological, computational and energetic constraints incorporates how these optimal goals can be achieved through heuristic approximation.++{{:reprints:gardner-2019-nature-neuroscience.pdf|pdf}} Birman, D., and Gardner, J. L. (2018) A quantitative framework for motion visibility in human cortex. //Journal of Neurophysiology// 120:1824-1839. [[https://www.physiology.org/doi/abs/10.1152/jn.00433.2018|DOI]] [[https://osf.io/s7j9p/|DATA]]++Abstract| \\ \\ Despite the central use of motion visibility to reveal the neural basis of perception, perceptual decision making, and sensory inference there exists no comprehensive quantitative framework establishing how motion visibility parameters modulate human cortical response. Random-dot motion stimuli can be made less visible by reducing image contrast or motion coherence, or by shortening the stimulus duration. Because each of these manipulations modulates the strength of sensory neural responses they have all been extensively used to reveal cognitive and other non-sensory phenomenon such as the influence of priors, attention, and choice-history biases. However, each of these manipulations is thought to influence response in different ways across different cortical regions and a comprehensive study is required to interpret this literature. Here, human participants observed random-dot stimuli varying across a large range of contrast, coherence, and stimulus durations as we measured blood-oxygen-level dependent responses. We developed a framework for modeling these responses which quantifies their functional form and sensitivity across areas. Our framework demonstrates the sensitivity of all visual areas to each parameter, with early visual areas V1-V4 showing more parametric sensitivity to changes in contrast and V3A and MT to coherence. Our results suggest that while motion contrast, coherence, and duration share cortical representation, they are encoded with distinct functional forms and sensitivity. Thus, our quantitative framework serves as a reference for interpretation of the vast perceptual literature manipulating these parameters and shows that different manipulations of visibility will have different effects across human visual cortex and need to be interpreted accordingly.++{{:reprints:birman_jnphys_2018.pdf|pdf}} Dobs, K., Schultz, J., Bulthoff, I., and Gardner, J. L. (2018) Task-dependent enhancement of facial expression and identity representations in human cortex. //Neuroimage// 10:689-702. [[https://doi.org/10.1016/j.neuroimage.2018.02.013|DOI]]++Abstract| \\ \\ What cortical mechanisms allow humans to easily discern the expression or identity of a face? Subjects detected changes in expression or identity of a stream of dynamic faces while we measured BOLD responses from topo-graphically and functionally defined areas throughout the visual hierarchy. Responses in dorsal areas increasedduring the expression task, whereas responses in ventral areas increased during the identity task, consistent with previous studies. Similar to ventral areas, early visual areas showed increased activity during the identity task. If visual responses are weighted by perceptual mechanisms according to their magnitude, these increased responses would lead to improved attentional selection of the task-appropriate facial aspect. Alternatively, increased responses could be a signature of a sensitivity enhancement mechanism that improves representations of the attended facial aspect. Consistent with the latter sensitivity enhancement mechanism, attending to expression led to enhanced decoding of exemplars of expression both in early visual and dorsal areas relative to attending identity. Similarly, decoding identity exemplars when attending to identity was improved in dorsal and ventral areas. We conclude that attending to expression or identity of dynamic faces is associated with increased selectivity in representations consistent with sensitivity enhancement.++{{:reprints:dobs_et_al.pdf|pdf}} Laquitaine, S. and Gardner, J. L. (2018) A switching observer for human perceptual estimation. //Neuron// 97(2): 462-474. [[https://dx.doi.org/10.1016/j.neuron.2017.12.011|DOI]] ++Abstract| \\ \\ Human perceptual inference has been fruitfully characterized as a normative Bayesian process in which sensory evidence and priors are multiplicatively combined to form posteriors from which sensory estimates can be optimally read-out. We tested whether this basic Bayesian framework could explain human subjects’ behavior in two estimation tasks in which we varied the strength of sensory evidence (motion coherence or contrast) and priors (set of directions or orientations). We found that despite excellent agreement of estimates mean and variability with a Basic Bayesian observer model, the estimate distributions were bimodal with unpredicted modes near the prior and the likelihood. We developed a model that switched between prior and sensory evidence rather than integrating the two, that better explained the data than the Basic and several other Bayesian observers. Our data suggest that humans can approximate Bayesian optimality with a switching heuristic that forgoes multiplicative combination of priors and likelihoods.++{{:reprints:switching.pdf|pdf}} Liu, T., Cable, D., and Gardner, J. L. (2018) Inverted encoding models of human population response conflate noise and neural tuning width. //The Journal of Neuroscience// 38(2): 398-408. [[https://doi.org/10.1523/JNEUROSCI.2453-17.2017|DOI]] [[https://doi.org/10.17605/OSF.IO/9D3EX|DATA]] ++Abstract| \\ \\ Channel encoding models offer the ability to bridge different scales of neuronal measurement by interpreting population responses, typically measured with BOLD imaging in humans, as linear sums of groups of neurons (channels) tuned for visual stimulus properties. Inverting these models to form predicted channel responses from population measurements in humans seemingly offers the potential to infer neuronal tuning properties. Here, we test the ability to make inferences about neural tuning width from inverted encoding models. We examined contrast invariance of orientation selectivity in human V1 (both sexes) and found that inverting the encoding model resulted in channel response functions that became broader with lower contrast, thus, apparently, violating contrast invariance. Simulations showed that this broadening could be explained by contrast-invariant single-unit tuning with the measured decrease in response amplitude at lower contrast. The decrease in response lowers the signal-to-noise ratio of population responses that results in poorer population representation of orientation. Simulations further showed that increasing signal-to-noise makes channel response functions less sensitive to underlying neural tuning width, and in the limit of zero noise will reconstruct the channel function assumed by the model regardless of the bandwidth of single-units. We conclude that our data are consistent with contrast invariant orientation tuning in human V1. More generally, our results demonstrate that population selectivity measures obtained by encoding models can deviate substantially from the behavior of single-units because they conflate neural tuning width and noise and are therefore better used to estimate the uncertainty of decoded stimulus properties. ++{{:reprints:cinvor.pdf|pdf}} Abrahamyan, A., Silva, L. L., Dakin, S. C., Carandini, M. and Gardner, J. L. (2016) Adaptable history biases in human perceptual decisions. //Proceedings of the National Academy of Sciences// 113.25: E3548-E3557 [[http://dx.doi.org/10.1073/pnas.1518786113|DOI]] ++Abstact| \\ \\ When making choices under conditions of perceptual uncertainty, past experience can play a vital role. However, it can also lead to biases that worsen decisions. Consistent with previous observations, we found that human choices are influenced by the success or failure of past choices even in a standard two-alternative detection task, where choice history is irrelevant. The typical bias was one that made the subject switch choices after a failure. These choice-history biases led to poorer performance and were similar for observers in different countries. They were well captured by a simple logistic regression model that had been previously applied to describe psychophysical performance in mice. Such irrational biases seem at odds with the principles of reinforcement learning, which would predict exquisite adaptability to choice history. We therefore asked whether subjects could adapt their irrational biases following changes in trial order statistics. Adaptability was strong in the direction that confirmed a subject’s default biases, but weaker in the opposite direction, so that existing biases could not be eradicated. We conclude that humans can adapt choice history biases, but cannot easily overcome existing biases even if irrational in the current context: adaptation is more sensitive to confirmatory than contradictory statistics.++{{:reprints:pnas-2016-abrahamyan.pdf|pdf}} Birman, D. & Gardner, J. L. (2016) Parietal and prefrontal: categorical differences? //Nature Neuroscience// 19: 5-7 [[http://dx.doi.org/10.1038/nn.4204|DOI]] ++Abstact| \\ \\ A working memory representation goes missing in monkey parietal cortex during categorization learning, but is still found in the prefrontal cortex. ++{{:reprints:nn4204.pdf|pdf}} Gardner, J. L. (2015) A case for human systems neuroscience. //Neuroscience// 296: 130-137 [[http://dx.doi.org/10.1016/j.neuroscience.2014.06.052|DOI]] ++Abstact| \\ \\ Can the human brain itself serve as a model for a systems neuroscience approach to understanding the human brain? After all, how the brain is able to create the richness and complexity of human behavior is still largely mysterious. What better choice to study that complexity than to study it in humans? However, measurements of brain activity typically need to be made non-invasively which puts severe constraints on what can be learned about the internal workings of the brain. Our approach has been to use a combination of psychophysics in which we can use human behavioral flexibility to make quantitative measure- ments of behavior and link those through computational models to measurements of cortical activity through mag- netic resonance imaging. In particular, we have tested vari- ous computational hypotheses about what neural mechanisms could account for behavioral enhancement with spatial attention (Pestilli et al., 2011). Resting both on quantitative measurements and considerations of what is known through animal models, we concluded that weighting of sensory signals by the magnitude of their response is a neural mechanism for efficient selection of sensory signals and consequent improvements in behavioral performance with attention. While animal models have many technical advantages over studying the brain in humans, we believe that human systems neuroscience should endeavor to validate, replicate and extend basic knowledge learned from animal model systems and thus form a bridge to understanding how the brain creates the complex and rich cognitive capacities of humans. ++{{:reprints:gardner_neuroscience_2014.pdf|pdf}} Hara, Y. and Gardner, J. L. (2014) Encoding of graded changes in spatial specificity of prior cues in human visual cortex. //Journal of Neurophysiology// 112:2834-49. [[http://dx.doi.org/10.1152/jn.00729.2013|DOI]]. ++Abstract| \\ \\ Prior information about the relevance of spatial locations can vary in specificity; a single location, a subset of locations or all locations may be of potential importance. Using a contrast-discrimination task with 4 possible targets, we asked whether performance benefits are graded with the spatial specificity of a prior cue and whether we could quantitatively account for behavioral performance with cortical activity changes measured by blood oxygenation level dependent (BOLD) imaging. Thus we changed the prior probability that each location contained the target from 100 to 50 to 25% by cueing in advance 1, 2 or 4 of the possible locations. We found that behavioral performance (discrimination thresholds) improved in a graded fashion with spatial specificity. However, concurrently measured cortical responses from retinotopically-defined visual areas were not strictly graded; response magnitude decreased when all four locations were cued (25% prior probability) relative to the 100 and 50% prior probability conditions, but no significant difference in response magnitude was found between the 100 and 50% prior probability conditions for either cued or uncued locations. Also, while cueing locations increased responses relative to non-cueing, this cue-sensitivity was not graded with prior probability. Further, contrast-sensitivity of cortical responses, which could improve contrast discrimination performance, was not graded. Instead, an efficient-selection model showed that even if sensory responses do not strictly scale with prior probability, selection of sensory responses by weighting larger responses more can result in graded behavioral performance benefits with increasing spatial specificity of prior information.++{{:reprints:haragardnerjnp2014.pdf|pdf}} Vintch, B. and Gardner, J. L. (2014) Cortical correlates of human motion perception biases. //The Journal of Neuroscience// 34: 2592–2604. [[http://dx.doi.org/10.1523/JNEUROSCI.2809-13.2014|DOI]] ++Abstract| \\ \\ Human sensory perception is not a faithful reproduction of the sensory environment. For example, at low contrast, objects appear to move slower and flicker faster than veridical. While these biases have been robustly observed, their neural underpinning are unknown, thus suggesting a possible disconnect of the well established link between motion perception and cortical responses. We used functional imaging to examine the encoding of speed in the human cortex at the scale of neuronal populations and asked where and how these biases are encoded. Decoding, voxel population and forward- encoding analyses revealed biases towards slow speeds and high temporal frequencies at low contrast in the earliest visual cortical regions, matching perception. These findings thus offer a resolution to the disconnect between cortical responses and motion perception in humans. Moreover, biases in speed perception are considered a leading example of Bayesian inference as they can be interpreted as a prior for slow speeds. Our data therefore suggest that perceptual priors of this sort can be encoded by neural populations in the same early cortical areas that provide sensory evidence.++{{:reprints:2014vintch.pdf|pdf}} Hara Y., Pestilli F. and Gardner J. L. (2014). Differing effects of attention in single-units and populations are well predicted by heterogeneous tuning and the normalization model of attention. //Frontiers in Computational Neuroscience// 8:12. [[http://dx.doi.org/10.3389/fncom.2014.00012|DOI]] ++Abstract| \\ \\ Single-unit measurements have reported many different effects of attention on contrast-response (e.g. contrast-gain, response-gain, additive-offset dependent on visibility), while functional imaging measurements have more uniformly reported increases in response across all contrasts (additive-offset). The normalization model of attention elegantly predicts the diversity of effects of attention reported in single-units well-tuned to the stimulus, but what predictions does it make for more realistic populations of neurons with heterogeneous tuning? Are predictions in accordance with population-scale measurements? We used functional imaging data from humans to determine a realistic ratio of attention-field to stimulus-drive size (a key parameter for the model) and predicted effects of attention in a population of model neurons with heterogeneous tuning. We found that within the population, neurons well-tuned to the stimulus showed a response-gain effect, while less-well-tuned neurons showed a contrast-gain effect. Averaged across the population, these disparate effects of attention gave rise to additive-offsets in contrast-response, similar to reports in human functional imaging as well as population averages of single-units. Differences in predictions for single-units and populations were observed across a wide range of model parameters (ratios of attention-field to stimulus-drive size and the amount of baseline response modifiable by attention), offering an explanation for disparity in physiological reports. Thus, by accounting for heterogeneity in tuning of realistic neuronal populations, the normalization model of attention can not only predict responses of well-tuned neurons, but also the activity of large populations of neurons. More generally, computational models can unify physiological findings across different scales of measurement, and make links to behavior, but only if factors such as heterogeneous tuning within a population are properly accounted for.++ \\ {{reprints:hara_pestilli_gardner_2014.pdf|pdf}} Costagli, M., Ueno, K., Sun, P., Gardner, J. L., Wan, X., Ricciardi, E., Pietrini, P., Tanaka, K., and Cheng, K. (2014) Functional signalers of changes in visual stimuli: Cortical responses to increments and decrements in motion coherence. //Cerebral Cortex// 24:110–118 [[http://dx.doi.org/10.1093/cercor/bhs294|DOI]]++Abstract| \\ \\ How does our brain detect changes in a natural scene? While changes by increments of specific visual attributes, such as contrast or motion coherence, can be signaled by an increase in neuronal activity in early visual areas, like the primary visual cortex (V1) or the human middle temporal complex (hMT+), respectively, the mechanisms for signaling changes resulting from decrements in a stimulus attribute are largely unknown. We have discovered opposing patterns of cortical responses to changes in motion coherence: unlike areas hMT+, V3A and parieto-occipital complex (V6+) that respond to changes in the level of motion coherence monotonically, human areas V4 (hV4), V3B, and ventral occipital always respond positively to both transient increments and decrements. This pattern of responding always positively to stimulus changes can emerge in the presence of either coherence-selective neuron populations, or neurons that are not tuned to particular coherences but adapt to a particular coherence level in a stimulus-selective manner. Our findings provide evidence that these areas possess physiological properties suited for signaling increments and decrements in a stimulus and may form a part of cortical vigilance system for detecting salient changes in the environment.++ \\ {{reprints:2012costagli.pdf|pdf}} Merriam, E. P., Gardner, J. L., Movshon, J. A., and Heeger, D. J. (2013) Modulation of visual responses by gaze direction in human visual cortex. //The Journal of Neuroscience// 33: 9879-9889 [[http://dx.doi.org/10.1523/JNEUROSCI.0500-12.2013|DOI]] ++Abstract| \\ \\ To locate visual objects, the brain combines information about retinal location and direction of gaze. Studies in monkeys have demonstrated that eye position modulates the gain of visual signals with “gain fields,” so that single neurons represent both retinotopic location and eye position. We wished to know whether eye position and retinotopic stimulus location are both represented in human visual cortex. Using functional magnetic resonance imaging, we measured separately for each of several different gaze positions cortical responses to stimuli that varied periodically in retinal locus. Visually evoked responses were periodic following the periodic retinotopic stimulation. Only the response amplitudes depended on eye position; response phases were indistinguishable across eye positions. We used multivoxel pattern analysis to decode eye position from the spatial pattern of response amplitudes. The decoder reliably discriminated eye position in five of the early visual cortical areas by taking advantage of a spatially heterogeneous eye position-dependent modulation of cortical activity. We conclude that responses in retinotopically organized visual cortical areas are modulated by gain fields qualitatively similar to those previously observed neurophysiologically.++ \\ {{reprints:2013merriam.pdf|pdf}} Sun, P., Gardner, J. L., Costagli, M., Ueno, K., Waggoner, R. A., Tanaka, K., and Cheng K. (2013) Demonstration of tuning to stimulus orientation in the human visual cortex: A high-resolution fMRI study with a novel continuous and periodic stimulation paradigm. //Cerebral Cortex// 23: 1618–1629 [[http://dx.doi.org/10.1093/cercor/bhs149|DOI]] [[http://gru.brain.riken.jp/lib/exe/fetch.php/reprints/2012sun_supp.pdf?id=shared%3Apublications&cache=cache|SI]]++Abstract| \\ \\ Cells in the animal early visual cortex are sensitive to contour orientations and form repeated structures known as orientation columns. At the behavioral level, there exist 2 well-known global biases in orientation perception (oblique effect and radial bias) in both animals and humans. However, their neural bases are still under debate. To unveil how these behavioral biases are achieved in the early visual cortex, we conducted high-resolution functional magnetic resonance imaging experiments with a novel continuous and periodic stimulation paradigm. By inserting resting recovery periods between successive stimulation periods and introducing a pair of orthogonal stimulation conditions that differed by 90° continuously, we focused on analyzing a blood oxygenation level-dependent response modulated by the change in stimulus orientation and reliably extracted orientation preferences of single voxels. We found that there are more voxels preferring horizontal and vertical orientations, a physiological substrate underlying the oblique effect, and that these over-representations of horizontal and vertical orientations are prevalent in the cortical regions near the horizontal- and vertical-meridian representations, a phenomenon related to the radial bias. Behaviorally, we also confirmed that there exists perceptual superiority for horizontal and vertical orientations around horizontal and vertical meridians, respectively. Our results, thus, refined the neural mechanisms of these 2 global biases in orientation perception.++ \\ {{reprints:2012sun.pdf|pdf}} Suzuki, S., Harasawa, N., Ueno, K., Gardner, J. L., Ichinohe, N., Haruno, M., Cheng, K., and Nakahara H. (2012) Learning to simulate others' decisions. //Neuron// 74:1125-37 [[http://dx.doi.org/10.1016/j.neuron.2012.04.030|DOI]] [[http://gru.brain.riken.jp/lib/exe/fetch.php/reprints/2012suzuki_supp.pdf?id=shared%3Apublications&cache=cache|SI]]++Abstract| \\ \\ A fundamental challenge in social cognition is how humans learn another person's values to predict their decision-making behavior. This form of learning is often assumed to require simulation of the other by direct recruitment of one's own valuation process to model the other's process. However, the cognitive and neural mechanism of simulation learning is not known. Using behavior, modeling, and fMRI, we show that simulation involves two learning signals in a hierarchical arrangement. A simulated-other's reward prediction error processed in ventromedial prefrontal cortex mediated simulation by direct recruitment, being identical for valuation of the self and simulated-other. However, direct recruitment was insufficient for learning, and also required observation of the other's choices to generate a simulated-other's action prediction error encoded in dorsomedial/dorsolateral prefrontal cortex. These findings show that simulation uses a core prefrontal circuit for modeling the other's valuation to generate prediction and an adjunct circuit for tracking behavioral variation to refine prediction.++ \\ {{reprints:2012suzuki.pdf|pdf}} Pestilli, F., Carrasco, M., Heeger, D. J. and Gardner, J. L. (2011) Attentional enhancement via selection and pooling of early sensory responses in human visual cortex. //Neuron// 72:832-46 [[http://dx.doi.org/10.1016/j.neuron.2011.09.025|DOI]] [[http://gru.brain.riken.jp/lib/exe/fetch.php/reprints/attentionselectionsi.pdf?id=shared%3Apublications&cache=cache|SI]] <[[http://dx.doi.org/10.1016/j.neuron.2011.11.005|Preview by John T. Serences]]>++Abstract| \\ \\ To characterize the computational processes by which attention improves behavioral performance, we measured activity in visual cortex with functional magnetic resonance imaging as humans performed a contrast-discrimination task with focal and distributed attention. Focal attention yielded robust improvements in behavioral performance that were accompanied by increases in cortical responses. Using a quantitative analysis, we determined that if performance were limited only by the sensitivity of the measured sensory signals, the improvements in behavioral performance would have corresponded to an unrealistically large (approximately 400%) reduction in response variability. Instead, behavioral performance was well characterized by a pooling and selection process for which the largest sensory responses, those most strongly modulated by attention, dominated the perceptual decision. This characterization predicts that high contrast distracters that evoke large sensory responses should have a negative impact on behavioral performance. We tested and confirmed this prediction. We conclude that attention enhanced behavioral performance predominantly by enabling efficient selection of the behaviorally relevant sensory signals.++{{reprints:attentionselection.pdf|pdf}} Liu, T., Hospadaruk, L., Zhu, D., and Gardner, J. L. (2011) Feature-specific attentional priority signals in human cortex. //The Journal of Neuroscience// 31:4484-95 [[http://dx.doi.org/10.1523/JNEUROSCI.5745-10.2011|DOI]]++Abstract| \\ \\ Human can flexibly attend to a variety of stimulus dimensions, including spatial location and various features such as color and direction of motion. While the locus of spatial attention has been hypothesized to be represented by priority maps encoded in several dorsal frontal and parietal areas, it is unknown how the brain represents attended features. Here we examined the distribution and organization of neural signals related to deployment of feature-based attention. Subjects viewed a compound stimulus containing two superimposed motion directions (or colors), and were instructed to perform an attention-demanding task on one of the directions (or colors). We found elevated and sustained fMRI response for the attention task compared to a neutral condition, without reliable differences in overall response amplitude between attending to different features. However, using multi-voxel pattern analysis, we were able to decode the attended feature in both early visual areas (V1 to hMT+) and frontal and parietal areas (e.g., IPS1-4 and FEF) that are commonly associated with spatial attention. Furthermore, analysis of the classifier weight maps showed that attending to motion and color evoked different patterns of activity, suggesting different neuronal subpopulations in these regions are recruited for attending to different feature dimensions. Thus, our finding suggests that rather than a purely spatial representation of priority, frontal and parietal cortical areas also contain multiplexed signals related to the priority of different non-spatial features. ++{{:reprints:liu_hospadaruk_zhu_gardner_jn_2011.pdf|pdf}} Gardner, J. L. (2010) Is cortical vasculature functionally organized? //Neuroimage// 49:1953-6. [[http://dx.doi.org/10.1016/j.neuroimage.2009.07.004|DOI]] ++Abstact| \\ \\ The cortical vasculature is a well-structured and organized system, but the extent to which it is organized with respect to the neuronal functional architecture is unknown. In particular, does vasculature follow the same functional organization as cortical columns? In principle, cortical columns that share tuning for stimulus features like orientation may often be active together and thus require oxygen and metabolic nutrients together. If the cortical vasculature is built to serve these needs, it may also tend to aggregate and amplify orientation specific signals and explain why they are available in fMRI data at very low resolution. ++{{:reprints:jg_commentary_neuroimage.pdf|pdf}} Offen S, Gardner, J. L., Schluppeck D and Heeger, D.J. (2010) Differential roles for frontal eye fields (FEFs) and intraparietal sulcus (IPS) in visual working memory and visual attention. //Journal of Vision// 10:1-14 [[http://dx.doi.org/10.1167/10.11.28|DOI]] ++Abstract| \\ \\ Cortical activity was measured with functional magnetic resonance imaging to probe the involvement of the superior precentral sulcus (including putative human frontal eye fields, FEFs) and the intraparietal sulcus (IPS) in visual short-term memory and visual attention. In two experimental tasks, human subjects viewed two visual stimuli separated by a variable delay period. The tasks placed differential demands on short-term memory and attention, but the stimuli were visually identical until after the delay period. An earlier study (S. Offen, D. Schluppeck, & D. J. Heeger, 2009) had found a dissociation in early visual cortex that suggested different computational mechanisms underlying the two processes. In contrast, the results reported here show that the patterns of activation in prefrontal and parietal cortex were different from one another but were similar for the two tasks. In particular, the FEF showed evidence for sustained delay period activity for both the working memory and the attention task, while the IPS did not show evidence for sustained delay period activity for either task. The results imply differential roles for the FEF and IPS in these tasks; the results also suggest that feedback of sustained activity from frontal cortex to visual cortex might be gated by task demands. ++{{reprints:offen.pdf|pdf}} Dinstein I, Gardner, J. L., Jazayeri, M and Heeger, D.J. (2008) Executed and observed movements have different distributed representations in human aIPS. //The Journal of Neuroscience// 28:11231-11239 [[http://dx.doi.org/10.1523/JNEUROSCI.3585-08.2008|DOI]] ++Abstract| \\ \\ How similar are the representations of executed and observed hand movements in the human brain? We used functional magnetic resonance imaging (fMRI) and multivariate pattern classification analysis to compare spatial distributions of cortical activity in response to several observed and executed movements. Subjects played the rock-paper-scissors game against a videotaped opponent, freely choosing their movement on each trial and observing the opponent's hand movement after a short delay. The identities of executed movements were correctly classified from fMRI responses in several areas of motor cortex, observed movements were classified from responses in visual cortex, and both observed and executed movements were classified from responses in either left or right anterior intraparietal sulcus (aIPS). We interpret above chance classification as evidence for reproducible, distributed patterns of cortical activity that were unique for execution and/or observation of each movement. Responses in aIPS enabled accurate classification of movement identity within each modality (visual or motor), but did not enable accurate classification across modalities (i.e., decoding observed movements from a classifier trained on executed movements and vice versa). These results support theories regarding the central role of aIPS in the perception and execution of movements. However, the spatial pattern of activity for a particular observed movement was distinctly different from that for the same movement when executed, suggesting that observed and executed movements are mostly represented by distinctly different subpopulations of neurons in aIPS. ++{{reprints:dinstein.pdf|pdf}} Gardner, J. L. , Merriam, E. P., Movshon, J. A., and Heeger, D.J. (2008) Maps of visual space in human occipital cortex are retinotopic, not spatiotopic. //The Journal of Neuroscience// 28:3988-3999 [[http://dx.doi.org/10.1523/JNEUROSCI.5476-07.2008|DOI]] [[http://www.jneurosci.org/content/28/15/i.full|This Week in the Journal]] ++Abstract| \\ \\ We experience the visual world as phenomenally invariant to eye position, but almost all cortical maps of visual space in monkeys use a retinotopic reference frame, that is, the cortical representation of a point in the visual world is different across eye positions. It was recently reported that human cortical area MT (unlike monkey MT) represents stimuli in a reference frame linked to the position of stimuli in space, a "spatiotopic" reference frame. We used visuotopic mapping with blood oxygen level-dependent functional magnetic resonance imaging signals to define 12 human visual cortical areas, and then determined whether the reference frame in each area was spatiotopic or retinotopic. We found that all 12 areas, including MT, represented stimuli in a retinotopic reference frame. Although there were patches of cortex in and around these visual areas that were ostensibly spatiotopic, none of these patches exhibited reliable stimulus-evoked responses. We conclude that the early, visuotopically organized visual cortical areas in the human brain (like their counterparts in the monkey brain) represent stimuli in a retinotopic reference frame. ++{{reprints:retinotopic.pdf|pdf}} Sun, P., Ueno K., Waggoner, R. A., Gardner, J. L. , Tanaka, K., and Cheng K. (2007) A temporal frequency-dependent functional architecture in human V1 revealed by high-resolution fmri. //Nature Neuroscience// 10:1404-1406 [[http://dx.doi.org/10.1038/nn1983|DOI]] ++Abstract| \\ \\ Although cortical neurons with similar functional properties often cluster together in a columnar organization, only ocular dominance columns, the columnar structure representing segregated anatomical input (from one of the two eyes), have been found in human primary visual cortex (V1). It has yet to be shown whether other columnar organizations that arise only from differential responses to stimulus properties also exist in human V1. Using high-resolution functional magnetic resonance imaging, we have found such a functional architecture containing domains that respond preferentially to either low or high temporal frequency. ++{{reprints:tfdomains.pdf|pdf}} Gardner, J. L. , Sun, P., Waggoner, R. A., Ueno K., Tanaka, K., and Cheng K. (2005) Contrast adaptation and representation in human early visual cortex. //Neuron// 47:607-620 [[http://dx.doi.org/10.1016/j.neuron.2005.07.016|DOI]] <[[http://dx.doi.org/10.1016/j.neuron.2005.08.003|Preview by Geoffrey M. Boynton]]>++Abstact| \\ \\ The human visual system can distinguish variations in image contrast over a much larger range than measurements of the static relationship between contrast and response in visual cortex would suggest. This discrepancy may be explained if adaptation serves to re-center contrast response functions around the ambient contrast, yet experiments on humans have yet to report such an effect. By using event-related fMRI and a data-driven analysis approach, we found that contrast response functions in V1, V2, and V3 shift to approximately center on the adapting contrast. Furthermore, we discovered that, unlike earlier areas, human V4 (hV4) responds positively to contrast changes, whether increments or decrements, suggesting that hV4 does not faithfully represent contrast, but instead responds to salient changes. These findings suggest that the visual system discounts slow uninformative changes in contrast with adaptation, yet remains exquisitely sensitive to changes that may signal important events in the environment. ++{{reprints:cadapt.pdf|pdf}} Gardner, J. L. , Tokiyama, S., and Lisberger, S. G. (2004) A population decoding framework for motion aftereffects on smooth pursuit eye movements. //The Journal of Neuroscience// 24:9035-9048 [[http://dx.doi.org/10.1523/JNEUROSCI.0337-04.2004|DOI]] ++Abstract| \\ \\ Both perceptual and motor systems must decode visual information from the distributed activity of large populations of cortical neurons. We have sought a common framework for understanding decoding strategies for visually guided movement and perception by asking whether the strong motion aftereffects seen in the perceptual domain lead to similar expressions in motor output. We found that motion adaptation indeed has strong sequelae in the direction and speed of smooth pursuit eye movements. After adaptation with a stimulus that moves in a given direction for 7 sec, the direction of pursuit is repelled from the direction of pursuit targets that move within 90 degrees of the adapting direction. The speed of pursuit decreases for targets that move at the direction and speed of the adapting stimulus and is repelled from the adapting speed in the sense that the decrease either becomes greater or smaller (eventually turning to an increase) when tracking targets move slower or faster than the adapting speed. The effects of adaptation are spatially specific and fixed to the retinal location of the adapting stimulus. The magnitude of adaptation of pursuit speed and direction is uncorrelated, suggesting that the two parameters are decoded independently. Computer simulation of motion adaptation in the middle temporal visual area (MT) shows that vector-averaging decoding of the population response in MT can account for the effects of adaptation on the direction of pursuit. Our results suggest a unified framework for thinking, in terms of population decoding, about motion adaptation for both perception and action. ++{{reprints:aftereffects.pdf|pdf}} Churchland, A. K., Gardner, J. L., Chou, I. H., Priebe, N. J., and Lisberger, S. G. (2003) Directional anisotropies reveal a functional segregation of visual motion processing for perception and action. //Neuron// 37:1001-1011 [[http://dx.doi.org/10.1016/S0896-6273(03)00145-4|DOI]] ++Abstract| \\ \\ Human exhibits an anisotropy in direction perception: discrimination is superior when motion is around horizontal or vertical rather than diagonal axes. In contrast to the consistent directional anisotropy in perception, we found only small idiosyncratic anisotropies in smooth pursuit eye movements, a motor action requiring accurate discrimination of visual motion direction. Both pursuit and perceptual direction discrimination rely on signals from the middle temporal visual area (MT), yet analysis of multiple measures of MT neuronal responses in the macaque failed to provide evidence of a directional anisotropy. We conclude that MT represents different motion directions uniformly, and subsequent processing creates a directional anisotropy in pathways unique to perception. Our data support the hypothesis that, at least for visual motion, perception and action are guided by inputs from separate sensory streams. The directional anisotropy of perception appears to originate after the two streams have segregated and downstream from area MT. ++{{reprints:dprime.pdf|pdf}} Gardner, J. L., and Lisberger, S. G. (2002) Serial linkage of target selection for orienting and tracking eye movements. //Nature Neuroscience// 5:892-899 [[http://dx.doi.org/10.1038/nn897|DOI]] <[[http://dx.doi.org/10.1038/nn0902-819|News and Views by Michael N. Shadlen]]> ++Abstract| \\ \\ Many natural actions require the coordination of two different kinds of movements. How are targets chosen under these circumstances: do central commands instruct different movement systems in parallel, or does the execution of one movement activate a serial chain that automatically chooses targets for the other movement? We examined a natural eye tracking action that consists of orienting saccades and tracking smooth pursuit eye movements, and found strong physiological evidence for a serial strategy. Monkeys chose freely between two identical spots that appeared at different sites in the visual field and moved in orthogonal directions. If a saccade was evoked to one of the moving targets by microstimulation in either the frontal eye field (FEF) or the superior colliculus (SC), then the same target was automatically chosen for pursuit. Our results imply that the neural signals responsible for saccade execution can also act as an internal command of target choice for other movement systems. ++{{reprints:stimsac.pdf|pdf}} Gardner, J. L., and Lisberger, S. G. (2001) Linked target selection for saccadic and smooth pursuit eye movements. //The Journal of Neuroscience// 21(6):2075-2084 [[http://www.jneurosci.org/content/21/6/2075.short|link]] ++Abstract| \\ \\ In natural situations, motor activity must often choose a single target when multiple distractors are present. The present paper asks how primate smooth pursuit eye movements choose targets, by analysis of a natural target-selection task. Monkeys tracked two targets that started 1.5 degrees eccentric and moved in different directions (up, right, down, and left) toward the position of fixation. As expected from previous results, the smooth pursuit before the first saccade reflected a vector average of the responses to the two target motions individually. However, post-saccadic smooth eye velocity showed enhancement that was spatially selective for the motion at the endpoint of the saccade. If the saccade endpoint was close to one of the two targets, creating a targeting saccade, then pursuit was selectively enhanced for the visual motion of that target and suppressed for the other target. If the endpoint landed between the two targets, creating an averaging saccade, then post-saccadic smooth eye velocity also reflected a vector average of the two target motions. Saccades with latencies >200 msec were almost always targeting saccades. However, pursuit did not transition from vector-averaging to target-selecting until the occurrence of a saccade, even when saccade latencies were >300 msec. Thus, our data demonstrate that post-saccadic enhancement of pursuit is spatially selective and that noncued target selection for pursuit is time-locked to the occurrence of a saccade. This raises the possibility that the motor commands for saccades play a causal role, not only in enhancing visuomotor transmission for pursuit but also in choosing a target for pursuit. ++{{reprints:pursac.pdf|pdf}} Gardner, J. L., Anzai, A., Ohzawa. I., and Freeman, R. D. (1999) Linear and nonlinear contributions to orientation tuning of simple cells in the cat's striate cortex. //Visual Neuroscience// 16:1115-1121 [[http://dx.doi.org/10.1017/S0952523899166112|DOI]] ++Abstract| \\ \\ Orientation selectivity is one of the most conspicuous receptive-field (RF) properties that distinguishes neurons in the striate cortex from those in the lateral geniculate nucleus (LGN). It has been suggested that orientation selectivity arises from an elongated array of feedforward LGN inputs (Hubel & Wiesel, 1962). Others have argued that cortical mechanisms underlie orientation selectivity (e.g. Sillito, 1975; Somers et al., 1995). However, isolation of each mechanism is experimentally difficult and no single study has analyzed both processes simultaneously to address their relative roles. An alternative approach, which we have employed in this study, is to examine the relative contributions of linear and nonlinear mechanisms in sharpening orientation tuning. Since the input stage of simple cells is remarkably linear, the nonlinear contribution can be attributed solely to cortical factors. Therefore, if the nonlinear component is substantial compared to the linear contribution, it can be concluded that cortical factors play a prominent role in sharpening orientation tuning. To obtain the linear contribution, we first measure RF profiles of simple cells in the cat's striate cortex using a binary m-sequence noise stimulus. Then, based on linear spatial summation of the RF profile, we obtain a predicted orientation-tuning curve, which represents the linear contribution. The nonlinear contribution is estimated as the difference between the predicted tuning curve and that measured with drifting sinusoidal gratings. We find that measured tuning curves are generally more sharply tuned for orientation than predicted curves, which indicates that the linear mechanism is not enough to account for the sharpness of orientation-tuning. Therefore, cortical factors must play an important role in sharpening orientation tuning of simple cells. We also examine the relationship of RF shape (subregion aspect ratio) and size (subregion length and width) to orientation-tuning halfwidth. As expected, predicted tuning halfwidths are found to depend strongly on both subregion length and subregion aspect ratio. However, we find that measured tuning halfwidths show only a weak correlation with subregion aspect ratio, and no significant correlation with RF length and width. These results suggest that cortical mechanisms not only serve to sharpen orientation tuning, but also serve to make orientation tuning less dependent on the size and shape of the RF. This ensures that orientation is represented equally well regardless of RF size and shape. ++{{reprints:orient.pdf|pdf}}