Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Thaddeus B. Czuba is active.

Publication


Featured researches published by Thaddeus B. Czuba.


The Journal of Neuroscience | 2014

Area MT Encodes Three-Dimensional Motion

Thaddeus B. Czuba; Alexander C. Huk; Lawrence K. Cormack; Adam Kohn

We use visual information to determine our dynamic relationship with other objects in a three-dimensional (3D) world. Despite decades of work on visual motion processing, it remains unclear how 3D directions—trajectories that include motion toward or away from the observer—are represented and processed in visual cortex. Area MT is heavily implicated in processing visual motion and depth, yet previous work has found little evidence for 3D direction sensitivity per se. Here we use a rich ensemble of binocular motion stimuli to reveal that most neurons in area MT of the anesthetized macaque encode 3D motion information. This tuning for 3D motion arises from multiple mechanisms, including different motion preferences in the two eyes and a nonlinear interaction of these signals when both eyes are stimulated. Using a novel method for functional binocular alignment, we were able to rule out contributions of static disparity tuning to the 3D motion tuning we observed. We propose that a primary function of MT is to encode 3D motion, critical for judging the movement of objects in dynamic real-world environments.


Journal of Vision | 2011

Three-dimensional motion aftereffects reveal distinct direction-selective mechanisms for binocular processing of motion through depth.

Thaddeus B. Czuba; Bas Rokers; Kyle R. Guillet; Alexander C. Huk; Lawrence K. Cormack

Motion aftereffects are historically considered evidence for neuronal populations tuned to specific directions of motion. Despite a wealth of motion aftereffect studies investigating 2D (frontoparallel) motion mechanisms, there is a remarkable dearth of psychophysical evidence for neuronal populations selective for the direction of motion through depth (i.e., tuned to 3D motion). We compared the effects of prolonged viewing of unidirectional motion under dichoptic and monocular conditions and found large 3D motion aftereffects that could not be explained by simple inheritance of 2D monocular aftereffects. These results (1) demonstrate the existence of neurons tuned to 3D motion as distinct from monocular 2D mechanisms, (2) show that distinct 3D direction selectivity arises from both interocular velocity differences and changing disparities over time, and (3) provide a straightforward psychophysical tool for further probing 3D motion mechanisms.


Journal of Vision | 2012

To CD or not to CD: Is there a 3D motion aftereffect based on changing disparities?

Thaddeus B. Czuba; Bas Rokers; Alexander C. Huk; Lawrence K. Cormack

Recently, T. B. Czuba, B. Rokers, K. Guillet, A. C. Huk, and L. K. Cormack, (2011) and Y. Sakano, R. S. Allison, and I. P. Howard (2012) published very similar studies using the motion aftereffect to probe the way in which motion through depth is computed. Here, we compare and contrast the findings of these two studies and incorporate their results with a brief follow-up experiment. Taken together, the results leave no doubt that the human visual system incorporates a mechanism that is uniquely sensitive to the difference in velocity signals between the two eyes, but--perhaps surprisingly--evidence for a neural representation of changes in binocular disparity over time remains elusive.


Annual Review of Vision Science | 2017

Binocular Mechanisms of 3D Motion Processing

Lawrence K. Cormack; Thaddeus B. Czuba; Jonas Knöll; Alexander C. Huk

The visual system must recover important properties of the external environment if its host is to survive. Because the retinae are effectively two-dimensional but the world is three-dimensional (3D), the patterns of stimulation both within and across the eyes must be used to infer the distal stimulus-the environment-in all three dimensions. Moreover, animals and elements in the environment move, which means the input contains rich temporal information. Here, in addition to reviewing the literature, we discuss how and why prior work has focused on purported isolated systems (e.g., stereopsis) or cues (e.g., horizontal disparity) that do not necessarily map elegantly on to the computations and complex patterns of stimulation that arise when visual systems operate within the real world. We thus also introduce the binoptic flow field (BFF) as a description of the 3D motion information available in realistic environments, which can foster the use of ecologically valid yet well-controlled stimuli. Further, it can help clarify how future studies can more directly focus on the computations and stimulus properties the visual system might use to support perception and behavior in a dynamic 3D world.


The Journal of Neuroscience | 2016

Separate Perceptual and Neural Processing of Velocity- and Disparity-Based 3D Motion Signals

Sung Jun Joo; Thaddeus B. Czuba; Lawrence K. Cormack; Alexander C. Huk

Although the visual system uses both velocity- and disparity-based binocular information for computing 3D motion, it is unknown whether (and how) these two signals interact. We found that these two binocular signals are processed distinctly at the levels of both cortical activity in human MT and perception. In human MT, adaptation to both velocity-based and disparity-based 3D motions demonstrated direction-selective neuroimaging responses. However, when adaptation to one cue was probed using the other cue, there was no evidence of interaction between them (i.e., there was no “cross-cue” adaptation). Analogous psychophysical measurements yielded correspondingly weak cross-cue motion aftereffects (MAEs) in the face of very strong within-cue adaptation. In a direct test of perceptual independence, adapting to opposite 3D directions generated by different binocular cues resulted in simultaneous, superimposed, opposite-direction MAEs. These findings suggest that velocity- and disparity-based 3D motion signals may both flow through area MT but constitute distinct signals and pathways. SIGNIFICANCE STATEMENT Recent human neuroimaging and monkey electrophysiology have revealed 3D motion selectivity in area MT, which is driven by both velocity-based and disparity-based 3D motion signals. However, to elucidate the neural mechanisms by which the brain extracts 3D motion given these binocular signals, it is essential to understand how—or indeed if—these two binocular cues interact. We show that velocity-based and disparity-based signals are mostly separate at the levels of both fMRI responses in area MT and perception. Our findings suggest that the two binocular cues for 3D motion might be processed by separate specialized mechanisms.


Journal of Neurophysiology | 2010

Speed and Eccentricity Tuning Reveal a Central Role for the Velocity-Based Cue to 3D Visual Motion

Thaddeus B. Czuba; Bas Rokers; Alexander C. Huk; Lawrence K. Cormack


Journal of Vision | 2011

Motion processing with two eyes in three dimensions

Bas Rokers; Thaddeus B. Czuba; Lawrence K. Cormack; Alexander C. Huk


Journal of Vision | 2013

Neuronal selectivity for directions of 3D motion in area MT

Thaddeus B. Czuba; Lawrence K. Cormack; Alexander C. Huk; Adam Kohn


Journal of Vision | 2011

Isolation of binocular 3D motion cues in human visual cortex

Thaddeus B. Czuba; Alexander C. Huk; Lawrence K. Cormack


Journal of Vision | 2010

Mechanisms of 3D motion: Integration of disparity and motion cues

Thaddeus B. Czuba; Bas Rokers; Lawrence K. Cormack; Alexander C. Huk

Collaboration


Dive into the Thaddeus B. Czuba's collaboration.

Top Co-Authors

Avatar

Alexander C. Huk

University of Texas at Austin

View shared research outputs
Top Co-Authors

Avatar

Lawrence K. Cormack

University of Texas at Austin

View shared research outputs
Top Co-Authors

Avatar

Bas Rokers

University of Wisconsin-Madison

View shared research outputs
Top Co-Authors

Avatar

Adam Kohn

Albert Einstein College of Medicine

View shared research outputs
Top Co-Authors

Avatar

Kathryn Bonnen

University of Texas at Austin

View shared research outputs
Top Co-Authors

Avatar

Kyle R. Guillet

University of Texas at Austin

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Austin Kuo

University of Texas at Austin

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Jake Whritner

University of Texas at Austin

View shared research outputs
Researchain Logo
Decentralizing Knowledge