Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Kamran Binaee is active.

Publication


Featured researches published by Kamran Binaee.


acm symposium on applied perception | 2016

Binocular eye tracking calibration during a virtual ball catching task using head mounted display

Kamran Binaee; Gabriel Diaz; Jeff B. Pelz; Flip Phillips

When tracking the eye movements of an active observer, the quality of the tracking data is continuously affected by physical shifts of the eye-tracker on an observers head. This is especially true for eye-trackers integrated within virtual-reality (VR) helmets. These configurations modify the weight and inertia distribution well beyond that of the eye-tracker alone. Despite the continuous nature of this degradation, it is common practice for calibration procedures to establish eye-to-screen mappings, fixed over the time-course of an experiment. Even with periodic recalibration, data quality can quickly suffer due to head motion. Here, we present a novel post-hoc calibration method that allows for continuous temporal interpolation between discrete calibration events. Analysis focuses on the comparison of fixed vs. continuous calibration schemes and their effects upon the quality of a binocular gaze data to virtual targets, especially with respect to depth. Calibration results were applied to binocular eye tracking data from a VR ball catching task and improved the tracking accuracy especially in the dynamic case.


Behavior Research Methods | 2018

Assessment of an augmented reality apparatus for the study of visually guided walking and obstacle crossing

Kamran Binaee; Gabriel Diaz

To walk through the cluttered natural environment requires visually guided and anticipatory adjustments to gait in advance of upcoming obstacles. However, scientific investigation of visual contributions to obstacle crossing have historically been limited by the practical issues involved with the repeated presentation of multiple obstacles upon a ground plane. This study evaluates an approach in which the perception of a 3D obstacle is generated from 2D projection onto the ground plane with perspective correction based on the subject’s motion-tracked head position. The perception of depth is further reinforced with the use of stereoscopic goggles. To evaluate the validity of this approach, behavior was compared between approaches to two types of obstacles in a blocked design: physical obstacles, and the augmented reality (AR) obstacles projected upon the ground plane. In addition, obstacle height, defined in units of leg length (LL), was varied on each trial (0.15, 0.25, 0.35 LL). Approaches to ended with collision on 0.8% of trials with physical obstacles per subject, and on 1.4% trials with AR obstacles. Collisions were signaled by auditory feedback. Linear changes in the height of both AR and physical obstacles produced linear changes in maximum step height, preserving a constant clearance magnitude across changes in obstacle height. However, for AR obstacles, approach speed was slower, the crossing step peaked higher above the obstacle, and there was greater clearance between the lead toe and the obstacle. These results suggest that subjects were more cautious when approaching and stepping over AR obstacles.


Proceedings of the Ninth Biennial ACM Symposium on Eye Tracking Research & Applications | 2016

Novel apparatus for investigation of eye movements when walking in the presence of 3D projected obstacles

Rakshit Kothari; Kamran Binaee; Jonathan Matthis; Reynold J. Bailey; Gabriel Diaz

The human gait cycle is incredibly efficient and stable largely because of the use of advance visual information to make intelligent selections of heading direction, foot placement, gait dynamics, and posture when faced with terrain complexity [Patla and Vickers 1997; Patla and Vickers 2003; Matthis and Fajen 2013; Matthis and Hayhoe 2015]. This is behaviorally demonstrated by a coupling between saccades and foot placement.


arXiv: Other Computer Science | 2018

Characterizing the Temporal Dynamics of Information in Visually Guided Predictive Control Using LSTM Recurrent Neural Networks.

Kamran Binaee; Anna Starynska; Jeff B. Pelz; Christopher Kanan; Gabriel Diaz


Journal of Vision | 2018

Classification and Statistics of Gaze In World Events

Rakshit Kothari; Zhizhuo Yang; Kamran Binaee; Reynold J. Bailey; Christopher Kanan; Jeff B. Pelz; Gabriel Diaz


Journal of Vision | 2018

Investigating the Differences in Predictive Oculomotor Strategies using Long Short-Term Memory Recurrent Neural Network Models

Kamran Binaee; Rakshit Kothari; Jeff B. Pelz; Gabriel Diaz


Journal of Vision | 2017

The contribution of visual pursuit to prediction in a naturalistic interception task

Kamran Binaee; Gabriel Diaz


Journal of Vision | 2017

Gaze-in-World movement Classification for Unconstrained Head Motion during Natural Tasks.

Rakshit Kothari; Kamran Binaee; Reynold J. Bailey; Christopher Kanan; Gabriel Diaz; Jeff B. Pelz


Journal of Vision | 2017

Modeling Hand-Eye Movements in a Virtual Ball Catching Setup using Deep Recurrent Neural Network

Kamran Binaee; Anna Starynska; Rakshit Kothari; Christopher Kanan; Jeff B. Pelz; Gabriel Diaz


Journal of Vision | 2016

The Influence of Biomechanics on Visual Attention while Walking

Rakshit Kothari; Gabriel Diaz; Kamran Binaee; Reynold J. Bailey; Johnatan Matthis

Collaboration


Dive into the Kamran Binaee's collaboration.

Top Co-Authors

Avatar

Gabriel Diaz

Rochester Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Rakshit Kothari

Rochester Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Jeff B. Pelz

Rochester Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Reynold J. Bailey

Rochester Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Christopher Kanan

Rochester Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Anna Starynska

Rochester Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Flip Phillips

Rochester Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Andrew Smith

Rochester Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Rahul Gopinathan

Rochester Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Zhizhuo Yang

Rochester Institute of Technology

View shared research outputs
Researchain Logo
Decentralizing Knowledge