Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Aayush Bansal is active.

Publication


Featured researches published by Aayush Bansal.


computer vision and pattern recognition | 2016

Marr Revisited: 2D-3D Alignment via Surface Normal Prediction

Aayush Bansal; Bryan C. Russell; Abhinav Gupta

We introduce an approach that leverages surface normal predictions, along with appearance cues, to retrieve 3D models for objects depicted in 2D still images from a large CAD object library. Critical to the success of our approach is the ability to recover accurate surface normals for objects in the depicted scene. We introduce a skip-network model built on the pre-trained Oxford VGG convolutional neural network (CNN) for surface normal prediction. Our model achieves state-of-the-art accuracy on the NYUv2 RGB-D dataset for surface normal prediction, and recovers fine object detail compared to previous methods. Furthermore, we develop a two-stream network over the input image and predicted surface normals that jointly learns pose and style for CAD model retrieval. When using the predicted surface normals, our two-stream network matches prior work using surface normals computed from RGB-D images on the task of pose prediction, and achieves state of the art when using RGB-D input. Finally, our two-stream network allows us to retrieve CAD models that better match the style and pose of a depicted object compared with baseline approaches.


european conference on computer vision | 2014

Towards Transparent Systems: Semantic Characterization of Failure Modes

Aayush Bansal; Ali Farhadi; Devi Parikh

Today’s computer vision systems are not perfect. They fail frequently. Even worse, they fail abruptly and seemingly inexplicably. We argue that making our systems more transparent via an explicit human understandable characterization of their failure modes is desirable. We propose characterizing the failure modes of a vision system using semantic attributes. For example, a face recognition system may say “If the test image is blurry, or the face is not frontal, or the person to be recognized is a young white woman with heavy make up, I am likely to fail.” This information can be used at training time by researchers to design better features, models or collect more focused training data. It can also be used by a downstream machine or human user at test time to know when to ignore the output of the system, in turn making it more reliable. To generate such a “specification sheet”, we discriminatively cluster incorrectly classified images in the semantic attribute space using L1-regularized weighted logistic regression. We show that our specification sheets can predict oncoming failures for face and animal species recognition better than several strong baselines. We also show that lay people can easily follow our specification sheets.


intelligent vehicles symposium | 2014

Understanding how camera configuration and environmental conditions affect appearance-based localization

Aayush Bansal; Hernán Badino; Daniel Huber

Localization is a central problem for intelligent vehicles. Visual localization can supplement or replace GPS-based localization approaches in situations where GPS is unavailable or inaccurate. Although visual localization has been demonstrated in a variety of algorithms and systems, the problem of how to best configure such a system remains largely an open question. Design choices, such as “where should the camera be placed?” and “how should it be oriented?” can have substantial effect on the cost and robustness of a fielded intelligent vehicle. This paper analyzes how different sensor configuration parameters and environmental conditions affect visual localization performance with the goal of understanding what causes certain configurations to perform better than others and providing general principles for configuring systems for visual localization. We ground the investigation using extensive field testing of a visual localization algorithm, and the data sets used for the analysis are made available for comparative evaluation.


international conference on computer vision | 2013

Which Edges Matter

Aayush Bansal; Adarsh Kowdle; Devi Parikh; Andrew C. Gallagher; Larry Zitnick

In this paper, we investigate the ability of humans to recognize objects using different types of edges. Edges arise in images because of several different physical phenomena, such as shadow boundaries, changes in material albedo or reflectance, changes to surface normals, and occlusion boundaries. By constructing synthetic photo realistic scenes, we control which edges are visible in a rendered image to investigate the relationship between human visual recognition and that edge type. We evaluate the information conveyed by each edge type through human studies on object recognition tasks. We find that edges related to surface normals and depth are the most informative edges, while texture and shadow edges can confuse recognition tasks. This work corroborates recent advances in practical vision systems where active sensors capture depth edges (e.g. Microsoft Kinect) as well as in edge detection where progress is being made towards finding object boundaries instead of just pixel gradients. Further, we evaluate seven standard and state-of-the-art edge detectors based on the types of edges they find by comparing the detected edges with known informative edges in the synthetic scene. We suggest that this evaluation method could lead to more informed metrics for gauging developments in edge detection, without requiring any human labeling. In summary, this work shows that human proficiency at object recognition is due to surface normal and depth edges and suggests that future research should focus on explicitly modeling edge types to increase the likelihood of finding informative edges.


Proceedings of SPIE | 2013

CANINE: a robotic mine dog

Brian A. Stancil; Jeffrey Hyams; Jordan Shelley; Kartik Babu; Hernán Badino; Aayush Bansal; Daniel Huber; Parag H. Batavia

Neya Systems, LLC competed in the CANINE program sponsored by the U.S. Army Tank Automotive Research Development and Engineering Center (TARDEC) which culminated in a competition held at Fort Benning as part of the 2012 Robotics Rodeo. As part of this program, we developed a robot with the capability to learn and recognize the appearance of target objects, conduct an area search amid distractor objects and obstacles, and relocate the target object in the same way that Mine dogs and Sentry dogs are used within military contexts for exploration and threat detection. Neya teamed with the Robotics Institute at Carnegie Mellon University to develop vision-based solutions for probabilistic target learning and recognition. In addition, we used a Mission Planning and Management System (MPMS) to orchestrate complex search and retrieval tasks using a general set of modular autonomous services relating to robot mobility, perception and grasping.


arXiv: Computer Vision and Pattern Recognition | 2017

PixelNet: Representation of the pixels, by the pixels, and for the pixels.

Aayush Bansal; Xinlei Chen; Bryan C. Russell; Abhinav Gupta; Deva Ramanan


arXiv: Computer Vision and Pattern Recognition | 2016

PixelNet: Towards a General Pixel-level Architecture.

Aayush Bansal; Xinlei Chen; Bryan C. Russell; Abhinav Gupta; Deva Ramanan


international conference on learning representations | 2018

PixelNN: Example-based Image Synthesis

Aayush Bansal; Yaser Sheikh; Deva Ramanan


arXiv: Computer Vision and Pattern Recognition | 2015

Mid-level Elements for Object Detection.

Aayush Bansal; Abhinav Shrivastava; Carl Doersch; Abhinav Gupta


european conference on computer vision | 2018

Recycle-GAN: Unsupervised Video Retargeting

Aayush Bansal; Shugao Ma; Deva Ramanan; Yaser Sheikh

Collaboration


Dive into the Aayush Bansal's collaboration.

Top Co-Authors

Avatar

Deva Ramanan

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar

Abhinav Gupta

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Daniel Huber

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar

Devi Parikh

Georgia Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Hernán Badino

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar

Victor Fragoso

University of California

View shared research outputs
Top Co-Authors

Avatar

Xinlei Chen

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar

Yaser Sheikh

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge