Ravi Garg
University of Adelaide
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Ravi Garg.
european conference on computer vision | 2016
Ravi Garg; B. G. Vijay Kumar; Gustavo Carneiro; Ian D. Reid
A significant weakness of most current deep Convolutional Neural Networks is the need to train them using vast amounts of manually labelled data. In this work we propose a unsupervised framework to learn a deep convolutional neural network for single view depth prediction, without requiring a pre-training stage or annotated ground-truth depths. We achieve this by training the network in a manner analogous to an autoencoder. At training time we consider a pair of images, source and target, with small, known camera motion between the two such as a stereo pair. We train the convolutional encoder for the task of predicting the depth map for the source image. To do so, we explicitly generate an inverse warp of the target image using the predicted depth and known inter-view displacement, to reconstruct the source image; the photometric error in the reconstruction is the reconstruction loss for the encoder. The acquisition of this training data is considerably simpler than for equivalent systems, requiring no manual annotation, nor calibration of depth sensor to camera. We show that our network trained on less than half of the KITTI dataset gives comparable performance to that of the state-of-the-art supervised methods for single view depth estimation.
international conference on robotics and automation | 2017
Chamara Saroj Weerasekera; Yasir Latif; Ravi Garg; Ian D. Reid
This paper presents an efficient framework for dense 3D scene reconstruction using input from a moving monocular camera. Visual SLAM (Simultaneous Localisation and Mapping) approaches based solely on geometric methods have proven to be quite capable of accurately tracking the pose of a moving camera and simultaneously building a map of the environment in real-time. However, most of them suffer from the 3D map being too sparse for practical use. The missing points in the generated map correspond mainly to areas lacking texture in the input images, and dense mapping systems often rely on hand-crafted priors like piecewise-planarity or piecewise-smooth depth. These priors do not always provide the required level of scene understanding to accurately fill the map. On the other hand, Convolutional Neural Networks (CNNs) have had great success in extracting high-level information from images and regressing pixel-wise surface normals, semantics, and even depth. In this work we leverage this high-level scene context learned by a deep CNN in the form of a surface normal prior. We show, in particular, that using the surface normal prior leads to better reconstructions than the weaker smoothness prior.
computer vision and pattern recognition | 2018
Huangying Zhan; Ravi Garg; Chamara Saroj Weerasekera; Kejie Li; Harsh Agarwal; Ian D. Reid
national conference on artificial intelligence | 2017
Anton Milan; S. Hamid Rezatofighi; Ravi Garg; Anthony R. Dick; Ian D. Reid
international conference on computer vision | 2017
Adrian Johnston; Ravi Garg; Gustavo Carneiro; Ian D. Reid
international conference on robotics and automation | 2018
Yasir Latif; Ravi Garg; Michael Milford; Ian D. Reid
arXiv: Computer Vision and Pattern Recognition | 2017
Chamara Saroj Weerasekera; Ravi Garg; Ian D. Reid
arXiv: Computer Vision and Pattern Recognition | 2017
Ravi Garg; Anders Eriksson; Ian D. Reid
international conference on robotics and automation | 2018
Chamara Saroj Weerasekera; Thanuja Dharmasiri; Ravi Garg; Tom Drummond; Ian D. Reid
arXiv: Computer Vision and Pattern Recognition | 2017
Pan Ji; Ian D. Reid; Ravi Garg; Hongdong Li; Mathieu Salzmann