Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Anurag Arnab is active.

Publication


Featured researches published by Anurag Arnab.


european conference on computer vision | 2016

Higher Order Conditional Random Fields in Deep Neural Networks

Anurag Arnab; Sadeep Jayasumana; Shuai Zheng; Philip H. S. Torr

We address the problem of semantic segmentation using deep learning. Most segmentation systems include a Conditional Random Field (CRF) to produce a structured output that is consistent with the images visual features. Recent deep learning approaches have incorporated CRFs into Convolutional Neural Networks (CNNs), with some even training the CRF end-to-end with the rest of the network. However, these approaches have not employed higher order potentials, which have previously been shown to significantly improve segmentation performance. In this paper, we demonstrate that two types of higher order potential, based on object detections and superpixels, can be included in a CRF embedded within a deep network. We design these higher order potentials to allow inference with the differentiable mean field algorithm. As a result, all the parameters of our richer CRF model can be learned end-to-end with our pixelwise CNN classifier. We achieve state-of-the-art segmentation performance on the PASCAL VOC benchmark with these trainable higher order potentials.


computer vision and pattern recognition | 2017

Pixelwise Instance Segmentation with a Dynamically Instantiated Network

Anurag Arnab; Philip H. S. Torr

Semantic segmentation and object detection research have recently achieved rapid progress. However, the former task has no notion of different instances of the same object, and the latter operates at a coarse, bounding-box level. We propose an Instance Segmentation system that produces a segmentation map where each pixel is assigned an object class and instance identity label. Most approaches adapt object detectors to produce segments instead of boxes. In contrast, our method is based on an initial semantic segmentation module, which feeds into an instance subnetwork. This subnetwork uses the initial category-level segmentation, along with cues from the output of an object detector, within an end-to-end CRF to predict instances. This part of our model is dynamically instantiated to produce a variable number of instances per image. Our end-to-end approach requires no post-processing and considers the image holistically, instead of processing independent proposals. Therefore, unlike some related work, a pixel cannot belong to multiple instances. Furthermore, far more precise segmentations are achieved, as shown by our substantial improvements at high APr thresholds.


british machine vision conference | 2016

Bottom-up Instance Segmentation using Deep Higher-Order CRFs.

Anurag Arnab; Philip H. S. Torr

Traditional Scene Understanding problems such as Object Detection and Semantic Segmentation have made breakthroughs in recent years due to the adoption of deep learning. However, the former task is not able to localise objects at a pixel level, and the latter task has no notion of different instances of objects of the same class. We focus on the task of Instance Segmentation which recognises and localises objects down to a pixel level. Our model is based on a deep neural network trained for semantic segmentation. This network incorporates a Conditional Random Field with end-to-end trainable higher order potentials based on object detector outputs. This allows us to reason about instances from an initial, category-level semantic segmentation. Our simple method effectively leverages the great progress recently made in semantic segmentation and object detection. The accurate instance-level segmentations that our network produces is reflected by the considerable improvements obtained over previous work.


british machine vision conference | 2015

Joint Object-Material Category Segmentation from Audio-Visual Cues.

Anurag Arnab; Michael Sapienza; Stuart Golodetz; Julien P. C. Valentin; Ondrej Miksik; Shahram Izadi; Philip H. S. Torr

It is not always possible to recognise objects and infer material properties for a scene from visual cues alone, since objects can look visually similar whilst being made of very different materials. In this paper, we therefore present an approach that augments the available dense visual cues with sparse auditory cues in order to estimate dense object and material labels. Since estimates of object class and material properties are mutually informative, we optimise our multi-output labelling jointly using a random-field framework. We evaluate our system on a new dataset with paired visual and auditory data that we make publicly available. We demonstrate that this joint estimation of object and material labels significantly outperforms the estimation of either category in isolation.


IEEE Signal Processing Magazine | 2018

Conditional Random Fields Meet Deep Neural Networks for Semantic Segmentation: Combining Probabilistic Graphical Models with Deep Learning for Structured Prediction

Anurag Arnab; Shuai Zheng; Sadeep Jayasumana; Bernardino Romera-Paredes; Måns Larsson; Alexander Kirillov; Bogdan Savchynskyy; Carsten Rother; Fredrik Kahl; Philip H. S. Torr

Semantic segmentation is the task of labeling every pixel in an image with a predefined object category. It has numerous applications in scenarios where the detailed understanding of an image is required, such as in autonomous vehicles and medical diagnosis. This problem has traditionally been solved with probabilistic models known as conditional random fields (CRFs) due to their ability to model the relationships between the pixels being predicted. However, deep neural networks (DNNs) recently have been shown to excel at a wide range of computer vision problems due to their ability to automatically learn rich feature representations from data, as opposed to traditional handcrafted features. The idea of combining CRFs and DNNs have achieved state-of-the-art results in a number of domains. We review the literature on combining the modeling power of CRFs with the representation-learning ability of DNNs, ranging from early work that combines these two techniques as independent stages of a common pipeline to recent approaches that embed inference of probabilistic models directly in the neural network itself. Finally, we summarize future research directions.


energy minimization methods in computer vision and pattern recognition | 2017

A Projected Gradient Descent Method for CRF Inference Allowing End-to-End Training of Arbitrary Pairwise Potentials

Måns Larsson; Anurag Arnab; Fredrik Kahl; Shuai Zheng; Philip H. S. Torr

Are we using the right potential functions in the Conditional Random Field models that are popular in the Vision community? Semantic segmentation and other pixel-level labelling tasks have made significant progress recently due to the deep learning paradigm. However, most state-of-the-art structured prediction methods also include a random field model with a hand-crafted Gaussian potential to model spatial priors, label consistencies and feature-based image conditioning. In this paper, we challenge this view by developing a new inference and learning framework which can learn pairwise CRF potentials restricted only by their dependence on the image pixel values and the size of the support. Both standard spatial and high-dimensional bilateral kernels are considered. Our framework is based on the observation that CRF inference can be achieved via projected gradient descent and consequently, can easily be integrated in deep neural networks to allow for end-to-end training. It is empirically demonstrated that such learned potentials can improve segmentation accuracy and that certain label class interactions are indeed better modelled by a non-Gaussian potential. In addition, we compare our inference method to the commonly used mean-field algorithm. Our framework is evaluated on several public benchmarks for semantic segmentation with improved performance compared to previous state-of-the-art CNN+CRF models.


arXiv: Computer Vision and Pattern Recognition | 2015

Higher Order Potentials in End-to-End Trainable Conditional Random Fields.

Anurag Arnab; Sadeep Jayasumana; Shuai Zheng; Philip H. S. Torr


computer vision and pattern recognition | 2018

On the Robustness of Semantic Segmentation Models to Adversarial Attacks

Anurag Arnab; Ondrej Miksik; Philip H. S. Torr


international conference on computer graphics and interactive techniques | 2015

SemanticPaint: interactive segmentation and learning of 3D worlds

Stuart Golodetz; Michael Sapienza; Julien P. C. Valentin; Vibhav Vineet; Ming-Ming Cheng; Victor Adrian Prisacariu; Olaf Kähler; Carl Yuheng Ren; Anurag Arnab; Stephen L. Hicks; David W. Murray; Shahram Izadi; Philip H. S. Torr


british machine vision conference | 2017

Holistic, Instance-level Human Parsing.

Qizhu Li; Anurag Arnab; Philip H. S. Torr

Collaboration


Dive into the Anurag Arnab's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Fredrik Kahl

Chalmers University of Technology

View shared research outputs
Top Co-Authors

Avatar

Måns Larsson

Chalmers University of Technology

View shared research outputs
Top Co-Authors

Avatar

Sadeep Jayasumana

Australian National University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge