Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Shenlong Wang is active.

Publication


Featured researches published by Shenlong Wang.


computer vision and pattern recognition | 2012

Relaxed collaborative representation for pattern classification

Meng Yang; Lei Zhang; David Zhang; Shenlong Wang

Regularized linear representation learning has led to interesting results in image classification, while how the object should be represented is a critical issue to be investigated. Considering the fact that the different features in a sample should contribute differently to the pattern representation and classification, in this paper we present a novel relaxed collaborative representation (RCR) model to effectively exploit the similarity and distinctiveness of features. In RCR, each feature vector is coded on its associated dictionary to allow flexibility of feature coding, while the variance of coding vectors is minimized to address the similarity among features. In addition, the distinctiveness of different features is exploited by weighting its distance to other features in the coding domain. The proposed RCR is simple, while our extensive experimental results on benchmark image databases (e.g., various face and flower databases) show that it is very competitive with state-of-the-art image classification methods.


user interface software and technology | 2016

Holoportation: Virtual 3D Teleportation in Real-time

Sergio Orts-Escolano; Christoph Rhemann; Sean Ryan Fanello; Wayne Chang; Adarsh Prakash Murthy Kowdle; Yury Degtyarev; David Kim; Philip Lindsley Davidson; Sameh Khamis; Mingsong Dou; Vladimir Tankovich; Charles T. Loop; Qin Cai; Philip A. Chou; Sarah Mennicken; Julien P. C. Valentin; Vivek Pradeep; Shenlong Wang; Sing Bing Kang; Pushmeet Kohli; Yuliya Lutchyn; Cem Keskin; Shahram Izadi

We present an end-to-end system for augmented and virtual reality telepresence, called Holoportation. Our system demonstrates high-quality, real-time 3D reconstructions of an entire space, including people, furniture and objects, using a set of new depth cameras. These 3D models can also be transmitted in real-time to remote users. This allows users wearing virtual or augmented reality displays to see, hear and interact with remote participants in 3D, almost as if they were present in the same physical space. From an audio-visual perspective, communicating and interacting with remote users edges closer to face-to-face communication. This paper describes the Holoportation technical system in full, its key interactive capabilities, the application scenarios it enables, and an initial qualitative study of using this new communication medium.


computer vision and pattern recognition | 2015

Holistic 3D scene understanding from a single geo-tagged image

Shenlong Wang; Sanja Fidler; Raquel Urtasun

In this paper we are interested in exploiting geographic priors to help outdoor scene understanding. Towards this goal we propose a holistic approach that reasons jointly about 3D object detection, pose estimation, semantic segmentation as well as depth reconstruction from a single image. Our approach takes advantage of large-scale crowd-sourced maps to generate dense geographic, geometric and semantic priors by rendering the 3D world. We demonstrate the effectiveness of our holistic model on the challenging KITTI dataset [13], and show significant improvements over the baselines in all metrics and tasks.


asian conference on computer vision | 2012

Nonlocal spectral prior model for low-level vision

Shenlong Wang; Lei Zhang; Yan Liang

Image nonlocal self-similarity has been widely adopted as natural image prior in various low-level vision tasks such as image restoration, while the low-rank matrix recovery theory has been drawing much attention to describe and utilize the image nonlocal self-similarities. However, whether the low-rank prior models exist to characterize the nonlocal self-similarity for a wide range of natural images is not clear yet. In this paper we investigate this issue by evaluating the heavy-tailed distributions of singular values of the matrices of nonlocal similar patches collected from natural images. A novel image prior model, namely nonlocal spectral prior (NSP) model, is then proposed to characterize the singular values of nonlocal similar patches. We consequently apply the NSP model to typical image restoration tasks, including denoising, super-resolution and deblurring, and the experimental results demonstrated the highly competitive performance of NSP in solving these low-level vision problems.


international conference on computer vision | 2015

Lost Shopping! Monocular Localization in Large Indoor Spaces

Shenlong Wang; Sanja Fidler; Raquel Urtasun

In this paper we propose a novel approach to localization in very large indoor spaces (i.e., 200+ store shopping malls) that takes a single image and a floor plan of the environment as input. We formulate the localization problem as inference in a Markov random field, which jointly reasons about text detection (localizing shops names in the image with precise bounding boxes), shop facade segmentation, as well as cameras rotation and translation within the entire shopping mall. The power of our approach is that it does not use any prior information about appearance and instead exploits text detections corresponding to the shop names. This makes our method applicable to a variety of domains and robust to store appearance variation across countries, seasons, and illumination conditions. We demonstrate the performance of our approach in a new dataset we collected of two very large shopping malls, and show the power of holistic reasoning.


computer vision and pattern recognition | 2016

HD Maps: Fine-Grained Road Segmentation by Parsing Ground and Aerial Images

Gellert Mattyus; Shenlong Wang; Sanja Fidler; Raquel Urtasun

In this paper we present an approach to enhance existing maps with fine grained segmentation categories such as parking spots and sidewalk, as well as the number and location of road lanes. Towards this goal, we propose an efficient approach that is able to estimate these fine grained categories by doing joint inference over both, monocular aerial imagery, as well as ground images taken from a stereo camera pair mounted on top of a car. Important to this is reasoning about the alignment between the two types of imagery, as even when the measurements are taken with sophisticated GPS+IMU systems, this alignment is not sufficiently accurate. We demonstrate the effectiveness of our approach on a new dataset which enhances KITTI [8] with aerial images taken with a camera mounted on an airplane and flying around the city of Karlsruhe, Germany.


international conference on computer vision | 2015

Enhancing Road Maps by Parsing Aerial Images Around the World

Gellert Mattyus; Shenlong Wang; Sanja Fidler; Raquel Urtasun

In recent years, contextual models that exploit maps have been shown to be very effective for many recognition and localization tasks. In this paper we propose to exploit aerial images in order to enhance freely available world maps. Towards this goal, we make use of OpenStreetMap and formulate the problem as the one of inference in a Markov random field parameterized in terms of the location of the road-segment centerlines as well as their width. This parameterization enables very efficient inference and returns only topologically correct roads. In particular, we can segment all OSM roads in the whole world in a single day using a small cluster of 10 computers. Importantly, our approach generalizes very well, it can be trained using only 1.5 km2 aerial imagery and produce very accurate results in any location across the globe. We demonstrate the effectiveness of our approach outperforming the state-of-the-art in two new benchmarks that we collect. We then show how our enhanced maps are beneficial for semantic segmentation of ground images.


computer vision and pattern recognition | 2016

The Global Patch Collider

Shenlong Wang; Sean Ryan Fanello; Christoph Rhemann; Shahram Izadi; Pushmeet Kohli

This paper proposes a novel extremely efficient, fully-parallelizable, task-specific algorithm for the computation of global point-wise correspondences in images and videos. Our algorithm, the Global Patch Collider, is based on detecting unique collisions between image points using a collection of learned tree structures that act as conditional hash functions. In contrast to conventional approaches that rely on pairwise distance computation, our algorithm isolates distinctive pixel pairs that hit the same leaf during traversal through multiple learned tree structures. The split functions stored at the intermediate nodes of the trees are trained to ensure that only visually similar patches or their geometric or photometric transformed versions fall into the same leaf node. The matching process involves passing all pixel positions in the images under analysis through the tree structures. We then compute matches by isolating points that uniquely collide with each other ie. fell in the same empty leaf in multiple trees. Our algorithm is linear in the number of pixels but can be made constant time on a parallel computation architecture as the tree traversal for individual image points is decoupled. We demonstrate the efficacy of our method by using it to perform optical flow matching and stereo matching on some challenging benchmarks. Experimental results show that not only is our method extremely computationally efficient, but it is also able to match or outperform state of the art methods that are much more complex.


international conference on robotics and automation | 2017

Find your way by observing the sun and other semantic cues

Wei-Chiu Ma; Shenlong Wang; Marcus A. Brubaker; Sanja Fidler; Raquel Urtasun

In this paper we present a robust, efficient and affordable approach to self-localization which requires neither GPS nor knowledge about the appearance of the world. Towards this goal, we utilize freely available cartographic maps and derive a probabilistic model that exploits semantic cues in the form of sun direction, presence of an intersection, road type, speed limit and ego-car trajectory to produce very reliable localization results. Our experimental evaluation shows that our approach can localize much faster (in terms of driving time) with less computation and more robustly than competing approaches, which ignore semantic information.


international conference on computational photography | 2014

Transductive Gaussian processes for image denoising

Shenlong Wang; Lei Zhang; Raquel Urtasun

In this paper we are interested in exploiting self-similarity information for discriminative image denoising. Towards this goal, we propose a simple yet powerful denoising method based on transductive Gaussian processes, which introduces self-similarity in the prediction stage. Our approach allows to build a rich similarity measure by learning hyper parameters defining multi-kernel combinations. We introduce perceptual-driven kernels to capture pixel-wise, gradient-based and local-structure similarities. In addition, our algorithm can integrate several initial estimates as input features to boost performance even further. We demonstrate the effectiveness of our approach on several benchmarks. The experiments show that our proposed denoising algorithm has better performance than competing discriminative denoising methods, and achieves competitive result with respect to the state-of-the-art.

Collaboration


Dive into the Shenlong Wang's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Lei Zhang

Hong Kong Polytechnic University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Hang Chu

University of Toronto

View shared research outputs
Top Co-Authors

Avatar

Wei-Chiu Ma

Massachusetts Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Bin Yang

University of Toronto

View shared research outputs
Researchain Logo
Decentralizing Knowledge