Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Ronald Clark is active.

Publication


Featured researches published by Ronald Clark.


international conference on robotics and automation | 2017

DeepVO: Towards end-to-end visual odometry with deep Recurrent Convolutional Neural Networks

Sen Wang; Ronald Clark; Hongkai Wen; Niki Trigoni

This paper studies monocular visual odometry (VO) problem. Most of existing VO algorithms are developed under a standard pipeline including feature extraction, feature matching, motion estimation, local optimisation, etc. Although some of them have demonstrated superior performance, they usually need to be carefully designed and specifically fine-tuned to work well in different environments. Some prior knowledge is also required to recover an absolute scale for monocular VO. This paper presents a novel end-to-end framework for monocular VO by using deep Recurrent Convolutional Neural Networks (RCNNs). Since it is trained and deployed in an end-to-end manner, it infers poses directly from a sequence of raw RGB images (videos) without adopting any module in the conventional VO pipeline. Based on the RCNNs, it not only automatically learns effective feature representation for the VO problem through Convolutional Neural Networks, but also implicitly models sequential dynamics and relations using deep Recurrent Neural Networks. Extensive experiments on the KITTI VO dataset show competitive performance to state-of-the-art methods, verifying that the end-to-end Deep Learning technique can be a viable complement to the traditional VO systems.


computer vision and pattern recognition | 2017

VidLoc: A Deep Spatio-Temporal Model for 6-DoF Video-Clip Relocalization

Ronald Clark; Sen Wang; Andrew Markham; Niki Trigoni; Hongkai Wen

Machine learning techniques, namely convolutional neural networks (CNN) and regression forests, have recently shown great promise in performing 6-DoF localization of monocular images. However, in most cases image-sequences, rather only single images, are readily available. To this extent, none of the proposed learning-based approaches exploit the valuable constraint of temporal smoothness, often leading to situations where the per-frame error is larger than the camera motion. In this paper we propose a recurrent model for performing 6-DoF localization of video-clips. We find that, even by considering only short sequences (20 frames), the pose estimates are smoothed and the localization error can be drastically reduced. Finally, we consider means of obtaining probabilistic pose estimates from our model. We evaluate our method on openly-available real-world autonomous driving and indoor localization datasets.


intelligent robots and systems | 2016

Keyframe based large-scale indoor localisation using geomagnetic field and motion pattern

Sen Wang; Hongkai Wen; Ronald Clark; Niki Trigoni

This paper studies indoor localisation problem by using low-cost and pervasive sensors. Most of existing indoor localisation algorithms rely on camera, laser scanner, floor plan or other pre-installed infrastructure to achieve sub-meter or sub-centimetre localisation accuracy. However, in some circumstances these required devices or information may be unavailable or too expensive in terms of cost or deployment. This paper presents a novel keyframe based Pose Graph Simultaneous Localisation and Mapping (SLAM) method, which correlates ambient geomagnetic field with motion pattern and employs low-cost sensors commonly equipped in mobile devices, to provide positioning in both unknown and known environments. Extensive experiments are conducted in large-scale indoor environments to verify that the proposed method can achieve high localisation accuracy similar to state-of-the-arts, such as vision based Google Project Tango.


information processing in sensor networks | 2015

Robust vision-based indoor localization

Ronald Clark; Niki Trigoni; Andrew Markham

Vision-based positioning has proven to be highly successful and popular in mobile robotics and computer vision applications. These methods have, however, not enjoyed the same popularity in the field of indoor localization. In this work we highlight some of the issues that arise when using vision-based methods for indoor localization. We then propose means of addressing these issues and implement a proof-of-concept visual inertial odometry system for a mobile device. Preliminary experiments have been carried out in a small library where sub-meter positioning accuracy was attained. Based on our proof-of-concept, we believe that visual inertial odometry techniques can provide the levels of positioning accuracy needed for widespread adoption.


The International Journal of Robotics Research | 2018

End-to-end, sequence-to-sequence probabilistic visual odometry through deep neural networks

Sen Wang; Ronald Clark; Hongkai Wen; Niki Trigoni

This paper studies visual odometry (VO) from the perspective of deep learning. After tremendous efforts in the robotics and computer vision communities over the past few decades, state-of-the-art VO algorithms have demonstrated incredible performance. However, since the VO problem is typically formulated as a pure geometric problem, one of the key features still missing from current VO systems is the capability to automatically gain knowledge and improve performance through learning. In this paper, we investigate whether deep neural networks can be effective and beneficial to the VO problem. An end-to-end, sequence-to-sequence probabilistic visual odometry (ESP-VO) framework is proposed for the monocular VO based on deep recurrent convolutional neural networks. It is trained and deployed in an end-to-end manner, that is, directly inferring poses and uncertainties from a sequence of raw images (video) without adopting any modules from the conventional VO pipeline. It can not only automatically learn effective feature representation encapsulating geometric information through convolutional neural networks, but also implicitly model sequential dynamics and relation for VO using deep recurrent neural networks. Uncertainty is also derived along with the VO estimation without introducing much extra computation. Extensive experiments on several datasets representing driving, flying and walking scenarios show competitive performance of the proposed ESP-VO to the state-of-the-art methods, demonstrating a promising potential of the deep learning technique for VO and verifying that it can be a viable complement to current VO systems.


ieee-ras international conference on humanoid robots | 2016

Increasing the efficiency of 6-DoF visual localization using multi-modal sensory data

Ronald Clark; Sen Wang; Hongkai Wen; Niki Trigoni; Andrew Markham

Localization is a key requirement for mobile robot autonomy and human-robot interaction. Vision-based localization is accurate and flexible, however, it incurs a high computational burden which limits its application on many resource-constrained platforms. In this paper, we address the problem of performing real-time localization in large-scale 3D point cloud maps of ever-growing size. While most systems using multi-modal information reduce localization time by employing side-channel information in a coarse manner (eg. WiFi for a rough prior position estimate), we propose to inter-weave the map with rich sensory data. This multi-modal approach achieves two key goals simultaneously. First, it enables us to harness additional sensory data to localise against a map covering a vast area in real-time; and secondly, it also allows us to roughly localise devices which are not equipped with a camera. The key to our approach is a localization policy based on a sequential Monte Carlo estimator. The localiser uses this policy to attempt point-matching only in nodes where it is likely to succeed, significantly increasing the efficiency of the localization process. The proposed multi-modal localization system is evaluated extensively in a large museum building. The results show that our multi-modal approach not only increases the localization accuracy but significantly reduces computational time.


european conference on computer vision | 2018

Learning to Solve Nonlinear Least Squares for Monocular Stereo

Ronald Clark; Michael Bloesch; Jan Czarnowski; Stefan Leutenegger; Andrew J. Davison

Sum-of-squares objective functions are very popular in computer vision algorithms. However, these objective functions are not always easy to optimize. The underlying assumptions made by solvers are often not satisfied and many problems are inherently ill-posed. In this paper, we propose LS-Net, a neural nonlinear least squares optimization algorithm which learns to effectively optimize these cost functions even in the presence of adversities. Unlike traditional approaches, the proposed solver requires no hand-crafted regularizers or priors as these are implicitly learned from the data. We apply our method to the problem of motion stereo ie. jointly estimating the motion and scene geometry from pairs of images of a monocular sequence. We show that our learned optimizer is able to efficiently and effectively solve this challenging optimization problem.


national conference on artificial intelligence | 2017

VINet: Visual Inertial Odometry as a Sequence to Sequence Learning Problem

Ronald Clark; Sen Wang; Hongkai Wen; Andrew Markham; Niki Trigoni


international conference on computer vision | 2017

3D Object Reconstruction from a Single Depth View with Adversarial Learning

Bo Yang; Hongkai Wen; Sen Wang; Ronald Clark; Andrew Markham; Niki Trigoni


Archive | 2017

VidLoc: 6-DoF Video-Clip Relocalization.

Ronald Clark; Sen Wang; Andrew Markham; Niki Trigoni; Hongkai Wen

Collaboration


Dive into the Ronald Clark's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Sen Wang

University of Oxford

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Bo Yang

University of Oxford

View shared research outputs
Researchain Logo
Decentralizing Knowledge