Zhipeng Deng
National University of Defense Technology
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Zhipeng Deng.
Sensors | 2017
Tianyu Tang; Shilin Zhou; Zhipeng Deng; Huanxin Zou; Lin Lei
Detecting vehicles in aerial imagery plays an important role in a wide range of applications. The current vehicle detection methods are mostly based on sliding-window search and handcrafted or shallow-learning-based features, having limited description capability and heavy computational costs. Recently, due to the powerful feature representations, region convolutional neural networks (CNN) based detection methods have achieved state-of-the-art performance in computer vision, especially Faster R-CNN. However, directly using it for vehicle detection in aerial images has many limitations: (1) region proposal network (RPN) in Faster R-CNN has poor performance for accurately locating small-sized vehicles, due to the relatively coarse feature maps; and (2) the classifier after RPN cannot distinguish vehicles and complex backgrounds well. In this study, an improved detection method based on Faster R-CNN is proposed in order to accomplish the two challenges mentioned above. Firstly, to improve the recall, we employ a hyper region proposal network (HRPN) to extract vehicle-like targets with a combination of hierarchical feature maps. Then, we replace the classifier after RPN by a cascade of boosted classifiers to verify the candidate regions, aiming at reducing false detection by negative example mining. We evaluate our method on the Munich vehicle dataset and the collected vehicle dataset, with improvements in accuracy and robustness compared to existing methods.
IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing | 2017
Zhipeng Deng; Hao Sun; Shilin Zhou; Juanping Zhao; Huanxin Zou
Vehicle detection in aerial images, being an interesting but challenging problem, plays an important role for a wide range of applications. Traditional methods are based on sliding-window search and handcrafted or shallow-learning-based features with heavy computational costs and limited representation power. Recently, deep learning algorithms, especially region-based convolutional neural networks (R-CNNs), have achieved state-of-the-art detection performance in computer vision. However, several challenges limit the applications of R-CNNs in vehicle detection from aerial images: 1) vehicles in large-scale aerial images are relatively small in size, and R-CNNs have poor localization performance with small objects; 2) R-CNNs are particularly designed for detecting the bounding box of the targets without extracting attributes; 3) manual annotation is generally expensive and the available manual annotation of vehicles for training R-CNNs are not sufficient in number. To address these problems, this paper proposes a fast and accurate vehicle detection framework. On one hand, to accurately extract vehicle-like targets, we developed an accurate-vehicle-proposal-network (AVPN) based on hyper feature map which combines hierarchical feature maps that are more accurate for small object detection. On the other hand, we propose a coupled R-CNN method, which combines an AVPN and a vehicle attribute learning network to extract the vehicles location and attributes simultaneously. For original large-scale aerial images with limited manual annotations, we use cropped image blocks for training with data augmentation to avoid overfitting. Comprehensive evaluations on the public Munich vehicle dataset and the collected vehicle dataset demonstrate the accuracy and effectiveness of the proposed method.
Remote Sensing | 2017
Tianyu Tang; Shilin Zhou; Zhipeng Deng; Lin Lei; Huanxin Zou
Vehicle detection with orientation estimation in aerial images has received widespread interest as it is important for intelligent traffic management. This is a challenging task, not only because of the complex background and relatively small size of the target, but also the various orientations of vehicles in aerial images captured from the top view. The existing methods for oriented vehicle detection need several post-processing steps to generate final detection results with orientation, which are not efficient enough. Moreover, they can only get discrete orientation information for each target. In this paper, we present an end-to-end single convolutional neural network to generate arbitrarily-oriented detection results directly. Our approach, named Oriented_SSD (Single Shot MultiBox Detector, SSD), uses a set of default boxes with various scales on each feature map location to produce detection bounding boxes. Meanwhile, offsets are predicted for each default box to better match the object shape, which contain the angle parameter for oriented bounding boxes’ generation. Evaluation results on the public DLR Vehicle Aerial dataset and Vehicle Detection in Aerial Imagery (VEDAI) dataset demonstrate that our method can detect both the location and orientation of the vehicle with high accuracy and fast speed. For test images in the DLR Vehicle Aerial dataset with a size of 5616 × 3744 , our method achieves 76.1% average precision (AP) and 78.7% correct direction classification at 5.17 s on an NVIDIA GTX-1060.
international conference on image vision and computing | 2017
Xiaolin Gu; Shilin Zhou; Lin Lei; Zhipeng Deng
For multi-object tracking in aerial videos which are acquired from moving cameras, the motion of cars are complicated by global camera movements and always unpredictable. To deal with such unexpected camera motion for online multi-vehicles tracking, structural motion context between objects has been used thanks to its robustness to camera motion. In this paper, we propose an effective data association method that exploits structural motion context in the presence of large camera motion. In addition, to further improve the robustness of algorithm against missing data due to the target being occluded behind other objects, an appearance context model is developed to represent appearance information of objects we need to track. The structural motion context and appearance context are then used to predict the location of the unobserved objects. Experimental results on VIVID datasets show the effectiveness of the proposed algorithm for multi-object tracking.
2017 International Workshop on Remote Sensing with Intelligent Processing (RSIP) | 2017
Zhipeng Deng; Lin Lei; Hao Sun; Huanxin Zou; Shilin Zhou; Juanping Zhao
Faster Region based convolutional neural networks (FRCN) has shown great success in object detection in recent years. However, its performance will degrade on densely packed objects in real remote sensing applications. To address this problem, an enhanced deep CNN based method is developed in this paper. Following the common pipeline of “CNN feature extraction + region proposal + Region classification”, our method is primarily based on the latest Residual Networks (ResNets) and consists of two sub-networks: an object proposal network and an object detection network. For detecting densely packed objects, the output of multi-scale layers are combined together to enhance the resolution of the feature maps. Our method is trained on the VHR-10 data set with limited samples and successfully tested on large-scale google earth images, such as aircraft boneyard or tank farm, containing a substantial number of densely packed objects.
2017 International Workshop on Remote Sensing with Intelligent Processing (RSIP) | 2017
Tianyu Tang; Zhipeng Deng; Shilin Zhou; Lin Lei; Huanxin Zou
Fast and accurate vehicle detection in unmanned aerial vehicle (UAV) images remains a challenge, due to its very high spatial resolution and very few annotations. Although numerous vehicle detection methods exist, most of them cannot achieve real-time detection for different scenes. Recently, deep learning algorithms has achieved fantastic detection performance in computer vision, especially regression based convolutional neural networks YOLOv2. Its good both at accuracy and speed, outperforming other state-of-the-art detection methods. This paper for the first time aims to investigate the use of YOLOv2 for vehicle detection in UAV images, as well as to explore the new method for data annotation. Our method starts with image annotation and data augmentation. CSK tracking method is used to help annotate vehicles in images captured from simple scenes. Subsequently, a regression based single convolutional neural network YOLOv2 is used to detect vehicles in UAV images. To evaluate our method, UAV video images were taken over several urban areas, and experiments were conducted on this dataset and Stanford Drone dataset. The experimental results have proven that our data preparation strategy is useful, and YOLOv2 is effective for real-time vehicle detection of UAV video images.
Image and Signal Processing for Remote Sensing XXI | 2015
Zhipeng Deng; Lin Lei; Shilin Zhou
Automatic image registration is a vital yet challenging task, particularly for non-rigid deformation images which are more complicated and common in remote sensing images, such as distorted UAV (unmanned aerial vehicle) images or scanning imaging images caused by flutter. Traditional non-rigid image registration methods are based on the correctly matched corresponding landmarks, which usually needs artificial markers. It is a rather challenging task to locate the accurate position of the points and get accurate homonymy point sets. In this paper, we proposed an automatic non-rigid image registration algorithm which mainly consists of three steps: To begin with, we introduce an automatic feature point extraction method based on non-linear scale space and uniform distribution strategy to extract the points which are uniform distributed along the edge of the image. Next, we propose a hybrid point matching algorithm using DaLI (Deformation and Light Invariant) descriptor and local affine invariant geometric constraint based on triangulation which is constructed by K-nearest neighbor algorithm. Based on the accurate homonymy point sets, the two images are registrated by the model of TPS (Thin Plate Spline). Our method is demonstrated by three deliberately designed experiments. The first two experiments are designed to evaluate the distribution of point set and the correctly matching rate on synthetic data and real data respectively. The last experiment is designed on the non-rigid deformation remote sensing images and the three experimental results demonstrate the accuracy, robustness, and efficiency of the proposed algorithm compared with other traditional methods.
world congress on intelligent control and automation | 2014
Zhipeng Deng; Lin Lei; Yi Hou; Shilin Zhou
Image registration is an important research topic in the field of computer vision. Traditional non-rigid image registration methods are based on the correctly matched corresponding landmarks, which usually needs artificial markers. It is a rather challenging and demanding task to locate the accurate position of the points and get the correspondance. In order to get the most correctly matched point set automatically, a new point matching method based on deformation invariant feature and local affine-invariant geometric constraint is proposed in this paper. Particularly mention should be the geodesic-intensity histogram (GIH), an interesting deformation invariant descriptor, which is introduced to describe the local feature of a point. In addition, the local affine invariant structure is employed as a geometric constraint. Therefore, an objective function that combines both local features and geometric constraint is formulated and computed by linear programming efficiently. Then, the correspondence is obtained and thin-plate spline (TPS) is employed for non-rigid registration. Our method is demonstrated with deliberately designed synthetic data and real data and the proposed method can better improve the accuracy as compared to the traditional registration techniques.
Isprs Journal of Photogrammetry and Remote Sensing | 2018
Zhipeng Deng; Hao Sun; Shilin Zhou; Juanping Zhao; Lin Lei; Huanxin Zou
international conference on image vision and computing | 2018
Xiaoli Zhao; Shilin Zhou; Lin Lei; Zhipeng Deng