Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Nick C. Tang is active.

Publication


Featured researches published by Nick C. Tang.


IEEE Transactions on Circuits and Systems for Video Technology | 2009

Exemplar-Based Video Inpainting Without Ghost Shadow Artifacts by Maintaining Temporal Continuity

Timothy K. Shih; Nick C. Tang; Jenq-Neng Hwang

Image inpainting or image completion is the technique that automatically restores/completes removed areas in an image. When dealing with a similar problem in video, not only should a robust tracking algorithm be used, but the temporal continuity among video frames also needs to be taken into account, especially when the video has camera motions such as zooming and tilting. In this paper, we extend an exemplar-based image inpainting algorithm by incorporating an improved patch matching strategy for video inpainting. In our proposed algorithm, different motion segments with different temporal continuity call for different candidate patches, which are used to inpaint holes after a selected video object is tracked and removed. The proposed new video inpainting algorithm produces very few ldquoghost shadows,rdquo which were produced by most image inpainting algorithms directly applied on video. Our experiments use different types of videos, including cartoon, video from games, and video from digital camera with different camera motions. Our demonstration at http://member.mine.tku.edu.tw/www/T_CSVT/web/shows the promising results.


acm multimedia | 2006

Video inpainting and implant via diversified temporal continuations

Timothy K. Shih; Nick C. Tang; Wei-Sung Yeh; Ta-Jen Chen; Wonjun Lee

Recent interesting issues in video inpainting are defect removal and object removal. We take one more step to replace the removed objects in a video sequence by implanting objects from another video. Before implant, we improve an exemplar-based image inpainting algorithm by using a new patch matching strategy which incorporates edge properties. The data term used in a priority computation of candidate patches is also redefined. We take varieties of temporal continuations of foreground and background into consideration. A motion compensated inpainting procedure is then proposed. The inpainted video backgrounds are visually pleasant with smooth transitions. A simple tracking algorithm is then used to produce a foreground video, which is implanted into the inpainted background video. Our results are available at http://www.mine.tku.edu.tw/inpainting.


IEEE Transactions on Multimedia | 2011

Video Inpainting on Digitized Vintage Films via Maintaining Spatiotemporal Continuity

Nick C. Tang; Chiou-Ting Hsu; Chih-Wen Su; Timothy K. Shih; Hong-Yuan Mark Liao

Video inpainting is an important video enhancement technique used to facilitate the repair or editing of digital videos. It has been employed worldwide to transform cultural artifacts such as vintage videos/films into digital formats. However, the quality of such videos is usually very poor and often contain unstable luminance and damaged content. In this paper, we propose a video inpainting algorithm for repairing damaged content in digitized vintage films, focusing on maintaining good spatiotemporal continuity. The proposed algorithm utilizes two key techniques. Motion completion recovers missing motion information in damaged areas to maintain good temporal continuity. Frame completion repairs damaged frames to produce a visually pleasing video with good spatial continuity and stabilized luminance. We demonstrate the efficacy of the algorithm on different types of video clips.


computer vision and pattern recognition | 2014

Depth and Skeleton Associated Action Recognition without Online Accessible RGB-D Cameras

Yen-Yu Lin; Ju-Hsuan Hua; Nick C. Tang; Min-Hung Chen; Hong-Yuan Mark Liao

The recent advances in RGB-D cameras have allowed us to better solve increasingly complex computer vision tasks. However, modern RGB-D cameras are still restricted by the short effective distances. The limitation may make RGB-D cameras not online accessible in practice, and degrade their applicability. We propose an alternative scenario to address this problem, and illustrate it with the application to action recognition. We use Kinect to offline collect an auxiliary, multi-modal database, in which not only the RGB videos but also the depth maps and skeleton structures of actions of interest are available. Our approach aims to enhance action recognition in RGB videos by leveraging the extra database. Specifically, it optimizes a feature transformation, by which the actions to be recognized can be concisely reconstructed by entries in the auxiliary database. In this way, the inter-database variations are adapted. More importantly, each action can be augmented with additional depth and skeleton images retrieved from the auxiliary database. The proposed approach has been evaluated on three benchmarks of action recognition. The promising results manifest that the augmented depth and skeleton features can lead to remarkable boost in recognition accuracy.


IEEE Transactions on Image Processing | 2015

Cross-Camera Knowledge Transfer for Multiview People Counting

Nick C. Tang; Yen-Yu Lin; Ming-Fang Weng; Hong-Yuan Mark Liao

We present a novel two-pass framework for counting the number of people in an environment, where multiple cameras provide different views of the subjects. By exploiting the complementary information captured by the cameras, we can transfer knowledge between the cameras to address the difficulties of people counting and improve the performance. The contribution of this paper is threefold. First, normalizing the perspective of visual features and estimating the size of a crowd are highly correlated tasks. Hence, we treat them as a joint learning problem. The derived counting model is scalable and it provides more accurate results than existing approaches. Second, we introduce an algorithm that matches groups of pedestrians in images captured by different cameras. The results provide a common domain for knowledge transfer, so we can work with multiple cameras without worrying about their differences. Third, the proposed counting system is comprised of a pair of collaborative regressors. The first one determines the people count based on features extracted from intracamera visual information, whereas the second calculates the residual by considering the conflicts between intercamera predictions. The two regressors are elegantly coupled and provide an accurate people counting system. The results of experiments in various settings show that, overall, our approach outperforms comparable baseline methods. The significant performance improvement demonstrates the effectiveness of our two-pass regression framework.


international conference on multimedia and expo | 2007

Ghost Shadow Removal in Multi-Layered Video Inpaintinga

Timothy K. Shih; Nick C. Tang; Jenq-Neng Hwang

Image in-painting or image completion removes objects from a photo and automatically produces a visually pleasant result. However, to remove objects from a video, the resulting video may have ghost shadows even each individual frame is in-painted properly. We use motion estimation algorithm to separate objects and backgrounds into several layers. Objects in separated layers are in-painted from back to front layers, with a consideration of the temporal continuity of motion segments among different frames. The resulting video is visually more pleasant with most ghost shadows removed. Interested readers are welcome to look at our demonstration Website at http://www.mine.tku.edu.tw/demo.


ieee international conference on ubi-media computing | 2008

Application of inpainting technology to video restoration

Rong-Chi Chang; Nick C. Tang; Chia Cheng Chao

Video inpainting is the unique technique that can automatically restore damaged or partially removed image.It is also the tool for filling in the missing part in a video sequence. Exploration of more advanced concepts in video painting, this paper is intended to develop a new video algorithm based on temporal continuations and exemplar-based image in painting techniques. This proposed algorithm involves the elements of removing the moving objects on stationary and non-stationary background. Therefore, the related experiments include a wide variety of temporal continuations of foreground and background. According to the results of our experiments, a motion-compensated in painting procedure is successfully developed and it can be further extended to implant into the inpainted background video. Ultimately, the inpainted video backgrounds are visually pleasant with smooth transitions.


IEEE Transactions on Image Processing | 2015

Robust Action Recognition via Borrowing Information Across Video Modalities

Nick C. Tang; Yen-Yu Lin; Ju-Hsuan Hua; Shih-En Wei; Ming-Fang Weng; Hong-Yuan Mark Liao

The recent advances in imaging devices have opened the opportunity of better solving the tasks of video content analysis and understanding. Next-generation cameras, such as the depth or binocular cameras, capture diverse information, and complement the conventional 2D RGB cameras. Thus, investigating the yielded multimodal videos generally facilitates the accomplishment of related applications. However, the limitations of the emerging cameras, such as short effective distances, expensive costs, or long response time, degrade their applicability, and currently make these devices not online accessible in practical use. In this paper, we provide an alternative scenario to address this problem, and illustrate it with the task of recognizing human actions. In particular, we aim at improving the accuracy of action recognition in RGB videos with the aid of one additional RGB-D camera. Since RGB-D cameras, such as Kinect, are typically not applicable in a surveillance system due to its short effective distance, we instead offline collect a database, in which not only the RGB videos but also the depth maps and the skeleton data of actions are available jointly. The proposed approach can adapt the interdatabase variations, and activate the borrowing of visual knowledge across different video modalities. Each action to be recognized in RGB representation is then augmented with the borrowed depth and skeleton features. Our approach is comprehensively evaluated on five benchmark data sets of action recognition. The promising results manifest that the borrowed information leads to remarkable boost in recognition accuracy.


IEEE Transactions on Visualization and Computer Graphics | 2016

Court Reconstruction for Camera Calibration in Broadcast Basketball Videos

Pei-Chih Wen; Wei-Chih Cheng; Yu-Shuen Wang; Hung-Kuo Chu; Nick C. Tang; Hong-Yuan Mark Liao

We introduce a technique of calibrating camera motions in basketball videos. Our method particularly transforms player positions to standard basketball court coordinates and enables applications such as tactical analysis and semantic basketball video retrieval. To achieve a robust calibration, we reconstruct the panoramic basketball court from a video, followed by warping the panoramic court to a standard one. As opposed to previous approaches, which individually detect the court lines and corners of each video frame, our technique considers all video frames simultaneously to achieve calibration; hence, it is robust to illumination changes and player occlusions. To demonstrate the feasibility of our technique, we present a stroke-based system that allows users to retrieve basketball videos. Our system tracks player trajectories from broadcast basketball videos. It then rectifies the trajectories to a standard basketball court by using our camera calibration method. Consequently, users can apply stroke queries to indicate how the players move in gameplay during retrieval. The main advantage of this interface is an explicit query of basketball videos so that unwanted outcomes can be prevented. We show the results in Figs. 1, 7, 9, 10 and our accompanying video to exhibit the feasibility of our technique.We introduce a technique of calibrating camera motions in basketball videos. Our method particularly transforms player positions to standard basketball court coordinates and enables applications such as tactical analysis and semantic basketball video retrieval. To achieve a robust calibration, we reconstruct the panoramic basketball court from a video, followed by warping the panoramic court to a standard one. As opposed to previous approaches, which individually detect the court lines and corners of each video frame, our technique considers all video frames simultaneously to achieve calibration; hence, it is robust to illumination changes and player occlusions. To demonstrate the feasibility of our technique, we present a stroke-based system that allows users to retrieve basketball videos. Our system tracks player trajectories from broadcast basketball videos. It then rectifies the trajectories to a standard basketball court by using our camera calibration method. Consequently, users can apply stroke queries to indicate how the players move in gameplay during retrieval. The main advantage of this interface is an explicit query of basketball videos so that unwanted outcomes can be prevented. We show the results in Figs. 1, 7, 9, 10 and our accompanying video to exhibit the feasibility of our technique.


IEEE Transactions on Multimedia | 2014

Example-Based Human Motion Extrapolation and Motion Repairing Using Contour Manifold

Nick C. Tang; Chiou-Ting Hsu; Ming-Fang Weng; Tsung-Yi Lin; Hong-Yuan Mark Liao

We propose a human motion extrapolation algorithm that synthesizes new motions of a human object in a still image from a given reference motion sequence. The algorithm is implemented in two major steps: contour manifold construction and object motion synthesis. Contour manifold construction searches for low-dimensional manifolds that represent the temporal-domain deformation of the reference motion sequence. Since the derived manifolds capture the motion information of the reference sequence, the representation is more robust to variations in shape and size. With this compact representation, we can easily modify and manipulate human motions through interpolation or extrapolation in the contour manifold space. In the object motion synthesis step, the proposed algorithm generates a sequence of new shapes of the input human object in the contour manifold space and then renders the textures of those shapes to synthesize a new motion sequence. We demonstrate the efficacy of the algorithm on different types of practical applications, namely, motion extrapolation and motion repair.

Collaboration


Dive into the Nick C. Tang's collaboration.

Top Co-Authors

Avatar

Timothy K. Shih

National Central University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Ming-Fang Weng

National Taiwan University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Chiou-Ting Hsu

National Tsing Hua University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Chun-Hong Huang

Lunghwa University of Science and Technology

View shared research outputs
Researchain Logo
Decentralizing Knowledge