Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Congcong Li is active.

Publication


Featured researches published by Congcong Li.


IEEE Journal of Selected Topics in Signal Processing | 2009

Aesthetic Visual Quality Assessment of Paintings

Congcong Li; Tsuhan Chen

This paper aims to evaluate the aesthetic visual quality of a special type of visual media: digital images of paintings. Assessing the aesthetic visual quality of paintings can be considered a highly subjective task. However, to some extent, certain paintings are believed, by consensus, to have higher aesthetic quality than others. In this paper, we treat this challenge as a machine learning problem, in order to evaluate the aesthetic quality of paintings based on their visual content. We design a group of methods to extract features to represent both the global characteristics and local characteristics of a painting. Inspiration for these features comes from our prior knowledge in art and a questionnaire survey we conducted to study factors that affect humans judgments. We collect painting images and ask human subjects to score them. These paintings are then used for both training and testing in our experiments. Experimental results show that the proposed work can classify high-quality and low-quality paintings with performance comparable to humans. This work provides a machine learning scheme for the research of exploring the relationship between aesthetic perceptions of human and the computational visual features extracted from paintings.


IEEE Transactions on Pattern Analysis and Machine Intelligence | 2012

Toward Holistic Scene Understanding: Feedback Enabled Cascaded Classification Models

Congcong Li; Adarsh Kowdle; Ashutosh Saxena; Tsuhan Chen

Scene understanding includes many related subtasks, such as scene categorization, depth estimation, object detection, etc. Each of these subtasks is often notoriously hard, and state-of-the-art classifiers already exist for many of them. These classifiers operate on the same raw image and provide correlated outputs. It is desirable to have an algorithm that can capture such correlation without requiring any changes to the inner workings of any classifier. We propose Feedback Enabled Cascaded Classification Models (FE-CCM), that jointly optimizes all the subtasks while requiring only a “black box” interface to the original classifier for each subtask. We use a two-layer cascade of classifiers, which are repeated instantiations of the original ones, with the output of the first layer fed into the second layer as input. Our training method involves a feedback step that allows later classifiers to provide earlier classifiers information about which error modes to focus on. We show that our method significantly improves performance in all the subtasks in the domain of scene understanding, where we consider depth estimation, scene categorization, event categorization, object detection, geometric labeling, and saliency detection. Our method also improves performance in two robotic applications: an object-grasping robot and an object-finding robot.


acm multimedia | 2010

Towards aesthetics: a photo quality assessment and photo selection system

Congcong Li; Alexander C. Loui; Tsuhan Chen

Automatic photo quality assessment and selection systems are helpful for managing the large mount of consumer photos. In this paper, we present such a system based on evaluating the aesthetic quality of consumer photos. The proposed system focuses on photos with faces, which constitute an important part of consumer photo albums. The system has three contributions: 1) We propose an aesthetics-based photo assessment algorithm, by considering different aesthetics-related factors, including the technical characteristics of the photo and the specific features related to faces; 2) Based on the aesthetic measurement, we propose a cropping-based photo editing algorithm, which differs from prior works by eliminating unimportant faces before optimizing photo composition; 3) We also incorporate the aesthetic evaluation with other metrics to select quintessential photos for a large collection of photos. The entire system is delivered by a web interface, which allows users to submit images or albums, and returns promising results for photo evaluation, editing recommendation, and photo selection.


international conference on image processing | 2010

Aesthetic quality assessment of consumer photos with faces

Congcong Li; Andrew C. Gallagher; Alexander C. Loui; Tsuhan Chen

Automatically assessing the subjective quality of a photo is a challenging area in visual computing. Previous works study the aesthetic quality assessment on a general set of photos regardless of the photos content and mainly use features extracted from the entire image. In this work, we focus on a specific genre of photos: consumer photos with faces. This group of photos constitutes an important part of consumer photo collections. We first conduct an online study on Mechanical Turk to collect ground-truth and subjective opinions for a database of consumer photos with faces. We then extract technical features, perceptual features, and social relationship features to represent the aesthetic quality of a photo, by focusing on face-related regions. Experiments show that our features perform well for categorizing or predicting the aesthetic quality.


computer vision and pattern recognition | 2012

Automatic discovery of groups of objects for scene understanding

Congcong Li; Devi Parikh; Tsuhan Chen

Objects in scenes interact with each other in complex ways. A key observation is that these interactions manifest themselves as predictable visual patterns in the image. Discovering and detecting these structured patterns is an important step towards deeper scene understanding. It goes beyond using either individual objects or the scene as a whole as the semantic unit. In this work, we promote groups of objects. They are high-order composites of objects that demonstrate consistent spatial, scale, and viewpoint interactions with each other. These groups of objects are likely to correspond to a specific layout of the scene. They can thus provide cues for the scene category and can also prime the likely locations of other objects in the scene. It is not feasible to manually generate a list of all possible groupings of objects we find in our visual world. Hence, we propose an algorithm that automatically discovers groups of arbitrary numbers of participating objects from a collection of images labeled with object categories. Our approach builds a 4-dimensional transform space of location, scale and viewpoint, and efficiently identifies all recurring compositions of objects across images. We then model the discovered groups of objects using the deformable parts-based model. Our experiments on a variety of datasets show that using groups of objects can significantly boost the performance of object detection and scene categorization.


international conference on computer vision | 2011

Extracting adaptive contextual cues from unlabeled regions

Congcong Li; Devi Parikh; Tsuhan Chen

Existing approaches to contextual reasoning for enhanced object detection typically utilize other labeled categories in the images to provide contextual information. As a consequence, they inadvertently commit to the granularity of information implicit in the labels. Moreover, large portions of the images may not belong to any of the manually-chosen categories, and these unlabeled regions are typically neglected. In this paper, we overcome both these drawbacks and propose a contextual cue that exploits unlabeled regions in images. Our approach adaptively determines the granularity (scene, inter-object, intra-object, etc.) at which contextual information is captured. In order to extract the proposed contextual cue, we consider a scene to be a structured configuration of objects and regions; just as an object is a composition of parts. We thus learn our proposed “contextual meta-objects” using any off-the-shelf object detector, which makes our proposed cue widely accessible to the community. Our results show that incorporating our proposed cue provides a relative improvement of 12% over a state-of-the-art object detector on the challenging PASCAL dataset.


Neurocomputing | 2007

A face and fingerprint identity authentication system based on multi-route detection

Jun Zhou; Guangda Su; Chunhong Jiang; Yafeng Deng; Congcong Li

This paper presents a novel face and fingerprint identity authentication system based on multi-route detection. To exclude the influence of pose on face recognition, a multi-route detection module is adopted, with parallel processing technology to speed the authentication process. Parallel processing technology is used to speed the authentication process. Fusion of face and fingerprint by support vector machine (SVM) which introduced a new normalization method improved the authentication accuracy. A new concept of optimal two-dimensional face is proposed to improve the performance of the dynamic face authentication system. Experiments on a real database showed that the proposed system achieved better results compared with face-only or fingerprint-only system.


international conference on image processing | 2009

Motion-focusing key frame extraction and video summarization for lane surveillance system

Congcong Li; Yi-Ta Wu; Shiaw-Shian Yu; Tsuhan Chen

This paper proposes a motion-focusing method to extract key frames and generate summarization synchronously for surveillance videos. Within each pre-segmented video shot, the proposed method focuses on one constant-speed motion and aligns the video frames by fixing this focused motion into a static situation. According to the relative motion theory, the other objects in the video are moving relatively to the selected kind of motion. This method finally generates a summary image containing all moving objects and embedded with spatial and motional information, together with key frames to provide details corresponding to the regions of interest in the summary image. We apply this method to the lane surveillance system and the results provide us a new way to understand the video efficiently.


international conference on machine learning and cybernetics | 2005

A high performance face recognition system based on a huge face database

Meng; Guangda Su; Congcong Li; Bo Fu; Jun Zhou

This paper presents a high performance face recognition system, in which the face database has a large amount of 2.5 million faces. Huge as the face database is, the recognition processes in ordinary ways meets with great difficulties: the identification rate of most algorithms may decline significantly; meanwhile, querying on a large-scale database may be quite time-consuming. In our system, a special distributed parallel architecture is proposed to speed up the computation. Furthermore, a multimodal part face recognition method based on principal component analysis (MMP-PCA) is adopted to perform the recognition task, and the MMX technology is introduced to accelerate the matching procedure. Practical results prove that this system has an excellent performance in recognition: when searching among 2,560,000 faces on 6 PC servers with Xeon 2.4 GHz CPU, the querying time is only 1.094 s and the identification rate is above 85% in most cases. Moreover, the greatest advantage of this system is not only increasing recognition speed but also breaking the upper limit of face data capacity. Consequently, the face data capability of this system can be extended to an arbitrarily large amount.


international conference on robotics and automation | 2011

FeCCM for scene understanding: Helping the robot to learn multiple tasks

Congcong Li; Tp Wong; Norris Xu; Ashutosh Saxena

Helping a robot to understand a scene can include many sub-tasks, such as scene categorization, object detection, geometric labeling, etc. Each sub-task is notoriously hard, and state-of-art classifiers exist for many sub-tasks. It is desirable to have an algorithm that can capture such correlation without requiring to make any changes to the inner workings of any classifier, and therefore make the perception for a robot better. We have recently proposed a generic model (Feedback Enabled Cascaded Classification Model) that enables us to easily take state-of-art classifiers as black-boxes and improve performance. In this video, we show that we can use our FeCCM model to quickly combine existing classifiers for various sub-tasks, and build a shoe finder robot in a day. The video shows our robot using FeCCM to find a shoe on request.

Collaboration


Dive into the Congcong Li's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Devi Parikh

Georgia Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Shiaw-Shian Yu

Industrial Technology Research Institute

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge