Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Feiyue Huang is active.

Publication


Featured researches published by Feiyue Huang.


international conference on computer graphics and interactive techniques | 2014

Data-driven face cartoon stylization

Yong Zhang; Weiming Dong; Oliver Deussen; Feiyue Huang; Ke Li; Bao-Gang Hu

This paper presents a data-driven framework for generating cartoon-like facial representations from a given portrait image. We solve our problem by an optimization that simultaneously considers a desired artistic style, image-cartoon relationships of facial components as well as automatic adjustment of the image composition. The stylization operation consists of two steps: a face parsing step to localize and extract facial components from the input image; a cartoon generation step to cartoonize the face according to the extracted information. The components of the cartoon are assembled from a database of stylized facial components. Quantifying the similarity between facial components of input and cartoon is done by image feature matching. We incorporate prior knowledge about photo-cartoon relationships and the optimal composition of cartoon facial components extracted from a set of cartoon faces to maintain a natural and attractive look of the results.


IEEE Transactions on Visualization and Computer Graphics | 2016

Measuring and Predicting Visual Importance of Similar Objects

Yan Kong; Weiming Dong; Xing Mei; Chongyang Ma; Tong-Yee Lee; Siwei Lyu; Feiyue Huang; Xiaopeng Zhang

Similar objects are ubiquitous and abundant in both natural and artificial scenes. Determining the visual importance of several similar objects in a complex photograph is a challenge for image understanding algorithms. This study aims to define the importance of similar objects in an image and to develop a method that can select the most important instances for an input image from multiple similar objects. This task is challenging because multiple objects must be compared without adequate semantic information. This challenge is addressed by building an image database and designing an interactive system to measure object importance from human observers. This ground truth is used to define a range of features related to the visual importance of similar objects. Then, these features are used in learning-to-rank and random forest to rank similar objects in an image. Importance predictions were validated on 5,922 objects. The most important objects can be identified automatically. The factors related to composition (e.g., size, location, and overlap) are particularly informative, although clarity and color contrast are also important. We demonstrate the usefulness of similar object importance on various applications, including image retargeting, image compression, image re-attentionizing, image admixture, and manipulation of blindness images.


Pattern Recognition | 2017

Centroid-aware local discriminative metric learning in speaker verification

Kekai Sheng; Weiming Dong; Wei Li; Joseph Razik; Feiyue Huang; Bao-Gang Hu

Abstract We propose a new mechanism to pave the way for efficient learning against class-imbalance and improve representation of identity vector (i-vector) in automatic speaker verification (ASV). The insight is to effectively exploit the inherent structure within ASV corpus — centroid priori. In particular: (1) to ensure learning efficiency against class-imbalance, the centroid-aware balanced boosting sampling is proposed to collect balanced mini-batch; (2) to strengthen local discriminative modeling on the mini-batches, neighborhood component analysis (NCA) and magnet loss (MNL) are adopted in ASV-specific modifications. The integration creates adaptive NCA (AdaNCA) and linear MNL (LMNL). Numerical results show that LMNL is a competitive candidate for low-dimensional projection on i-vector (EER=3.84% on SRE2008, EER=1.81% on SRE2010), enjoying competitive edge over linear discriminant analysis (LDA). AdaNCA (EER=4.03% on SRE2008, EER=2.05% on SRE2010) also performs well. Furthermore, to facilitate the future study on boosting sampling, connections between boosting sampling, hinge loss and data augmentation have been established, which help understand the behavior of boosting sampling further.


IEEE Transactions on Image Processing | 2017

Data-Driven Synthesis of Cartoon Faces Using Different Styles

Yong Zhang; Weiming Dong; Chongyang Ma; Xing Mei; Ke Li; Feiyue Huang; Bao-Gang Hu; Oliver Deussen

This paper presents a data-driven approach for automatically generating cartoon faces in different styles from a given portrait image. Our stylization pipeline consists of two steps: an offline analysis step to learn about how to select and compose facial components from the databases; a runtime synthesis step to generate the cartoon face by assembling parts from a database of stylized facial components. We propose an optimization framework that, for a given artistic style, simultaneously considers the desired image-cartoon relationships of the facial components and a proper adjustment of the image composition. We measure the similarity between facial components of the input image and our cartoon database via image feature matching, and introduce a probabilistic framework for modeling the relationships between cartoon facial components. We incorporate prior knowledge about image-cartoon relationships and the optimal composition of facial components extracted from a set of cartoon faces to maintain a natural, consistent, and attractive look of the results. We demonstrate generality and robustness of our approach by applying it to a variety of portrait images and compare our output with stylized results created by artists via a comprehensive user study.


Computer Graphics Forum | 2015

Evaluating the Quality of Face Alignment without Ground Truth

Kekai Sheng; Weiming Dong; Yan Kong; Xing Mei; Jilin Li; Chengjie Wang; Feiyue Huang; Bao-Gang Hu

The study of face alignment has been an area of intense research in computer vision, with its achievements widely used in computer graphics applications. The performance of various face alignment methods is often image‐dependent or somewhat random because of their own strategy. This study aims to develop a method that can select an input image with good face alignment results from many results produced by a single method or multiple ones. The task is challenging because different face alignment results need to be evaluated without any ground truth. This study addresses this problem by designing a feasible feature extraction scheme to measure the quality of face alignment results. The feature is then used in various machine learning algorithms to rank different face alignment results. Our experiments show that our method is promising for ranking face alignment results and is able to pick good face alignment results, which can enhance the overall performance of a face alignment method with a random strategy. We demonstrate the usefulness of our ranking‐enhanced face alignment algorithm in two practical applications: face cartoon stylization and digital face makeup.


international conference on computer graphics and interactive techniques | 2014

Animated construction of ink-wash paintings

Fan Tang; Yiping Meng; Weiming Dong; Xing Mei; Feiyue Huang; Xiaopeng Zhang

Chinese ink-wash painting is a unique and fascinating form of art. Revealing the drawing process of an ink-wash artwork is visually intriguing and useful for training of painting skills. Recording the sequences of strokes by camera during the creation of paintings is either laborious or even unavailable especially for some ancient paintings. In this work, we propose an effective solution for estimating an order given a number of 2D strokes extracted from ink-wash painting images (Figure 1). The understanding of art varies with people, so our objective is finding a reasonable solution that is plausible to human eyes, instead of searching for the best drawing sequence. We formulate the ordering of strokes as breadth first search on a graph encoding both the individual features of strokes (e.g., size, shade, position, etc.) and their spatial relationships. Compared with the work of recovering the drawing order of line artworks [Fu et al. 2011], our problem is more difficult since the art form of ink-wash painting is more complicated than line drawing.


acm multimedia | 2018

Dense Auto-Encoder Hashing for Robust Cross-Modality Retrieval

Hong Liu; Mingbao Lin; Shengchuan Zhang; Yongjian Wu; Feiyue Huang; Rongrong Ji

Cross-modality retrieval has been widely studied, which aims to search images as response to text queries or vice versa. When faced with large-scale dataset, cross-modality hashing serves as an efficient and effective solution, which learns binary codes to approximate the cross-modality similarity in the Hamming space. Most recent cross-modality hashing schemes focus on learning the hash functions from data instances with fully modalities. However, how to learn robust binary codes when facing incomplete modality (i.e., with one modality missed or partially observed), is left unexploited, which however widely occurs in real-world applications. In this paper, we propose a novel cross-modality hashing, termed Dense Auto-encoder Hashing (DAH), which can explicitly impute the missed modality and produce robust binary codes by leveraging the relatedness among different modalities. To that effect, we propose a novel Dense Auto-encoder Network (DAN) to impute the missing modalities, which densely connects each layer to every other layer in a feed-forward fashion. For each layer, a noisy auto-encoder block is designed to calculate the residue between the current prediction and original data. Finally, a hash-layer is added to the end of DAN, which serves as a special binary encoder model to deal with the incomplete modality input. Quantitative experiments on three cross-modality visual search benchmarks, i.e., the Wiki, NUS-WIDE, and FLICKR-25K, have shown that the proposed DAH has superior performance over the state-of-the-art approaches.


acm multimedia | 2018

Attention-based Multi-Patch Aggregation for Image Aesthetic Assessment

Kekai Sheng; Weiming Dong; Chongyang Ma; Xing Mei; Feiyue Huang; Bao-Gang Hu

Aggregation structures with explicit information, such as image attributes and scene semantics, are effective and popular for intelligent systems for assessing aesthetics of visual data. However, useful information may not be available due to the high cost of manual annotation and expert design. In this paper, we present a novel multi-patch (MP) aggregation method for image aesthetic assessment. Different from state-of-the-art methods, which augment an MP aggregation network with various visual attributes, we train the model in an end-to-end manner with aesthetic labels only (i.e., aesthetically positive or negative). We achieve the goal by resorting to an attention-based mechanism that adaptively adjusts the weight of each patch during the training process to improve learning efficiency. In addition, we propose a set of objectives with three typical attention mechanisms (i.e., average, minimum, and adaptive) and evaluate their effectiveness on the Aesthetic Visual Analysis (AVA) benchmark. Numerical results show that our approach outperforms existing methods by a large margin. We further verify the effectiveness of the proposed attention-based objectives via ablation studies and shed light on the design of aesthetic assessment systems.


Multimedia Tools and Applications | 2018

Joint face alignment and segmentation via deep multi-task learning

Yucheng Zhao; Fan Tang; Weiming Dong; Feiyue Huang; Xiaopeng Zhang

Face alignment and segmentation are challenging problems which have been extensively studied in the field of multimedia. These two tasks are closely related and their learning processes are supposed to benefit each other. Hence, we present a joint multi-task learning algorithm for both face alignment and segmentation using deep convolutional neural network (CNN). The proposed multi-task learning approach allows CNN model to simultaneously share visual knowledge between different tasks. With a carefully designed refinement residual module, the cross-layer features are fused in a collaborative manner. To the best of our knowledge, this is the first time that face alignment and segmentation are learned together via deep multi-task learning. Our experiments show that learning these two related tasks simultaneously builds a synergy between them, improves the performance of each individual task, and rivals recent approaches. Furthermore, we demonstrate the effectiveness of our model in two practical applications: virtual makeup and face swap.


international conference on computer graphics and interactive techniques | 2017

Content-based measure of image set diversity

Xingjia Pan; Juntao Ye; Fan Tang; Weiming Dong; Feiyue Huang; Xiaopeng Zhang

With the ubiquity of digital cameras and the growth of social media population, people share and upload millions of photos per day. To effectively manage or explore a series of shots of different scenes, people often hope to pick a few representative examples with various contents, in order to fastly have a global view of the whole set. Thus, it is important considering to evaluate the diversity of an image set.

Collaboration


Dive into the Feiyue Huang's collaboration.

Top Co-Authors

Avatar

Weiming Dong

Chinese Academy of Sciences

View shared research outputs
Top Co-Authors

Avatar

Bao-Gang Hu

Chinese Academy of Sciences

View shared research outputs
Top Co-Authors

Avatar

Xing Mei

Chinese Academy of Sciences

View shared research outputs
Top Co-Authors

Avatar

Xiaopeng Zhang

Chinese Academy of Sciences

View shared research outputs
Top Co-Authors

Avatar

Fan Tang

Chinese Academy of Sciences

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Kekai Sheng

Chinese Academy of Sciences

View shared research outputs
Top Co-Authors

Avatar

Yong Zhang

Chinese Academy of Sciences

View shared research outputs
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge