Quanshi Zhang
University of California, Los Angeles
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Quanshi Zhang.
international conference on computer vision | 2015
Quanshi Zhang; Ying Nian Wu; Song-Chun Zhu
This paper reformulates the theory of graph mining on the technical basis of graph matching, and extends its scope of applications to computer vision. Given a set of attributed relational graphs (ARGs), we propose to use a hierarchical And-Or Graph (AoG) to model the pattern of maximal-size common subgraphs embedded in the ARGs, and we develop a general method to mine the AoG model from the unlabeled ARGs. This method provides a general solution to the problem of mining hierarchical models from unannotated visual data without exhaustive search of objects. We apply our method to RGB/RGB-D images and videos to demonstrate its generality and the wide range of applicability. The code will be available at https://sites.google.com/site/quanshizhang/mining-and-or-graphs.
Journal of Zhejiang University Science C | 2018
Quanshi Zhang; Song-Chun Zhu
This paper reviews recent studies in understanding neural-network representations and learning neural networks with interpretable/disentangled middle-layer representations. Although deep neural networks have exhibited superior performance in various tasks, interpretability is always Achilles’ heel of deep neural networks. At present, deep neural networks obtain high discrimination power at the cost of a low interpretability of their black-box representations. We believe that high model interpretability may help people break several bottlenecks of deep learning, e.g., learning from a few annotations, learning via human–computer communications at the semantic level, and semantically debugging network representations. We focus on convolutional neural networks (CNNs), and revisit the visualization of CNN representations, methods of diagnosing representations of pre-trained CNNs, approaches for disentangling pre-trained CNN representations, learning of CNNs with disentangled representations, and middle-to-end learning based on model interpretability. Finally, we discuss prospective trends in explainable artificial intelligence.
computer vision and pattern recognition | 2017
Quanshi Zhang; Ruiming Cao; Ying Nian Wu; Song-Chun Zhu
Given a convolutional neural network (CNN) that is pre-trained for object classification, this paper proposes to use active question-answering to semanticize neural patterns in conv-layers of the CNN and mine part concepts. For each part concept, we mine neural patterns in the pre-trained CNN, which are related to the target part, and use these patterns to construct an And-Or graph (AOG) to represent a four-layer semantic hierarchy of the part. As an interpretable model, the AOG associates different CNN units with different explicit object parts. We use an active human-computer communication to incrementally grow such an AOG on the pre-trained CNN as follows. We allow the computer to actively identify objects, whose neural patterns cannot be explained by the current AOG. Then, the computer asks human about the unexplained objects, and uses the answers to automatically discover certain CNN patterns corresponding to the missing knowledge. We incrementally grow the AOG to encode new knowledge discovered during the active-learning process. In experiments, our method exhibits high learning efficiency. Our method uses about 1/6–1/3 of the part annotations for training, but achieves similar or better part-localization performance than fast-RCNN methods.
Computer Vision and Image Understanding | 2018
Quanshi Zhang; Ying Nian Wu; Song-Chun Zhu; Hao Zhang
Abstract This paper presents a cost-sensitive active Question-Answering (QA) framework for learning a nine-layer And-Or graph (AOG) from web images. The AOG explicitly represents object categories, poses/viewpoints, parts, and detailed structures within the parts in a compositional hierarchy. The QA framework is designed to minimize an overall risk, which trades off the loss and query costs. The loss is defined for nodes in all layers of the AOG, including the generative loss (measuring the likelihood of the images) and the discriminative loss (measuring the fitness to human answers). The cost comprises both the human labor of answering questions and the computational cost of model learning. The cost-sensitive QA framework iteratively selects different storylines of questions to update different nodes in the AOG. Experiments showed that our method required much less human supervision (e.g. labeling parts on 3–10 training objects for each category) and achieved better performance than baseline methods.
computer vision and pattern recognition | 2018
Quanshi Zhang; Ying Nian Wu; Song-Chun Zhu
national conference on artificial intelligence | 2017
Quanshi Zhang; Ruiming Cao; Ying Nian Wu; Song-Chun Zhu
national conference on artificial intelligence | 2018
Quanshi Zhang; Ruiming Cao; Feng Shi; Ying Nian Wu; Song-Chun Zhu
arXiv: Computer Vision and Pattern Recognition | 2018
Quanshi Zhang; Yu Yang; Ying Nian Wu; Song-Chun Zhu
arXiv: Computer Vision and Pattern Recognition | 2017
Quanshi Zhang; Ruiming Cao; Shengming Zhang; Mark Edmonds; Ying Nian Wu; Song-Chun Zhu
Archive | 2017
Quanshi Zhang; Ying Nian Wu; Song-Chun Zhu