Yujia Li
University of Toronto
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Yujia Li.
computer vision and pattern recognition | 2013
Yujia Li; Daniel Tarlow; Richard S. Zemel
When modeling structured outputs such as image segmentations, prediction can be improved by accurately modeling structure present in the labels. A key challenge is developing tractable models that are able to capture complex high level structure like shape. In this work, we study the learning of a general class of pattern-like high order potential, which we call Compositional High Order Pattern Potentials (CHOPPs). We show that CHOPPs include the linear deviation pattern potentials of Rother et al. [26] and also Restricted Boltzmann Machines (RBMs), we also establish the near equivalence of these two models. Experimentally, we show that performance is affected significantly by the degree of variability present in the datasets, and we define a quantitative variability measure to aid in studying this. We then improve CHOPPs performance in high variability datasets with two primary contributions: (a) developing a loss-sensitive joint learning procedure, so that internal pattern parameters can be learned in conjunction with other model potentials to minimize expected loss, and (b) learning an image-dependent mapping that encourages or inhibits patterns depending on image features. We also explore varying how multiple patterns are composed, and learning convolutional patterns. Quantitative results on challenging highly variable datasets show that the joint learning and image-dependent high order potentials can improve performance.
international conference on acoustics, speech, and signal processing | 2015
Yujia Li; Kaisheng Yao; Geoffrey Zweig
This paper presents a novel interactive method for recognizing handwritten words, using the inertial sensor data available on smart watches. The goal is to allow the user to write with a finger, and use the smart watch sensor signals to infer what the user has written. Past work has exploited the similarity of handwriting recognition to speech recognition in order to deploy HMM based methods. In contrast to speech recognition, however, in our scenario, the user can see the individual letters that are recognized on a sequential basis, and provide feedback or corrections after each letter. In this paper, we exploit this key difference to improve the input mechanism over a classical source-channel model. For a small increase in the amount of time required to input a word, we improve recognition accuracy from 59.6% to 91.4% with an implicit feedback mechanism, and to 100% with an explicit feedback mechanism.
arXiv: Learning | 2016
Yujia Li; Daniel Tarlow; Marc Brockschmidt; Richard S. Zemel
international conference on machine learning | 2015
Yujia Li; Kevin Swersky; Richard S. Zemel
international conference on learning representations | 2016
Christos Louizos; Kevin Swersky; Yujia Li; Max Welling; Richard S. Zemel
neural information processing systems | 2016
Wenjie Luo; Yujia Li; Raquel Urtasun; Richard S. Zemel
neural information processing systems | 2017
Sébastien Racanière; Theophane Weber; David P. Reichert; Lars Buesing; Arthur Guez; Danilo Jimenez Rezende; Adrià Puigdomènech Badia; Oriol Vinyals; Nicolas Heess; Yujia Li; Razvan Pascanu; Peter Battaglia; Demis Hassabis; David Silver; Daan Wierstra
international joint conference on artificial intelligence | 2013
Xuetao Ding; Xiaoming Jin; Yujia Li; Lianghao Li
international conference on machine learning | 2014
Yujia Li; Richard S. Zemel
international conference on learning representations | 2018
Yujia Li; Oriol Vinyals; Chris Dyer; Razvan Pascanu; Peter Battaglia
Collaboration
Dive into the Yujia Li's collaboration.
Dalle Molle Institute for Artificial Intelligence Research
View shared research outputs