Daniel J. Walter
University of Michigan
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Daniel J. Walter.
computer vision and pattern recognition | 2015
Zeynep Akata; Scott E. Reed; Daniel J. Walter; Honglak Lee; Bernt Schiele
Image classification has advanced significantly in recent years with the availability of large-scale image sets. However, fine-grained classification remains a major challenge due to the annotation cost of large numbers of fine-grained categories. This project shows that compelling classification performance can be achieved on such categories even without labeled training data. Given image and class embeddings, we learn a compatibility function such that matching embeddings are assigned a higher score than mismatching ones; zero-shot classification of an image proceeds by finding the label yielding the highest joint compatibility score. We use state-of-the-art image features and focus on different supervised attributes and unsupervised output embeddings either derived from hierarchies or learned from unlabeled text corpora. We establish a substantially improved state-of-the-art on the Animals with Attributes and Caltech-UCSD Birds datasets. Most encouragingly, we demonstrate that purely unsupervised output embeddings (learned from Wikipedia and improved with finegrained text) achieve compelling results, even outperforming the previous supervised state-of-the-art. By combining different output embeddings, we further improve results.Despite significant recent advances in image classification, fine-grained classification remains a challenge. In the present paper, we address the zero-shot and few-shot learning scenarios as obtaining labeled data is especially difficult for fine-grained classification tasks. First, we embed state-of-the-art image descriptors in a label embedding space using side information such as attributes. We argue that learning a joint embedding space, that maximizes the compatibility between the input and output embeddings, is highly effective for zero/few-shot learning. We show empirically that such embeddings significantly outperforms the current state-of-the-art methods on two challenging datasets (Caltech-UCSD Birds and Animals with Attributes). Second, to reduce the amount of costly manual attribute annotations, we use alternate output embeddings based on the word-vector representations, obtained from large text-corpora without any supervision. We report that such unsupervised embeddings achieve encouraging results, and lead to further improvements when combined with the supervised ones.
Nuclear Engineering and Design | 2016
Victor Petrov; Brian K. Kendrick; Daniel J. Walter; Annalisa Manera; Jeffrey Robert Secker
Annals of Nuclear Energy | 2015
Daniel J. Walter; Brian K. Kendrick; Victor Petrov; Annalisa Manera; Benjamin Collins; Thomas Downar
Mathematics and Computations, Supercomputing in Nuclear Applications and Monte Carlo International Conference, M and C+SNA+MC 2015 | 2015
Daniel J. Walter; Annalisa Manera
Unknown Journal | 2013
Brandon LaFleur; Daniel J. Walter; Annalisa Manera
LWR Fuel Performance Meeting, Top Fuel 2013 | 2013
Brian K. Kendrick; Victor Petrov; Daniel J. Walter; Annalisa Manera; Ben Collins; Thomas J. Downar; Jeffrey Seeker; Kenneth Belcourt
Progress in Nuclear Energy | 2016
Daniel J. Walter; Annalisa Manera
Physics of Reactors 2016: Unifying Theory and Experiments in the 21st Century, PHYSOR 2016 | 2016
Daniel J. Walter; Victor Petrov; Annalisa Manera; Brian K. Kendrick
Physics of Reactors 2016: Unifying Theory and Experiments in the 21st Century, PHYSOR 2016 | 2016
Daniel J. Walter; Victor Petrov; Annalisa Manera; Brian K. Kendrick
Annals of Nuclear Energy | 2016
Daniel J. Walter; Annalisa Manera