Jinseok Nam
Technische Universität Darmstadt
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Jinseok Nam.
european conference on machine learning | 2014
Jinseok Nam; Jungi Kim; Eneldo Loza Mencía; Iryna Gurevych; Johannes Fürnkranz
Neural networks have recently been proposed for multi-label classification because they are able to capture and model label dependencies in the output layer. In this work, we investigate limitations of BP-MLL, a neural network (NN) architecture that aims at minimizing pairwise ranking error. Instead, we propose to use a comparably simple NN approach with recently proposed learning techniques for large-scale multi-label text classification tasks. In particular, we show that BP-MLLs ranking loss minimization can be efficiently and effectively replaced with the commonly used cross entropy error function, and demonstrate that several advances in neural network training that have been developed in the realm of deep learning can be effectively employed in this setting. Our experimental results show that simple NN models equipped with advanced techniques such as rectified linear units, dropout, and AdaGrad perform as well as or even outperform state-of-the-art approaches on six large-scale textual datasets with diverse characteristics.
european conference on machine learning | 2015
Jinseok Nam; Eneldo Loza Mencía; Hyunwoo Kim; Johannes Fürnkranz
An important problem in multi-label classification is to capture label patterns or underlying structures that have an impact on such patterns. One way of learning underlying structures over labels is to project both instances and labels into the same space where an instance and its relevant labels tend to have similar representations. In this paper, we present a novel method to learn a joint space of instances and labels by leveraging a hierarchy of labels. We also present an efficient method for pretraining vector representations of labels, namely label embeddings, from large amounts of label co-occurrence patterns and hierarchical structures of labels. This approach also allows us to make predictions on labels that have not been seen during training. We empirically show that the use of pretrained label embeddings allows us to obtain higher accuracies on unseen labels even when the number of labels are quite large. Our experimental results also demonstrate qualitatively that the proposed method is able to learn regularities among labels by exploiting a label hierarchy as well as label co-occurrences.
north american chapter of the association for computational linguistics | 2016
Zheng Ma; Jinseok Nam; Karsten Weihe
In this paper, we introduce a novel approach to sentiment polarity classification of citations, which integrates data about the authors’ reputation. More specifically, our method extends the h-index with citation polarities and utilizes it in sentiment classification of citation sentences. Our computational results show that our method yields significant improvement in terms of classification performance.
european conference on machine learning | 2015
Jinseok Nam; Eneldo Loza Mencía; Hyunwoo Kim; Johannes Fürnkranz
Erratum to: Chapter 7 in: A. Appice et al. (Eds.) Machine Learning and Knowledge Discovery in Databases DOI: 10.1007/978-3-319-23528-8_7
international conference on computational linguistics | 2012
Jungi Kim; Jinseok Nam; Iryna Gurevych
national conference on artificial intelligence | 2016
Jinseok Nam; Eneldo Loza Mencía; Johannes Fürnkranz
neural information processing systems | 2017
Jinseok Nam; Eneldo Loza Mencía; Hyunwoo Kim; Johannes Fürnkranz
language resources and evaluation | 2016
Eneldo Loza Mencía; Gerard de Melo; Jinseok Nam
Archive | 2013
Eneldo Loza Mencía; Jinseok Nam; Dong-Hyun Lee
KONVENS | 2014
Jinseok Nam; Christian Kirschner; Zheng Ma; Nicolai Erbs; Susanne Neumann; Daniela Oelke; Steffen Remus; Chris Biemann; Judith Eckle-Kohler; Johannes Fürnkranz; Iryna Gurevych; Marc Rittberger; Karsten Weihe