Bryan Kisiel
Carnegie Mellon University
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Bryan Kisiel.
international acm sigir conference on research and development in information retrieval | 2003
Yiming Yang; Jian Zhang; Bryan Kisiel
Real-world applications of text categorization often require a system to deal with tens of thousands of categories defined over a large taxonomy. This paper addresses the problem with respect to a set of popular algorithms in text categorization, including Support Vector Machines, k-nearest neighbor, ridge regression, linear least square fit and logistic regression. By providing a formal analysis of the computational complexity of each classification method, followed by an investigation on the usage of different classifiers in a hierarchical setting of categorization, we show how the scalability of a method depends on the topology of the hierarchy and the category distributions. In addition, we are able to obtain tight bounds for the complexities by using the power law to approximate category distributions over a hierarchy. Experiments with kNN and SVM classifiers on the OHSUMED corpus are reported on, as concrete examples.
international acm sigir conference on research and development in information retrieval | 2005
Yiming Yang; Shinjae Yoo; Jian Zhang; Bryan Kisiel
This paper reports a cross-benchmark evaluation of regularized logistic regression (LR) and incremental Rocchio for adaptive filtering. Using four corpora from the Topic Detection and Tracking (TDT) forum and the Text Retrieval Conferences (TREC) we evaluated these methods with non-stationary topics at various granularity levels, and measured performance with different utility settings. We found that LR performs strongly and robustly in optimizing T11SU (a TREC utility function) while Rocchio is better for optimizing Ctrk (the TDT tracking cost), a high-recall oriented objective function. Using systematic cross-corpus parameter optimization with both methods, we obtained the best results ever reported on TDT5, TREC10 and TREC11. Relevance feedback on a small portion (0.05~0.2%) of the TDT5 test documents yielded significant performance improvements, measuring up to a 54% reduction in Ctrk and a 20.9% increase in T11SU (with b=0.1), compared to the results of the top-performing system in TDT2004 without relevance feedback information.
international conference on enterprise information systems | 2009
Yiming Yang; Abhimanyu Lad; Henry Shu; Bryan Kisiel; Chad M. Cumby; Rayid Ghani; Katharina Probst
In many practical applications, multiple interrelated tasks must be accomplished sequentially through user interaction with retrieval, classification and recommendation systems. The ordering of the tasks may have a significant impact on the overall utility (or performance) of the systems; hence optimal ordering of tasks is desirable. However, manual specification of optimal ordering is often difficult when task dependencies are complex, and exhaustive search for the optimal order is computationally intractable when the number of tasks is large. We propose a novel approach to this problem by using a directed graph to represent partialorder preferences among task pairs, and using link analysis (HITS and PageRank) over the graph as a heuristic to order tasks based on how important they are in reinforcing and propagating the ordering preference. These strategies allow us to find near-optimal solutions with efficient computation, scalable to large applications. We conducted a comparative evaluation of the proposed approach on a form-filling application involving a large collection of business proposals from the Accenture Consulting & Technology Company, using SVM classifiers to recommend keywords, collaborators, customers, technical categories and other related fillers for multiple fields in each proposal. With the proposed approach we obtained nearoptimal task orders that improved the utility of the recommendation system by 27% in macro-averaged F1, and 13% in micro-averaged F1, compared to the results obtained using arbitrarily chosen orders, and that were competitive against the best order suggested by domain experts.
national conference on artificial intelligence | 2010
Andrew Carlson; Justin Betteridge; Bryan Kisiel; Burr Settles; Estevam R. Hruschka; Tom M. Mitchell
national conference on artificial intelligence | 2015
Tom M. Mitchell; William W. Cohen; E. Hruschka; Partha Pratim Talukdar; Justin Betteridge; Andrew Carlson; Bhavana Dalvi; Matt Gardner; Bryan Kisiel; Jayant Krishnamurthy; Ni Lao; Kathryn Mazaitis; T. Mohamed; Ndapandula Nakashole; Emmanouil Antonios Platanios; Alan Ritter; Mehdi Samadi; Burr Settles; Richard C. Wang; Derry Tanti Wijaya; Abhinav Gupta; Xi Chen; A. Saparov; M. Greaves; J. Welling
empirical methods in natural language processing | 2013
Matt Gardner; Partha Pratim Talukdar; Bryan Kisiel; Tom M. Mitchell
international acm sigir conference on research and development in information retrieval | 2007
Yiming Yang; Abhimanyu Lad; Ni Lao; Abhay Harpale; Bryan Kisiel; Monica Rogati
conference on information and knowledge management | 2003
Yiming Yang; Bryan Kisiel
siam international conference on data mining | 2009
Abhimanyu Lad; Yiming Yang; Rayid Ghani; Bryan Kisiel
Theory and Applications of Categories | 2013
Bryan Kisiel; Justin Betteridge; Matt Gardner; Jayant Krishnamurthy; Ndapa Nakashole; Mehdi Samadi; Partha Pratim Talukdar; Derry Tanti Wijaya; Tom M. Mitchell