Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Taylor Berg-Kirkpatrick is active.

Publication


Featured researches published by Taylor Berg-Kirkpatrick.


meeting of the association for computational linguistics | 2016

Learning-Based Single-Document Summarization with Compression and Anaphoricity Constraints.

Greg Durrett; Taylor Berg-Kirkpatrick; Daniel Klein

We present a discriminative model for single-document summarization that integrally combines compression and anaphoricity constraints. Our model selects textual units to include in the summary based on a rich set of sparse features whose weights are learned on a large corpus. We allow for the deletion of content within a sentence when that deletion is licensed by compression rules; in our framework, these are implemented as dependencies between subsentential units of text. Anaphoricity constraints then improve cross-sentence coherence by guaranteeing that, for each pronoun included in the summary, the pronouns antecedent is included as well or the pronoun is rewritten as a full mention. When trained end-to-end, our final system outperforms prior work on both ROUGE as well as on human judgments of linguistic quality.


meeting of the association for computational linguistics | 2014

Sparser, Better, Faster GPU Parsing

David Leo Wright Hall; Taylor Berg-Kirkpatrick; Daniel Klein

Due to their origin in computer graphics, graphics processing units (GPUs) are highly optimized for dense problems, where the exact same operation is applied repeatedly to all data points. Natural language processing algorithms, on the other hand, are traditionally constructed in ways that exploit structural sparsity. Recently, Canny et al. (2013) presented an approach to GPU parsing that sacrifices traditional sparsity in exchange for raw computational power, obtaining a system that can compute Viterbi parses for a high-quality grammar at about 164 sentences per second on a mid-range GPU. In this work, we reintroduce sparsity to GPU parsing by adapting a coarse-to-fine pruning approach to the constraints of a GPU. The resulting system is capable of computing over 404 Viterbi parses per second—more than a 2x speedup—on the same hardware. Moreover, our approach allows us to efficiently implement less GPU-friendly minimum Bayes risk inference, improving throughput for this more accurate algorithm from only 32 sentences per second unpruned to over 190 sentences per second using pruning—nearly a 6x speedup.


empirical methods in natural language processing | 2015

An Empirical Analysis of Optimization for Max-Margin NLP

Jonathan K. Kummerfeld; Taylor Berg-Kirkpatrick; Daniel Klein

Despite the convexity of structured maxmargin objectives (Taskar et al., 2004; Tsochantaridis et al., 2004), the many ways to optimize them are not equally effective in practice. We compare a range of online optimization methods over a variety of structured NLP tasks (coreference, summarization, parsing, etc) and find several broad trends. First, margin methods do tend to outperform both likelihood and the perceptron. Second, for max-margin objectives, primal optimization methods are often more robust and progress faster than dual methods. This advantage is most pronounced for tasks with dense or continuous-valued features. Overall, we argue for a particularly simple online primal subgradient descent method that, despite being rarely mentioned in the literature, is surprisingly effective in relation to its alternatives.


international world wide web conferences | 2017

Tools for Automated Analysis of Cybercriminal Markets

Rebecca S. Portnoff; Sadia Afroz; Greg Durrett; Jonathan K. Kummerfeld; Taylor Berg-Kirkpatrick; Damon McCoy; Kirill Levchenko; Vern Paxson

Underground forums are widely used by criminals to buy and sell a host of stolen items, datasets, resources, and criminal services. These forums contain important resources for understanding cybercrime. However, the number of forums, their size, and the domain expertise required to understand the markets makes manual exploration of these forums unscalable. In this work, we propose an automated, top-down approach for analyzing underground forums. Our approach uses natural language processing and machine learning to automatically generate high-level information about underground forums, first identifying posts related to transactions, and then extracting products and prices. We also demonstrate, via a pair of case studies, how an analyst can use these automated approaches to investigate other categories of products and transactions. We use eight distinct forums to assess our tools: Antichat, Blackhat World, Carders, Darkode, Hack Forums, Hell, L33tCrew and Nulled. Our automated approach is fast and accurate, achieving over 80% accuracy in detecting post category, product, and prices.


The Journal of Experimental Biology | 2016

Using accelerometers to remotely and automatically characterize behavior in small animals.

Talisin T. Hammond; Dwight Springthorpe; Rachel E. Walsh; Taylor Berg-Kirkpatrick

ABSTRACT Activity budgets in wild animals are challenging to measure via direct observation because data collection is time consuming and observer effects are potentially confounding. Although tri-axial accelerometers are increasingly employed for this purpose, their application in small-bodied animals has been limited by weight restrictions. Additionally, accelerometers engender novel complications, as a system is needed to reliably map acceleration to behaviors. In this study, we describe newly developed, tiny acceleration-logging devices (1.5–2.5 g) and use them to characterize behavior in two chipmunk species. We collected paired accelerometer readings and behavioral observations from captive individuals. We then employed techniques from machine learning to develop an automatic system for coding accelerometer readings into behavioral categories. Finally, we deployed and recovered accelerometers from free-living, wild chipmunks. This is the first time to our knowledge that accelerometers have been used to generate behavioral data for small-bodied (<100 g), free-living mammals. Summary: Validation of the use of accelerometers for automated collection of behavioral data from two species of small-bodied, free-living animals.


meeting of the association for computational linguistics | 2014

Improved Typesetting Models for Historical OCR

Taylor Berg-Kirkpatrick; Daniel Klein

We present richer typesetting models that extend the unsupervised historical document recognition system of BergKirkpatrick et al. (2013). The first model breaks the independence assumption between vertical offsets of neighboring glyphs and, in experiments, substantially decreases transcription error rates. The second model simultaneously learns multiple font styles and, as a result, is able to accurately track italic and nonitalic portions of documents. Richer models complicate inference so we present a new, streamlined procedure that is over 25x faster than the method used by BergKirkpatrick et al. (2013). Our final system achieves a relative word error reduction of 22% compared to state-of-the-art results on a dataset of historical newspapers.


north american chapter of the association for computational linguistics | 2015

Unsupervised Code-Switching for Multilingual Historical Document Transcription

Dan Garrette; Hannah Alpert-Abrams; Taylor Berg-Kirkpatrick; Daniel Klein

Transcribing documents from the printing press era, a challenge in its own right, is more complicated when documents interleave multiple languages—a common feature of 16th century texts. Additionally, many of these documents precede consistent orthographic conventions, making the task even harder. We extend the state-of-the-art historical OCR model of Berg-Kirkpatrick et al. (2013) to handle word-level code-switching between multiple languages. Further, we enable our system to handle spelling variability, including now-obsolete shorthand systems used by printers. Our results show average relative character error reductions of 14% across a variety of historical texts.


meeting of the association for computational linguistics | 2017

Differentiable Scheduled Sampling for Credit Assignment

Kartik Goyal; Chris Dyer; Taylor Berg-Kirkpatrick

We demonstrate that a continuous relaxation of the argmax operation can be used to create a differentiable approximation to greedy decoding for sequence-to-sequence (seq2seq) models. By incorporating this approximation into the scheduled sampling training procedure (Bengio et al., 2015)--a well-known technique for correcting exposure bias--we introduce a new training objective that is continuous and differentiable everywhere and that can provide informative gradients near points where previous decoding decisions change their value. In addition, by using a related approximation, we demonstrate a similar approach to sampled-based training. Finally, we show that our approach outperforms cross-entropy training and scheduled sampling procedures in two sequence prediction tasks: named entity recognition and machine translation.


knowledge discovery and data mining | 2017

Efficient Correlated Topic Modeling with Topic Embedding

Junxian He; Zhiting Hu; Taylor Berg-Kirkpatrick; Ying Huang; Eric P. Xing

Correlated topic modeling has been limited to small model and problem sizes due to their high computational cost and poor scaling. In this paper, we propose a new model which learns compact topic embeddings and captures topic correlations through the closeness between the topic vectors. Our method enables efficient inference in the low-dimensional embedding space, reducing previous cubic or quadratic time complexity to linear w.r.t the topic size. We further speedup variational inference with a fast sampler to exploit sparsity of topic occurrence. Extensive experiments show that our approach is capable of handling model and data scales which are several orders of magnitude larger than existing correlation results, without sacrificing modeling quality by providing competitive or superior performance in document classification and retrieval.


meeting of the association for computational linguistics | 2017

Automatic Compositor Attribution in the First Folio of Shakespeare.

Maria Ryskina; Hannah Alpert-Abrams; Dan Garrette; Taylor Berg-Kirkpatrick

Compositor attribution, the clustering of pages in a historical printed document by the individual who set the type, is a bibliographic task that relies on analysis of orthographic variation and inspection of visual details of the printed page. In this paper, we introduce a novel unsupervised model that jointly describes the textual and visual features needed to distinguish compositors. Applied to images of Shakespeares First Folio, our model predicts attributions that agree with the manual judgements of bibliographers with an accuracy of 87%, even on text that is the output of OCR.

Collaboration


Dive into the Taylor Berg-Kirkpatrick's collaboration.

Top Co-Authors

Avatar

Daniel Klein

University of California

View shared research outputs
Top Co-Authors

Avatar

Greg Durrett

University of California

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Chris Dyer

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar

Eduard H. Hovy

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar

Graham Neubig

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar

Harsh Jhamtani

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Zhiting Hu

Carnegie Mellon University

View shared research outputs
Researchain Logo
Decentralizing Knowledge