Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Stephen H. Bach is active.

Publication


Featured researches published by Stephen H. Bach.


computer vision and pattern recognition | 2013

Collective Activity Detection Using Hinge-loss Markov Random Fields

Ben London; Sameh Khamis; Stephen H. Bach; Bert Huang; Lise Getoor; Larry S. Davis

We propose hinge-loss Markov random fields (HL-MRFs), a powerful class of continuous-valued graphical models, for high-level computer vision tasks. HL-MRFs are characterized by log-concave density functions, and are able to perform efficient, exact inference. Their templated hinge-loss potential functions naturally encode soft-valued logical rules. Using the declarative modeling language probabilistic soft logic, one can easily define HL-MRFs via familiar constructs from first-order logic. We apply HL-MRFs to the task of activity detection, using principles of collective classification. Our model is simple, intuitive and interpretable. We evaluate our model on two datasets and show that it achieves significant lift over the low-level detectors.


inductive logic programming | 2015

Statistical relational learning with soft quantifiers

Golnoosh Farnadi; Stephen H. Bach; Marjon Blondeel; Marie-Francine Moens; Lise Getoor; Martine De Cock

Quantification in statistical relational learning (SRL) is either existential or universal, however humans might be more inclined to express knowledge using soft quantifiers, such as “most” and “a few”. In this paper, we define the syntax and semantics of PSL\(^Q\), a new SRL framework that supports reasoning with soft quantifiers, and present its most probable explanation (MPE) inference algorithm. To the best of our knowledge, PSL\(^Q\) is the first SRL framework that combines soft quantifiers with first-order logic rules for modeling uncertain relational data. Our experimental results for link prediction in social trust networks demonstrate that the use of soft quantifiers not only allows for a natural and intuitive formulation of domain knowledge, but also improves the accuracy of inferred results.


Archive | 2015

Hinge-loss Markov random fields and probabilistic soft logic: A scalable approach to structured prediction

Stephen H. Bach

A fundamental challenge in developing high-impact machine learning technologies is balancing the need to model rich, structured domains with the ability to scale to big data. Many important problem areas are both richly structured and large scale, from social and biological networks, to knowledge graphs and the Web, to images, video, and natural language. In this paper, we introduce two new formalisms for modeling structured data, and show that they can both capture rich structure and scale to big data. The first, hinge-loss Markov random fields (HL-MRFs), is a new kind of probabilistic graphical model that generalizes different approaches to convex inference. We unite three approaches from the randomized algorithms, probabilistic graphical models, and fuzzy logic communities, showing that all three lead to the same inference objective. We then define HL-MRFs by generalizing this unified objective. The second new formalism, probabilistic soft logic (PSL), is a probabilistic programming language that makes HL-MRFs easy to define using a syntax based on first-order logic. We introduce an algorithm for inferring most-probable variable assignments (MAP inference) that is much more scalable than general-purpose convex optimization methods, because it uses message passing to take advantage of sparse dependency structures. We then show how to learn the parameters of HL-MRFs. The learned HL-MRFs are as accurate as analogous discrete models, but much more scalable. Together, these algorithms enable HL-MRFs and PSL to model rich, structured data at scales not previously possible.


neural information processing systems | 2012

A short introduction to probabilistic soft logic

Angelika Kimmig; Stephen H. Bach; Matthias Broecheler; Bert Huang; Lise Getoor


uncertainty in artificial intelligence | 2013

Hinge-loss Markov random fields: convex inference for structured prediction

Stephen H. Bach; Bert Huang; Ben London; Lise Getoor


Journal of Machine Learning Research | 2017

Hinge-Loss Markov Random Fields and Probabilistic Soft Logic

Stephen H. Bach; Matthias Broecheler; Bert Huang; Lise Getoor


neural information processing systems | 2012

Scaling MPE Inference for Constrained Continuous Markov Random Fields with Consensus Optimization

Stephen H. Bach; Matthias Broecheler; Lise Getoor; Dianne P. O'Leary


Archive | 2012

Social Group Modeling with Probabilistic Soft Logic

Bert Huang; Stephen H. Bach; Eric Norris; Jay Pujara; Lise Getoor


uncertainty reasoning for the semantic web | 2012

Graph summarization in annotated data using probabilistic soft logic

Alex Memory; Angelika Kimmig; Stephen H. Bach; Louiqa Raschid; Lise Getoor


international conference on machine learning | 2017

Learning the Structure of Generative Models without Labeled Data

Stephen H. Bach; Bryan D. He; Alexander Ratner; Christopher Ré

Collaboration


Dive into the Stephen H. Bach's collaboration.

Top Co-Authors

Avatar

Lise Getoor

University of California

View shared research outputs
Top Co-Authors

Avatar

Angelika Kimmig

Katholieke Universiteit Leuven

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Marie-Francine Moens

Katholieke Universiteit Leuven

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Marjon Blondeel

Vrije Universiteit Brussel

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Bryan D. He

California Institute of Technology

View shared research outputs
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge