You need to log in to edit.

You can create a new account if you don't have one.

Or, discuss a change on Slack.

You can create a new account if you don't have one.

Or, discuss a change on Slack.

1 code implementation • 20 Oct 2021 • Gemma E. Moran, Dhanya Sridhar, Yixin Wang, David M. Blei

The model is sparse in the sense that each feature of the dataset (i. e., each dimension) depends on a small subset of the latent factors.

1 code implementation • 24 Sep 2021 • Mingzhang Yin, Yixin Wang, David M. Blei

This paper presents a new optimization approach to causal estimation.

1 code implementation • EMNLP 2021 • Keyon Vafa, Yuntian Deng, David M. Blei, Alexander M. Rush

Compared to existing baselines, greedy rationalization is best at optimizing the combinatorial objective and provides the most faithful rationales.

1 code implementation • ACL 2020 • Keyon Vafa, Suresh Naidu, David M. Blei

In this paper, we introduce the text-based ideal point model (TBIP), an unsupervised probabilistic topic model that analyzes texts to quantify the political positions of its authors.

no code implementations • 10 Mar 2020 • Yixin Wang, David M. Blei

Wang and Blei (2019) studies multiple causal inference and proposes the deconfounder algorithm.

1 code implementation • NeurIPS 2019 • Aaron Schein, Scott W. Linderman, Mingyuan Zhou, David M. Blei, Hanna Wallach

This paper presents the Poisson-randomized gamma dynamical system (PRGDS), a model for sequentially observed count tensors that encodes a strong inductive bias toward sparsity and burstiness.

no code implementations • 15 Oct 2019 • Yixin Wang, David M. Blei

Ogburn et al. (2019, arXiv:1910. 05438) discuss "The Blessings of Multiple Causes" (Wang and Blei, 2018, arXiv:1805. 06826).

2 code implementations • 9 Oct 2019 • Adji B. Dieng, Francisco J. R. Ruiz, David M. Blei, Michalis K. Titsias

Generative adversarial networks (GANs) are a powerful approach to unsupervised learning.

Ranked #2 on Image Generation on Stacked MNIST

no code implementations • 2 Aug 2019 • Rajesh Ranganath, David M. Blei

Pop-PCs, in contrast, compare the posterior predictive distribution to the population distribution of the data.

1 code implementation • 12 Jul 2019 • Adji B. Dieng, Francisco J. R. Ruiz, David M. Blei

Topic modeling analyzes documents to learn meaningful patterns of words.

9 code implementations • TACL 2020 • Adji B. Dieng, Francisco J. R. Ruiz, David M. Blei

To this end, we develop the Embedded Topic Model (ETM), a generative model of documents that marries traditional topic models with word embeddings.

1 code implementation • 10 Jun 2019 • Wesley Tansey, Christopher Tosh, David M. Blei

The goal in each paired (cell line, drug) experiment is to map out the dose-response curve of the cell line as the dose level of the drug increases.

1 code implementation • NeurIPS 2019 • Claudia Shi, David M. Blei, Victor Veitch

We propose two adaptations based on insights from the statistical literature on the estimation of treatment effects.

no code implementations • 30 May 2019 • Yixin Wang, David M. Blei

Our results expand the theory in Wang & Blei (2018), justify the deconfounder for causal graphs, and extend the settings where it can be used.

4 code implementations • 29 May 2019 • Victor Veitch, Dhanya Sridhar, David M. Blei

To address this challenge, we develop causally sufficient embeddings, low-dimensional document representations that preserve sufficient information for causal identification and allow for efficient estimation of causal effects.

no code implementations • 26 May 2019 • Yixin Wang, Dhanya Sridhar, David M. Blei

Machine learning (ML) can automate decision-making by learning to predict decisions from historical data.

1 code implementation • NeurIPS 2019 • Yixin Wang, David M. Blei

As a consequence of these results, we find that the model misspecification error dominates the variational approximation error in VB posterior predictive distributions.

no code implementations • 3 Apr 2019 • Linying Zhang, Yixin Wang, Anna Ostropolets, Jami J. Mulgrave, David M. Blei, George Hripcsak

To adjust for unobserved confounders, we develop the medical deconfounder, a machine learning algorithm that unbiasedly estimates treatment effects from EHRs.

2 code implementations • NeurIPS 2019 • Victor Veitch, Yixin Wang, David M. Blei

We validate the method with experiments on a semi-synthetic social network dataset.

1 code implementation • 13 Dec 2018 • Wesley Tansey, Kathy Li, Haoran Zhang, Scott W. Linderman, Raul Rabadan, David M. Blei, Chris H. Wiggins

Personalized cancer treatments based on the molecular profile of a patient's tumor are an emerging and exciting class of treatments in oncology.

Applications

no code implementations • 1 Dec 2018 • Andrew C. Miller, Ziad Obermeyer, David M. Blei, John P. Cunningham, Sendhil Mullainathan

An electrocardiogram (EKG) is a common, non-invasive test that measures the electrical activity of a patient's heart.

3 code implementations • 1 Nov 2018 • Wesley Tansey, Victor Veitch, Haoran Zhang, Raul Rabadan, David M. Blei

We propose the holdout randomization test (HRT), an approach to feature selection using black box predictive models.

Methodology

no code implementations • 27 Sep 2018 • Adji B. Dieng, Kyunghyun Cho, David M. Blei, Yann Lecun

Furthermore, the reflective likelihood objective prevents posterior collapse when used to train stochastic auto-encoders with amortized inference.

no code implementations • 20 Aug 2018 • Yixin Wang, Dawen Liang, Laurent Charlin, David M. Blei

To this end, we develop a causal approach to recommendation, one where watching a movie is a "treatment" and a user's rating is an "outcome."

no code implementations • 12 Jul 2018 • Adji B. Dieng, Yoon Kim, Alexander M. Rush, David M. Blei

VAEs can capture complex distributions, but they can also suffer from an issue known as "latent variable collapse," especially if the likelihood model is powerful.

1 code implementation • 27 Jun 2018 • Victor Veitch, Morgane Austern, Wenda Zhou, David M. Blei, Peter Orbanz

We solve this problem using recent ideas from graph sampling theory to (i) define an empirical risk for relational data and (ii) obtain stochastic gradients for this empirical risk that are automatically unbiased.

no code implementations • ICML 2018 • Wesley Tansey, Yixin Wang, David M. Blei, Raul Rabadan

BB-FDR learns a series of black box predictive models to boost power and control the false discovery rate (FDR) at two stages of study analysis.

2 code implementations • 17 May 2018 • Yixin Wang, David M. Blei

Causal inference from observational data often assumes "ignorability," that all confounders are observed.

no code implementations • ICML 2018 • Adji B. Dieng, Rajesh Ranganath, Jaan Altosaar, David M. Blei

On the Penn Treebank, the method with Noisin more quickly reaches state-of-the-art performance.

no code implementations • 24 Mar 2018 • Kriste Krstovski, David M. Blei

Qualitatively, we found that equation embeddings provide coherent semantic representations of equations and can capture semantic similarity to other equations and to words.

1 code implementation • ICML 2018 • Francisco J. R. Ruiz, Michalis K. Titsias, Adji B. Dieng, David M. Blei

It maximizes a lower bound on the marginal likelihood of the data.

no code implementations • ICLR 2018 • Adji B. Dieng, Jaan Altosaar, Rajesh Ranganath, David M. Blei

We develop a noise-based regularization method for RNNs.

2 code implementations • 9 Nov 2017 • Francisco J. R. Ruiz, Susan Athey, David M. Blei

We develop SHOPPER, a sequential probabilistic model of shopping data.

no code implementations • ICLR 2018 • Dustin Tran, David M. Blei

For the first, we describe implicit causal models, a class of causal models that leverages neural architectures with an implicit density.

no code implementations • ICML 2017 • Alp Kucukelbir, Yixin Wang, David M. Blei

We propose to evaluate a model through posterior dispersion.

no code implementations • ICML 2017 • Li-Ping Liu, David M. Blei

In this paper, we develop zero-inflated embeddings, a new embedding method that is designed to learn from sparse observations.

1 code implementation • 31 May 2017 • Christian A. Naesseth, Scott W. Linderman, Rajesh Ranganath, David M. Blei

The success of variational approaches depends on (i) formulating a flexible parametric family of distributions, and (ii) optimizing the parameters to find the member of this family that most closely approximates the exact posterior.

1 code implementation • 24 May 2017 • Jaan Altosaar, Rajesh Ranganath, David M. Blei

Consequently, PVI is less sensitive to initialization and optimization quirks and finds better local optima.

no code implementations • 9 May 2017 • Yixin Wang, David M. Blei

The theorem leverages the theoretical characterizations of frequentist variational approximations to understand asymptotic properties of VB.

1 code implementation • 13 Apr 2017 • Stephan Mandt, Matthew D. Hoffman, David M. Blei

Specifically, we show how to adjust the tuning parameters of constant SGD to best match the stationary distribution to a posterior, minimizing the Kullback-Leibler divergence between these two distributions.

no code implementations • NeurIPS 2017 • Dustin Tran, Rajesh Ranganath, David M. Blei

Implicit probabilistic models are a flexible class of models defined by a simulation process for data.

no code implementations • 13 Jan 2017 • Dustin Tran, Matthew D. Hoffman, Rif A. Saurous, Eugene Brevdo, Kevin Murphy, David M. Blei

By treating inference as a first class citizen, on a par with modeling, we show that probabilistic programming can be as flexible and computationally efficient as traditional deep learning.

no code implementations • 1 Nov 2016 • Adji B. Dieng, Dustin Tran, Rajesh Ranganath, John Paisley, David M. Blei

In this paper we propose CHIVI, a black-box variational inference algorithm that minimizes $D_{\chi}(p || q)$, the $\chi$-divergence from $p$ to $q$.

no code implementations • 31 Oct 2016 • Dustin Tran, Alp Kucukelbir, Adji B. Dieng, Maja Rudolph, Dawen Liang, David M. Blei

Probabilistic modeling is a powerful approach for analyzing empirical information.

no code implementations • NeurIPS 2016 • Rajesh Ranganath, Jaan Altosaar, Dustin Tran, David M. Blei

Though this divergence has been widely used, the resultant posterior approximation can suffer from undesirable statistical properties.

1 code implementation • 26 Oct 2016 • Scott W. Linderman, Andrew C. Miller, Ryan P. Adams, David M. Blei, Liam Paninski, Matthew J. Johnson

Many natural systems, such as neurons firing in the brain or basketball teams traversing a court, give rise to time series data with complex, nonlinear dynamics.

2 code implementations • 18 Oct 2016 • Christian A. Naesseth, Francisco J. R. Ruiz, Scott W. Linderman, David M. Blei

Variational inference using the reparameterization trick has enabled large-scale approximate Bayesian inference in complex probabilistic models, leveraging stochastic optimization to sidestep intractable expectations.

no code implementations • NeurIPS 2016 • Francisco J. R. Ruiz, Michalis K. Titsias, David M. Blei

The reparameterization gradient has become a widely used method to obtain Monte Carlo gradients to optimize the variational objective.

no code implementations • NeurIPS 2016 • Maja R. Rudolph, Francisco J. R. Ruiz, Stephan Mandt, David M. Blei

In this paper, we develop exponential family embeddings, a class of methods that extends the idea of word embeddings to other types of high-dimensional data.

1 code implementation • ICML 2017 • Yixin Wang, Alp Kucukelbir, David M. Blei

We propose a way to systematically detect and mitigate mismatch of a large class of probabilistic models.

1 code implementation • 6 Jun 2016 • Aaron Schein, Mingyuan Zhou, David M. Blei, Hanna Wallach

We introduce Bayesian Poisson Tucker decomposition (BPTD) for modeling country--country interaction event data.

no code implementations • 24 May 2016 • Alp Kucukelbir, David M. Blei

We propose to evaluate a model through posterior dispersion.

no code implementations • 3 Mar 2016 • Francisco J. R. Ruiz, Michalis K. Titsias, David M. Blei

Instead of taking samples from the variational distribution, we use importance sampling to take samples from an overdispersed distribution in the same exponential family as the variational approximation.

4 code implementations • 2 Mar 2016 • Alp Kucukelbir, Dustin Tran, Rajesh Ranganath, Andrew Gelman, David M. Blei

Probabilistic modeling is iterative.

no code implementations • 8 Feb 2016 • Stephan Mandt, Matthew D. Hoffman, David M. Blei

With constant learning rates, it is a stochastic process that, after an initial phase of convergence, generates samples from a stationary distribution.

5 code implementations • 4 Jan 2016 • David M. Blei, Alp Kucukelbir, Jon D. McAuliffe

One of the core problems of modern statistics is to approximate difficult-to-compute probability densities.

no code implementations • 20 Nov 2015 • Dustin Tran, Rajesh Ranganath, David M. Blei

Variational inference is a powerful tool for approximate inference, and it has been recently applied for representation learning with deep generative models.

1 code implementation • 7 Nov 2015 • Rajesh Ranganath, Dustin Tran, David M. Blei

We study HVMs on a variety of deep discrete latent variable models.

1 code implementation • 23 Oct 2015 • Dawen Liang, Laurent Charlin, James McInerney, David M. Blei

The exposure is modeled as a latent variable and the model infers its value from data.

no code implementations • 17 Oct 2015 • Chong Wang, David M. Blei

Robust Bayesian models are appealing alternatives to standard models, providing protection from data that contains outliers or other departures from the model assumptions.

no code implementations • 15 Sep 2015 • Laurent Charlin, Rajesh Ranganath, James McInerney, David M. Blei

Models for recommender systems use latent factors to explain the preferences and behaviors of users with respect to a set of items (e. g., movies, books, academic papers).

2 code implementations • 19 Jul 2015 • James McInerney, Rajesh Ranganath, David M. Blei

Many modern data analysis problems involve inferences from streaming data.

no code implementations • 24 Jun 2015 • Maja R. Rudolph, Joseph G. Ellis, David M. Blei

In this paper, we develop a probabilistic method to learn a profitable strategy to set the reserve price.

1 code implementation • 10 Jun 2015 • Aaron Schein, John Paisley, David M. Blei, Hanna Wallach

We demonstrate that our model's predictive performance is better than that of standard non-negative tensor factorization methods.

no code implementations • NeurIPS 2015 • Alp Kucukelbir, Rajesh Ranganath, Andrew Gelman, David M. Blei

With ADVI we can use variational inference on any model we write in Stan.

no code implementations • NeurIPS 2015 • Dustin Tran, David M. Blei, Edoardo M. Airoldi

We develop a general variational inference method that preserves dependency among the latent variables.

no code implementations • 10 Nov 2014 • Rajesh Ranganath, Linpeng Tang, Laurent Charlin, David M. Blei

We describe \textit{deep exponential families} (DEFs), a class of latent variable models that are inspired by the hidden structures used in deep neural networks.

1 code implementation • 2 Nov 2014 • Alp Kucukelbir, David M. Blei

We develop population empirical Bayes (POP-EB), a hierarchical framework that explicitly models the empirical population distribution as part of Bayesian analysis.

no code implementations • 16 Apr 2014 • Matthew D. Hoffman, David M. Blei

Stochastic variational inference makes it possible to approximate posterior distributions induced by large datasets quickly using stochastic optimization.

2 code implementations • 31 Dec 2013 • Rajesh Ranganath, Sean Gerrish, David M. Blei

We evaluate our method against the corresponding black box sampling based methods.

2 code implementations • 7 Nov 2013 • Prem Gopalan, Jake M. Hofman, David M. Blei

This is an efficient algorithm that iterates over the observed entries and adjusts an approximate posterior over the user/item representations.

1 code implementation • PNAS 2013 2013 • Prem K. Gopalan, David M. Blei

Our approach is based on a Bayesian model of networks that allows nodes to participate in multiple communities, and a corresponding algorithm that naturally interleaves subsampling from the network and updating an estimate of its communities.

no code implementations • NeurIPS 2012 • Sean Gerrish, David M. Blei

We develop a probabilistic model of legislative data that uses the text of the bills to uncover lawmakers' positions on specific political issues.

no code implementations • NeurIPS 2012 • Chong Wang, David M. Blei

We present a truncation-free online variational inference algorithm for Bayesian nonparametric models.

no code implementations • NeurIPS 2012 • Prem K. Gopalan, Sean Gerrish, Michael Freedman, David M. Blei, David M. Mimno

We develop a scalable algorithm for posterior inference of overlapping communities in large networks.

no code implementations • 25 Oct 2012 • John Paisley, Chong Wang, David M. Blei, Michael. I. Jordan

We develop a nested hierarchical Dirichlet process (nHDP) for hierarchical topic modeling.

2 code implementations • 29 Jun 2012 • Matt Hoffman, David M. Blei, Chong Wang, John Paisley

We develop stochastic variational inference, a scalable algorithm for approximating posterior distributions.

no code implementations • NeurIPS 2011 • Soumya Ghosh, Andrei B. Ungureanu, Erik B. Sudderth, David M. Blei

The distance dependent Chinese restaurant process (ddCRP) was recently introduced to accommodate random partitions of non-exchangeable data.

no code implementations • NeurIPS 2010 • Lauren Hannah, Warren Powell, David M. Blei

Those similar to the current state are used to create a convex, deterministic approximation of the objective function.

no code implementations • NeurIPS 2010 • Abhinav Gupta, Martial Hebert, Takeo Kanade, David M. Blei

There has been a recent push in extraction of 3D spatial layout of scenes.

no code implementations • NeurIPS 2010 • Matthew Hoffman, Francis R. Bach, David M. Blei

We develop an online variational Bayes (VB) algorithm for Latent Dirichlet Allocation (LDA).

1 code implementation • NeurIPS 2007 • David M. Blei, Jon D. McAuliffe

We introduce supervised latent Dirichlet allocation (sLDA), a statistical model of labelled documents.

no code implementations • NeurIPS 2009 • Richard Socher, Samuel Gershman, Per Sederberg, Kenneth Norman, Adler J. Perotte, David M. Blei

We develop a probabilistic model of human memory performance in free recall experiments.

no code implementations • NeurIPS 2009 • Chong Wang, David M. Blei

The nested Chinese restaurant process (nCRP) is a powerful nonparametric Bayesian model for learning tree-based hierarchies from data.

no code implementations • NeurIPS 2009 • Chong Wang, David M. Blei

We present a nonparametric hierarchical Bayesian model of document collections that decouples sparsity and smoothness in the component distributions (i. e., the ``topics).

no code implementations • NeurIPS 2008 • Jordan L. Boyd-Graber, David M. Blei

We develop \name\ (STM), a nonparametric Bayesian model of parsed documents.

no code implementations • NeurIPS 2008 • Indraneel Mukherjee, David M. Blei

In this paper we provide the beginnings of such understanding.

no code implementations • NeurIPS 2008 • Edo M. Airoldi, David M. Blei, Stephen E. Fienberg, Eric P. Xing

Observations consisting of measurements on relationships for pairs of objects arise in many settings, such as protein interaction and gene regulatory networks, collections of author-recipient email, and social networks.

no code implementations • 27 Aug 2007 • David M. Blei, John D. Lafferty

This limitation stems from the use of the Dirichlet distribution to model the variability among the topic proportions.

Applications

2 code implementations • 1 Jan 2003 • David M. Blei, Andrew Y. Ng, Michael I. Jordan

Each topic is, in turn, modeled as an infinite mixture over an underlying set of topic probabilities.

Cannot find the paper you are looking for? You can
Submit a new open access paper.

Contact us on:
hello@paperswithcode.com
.
Papers With Code is a free resource with all data licensed under CC-BY-SA.