Vikash Mansinghka | Massachusetts Institute of Technology (MIT) (original) (raw)

Papers by Vikash Mansinghka

Research paper thumbnail of Beyond calculation: Probabilistic Computing Machines and Universal Stochastic Inference

Research paper thumbnail of Nonparametric Bayesian methods for supervised and unsupervised learning

I introduce two nonparametric Bayesian methods for solving problems of supervised and unsupervise... more I introduce two nonparametric Bayesian methods for solving problems of supervised and unsupervised learning. The first method simultaneously learns causal networks and causal theories from data. For example, given synthetic co-occurrence data from a simple causal model for the medical domain, it can learn relationships like "having a flu causes coughing", while also learning that observable quantities can be usefully grouped into categories like diseases and symptoms, and that diseases tend to cause symptoms, not the other way around. The second method is an online algorithm for learning a prototype-based model for categorial concepts, and can be used to solve problems of multiclass classification with missing features. I apply it to problems of categorizing newsgroup posts and recognizing handwritten digits.

Research paper thumbnail of Efficient Monte Carlo inference for infinite relational models

… Systems: presented at …, Jan 1, 2007

... Navia Systems, Inc. Joint work with: Keith Bonawitz (MIT CSAIL, Navia) Eric Jonas (MIT BCS, N... more ... Navia Systems, Inc. Joint work with: Keith Bonawitz (MIT CSAIL, Navia) Eric Jonas (MIT BCS, Navia) Josh Tenenbaum (MIT BCS/CSAIL) Page 2. IRM Review Intuition ... parallel ● Analysis for IRM CRP Gibbs ● Particle filter for IRM ● Parallel Tempering Transform ...

Research paper thumbnail of A Bayesian framework for modeling intuitive dynamics

Proceedings of the 31st …, Jan 1, 2009

People have strong intuitions about the masses of objects and the causal forces that they exert u... more People have strong intuitions about the masses of objects and the causal forces that they exert upon one another. These intuitions have been explored through a variety of tasks, in particular judging the relative masses of objects involved in collisions and evaluating whether one object caused another to move. We present a single framework for explaining two types of judgments that people make about the dynamics of objects, based on Bayesian inference. In this framework, we define a particular model of dynamics -essentially Newtonian physics plus Gaussian noise -which makes predictions about the trajectories of objects following collisions based on their masses. By applying Bayesian inference, it becomes possible to reason from trajectories back to masses, and to reason about whether one object caused another to move. We use this framework to predict human causality judgments using data collected from a mass judgment task.

Research paper thumbnail of Cross-categorization: A method for discovering multiple overlapping clusterings

… Bayes Workshop at …, Jan 1, 2009

Model-based clustering techniques, including inference in Dirichlet process mixture models, have ... more Model-based clustering techniques, including inference in Dirichlet process mixture models, have difficulty when different dimensions are best explained by very different clusterings. We introduce cross-categorization, an unsupervised learning technique that overcomes this basic limitation. Based on MCMC inference in a novel nonparametric Bayesian model, cross-categorization automatically discovers the number of independent nonparametric Bayesian models needed to explain the data, using a separate Dirichlet process mixture model for each group in an inferred partition of the dimensions. Unlike a DP mixture, our model is exchangeable over both the rows of a heterogeneous data array (the samples) and the columns (new dimensions), and can model any dataset as the number of samples and dimensions both go to infinity. We demonstrate the efficiency and robustness of our algorithm, including experiments on the full Dartmouth Health Atlas dataset without any preprocessing, showing that it finds veridical causal structure.

Research paper thumbnail of A probabilistic model of cross-categorization

Cognition, Jan 1, 2011

Most natural domains can be represented in multiple ways: we can categorize foods in terms of the... more Most natural domains can be represented in multiple ways: we can categorize foods in terms of their nutritional content or social role, animals in terms of their taxonomic groupings or their ecological niches, and musical instruments in terms of their taxonomic categories or social uses. Previous approaches to modeling human categorization have largely ignored the problem of cross-categorization, focusing on learning just a single system of categories that explains all of the features. Cross-categorization presents a difficult problem: how can we infer categories without first knowing which features the categories are meant to explain? We present a novel model that suggests that human cross-categorization is a result of joint inference about multiple systems of categories and the features that they explain. We also formalize two commonly proposed alternative explanations for cross-categorization behavior: a features-first and an objects-first approach. The features-first approach suggests that cross-categorization is a consequence of attentional processes, where features are selected by an attentional mechanism first and categories are derived second. The objects-first approach suggests that cross-categorization is a consequence of repeated, sequential attempts to explain features, where categories are derived first, then features that are poorly explained are recategorized. We present two sets of simulations and experiments testing the models' predictions about human categorization. We find that an approach based on joint inference provides the best fit to human categorization behavior, and we suggest that a full account of human category learning will need to incorporate something akin to these capabilities.

Research paper thumbnail of Natively probabilistic computation

I introduce a new set of natively probabilistic computing abstractions, including probabilistic g... more I introduce a new set of natively probabilistic computing abstractions, including probabilistic generalizations of Boolean circuits, backtracking search and pure Lisp. I show how these tools let one compactly specify probabilistic generative models, generalize and parallelize widely used sampling algorithms like rejection sampling and Markov chain Monte Carlo, and solve difficult Bayesian inference problems.

Research paper thumbnail of AClass: A simple, online, parallelizable algorithm for probabilistic classification

Journal of Machine Learning Research, Jan 1, 2007

Research paper thumbnail of Learning grounded causal models

Proceedings of the Twenty- …, Jan 1, 2007

We address the problem of learning grounded causal models: systems of concepts that are connected... more We address the problem of learning grounded causal models: systems of concepts that are connected by causal relations and explicitly grounded in perception. We present a Bayesian framework for learning these models-both a causal Bayesian network structure over variables and the consequential region of each variable in perceptual space-from dynamic perceptual evidence. Using a novel experimental paradigm we show that humans are able to learn grounded causal models, and that the Bayesian model accounts well for human performance.

Research paper thumbnail of Learning cross-cutting systems of categories

Proceedings of the …, Jan 1, 2006

Most natural domains can be represented in multiple ways: animals may be thought of in terms of t... more Most natural domains can be represented in multiple ways: animals may be thought of in terms of their taxonomic groupings or their ecological niches and foods may be thought of in terms of their nutritional content or social role. We present a computational framework that discovers multiple systems of categories given information about a domain of objects and their properties. Each system of object categories accounts for a distinct and coherent subset of the features. A first experiment shows that our CrossCat model predicts human learning in an artificial category learning task. A second experiment shows that the model discovers important structure in two real-world domains. Traditional models of categorization usually search for a single system of categories: we suggest that these models do not predict human performance in our task, and miss important structure in our real world examples.

Research paper thumbnail of Exact and Approximate Sampling by Systematic Stochastic Search

Journal of Machine Learning Research, Jan 1, 2009

Research paper thumbnail of Learning annotated hierarchies from relational data

Advances in neural …, Jan 1, 2007

Research paper thumbnail of Intuitive Theories of Mind: A Rational Approach to False Belief

Proceedings of the …, Jan 1, 2006

We propose a rational analysis of children's false belief reasoning. Our analysis realizes a cont... more We propose a rational analysis of children's false belief reasoning. Our analysis realizes a continuous, evidencedriven transition between two causal Bayesian models of false belief. Both models support prediction and explanation; however, one model is less complex while the other has greater explanatory resources. Because of this explanatory asymmetry, unexpected outcomes weigh more heavily against the simpler model. We test this account empirically by showing children the standard outcome of the false belief task and a novel "psychic" outcome. As expected, we find children whose explanations and predictions are consistent with each model, and an interaction between prediction and explanation. Critically, we find unexpected outcomes only induce children to move from predictions consistent with the simpler model to those consistent with the more complex one, never the reverse.

Research paper thumbnail of Structured Priors for Structure Learning

Proceedings of the …, Jan 1, 2006

Traditional approaches to Bayes net structure learning typically assume little regularity in grap... more Traditional approaches to Bayes net structure learning typically assume little regularity in graph structure other than sparseness. However, in many cases, we expect more systematicity: variables in real-world systems often group into classes that predict the kinds of probabilistic dependencies they participate in. Here we capture this form of prior knowledge in a hierarchical Bayesian framework, and exploit it to enable structure learning and type discovery from small datasets. Specifically, we present a nonparametric generative model for directed acyclic graphs as a prior for Bayes net structure learning. Our model assumes that variables come in one or more classes and that the prior probability of an edge existing between two variables is a function only of their classes. We derive an MCMC algorithm for simultaneous inference of the number of classes, the class assignments of variables, and the Bayes net structure over variables. For several realistic, sparse datasets, we show that the bias towards systematicity of connections provided by our model can yield more accurate learned networks than the traditional approach of using a uniform prior, and that the classes found by our model are appropriate.

Research paper thumbnail of Church: a language for generative models

Uncertainty in Artificial …, Jan 1, 2008

Research paper thumbnail of Beyond calculation: Probabilistic Computing Machines and Universal Stochastic Inference

Research paper thumbnail of Nonparametric Bayesian methods for supervised and unsupervised learning

I introduce two nonparametric Bayesian methods for solving problems of supervised and unsupervise... more I introduce two nonparametric Bayesian methods for solving problems of supervised and unsupervised learning. The first method simultaneously learns causal networks and causal theories from data. For example, given synthetic co-occurrence data from a simple causal model for the medical domain, it can learn relationships like "having a flu causes coughing", while also learning that observable quantities can be usefully grouped into categories like diseases and symptoms, and that diseases tend to cause symptoms, not the other way around. The second method is an online algorithm for learning a prototype-based model for categorial concepts, and can be used to solve problems of multiclass classification with missing features. I apply it to problems of categorizing newsgroup posts and recognizing handwritten digits.

Research paper thumbnail of Efficient Monte Carlo inference for infinite relational models

… Systems: presented at …, Jan 1, 2007

... Navia Systems, Inc. Joint work with: Keith Bonawitz (MIT CSAIL, Navia) Eric Jonas (MIT BCS, N... more ... Navia Systems, Inc. Joint work with: Keith Bonawitz (MIT CSAIL, Navia) Eric Jonas (MIT BCS, Navia) Josh Tenenbaum (MIT BCS/CSAIL) Page 2. IRM Review Intuition ... parallel ● Analysis for IRM CRP Gibbs ● Particle filter for IRM ● Parallel Tempering Transform ...

Research paper thumbnail of A Bayesian framework for modeling intuitive dynamics

Proceedings of the 31st …, Jan 1, 2009

People have strong intuitions about the masses of objects and the causal forces that they exert u... more People have strong intuitions about the masses of objects and the causal forces that they exert upon one another. These intuitions have been explored through a variety of tasks, in particular judging the relative masses of objects involved in collisions and evaluating whether one object caused another to move. We present a single framework for explaining two types of judgments that people make about the dynamics of objects, based on Bayesian inference. In this framework, we define a particular model of dynamics -essentially Newtonian physics plus Gaussian noise -which makes predictions about the trajectories of objects following collisions based on their masses. By applying Bayesian inference, it becomes possible to reason from trajectories back to masses, and to reason about whether one object caused another to move. We use this framework to predict human causality judgments using data collected from a mass judgment task.

Research paper thumbnail of Cross-categorization: A method for discovering multiple overlapping clusterings

… Bayes Workshop at …, Jan 1, 2009

Model-based clustering techniques, including inference in Dirichlet process mixture models, have ... more Model-based clustering techniques, including inference in Dirichlet process mixture models, have difficulty when different dimensions are best explained by very different clusterings. We introduce cross-categorization, an unsupervised learning technique that overcomes this basic limitation. Based on MCMC inference in a novel nonparametric Bayesian model, cross-categorization automatically discovers the number of independent nonparametric Bayesian models needed to explain the data, using a separate Dirichlet process mixture model for each group in an inferred partition of the dimensions. Unlike a DP mixture, our model is exchangeable over both the rows of a heterogeneous data array (the samples) and the columns (new dimensions), and can model any dataset as the number of samples and dimensions both go to infinity. We demonstrate the efficiency and robustness of our algorithm, including experiments on the full Dartmouth Health Atlas dataset without any preprocessing, showing that it finds veridical causal structure.

Research paper thumbnail of A probabilistic model of cross-categorization

Cognition, Jan 1, 2011

Most natural domains can be represented in multiple ways: we can categorize foods in terms of the... more Most natural domains can be represented in multiple ways: we can categorize foods in terms of their nutritional content or social role, animals in terms of their taxonomic groupings or their ecological niches, and musical instruments in terms of their taxonomic categories or social uses. Previous approaches to modeling human categorization have largely ignored the problem of cross-categorization, focusing on learning just a single system of categories that explains all of the features. Cross-categorization presents a difficult problem: how can we infer categories without first knowing which features the categories are meant to explain? We present a novel model that suggests that human cross-categorization is a result of joint inference about multiple systems of categories and the features that they explain. We also formalize two commonly proposed alternative explanations for cross-categorization behavior: a features-first and an objects-first approach. The features-first approach suggests that cross-categorization is a consequence of attentional processes, where features are selected by an attentional mechanism first and categories are derived second. The objects-first approach suggests that cross-categorization is a consequence of repeated, sequential attempts to explain features, where categories are derived first, then features that are poorly explained are recategorized. We present two sets of simulations and experiments testing the models' predictions about human categorization. We find that an approach based on joint inference provides the best fit to human categorization behavior, and we suggest that a full account of human category learning will need to incorporate something akin to these capabilities.

Research paper thumbnail of Natively probabilistic computation

I introduce a new set of natively probabilistic computing abstractions, including probabilistic g... more I introduce a new set of natively probabilistic computing abstractions, including probabilistic generalizations of Boolean circuits, backtracking search and pure Lisp. I show how these tools let one compactly specify probabilistic generative models, generalize and parallelize widely used sampling algorithms like rejection sampling and Markov chain Monte Carlo, and solve difficult Bayesian inference problems.

Research paper thumbnail of AClass: A simple, online, parallelizable algorithm for probabilistic classification

Journal of Machine Learning Research, Jan 1, 2007

Research paper thumbnail of Learning grounded causal models

Proceedings of the Twenty- …, Jan 1, 2007

We address the problem of learning grounded causal models: systems of concepts that are connected... more We address the problem of learning grounded causal models: systems of concepts that are connected by causal relations and explicitly grounded in perception. We present a Bayesian framework for learning these models-both a causal Bayesian network structure over variables and the consequential region of each variable in perceptual space-from dynamic perceptual evidence. Using a novel experimental paradigm we show that humans are able to learn grounded causal models, and that the Bayesian model accounts well for human performance.

Research paper thumbnail of Learning cross-cutting systems of categories

Proceedings of the …, Jan 1, 2006

Most natural domains can be represented in multiple ways: animals may be thought of in terms of t... more Most natural domains can be represented in multiple ways: animals may be thought of in terms of their taxonomic groupings or their ecological niches and foods may be thought of in terms of their nutritional content or social role. We present a computational framework that discovers multiple systems of categories given information about a domain of objects and their properties. Each system of object categories accounts for a distinct and coherent subset of the features. A first experiment shows that our CrossCat model predicts human learning in an artificial category learning task. A second experiment shows that the model discovers important structure in two real-world domains. Traditional models of categorization usually search for a single system of categories: we suggest that these models do not predict human performance in our task, and miss important structure in our real world examples.

Research paper thumbnail of Exact and Approximate Sampling by Systematic Stochastic Search

Journal of Machine Learning Research, Jan 1, 2009

Research paper thumbnail of Learning annotated hierarchies from relational data

Advances in neural …, Jan 1, 2007

Research paper thumbnail of Intuitive Theories of Mind: A Rational Approach to False Belief

Proceedings of the …, Jan 1, 2006

We propose a rational analysis of children's false belief reasoning. Our analysis realizes a cont... more We propose a rational analysis of children's false belief reasoning. Our analysis realizes a continuous, evidencedriven transition between two causal Bayesian models of false belief. Both models support prediction and explanation; however, one model is less complex while the other has greater explanatory resources. Because of this explanatory asymmetry, unexpected outcomes weigh more heavily against the simpler model. We test this account empirically by showing children the standard outcome of the false belief task and a novel "psychic" outcome. As expected, we find children whose explanations and predictions are consistent with each model, and an interaction between prediction and explanation. Critically, we find unexpected outcomes only induce children to move from predictions consistent with the simpler model to those consistent with the more complex one, never the reverse.

Research paper thumbnail of Structured Priors for Structure Learning

Proceedings of the …, Jan 1, 2006

Traditional approaches to Bayes net structure learning typically assume little regularity in grap... more Traditional approaches to Bayes net structure learning typically assume little regularity in graph structure other than sparseness. However, in many cases, we expect more systematicity: variables in real-world systems often group into classes that predict the kinds of probabilistic dependencies they participate in. Here we capture this form of prior knowledge in a hierarchical Bayesian framework, and exploit it to enable structure learning and type discovery from small datasets. Specifically, we present a nonparametric generative model for directed acyclic graphs as a prior for Bayes net structure learning. Our model assumes that variables come in one or more classes and that the prior probability of an edge existing between two variables is a function only of their classes. We derive an MCMC algorithm for simultaneous inference of the number of classes, the class assignments of variables, and the Bayes net structure over variables. For several realistic, sparse datasets, we show that the bias towards systematicity of connections provided by our model can yield more accurate learned networks than the traditional approach of using a uniform prior, and that the classes found by our model are appropriate.

Research paper thumbnail of Church: a language for generative models

Uncertainty in Artificial …, Jan 1, 2008