This list is intended to introduce some of the tools of Bayesian statistics and machine learning that can be useful to computational research in cognitive science. The first section mentions several useful general references, and the others provide supplementary readings on specific topics. If you would like to suggest some additions to the list, contact Tom Griffiths.
The sections covered in this list are:
There are no comprehensive treatments of the relevance of Bayesian methods to cognitive science. However, Trends in Cognitive Sciences recently ran a special issue (Volume 10, Issue 7) on probabilistic models of cognition that has a number of relevant papers. You can also check out the IPAM graduate summer school on probabilistic models of cognition at which many of the authors of these papers gave presentations.
The slides from three tutorials on Bayesian methods presented at the Annual Meeting of the Cognitive Science Society might also be of interest:
The 2006, 2008, and 2010 tutorials were based on material appearing in three papers:
Modern artificial intelligence uses a lot of statistical notions, and one of the best places to learn about some of these ideas and their relevance to topics related to cognition is
Radford Neal gave a tutorial presentation at NIPS 2004 on Bayesian machine learning, which outlines some of the philosophy of Bayesian inference, its relevance to the study of learning, and some fundamental methods.
David Mackay has written an excellent introduction to information theory and statistical inference which covers many topics relevant to cognitive science:
Two introductory books on Bayesian statistics (as statistics, rather than the basis for AI, machine learning, or cognitive science) that assume only a basic background, are
There are several advanced texts on Bayesian statistics motivated by statistical decision theory:
The latter is more recent and covers computational methods relevant to Bayesian statistics. The relevance of statistical decision theory to human and machine learning is illustrated in the early chapters of
which are largely reproduced in the second edition
The subjective interpretation of probability motivates other advanced texts:
The former builds on the work of De Finetti, exploring its consequences in a range of situations. The latter comes out of the approach taken by E. T. Jaynes in statistical physics.
Finally, there are also several advanced texts motivated by statistical applications and data analysis:
The former is a classic, illustrating how frequentist methods can be understood from a Bayesian perspective and then going far beyond them. The latter considers the practical problems that can be addressed using Bayesian models, and has chapters on modern computational techniques.
Tom Minka has a number of tutorial papers that apply these ideas in several important cases, including inferring a gaussian distribution, inference about the uniform distribution, and Bayesian linear regression.
Classics on the interpretation of probability
De Finetti gives a detailed account of the structure and consequences of subjective probability. Jeffreys discusses the idea of uninformative priors, and defines the approach to choosing priors that bears his name. Savage is the classic text on the decision-theoretic approach to probability.
Model selection and model averaging
A number of papers on model selection and model averaging by Raftery and colleagues are available here. There is also a webpage listing research on Bayesian model averaging. Some good reviews of both topics are:
Mackay gives a detailed account of how these methods can be applied in artificial neural networks:
The EM algorithmA general introduction to the EM algorithm and its applications is given by Ghahramani and Jordan. Some of the motivation behind EM is explored by Neal and Hinton and in a tutorial by Minka.
Monte Carlo methods
Mackay motivates and explains several Monte Carlo methods. Neal gives a detailed introduction to Markov chain Monte Carlo. The other two books give examples of how these methods can be used in Bayesian models.
The classic reference on graphical models in artificial intelligence is
This is supplemented by Pearl's more recent book, which considers how graphical models can be used to understand causality. In both books, the first two chapters introduce and motivate the ideas involved, while the later chapters explore the consequences of these ideas.
Kevin Murphy has both a toolbox for simulating Bayesian networks in Matlab and a detailed tutorial on the subject, including an extensive reading list. Introductions to inference and learning in Bayesian networks are provided by Jordan and Weiss and Heckerman.
Hidden Markov models and DBNsKevin Murphy has an excellent toolbox for HMMs, as well as a recently written chapter on dynamic Bayesian networks. The classic reference on HMMs is:
Bayesian methods and neural networksMacKay has written a number of papers integrating Bayesian methods with artifical neural networks. Some of the connections between neural networks and probability are explored by Jordan and Neal.