WHAT MY DEEP MODEL DOESN'T KNOW...
I recently spent some time trying to understand why dropout deep learning models work so well – trying to relate them to new research from the last couple of years. I was quite surprised to see how close these were to Gaussian processes. I was even more surprised to see that we can get uncertainty information from these deep learning models for free – without changing a thing.
[Post]DROPOUT AS A BAYESIAN APPROXIMATION
05/06/2015
2 new papers on dropout as a Bayesian approximation, with applications to model uncertainty in deep learning [1], and Bayesian convolutional neural networks [2]have been added to the publications
SOFTWARE
OPEN SOURCE PROJECTS I'M CURRENTLY WORKING ON
VSSGP
An implementation of the Variational Sparse Spectrum Gaussian Process using Theano (a Python package for symbolic differentiation).
[Software] [Paper]
CLGP
An implementation of the Categorical Latent Gaussian Process using Theano (a Python package for symbolic differentiation).
[Software] [Paper]
GPARML
A light-weight and minimal Python implementation of parallel inference for the Bayesian Gaussian process latent variable model and GP regression.
[Software] [Paper]
UNIVERSITY OF CAMBRIDGE PRESENTATION TEMPLATE
This is a presentation template with the colour scheme of the University of Cambridge. The beamer template is based oncambridge-beamer with changes to the colour scheme and page layout.
[Software] [Example]
GIZA#
An optimised C++ extension of Giza++ (a word alignment software package) implementing the hierarchical Pitman-Yor process alignment models.
[Software] [Paper]
發(fā)表論文
BAYESIAN CONVOLUTIONAL NEURAL NETWORKS WITH BERNOULLI APPROXIMATE VARIATIONAL INFERENCE
We present an efficient Bayesian convolutional neural network (convnet). The model offers better robustness to over-fitting on small data and achieves a considerable improvement in classification accuracy compared to previous approaches. We give state-of-the-art results on CIFAR-10 following our insights.
Yarin Gal, Zoubin GhahramaniIn submission, 2015 [arXiv] [BibTex]
DROPOUT AS A BAYESIAN APPROXIMATION: REPRESENTING MODEL UNCERTAINTY IN DEEP LEARNING
We show that dropout in multilayer perceptron models (MLPs) can be interpreted as a Bayesian approximation. Results are obtained for modelling uncertainty for dropout MLP models - extracting information that has been thrown away so far, from existing models. This mitigates the problem of representing uncertainty in deep learning without sacrificing computational performance or test accuracy.
Yarin Gal, Zoubin GhahramaniIn submission, 2015 [arXiv] [Appendix] [BibTex]
DROPOUT AS A BAYESIAN APPROXIMATION: INSIGHTS AND APPLICATIONS
Deep learning techniques lack the ability to reason about uncertainty over the features. We show that a multilayer perceptron (MLP) with arbitrary depth and non-linearities, with dropout applied after every weight layer, is mathematically equivalent to an approximation to a well known Bayesian model. This paper is a short version of the appendix of "Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning".
Yarin Gal, Zoubin GhahramaniDeep Learning Workshop, ICML, 2015 [PDF] [BibTex]
AN INFINITE PRODUCT OF SPARSE CHINESE RESTAURANT PROCESSES
We define a new process that gives a natural generalisation of the Indian buffet process (used for binary feature allocation) into categorical latent features. For this we take advantage of different limit parametrisations of the Dirichlet process and its generalisation the Pitman–Yor process.
Yarin Gal, Tomoharu Iwata, Zoubin Ghahramani10th Conference on Bayesian Nonparametrics (BNP), 2015 [Talk] [BibTex] We thank BNP for the travel award.
**
IMPROVING THE GAUSSIAN PROCESS SPARSE SPECTRUM APPROXIMATION BY REPRESENTING UNCERTAINTY IN FREQUENCY INPUTS
Standard sparse pseudo-input approximations to the Gaussian process (GP) cannot handle complex functions well. Sparse spectrum alternatives attempt to answer this but are known to over-fit. We use variational inference for the sparse spectrum approximation to avoid both issues. We extend the approximate inference to the distributed and stochastic domains.
Yarin Gal, Richard TurnerICML, 2015** [PDF] [Software] [BibTex]
LATENT GAUSSIAN PROCESSES FOR DISTRIBUTION ESTIMATION OF MULTIVARIATE CATEGORICAL DATA
Multivariate categorical data occur in many applications of machine learning. One of the main difficulties with these vectors of categorical variables is sparsity. The number of possible observations grows exponentially with vector length, but dataset diversity might be poor in comparison. Recent models have gained significant improvement in supervised tasks with this data. These models embed observations in a continuous space to capture similarities between them. Building on these ideas we propose a Bayesian model for the unsupervised task of distribution estimation of multivariate categorical data.
Yarin Gal, Yutian Chen, Zoubin GhahramaniWorkshop on Advances in Variational Inference, NIPS, 2014 [PDF] [Poster] [Presentation] [BibTex] ICML, 2015 [PDF] [Software] [BibTex] We thank Google DeepMind for the travel award.
**
DISTRIBUTED VARIATIONAL INFERENCE IN SPARSE GAUSSIAN PROCESS REGRESSION AND LATENT VARIABLE MODELS
We develop parallel inference for sparse Gaussian process regression and latent variable models. These processes are used to model functions in a principled way and for non-linear dimensionality reduction in linear time complexity. Using parallel inference we allow the models to work on much larger datasets than before.
Yarin Gal, Mark van der Wilk, Carl E. RasmussenWorkshop on New Learning Models and Frameworks for Big Data, ICML, 2014 [arXiv] [Presentation] [Software] [BibTex] NIPS, 2014 [PDF] [BibTex] We thank NIPS for the travel award.
**
FEATURE PARTITIONS AND MULTI-VIEW CLUSTERINGS
We define a new combinatorial structure that unifies Kingman's random partitions and Broderick, Pitman, and Jordan's feature frequency models. This structure underlies non-parametric multi-view clustering models, where data points are simultaneously clustered into different possible clusterings. The de Finetti measure is a product of paintbox constructions. Studying the properties of feature partitions allows us to understand the relations between the models they underlie and share algorithmic insights between them.
Yarin Gal, Zoubin GhahramaniInternational Society for Bayesian Analysis (ISBA), 2014[Link] [Poster]We thank ISBA for the travel award.
**
DIRICHLET FRAGMENTATION PROCESSES
We introduce a new class of models over trees based on the theory of fragmentation processes. The Dirichlet Fragmentation Process Mixture Model is an example model derived from this new class. This model has efficient and simple inference, and significantly outperforms existing approaches for hierarchical clustering and density modelling.
Hong Ge, Yarin Gal, Zoubin GhahramaniIn submission, 2014[PDF] [BibTex]
PITFALLS IN THE USE OF PARALLEL INFERENCE FOR THE DIRICHLET PROCESS
We show that the recently suggested parallel inference for the Dirichlet process is conceptually invalid. The Dirichlet process is important for many fields such as natural language processing. However the suggested inference would not work in most real-world applications.
Yarin Gal, Zoubin GhahramaniWorkshop on Big Learning, NIPS, 2013[PDF] [Presentation] [BibTex] ICML, 2014[PDF] [Talk] [Presentation] [Poster] [BibTex]
VARIATIONAL INFERENCE IN THE GAUSSIAN PROCESS LATENT VARIABLE MODEL AND SPARSE GP REGRESSION – A GENTLE TUTORIAL
We present an in-depth and self-contained tutorial for sparse Gaussian Process (GP) regression. We also explain GP latent variable models, a tool for non-linear dimensionality reduction. The sparse approximation reduces the time complexity of the models from cubic to linear but its development is scattered across the literature. The various results are collected here.
Yarin Gal, Mark van der WilkTutorial, 2014[arXiv] [BibTex]
SEMANTICS, MODELLING, AND THE PROBLEM OF REPRESENTATION OF MEANING – A BRIEF SURVEY OF RECENT LITERATURE
Over the past 50 years many have debated what representation should be used to capture the meaning of natural language utterances. Recently new needs of such representations have been raised in research. Here I survey some of the interesting representations suggested to answer for these new needs.
Yarin GalLiterature survey, 2013[arXiv] [BibTex]
A SYSTEMATIC BAYESIAN TREATMENT OF THE IBM ALIGNMENT MODELS
We used a non-parametric process in models that align words between pairs of sentences. These alignment models are used at the core of all machine translation systems. We obtained a significant improvement in translation using the process.
Yarin Gal, Phil BlunsomAssociation for Computational Linguistics (NA-ACL), 2013[PDF] [Presentation] [BibTex]
RELAXING HMM ALIGNMENT MODEL ASSUMPTIONS FOR MACHINE TRANSLATION USING A BAYESIAN APPROACH
We used a non-parametric process to relax some of the restricting assumptions often used in machine translation. When a long history of translation words is not available the process falls-back onto shorter histories in a principled way.
Yarin GalMaster's Dissertation, 2012[PDF] [BibTex]
OVERCOMING ALPHA-BETA LIMITATIONS USING EVOLVED ARTIFICIAL NEURAL NETWORKS
We trained a feed-forward neural network to play checkers. The network acts as both the value function for a min-max algorithm and a heuristic for pruning tree branches in a reinforcement learning setting. We used no supervised signal for training - a set of networks was assessed by playing against each-other and the winning networks' weights where changed slightly.
Yarin Gal, Mireille AvigalMachine Learning and Applications (IEEE), 2010[PDF] [BibTex]
TALKS
LATENT GAUSSIAN PROCESSES FOR DISTRIBUTION ESTIMATION OF MULTIVARIATE CATEGORICAL DATA
We discuss the issues with representing high dimensional vectors of discrete variables, and existing models that attempt to estimate the distribution of such. We then present our approach which relies on a continuous latent representation for the discrete data.
Yarin GalInvited talk: Microsoft Research, Cambridge, 2015Invited talk: NTT Labs, Kyoto, Japan, 2015[Presentation] [Video]
REPRESENTATIONS OF MEANING
We discuss various formal representations of meaning, including Gentzen sequent calculus, vector spaces over the real numbers, and symmetric closed monoidal categories.
Yarin GalInvited talk: Trinity College Mathematical Society, University of Cambridge, 2015[Presentation]
SYMBOLIC DIFFERENTIATION FOR RAPID MODEL PROTOTYPING IN MACHINE LEARNING AND DATA ANALYSIS – A HANDS-ON TUTORIAL
We talk about the theory of symbolic differentiation and demonstrate its use through the Theano Python package. We give two example models: logistic regression and a deep net, and continue to talk about rapid prototyping of probabilistic models with SVI. The talk is based in part on the Theano online tutorial.
Yarin GalMLG Seminar, 2014[Presentation]
RAPID PROTOTYPING OF PROBABILISTIC MODELS USING STOCHASTIC VARIATIONAL INFERENCE
In data analysis we have to develop new models which can often be a lengthy process. We need to derive appropriate inference which often involves cumbersome implementation which changes regularly. Rapid prototyping answers similar problems in manufacturing, where it is used for quick fabrication of scale models of physical parts. We present Stochastic Variational Inference (SVI) as a tool for rapid prototyping of probabilistic models.
Yarin GalShort talk, 2014[Presentation]
DISTRIBUTED INFERENCE IN BAYESIAN NONPARAMETRICS – THE DIRICHLET PROCESS AND THE GAUSSIAN PROCESS
I present distributed inference methodologies for two major processes in Bayesian nonparametrics. Pitfalls in the use of parallel inference for the Dirichlet process are discussed, and distributed variational inference in sparse Gaussian process regression and latent variable models is presented.
Yarin GalInvited talk: NTT Labs, Kyoto, Japan, 2014
EMERGENT COMMUNICATION FOR COLLABORATIVE REINFORCEMENT LEARNING
Slides from a seminar introducing collaborative reinforcement learning and how learning communication can improve collaboration. We use game theory to motivate the use of collaboration in a multi-agent setting. We then define multi-agent and decentralised multi-agent Markov decision processes. We discuss issues with these definitions and possible ways to overcome them. We then transition to emergent languages. We explain how the use of an emergent communication protocol could aid in collaborative reinforcement learning. Reviewing a range of emergent communication models developed from a linguistic motivation to a pragmatic view, we finish with an assessment of the problems left unanswered in the field.
Yarin Gal, Rowan McAllisterMLG Seminar, 2014[Presentation]
THE BOREL–KOLMOGOROV PARADOX
Slides from a short talk explaining the the Borel–Kolmogorov paradox, alluding to possible pitfalls in probabilistic modelling. The slides are partly based on Jaynes, E.T. (2003) "Probability Theory: The Logic of Science".
Yarin GalShort talk, 2014[Presentation]
BAYESIAN NONPARAMETRICS IN REAL-WORLD APPLICATIONS: STATISTICAL MACHINE TRANSLATION AND LANGUAGE MODELLING ON BIG DATASETS
Slides from a seminar introducing statistical machine translation and language modelling as real-world applications of Bayesian nonparametrics. We give a friendly introduction to statistical machine translation and language modelling, and then describe how recent developments in the field of Bayesian nonparametrics can be exploited for these tasks. The first part of the presentation is based on the lecture notes by Dr Phil Blunsom.
Yarin GalMLG Seminar, 2013[Presentation]