title | author | date | |||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
DRAFT SLIDES: Deep Gaussian Processes |
|
1970-01-01 |
\include{../talk-macros.tex}
\include{_deepgp/includes/deep_nn_gp.md} \include{_gp/includes/gp_extremely_short.md}
\newcommand{\hiddenScalar}{f} \newcommand{\latentScalar}{x}
\include{_deepgp/includes/deeptheory.md} \include{_gp/includes/gp-variational-complexity.md}
\include{_gp/includes/bottleneck.md} \include{_gp/includes/low-rank-motivation.md}
[Everything we want to do with a GP involves marginalising
-
Predictions
-
Marginal likelihood
-
Estimating covariance parameters
The posterior of
\include{_gp/includes/nystrom.md} \include{_gp/includes/inducing-notation.md} \include{_gp/includes/inducing-introduction.md}
[Instead of doing]{}
\include{_gp/includes/larger-graph-intro.md} \include{_gp/includes/larger-variational.md} \include{_gp/includes/larger-factorize.md}
\LARGE$$\mappingFunctionVector, \inducingVector \sim \gaussianSamp{\mathbf{0}}{\begin{bmatrix}\Kff & \Kfu\\Kuf & \Kuu\end{bmatrix}}$$
\include{_gp/includes/variational-compression.md} \include{_gp/includes/low-rank-variational.md} \include{_gplvm/includes/bayes-gplvm-intro.md} \include{_gplvm/includes/variational-bayes-gplvm-long.md} \include{_gplvm/includes/nested-variational-compression.md} \include{_gp/includes/larger-gaussian.md}
\include{_gplvm/includes/ard-gplvm.md} \include{_gplvm/includes/bayes-gplvm-intro.md} \include{_gplvm/includes/variational-bayes-gplvm-long.md}
\include{_gp/includes/gp-big-data-technical.md} \include{_gp/includes/gp-big-data.md}
\include{_deepgp/includes/deep-gps.md}
\include{_deepgp/includes/deep-step-function.md} \include{_deepgp/includes/deep-loop-detection.md}
\newcommand{\latentScalar}{f}
\include{_health/includes/deep-health-model.md}
\include{_gplvm/includes/ard_model.md} \include{_gplvm/includes/ard_results.md}
\include{_gplvm/includes/gpds.md}
\include{_gplvm/includes/mrd-gplvm.md} \include{_deepgp/includes/stack-gp-intro.md} \include{_deepgp/includes/stacked-pca.md} \include{_deepgp/includes/stacked-gp.md} \include{_deepgp/includes/deep-pathologies.md} \include{_deepgp/includes/deep-results.md}
\section{What Can We Do that Google Can’t?}
-
Google’s resources give them access to volumes of data (or Facebook, or Microsoft, or Amazon).
-
Is there anything for Universities to contribute?
-
Assimilation of multiple views of the patient: each perhaps from a different patient.
-
This may be done by small companies (with support of Universities).
-
A Facebook app for your personalised health.
-
These methodologies are part of that picture.
\include{_health/includes/deep-health-model.md} \include{_health/includes/deep-health-rangers.md}
\section{Summary}
-
Deep Gaussian Processes allow unsupervised and supervised deep learning.
-
They can be easily adapted to handle multitask learning.
-
Data dimensionality turns out to not be a computational bottleneck.
-
Variational compression algorithms show promise for scaling these models to massive data sets.
\references
\thanks