Deep kernel processes

Date:

March 4, 2021

Author:

Hrvoje Stojic



Abstract

Neural networks have taught us that effective performance on difficult tasks requires deep models with flexible top-layer representations. However, inference over intermediate layer features in DGPs or weights in Bayesian NNs is very difficult, with current approaches being highly approximate. Instead, we note that DGPs can be written entirely in terms of positive semi-definite Gram matrices formed by taking the inner product of features with themselves, because the Gram matrices are Wishart distributed, and the next-layer kernel can often be written directly in terms of the Gram matrix. Inference over Gram matrices is much more tractable than inference over weights or features, with joint posterior even being unimodal. We define a tractable deep kernel process, the deep inverse Wishart process, and give a doubly-stochastic inducing-point variational inference scheme that operates on the Gram matrices, not on the features, as in DGPs. We show that the deep inverse Wishart process gives superior performance to DGPs and infinite BNNs on standard fully-connected baselines. Finally, we give motivation additional motivation for this approach, by considering the differences between finite (https://arxiv.org/abs/1910.08013) and infinite (https://arxiv.org/abs/1808.05587) neural networks.


Notes


Share on social media

Share on social media

Share on social media

Share on social media

Related Seminars

Leveraging replication in active learning

Mickael Binois - INRIA Sophia Antipolis - Méditerranée

Jun 24, 2024

Leveraging replication in active learning

Mickael Binois - INRIA Sophia Antipolis - Méditerranée

Jun 24, 2024

Leveraging replication in active learning

Mickael Binois - INRIA Sophia Antipolis - Méditerranée

Jun 24, 2024

Leveraging replication in active learning

Mickael Binois - INRIA Sophia Antipolis - Méditerranée

Jun 24, 2024

From data to confident decisions

Ilija Bogunovic - University College London

Jun 13, 2024

From data to confident decisions

Ilija Bogunovic - University College London

Jun 13, 2024

From data to confident decisions

Ilija Bogunovic - University College London

Jun 13, 2024

From data to confident decisions

Ilija Bogunovic - University College London

Jun 13, 2024

Preference learning with Gaussian processes

Dario Azzimonti - IDSIA

May 23, 2024

Preference learning with Gaussian processes

Dario Azzimonti - IDSIA

May 23, 2024

Preference learning with Gaussian processes

Dario Azzimonti - IDSIA

May 23, 2024

Preference learning with Gaussian processes

Dario Azzimonti - IDSIA

May 23, 2024

Optimal experiment design in Markov chains

Mojmír Mutný - ETH Zurich

Mar 28, 2024

Optimal experiment design in Markov chains

Mojmír Mutný - ETH Zurich

Mar 28, 2024

Optimal experiment design in Markov chains

Mojmír Mutný - ETH Zurich

Mar 28, 2024

Optimal experiment design in Markov chains

Mojmír Mutný - ETH Zurich

Mar 28, 2024