Check out the paper π arxiv.org/pdf/2602.087...
Looking forward to presenting this work in Rio, and many thanks to @vincefort.bsky.social for his supervision!
Check out the paper π arxiv.org/pdf/2602.087...
Looking forward to presenting this work in Rio, and many thanks to @vincefort.bsky.social for his supervision!
Are humble Gaussian priors enough for BNNs to model highly complex stochastic processes? Do well-specified BNN priors remove the need for more costly approximate inference algorithms?
We provide answers in the paper!
2. It turns BNNs into flexible generative models (i.e., sampling from learned priors.
3. It enables capabilities that have been difficult for neural processes so far, including:
β’ Within-task minibatching
β’ Meta-learning in extremely data-scarce regimes.
Why this matters:
1. It lets us study BNNs under well-specified, data-driven priors rather than the usual isotropic guff.
3. The resulting model can be viewed as a neural process whose latent variable is the weights of a BNN, with the network itself acting as the decoder.
2. This is achieved via per-dataset amortised variational inference, allowing the model to infer dataset-specific posteriors while learning a shared, well-specified prior.
What we do:
1. We propose a way to learn a prior over neural network weights from data, using a collection of related datasets.
Bayesian neural network (BNN) practitioners have to specify priors over weights, but doing so is often unclear or ad hoc. In this paper, we bridge Bayesian deep learning and probabilistic meta-learning to offer a concrete answer.
The work tackles a fairly fundamental question in Bayesian deep learning:
"how can we be Bayesian if we donβt have any meaningful prior beliefs in the first place?"
Iβm pleased to share that our latest paper, βAmortising Inference and Meta-Learning Priors in Neural Networksβ, has been accepted to ICLR 2026 in Rio!
Are bitterns as fiendishly difficult to spot in Singapore as they are in Europe?
Arxiv link: arxiv.org/pdf/2504.01650
Itβs nice to be able to get the ball rolling on my PhD with this paper, and a nice achievement to have published my first non-workshop paper. A big thanks to @vincefort.bsky.social for his supervision on this project!
1.) you want/need GP levels of interpretability
2.) you donβt have that many training tasks, so need SOTA data efficiency (at the meta-level)
3.) you have accurate domain knowledge (in GP-prior form)
4.) each task has too many observations for exact GP inference
If you need probabilistic predictions across multiple related tasks/datasets, you should use this model if any combination of the following hold:
We introduce the ability to meta-learn sparse variational Gaussian process inference, resulting in a new type of neural process that is amenable to prior elicitation.
Very pleased to share that our new paper βSparse Gaussian Neural Processesβ has been accepted under the proceedings track at AABI 2025! π (1/n)
I've seen things you people wouldn't believe.
Attacks from reviewers on fire off the shoulders of #OpenReview.
I watched logic fallacies glitter in the dark near @iclr-conf.bsky.social
All those moments will be lost in time, like tears in the next resubmission.β―
Time to die.
#ML #Ai #PhDlife
πββοΈ
Thanks for putting this together - keen to be added!