Philipp Hennig

@PhilippHennig
7 Followers
64 Following
22 Posts
Professor for the Methods of Machine Learning at the University of Tübingen
Websitehttp://mml.inf.uni-tuebingen.de
Bookhttps://probabilistic-numerics.org/textbooks
YouTubehttps://YouTube.com/c/TübingenML
Twitterhttps://twitter.com/PhilippHennig5

Wonderfully insightful talk by Jimmy Ba in our “Has it Trained Yet” Workshop (hity-workshop.github.io/NeurIPS2022/).

Find us in Theatre B now for the next talk, by Susan Zhang

And you can still do our questionnaire at http://forms.gle/oxs9nUHzEkJPcZxH7!

HITY Workshop Poll

With this survey, we want to feel the pulse of how neural networks are currently being trained. Please answer the following questions with how you would typically train and what your setup usually looks like. The aggregated results of this poll will be presented at the HITY workshop at NeurIPS 2022.

Google Docs
Poster: Wednesday, 16-18, Hall J, poster 313
Paper: openreview.net/forum?id=Zzi8O…
Video: youtu.be/xsXifiz2Zfo
Code: github.com/JonathanWenger/itergp

This may seem like a little niche GP paper. But linear algebra / least squares is everywhere. If you agree that each least-squares fit should be a GP, then every run of a linear solver should also be a GP. For all the same reasons.

It’s also an example for how computationally self-aware agents should work: Keeps track not just of the fact you've only seen finite data so far, but also only done a subset of the necessary computation for training so far. We should want that everywhere.

It turns out you can actually keep track of the error of the linear solver with minimal overhead (just like you can keep track of the GP covariance in addition to the posterior mean, at minimal overhead). And the way to do it applies pretty much across all LinAlg Methods.

(Caveat: one can keep track of the _combined_ error of finite data and finite compute. That’s what you actually want, right? Tracking either of them individually costs extra. I’ll leave this 🔫 lying right here on the table…)

That’s _exactly_ like the data itself: If you have N data, but only use M<N, the posterior mean is “off”.

For the data, we all agree how to deal with this: That’s what the posterior covariance is for, right? Finite data -> finite error, modelled by the covariance.

Why not do the same for compute? Finite compute -> finite error, modelled by … well … the covariance!

Say you have a set of N data. To do GP inference, you call a linear algebra method: Cholesky, Conjugate Gradients, gradient descent (don’t!), inducing points. All these algorithms — even Cholesky! — iterate over an atomic step that amounts to a linear projection (mat-vec product)

You can stop the algorithm before it converges (unless N<1e4, you probably do). If you do, its estimate for the GP (posterior mean _and_ covariance) are not yet right. They have an error.

I’m super excited for one of our #NeurIPS #probnum poster tomorrow!

You know the textbook plot for Gaussian Processes, where uncertainty drops as the model adapts to each new datum?

In this paper, we do this iterative update for the (linear algebra) *computation* in the GP. A 🧵

I'm stoked for our upcoming NeurIPS workshop on Deep Training (https://hity-workshop.github.io/NeurIPS2022/).

In preparation, we're asking the community: How do *you* train your deep networks?

Share your training protocol, secrets, and frustrations in this short poll: http://forms.gle/oxs9nUHzEkJPcZxH7

Next Friday in Theatre B of the New Orleans Convention Center, we'll all compare notes, with experts from industry and academia.

Home

We all think we know how to train neural nets, but we seem to have different ideas. Let’s discuss which methods truly speed up training!

Has It Trained Yet?

We are now on mastodon! Follow us for news about our Cluster of Excellence "Machine Learning: New Perspectives for Science", the work of our scientists and their latest publications.

On our blog you'll find articles on latest research, opinion pieces on #MachineLearning (ML) and its impact on science and society and you will also get more insight into the scientists and their motivation to do research in #ML. #introduction #AI #science

Check out our blog now:
www.machinelearningforscience.de

Oh, and there’ll be socials, too! With the wonderful crowd of students and colleagues in Tübingen.

Particular thanks to the Excellence Cluster ML4Science for financial support!

Sign up now, spaces are filling up fast :).