Julius Berner

90 Followers
34 Following
11 Posts
PhD student @ UniVie | research intern @ MetaAI | former research intern @ NVIDIA | bridging theory and practice in deep learning
Websitehttps://jberner.info
Twitterhttps://twitter.com/julberner
LinkedInhttps://linkedin.com/in/julius-berner
GitHubhttps://github.com/juliusberner

🚨 Internship opportunity🚨

If you are interested in information theory, generative modeling, and AI4 Science let’s learn and explore together @ FAIR New York.

I got one internship spot for 2023. Apply until Jan 12 via the website, and mind the minimum requirements!

Link to the application form -> https://www.metacareers.com/jobs/901899764520819/
Additionally send me an email (karenu@) with subject: [Internship 2023] YOUR NAME
Talk about who you are and what we could work on together. Please no DMs.

Research Scientist Intern, AI Core Machine Learning (PhD)

Meta's mission is to give people the power to build community and bring the world closer together. Together, we can help people build stronger communities - join us.

Meta Careers

Had a fantastic week at #NeurIPS and look forward to seeing you all again soon!

🗒️ More details on our work: https://sigmoid.social/@jberner/109314075436493448
Thanks to my great collaborators Lorenz Richter & @karen_ullrich.

✈️ I am also grateful to G-Research for the travel grant.

Julius Berner (@[email protected])

Attached: 1 image Explore the connection between diffusion models and optimal control 🔥 🎙️ Come to our oral at the #NeurIPS workshop on score-based methods and let’s discuss how one field can benefit from the other. 📖 http://bit.ly/3UAhena Great work with Lorenz Richter and @karen_ullrich

Sigmoid Social

Bits-Back coding (with asymmetric numeral systems) can be used for lossless compression with latent variable models at a near optimal rate: https://arxiv.org/abs/1901.04866

For extensions to hierarchical latent variable models, such as diffusion models, see:
1️⃣ https://arxiv.org/abs/1912.09953 (this is the method we build upon)
2️⃣ https://arxiv.org/abs/1905.06845 (this is the Bit-Swap method used in “Variational Diffusion Models” https://arxiv.org/abs/2107.00630)

Practical Lossless Compression with Latent Variables using Bits Back Coding

Deep latent variable models have seen recent success in many data domains. Lossless compression is an application of these models which, despite having the potential to be highly useful, has yet to be implemented in a practical manner. We present `Bits Back with ANS' (BB-ANS), a scheme to perform lossless compression with latent variable models at a near optimal rate. We demonstrate this scheme by using it to compress the MNIST dataset with a variational auto-encoder model (VAE), achieving compression rates superior to standard methods with only a simple VAE. Given that the scheme is highly amenable to parallelization, we conclude that with a sufficiently high quality generative model this scheme could be used to achieve substantial improvements in compression rate with acceptable running time. We make our implementation available open source at https://github.com/bits-back/bits-back .

arXiv.org

📉The plot shows the avg. effective compression rate and terms of the (negative) ELBO over the time-steps of the diffusion model for CIFAR-10.

💻Our implementation supports ImageNet out of the box and can be extended to other datasets and diffusion models!

⏩ [2/3]

📢 📢 New Feature in #NeuralCompression repo: Bits-Back compression for diffusion models!
Compress image data 🖼️ using diffusion models at an effective rate close to the (negative) ELBO.

See: https://github.com/facebookresearch/NeuralCompression/tree/main/projects/bits_back_diffusion

Some context ⏩ [1/3]

NeuralCompression/projects/bits_back_diffusion at main · facebookresearch/NeuralCompression

A collection of tools for neural compression enthusiasts. - NeuralCompression/projects/bits_back_diffusion at main · facebookresearch/NeuralCompression

GitHub
Highlights:
1️⃣ Log-density of the underlying SDE satisfies a HJB equation.
2️⃣ ELBO follows directly from the verification theorem.
3️⃣ Diffusion-based approach to sample from (unnormalized) densities.
...and more to come!
Explore the connection between diffusion models and optimal control 🔥
🎙️ Come to our oral at the #NeurIPS workshop on score-based methods and let’s discuss how one field can benefit from the other.
📖 http://bit.ly/3UAhena
Great work with Lorenz Richter and @karen_ullrich

Hi all and thanks to @thegradient for setting up this server.

#introduction
I'm in the process of finishing my PhD, where I've been doing research on DL theory and neural solvers for PDEs. During my ongoing internship at MetaAI with @karen_ullrich, I have been working on diffusion models and neural compression.