New blog post. I tried something new, got good results, and learned some surprising nuances of GPU programming.
https://probablymarcus.com/blocks/2023/10/19/vectorizing-wide-pytorch-expressions.html
Co-founded & running ML Collective (https://mlcollective.org). Doing #MachineLearning #DeepLearning research at Google Brain->DeepMind. ICLR 2022-2024 DEI Chair. NeurIPS 2023 DIA Chair. Writing at https://deardeuce.com
Full bio here: https://rosanneliu.com/bio/
| Website | https://rosanneliu.com |
| https://twitter.com/savvyRL |
New blog post. I tried something new, got good results, and learned some surprising nuances of GPU programming.
https://probablymarcus.com/blocks/2023/10/19/vectorizing-wide-pytorch-expressions.html
The most popular Arxiv link yesterday (via noahconst@twitter):
Want your image generation model to stop misspelling everything? Try giving it access to character-level input features! https://t.co/AH6pI8VN47 https://t.co/YTOUq95F5x
Current image generation models struggle to reliably produce well-formed visual text. In this paper, we investigate a key contributing factor: popular text-to-image models lack character-level input features, making it much harder to predict a word's visual makeup as a series of glyphs. To quantify this effect, we conduct a series of experiments comparing character-aware vs. character-blind text encoders. In the text-only domain, we find that character-aware models provide large gains on a novel spelling task (WikiSpell). Applying our learnings to the visual domain, we train a suite of image generation models, and show that character-aware variants outperform their character-blind counterparts across a range of novel text rendering tasks (our DrawText benchmark). Our models set a much higher state-of-the-art on visual spelling, with 30+ point accuracy gains over competitors on rare words, despite training on far fewer examples.
For those who know me, the 2020 Uber layoff that affected me had, in retrospect, changed my life for the better. But I wouldn't have known it then. And it takes daily reminders to see it as an "opening" other than "failure."
Sharing my friend's post:
https://www.linkedin.com/posts/arpithahanumanth_amazonlayoffs-opentowork-tech-activity-7000484010917302272-0_nZ