Marcus A Brubaker

149 Followers
227 Following
29 Posts

Interested in #ComputerVision, #MachineLearning and #Statistics.

Associate Professor of Computer Science at York University in Toronto.

Faculty Affiliate of the Vector Institute.

Status-only Professor at University of Toronto.

Research Consulting work for Samsung AI Centre, Toronto and Borealis AI

Websitehttps://mbrubake.github.io
Google Scholarhttps://scholar.google.ca/citations?user=x2wyjkAAAAAJ&hl=en
At #ICCV2023 this week and looking for a postdoc in the near future? Shoot me a message and we'll find a time to chat.

"Are Emergent Abilities of Large Language Models a Mirage?"

https://arxiv.org/abs/2304.15004

Are Emergent Abilities of Large Language Models a Mirage?

Recent work claims that large language models display emergent abilities, abilities not present in smaller-scale models that are present in larger-scale models. What makes emergent abilities intriguing is two-fold: their sharpness, transitioning seemingly instantaneously from not present to present, and their unpredictability, appearing at seemingly unforeseeable model scales. Here, we present an alternative explanation for emergent abilities: that for a particular task and model family, when analyzing fixed model outputs, emergent abilities appear due to the researcher's choice of metric rather than due to fundamental changes in model behavior with scale. Specifically, nonlinear or discontinuous metrics produce apparent emergent abilities, whereas linear or continuous metrics produce smooth, continuous predictable changes in model performance. We present our alternative explanation in a simple mathematical model, then test it in three complementary ways: we (1) make, test and confirm three predictions on the effect of metric choice using the InstructGPT/GPT-3 family on tasks with claimed emergent abilities; (2) make, test and confirm two predictions about metric choices in a meta-analysis of emergent abilities on BIG-Bench; and (3) show to choose metrics to produce never-before-seen seemingly emergent abilities in multiple vision tasks across diverse deep networks. Via all three analyses, we provide evidence that alleged emergent abilities evaporate with different metrics or with better statistics, and may not be a fundamental property of scaling AI models.

arXiv.org
RT @jakevdp
Let's talk about JAX's vmap! It's a transformation that can automatically create vectorized, batched versions of your functions... but what exactly it does is sometimes misunderstood. So let's dig-in!

The list of papers accepted at SaTML 2023:

https://satml.org/accepted-papers/

I'd like to extend a big thank you to all the PC members whose work made putting together this exciting program possible!

Thank you to all who submitted their work: we hope the reviews provided useful feedback.

Accepted Papers

@at I'd suggest contacting the editor in chief of the journal about this. As the field has grown, many people have become editors (and ACs) without really understanding the job. I had a similar experience with a paper and after I raised it with the EIC they admitted the editor screwed up.
Here's my attempt to synthesize historical trends and patterns when a new technology came along that seemed to automate art, like photography and computer animation, and what this might tell us about "AI" art.
https://aaronhertzmann.com/2022/12/17/when-tech-changes-art.html
These ideas and text are still a work-in-progress.
1/
#AI #AIart #photography #innovation #technology #disruption
When Machines Change Art

At a few times in history, new technologies came along that changed the way we make art. Machines, chemicals, and/or algorithms replaced some of the steps that artists did, changing how we made art—and, sometimes, radically transforming what we thought think art is.

Aaron Hertzmann’s blog
From free speech to fascism in 7.5 weeks

Public Service Warning

Mastodon has a very big surge of new users right now. There's no way to tell if it will be sustained, but at this early point it looks similar to Nov 18 when Musk pulled the employee purge.

It is very challenging for system administration to accommodate so many new users. If your server starts to struggle, it is not broken and will get sorted out.

We're all in this together. It's our social network. Be patient. What we are building is amazing.

#twittermigration

Normalise being wrong. Normalise admitting to mistakes and changing your opinion as you learn more. It's how we learn. It's part of being human.
This is not a chart depicting Moore’s Law.

This chart is the CRAYOLA COROLLARY showing that the number of Crayola crayon colors doubles approximately every 18 years.