Tom Goldstein

311 Followers
39 Following
39 Posts
Associate Professor at Maryland. AI safety, security, and privacy. Fundamental machine learning theory and science. Optimization, applied math, stupid puns.
https://www.cs.umd.edu/~tomg

Marie Curie's notebooks, which are radioactive and must be stored in a lead-lined box in the Bibliothèque Nationale. Curie’s corpse is also radioactive. Her coffin is lined in an inch of lead. Both will remain radioactive for 1,500+ years.

More info: https://www.sciencealert.com/these-personal-effects-of-marie-curie-will-be-radioactive-for-another-1-500-years

#HistSci #ScienceMastodon #MedMastodon #science #scientist #HistoryOfScience #HistMed #WomenInStem #TIL #TheMoreYouKnow #WeirdHistory

Marie Curie's Belongings Will Be Radioactive For Another 1,500 Years

Marie Curie, known as the 'mother of modern physics', died from aplastic anaemia, a rare condition linked to high levels of exposure to her famed discoveries, the radioactive elements polonium and radium.

ScienceAlert

The cost to OpenAI may be lower though, because of its partnership with Microsoft.

Either way, that ain't cheap. Some say it's wasteful to pour these kinds of resources (and carbon) into a demo. But hey, it's not the worst use of Elon's money that we've seen of late 💸💸💸

I estimate the cost of running ChatGPT is $100K per day, or $3M per month. This is a back-of-the-envelope calculation. I assume nodes are always in use with a batch size of 1. In reality they probably batch during high volume, but have GPUs sitting fallow during low volume.

The real costs for a typical organization would almost certainly be higher than this because parallelization is not 100% efficient, GPUs are not 100% utilized, and my runtime estimate is optimistic.

How many queries are there?
ChatGPT acquired 1M users within its first 5 days of operation. If an average user has made 10 queries per day, I think it’s reasonable to estimate that ChatGPT serves ~10M queries per day.
But it generates a lot of words! The model usually responds to my queries with ~30 words, which adds up to about 1 cent per query.

Of course, you could never fit ChatGPT on a single GPU. You would need 5 80Gb A100 GPUs just to load the model and text. ChatGPT cranks out about 15-20 words per second. If it uses A100s, that could be done on an 8-GPU server (a likely choice on Azure cloud).

So what would this cost to host? On Azure cloud, each A100 card costs about $3 an hour. That's $0.0003 per word generated.

We don’t know the exact architecture of ChatGPT, but OpenAI has said that it is fine-tuned from a variant of GPT-3.5, so it probably has 175B parameters. That's pretty big.

How fast could it run? A 3-billion parameter model can generate a token in about 6ms on an A100 GPU (using half precision+tensorRT+activation caching). If we scale that up to the size of ChatGPT, it should take 350ms secs for an A100 GPU to print out a single word.

How many GPUs does it take to run ChatGPT? And how expensive is it for OpenAI? Let’s find out! 🧵🤑
“Max Rockatansky from the 1979 post-apocalyptic film Muppet Mad Max, 35mm Eastmancolor” #midjourney #AIArt
One suggestion: It would be great to have an MLM version of Galactica. This would make it ideal for matching papers to reviewers, searching databases for similar papers, and fixing grammatical problems for non-native speakers. As a bonus, it would be perceived as safer.