Furthermore, the use of the model also has a considerable footprint. An estimate is given here:
https://medium.com/@chrispointon/the-carbon-footprint-of-chatgpt-e1bc14e4cc2a
3.82 tCO₂e per day assuming a million users with 10 queries per day. It means that after about half a year, the user footprint exceeds the training footprint.
Of course 1M users with 10 queries is a very small number, and likely to grow steeply.
Just to be clear, in the big picture (total global ICT emissions of 5 GtCO₂e) this kind of footprint is still very small. But with a billion users (in practice, this will be other software, not people) with 100 queries per day (a very low estimate I think), we're at 15MtCO₂e, and if there would be 10 such models for different purposes, we can see that it starts to add up quickly.
If that 15MtCO₂e sounds improbably, consider this: currently, Bing and Google each process 10 billion queries per day.
If each of these queries would require a call to an LLM of the same complexity as GPT-3, as seems to be their intention, then that is already 3MtCO₂e purely from Bing and Google searches alone, without any growth or any other applications.
not to forget that developing these things means training them many
times. train, rejigg the network, train, fix a bug, train, many
times. this is not usually reported. most likely it is a couple of
orders of magnitude more than 0.5kT to produce a static trained model.
mostly they don't seem to be continuously feeding these things. not
yet anyways.
As for the As for the "less carbon intensive sources": renewables are only 20% of total US electricity generation. So if Microsoft or OpenAI or Google buys "100% renewables" they simply offload their emissions onto others who can't use renewables. So we should use the average CO2 intensity for the US, or maybe even for the whole world.
Buying electricity from renewables does not reduce global emissions.