@codinghorror @thomasfuchs this is incorrect. Weather models do understand weather and simulate the atmosphere to achieve their predictions. Randomness is sometimes introduced to see different outcomes if the initial measured conditions aren’t quite right
It’s not statistical in the sense of a language model. Also putting in “looking at studies” in your prompts might just be cluing the generator to make a more serious sounding answer that matches you language
@cjensen @codinghorror @thomasfuchs Just to clarify here, you're saying that "weather models do understand weather" in a way that LLMs don't?
What is the definition of "understanding" that is being used here?
@paul @codinghorror @thomasfuchs good question!
Weather models simulate the atmosphere. They do this by subdividing the map into sections, simulating the section, then integrating the result. Each loop of simulation is for a period of time (eg an hour of simulated time). Loop for however many hours into the future you want to predict.
1/
@paul @codinghorror @thomasfuchs
So in this sense the weather model knows what atmosphere is how to simulate it
By contrast, language models don’t understand what paragraphs, sentences, or words mean. They are just statistical models which pick a sequence of words and punctuation that closely match what a likely response to your prompt should look like.
2/
@paul @codinghorror @thomasfuchs
It’s a miracle this works as well as it does.
But knowing how it works, you can see where the “hallucinations” come from: if you ask for a legal brief and there is no precedent in the stats for the argument, it will generate words that statistically best fit the question, including fake cites if that statistically seems to fit
@cjensen @codinghorror @thomasfuchs
Packing an LLM's context window with a request for a nonexistent legal brief seems more like putting fictional synthetic initial conditions into a weather model. Garbage in, garbage out.
Also, it looks like weather labs are rapidly adopting transformer architectures of one kind or another:
https://www.alcf.anl.gov/news/argonne-develops-new-kind-ai-model-weather-prediction
https://e360.yale.edu/features/artificial-intelligence-weather-forecasting
@cjensen @codinghorror @thomasfuchs
For understandable overviews of how weather models are undergirded by the atmospheric physics, see this question. Forecasts have improved mostly because we are doing a better job collecting weather observations and faster computers mean we can make the grids smaller.
The statistical models come into play after running the physics models. They can help forecasters understand the spread of possible outcomes and refine the forecast using the history of actual weather given certain model conditions.
While it is true that LLMs use similar mathematical tools to produce results, the physics models do the heavy lifting in weather forecasting.
While that article does cite “more data”, it’s not as used in the LLM sense. It’s the inputs to the physical models. Where before the models had to simulate the flow of air and water based on a 90km grid, now they can work on a 1.6km grid (why do I think the original source was written in terms of miles?).
That said, there’s been a fair amount of success on using ML to simulate solutions of Navier-Stokes and friends to speed up such calculations in a pattern recognition sense, though I don’t think the work has been widely applied in meteorology.