@lexpedite @ltmccarty — Two ways this could play out:
1. OpenAI feels threatened by the many free + open-source competitors (e.g., Eleuther, Dolly-2), and wonders whether the time/expense of generating a Foundational Model is worth it — when they're competing with "free."
2. OpenAI — with Microsoft money — takes a run at improving the existing GPT-4 model incrementally. Like they did with the davinci releases of GPT3, GPT3.5, etc.
Seems like they're choosing Option 2. Long Microsoft runway
Here is another article on those comments by Sam Altman:
https://www-wired-com.cdn.ampproject.org/c/s/www.wired.com/story/openai-ceo-sam-altman-the-age-of-giant-ai-models-is-already-over/amp
@ltmccarty @lexpedite Yes, that's really helpful. Thanks, Thorne.
I wonder if this comes from the lack of high quality data sources. There are only so many human-created words. Reddit will only get you so far.
Last bastion of high quality legal data: Law? Judicial, statutory, and regulatory text seems like an evergreen source.
This is a "compared to what" and "what goal" problem.
Compared to Reddit and Twitter? Judicial writing is pretty high quality.
Gauging the law's current state (e.g., Roe v. Wade as no longer current law)? Pretty high quality.
Prediction? For that, is there *any* high quality source?