@lexpedite @ltmccarty — Two ways this could play out:
1. OpenAI feels threatened by the many free + open-source competitors (e.g., Eleuther, Dolly-2), and wonders whether the time/expense of generating a Foundational Model is worth it — when they're competing with "free."
2. OpenAI — with Microsoft money — takes a run at improving the existing GPT-4 model incrementally. Like they did with the davinci releases of GPT3, GPT3.5, etc.
Seems like they're choosing Option 2. Long Microsoft runway
Here is another article on those comments by Sam Altman:
https://www-wired-com.cdn.ampproject.org/c/s/www.wired.com/story/openai-ceo-sam-altman-the-age-of-giant-ai-models-is-already-over/amp
@ltmccarty @lexpedite Yes, that's really helpful. Thanks, Thorne.
I wonder if this comes from the lack of high quality data sources. There are only so many human-created words. Reddit will only get you so far.
Last bastion of high quality legal data: Law? Judicial, statutory, and regulatory text seems like an evergreen source.
You should talk to John Nay about what he's planning to build. 😏
LLM researcher affiliated with NYU and Stanford:
https://law.stanford.edu/directory/john-nay/
https://arxiv.org/a/nay_j_1.html
He's cooking something that should be good.