Large Language Model Performance Doubles Every 7 Months
Large Language Model Performance Doubles Every 7 Months
My son has doubled in size every month for the last few months. At this rate he’ll be fifty foot tall by the time he’s seven years old.
Yeah, it’s a stupid claim to make on the face of it. It also ignores practical realities. The first is those is training data, and the second is context windows. The idea that AI will successfully write a novel or code a large scale piece of software like a video game would require them to be able to hold that entire thing in their context window at once. Context windows are strongly tied to hardware usage, so scaling them to the point where they’re big enough for an entire novel may not ever be feasible (at least from a cost/benefit perspective).
I think there’s also the issue of how you define “success” for the purpose of a study like this. The article claims that AI may one day write a novel, but how do you define “successfully” writing a novel? Is the goal here that one day we’ll have a machine that can produce algorithmically mediocre works of art? What’s the value in that?
Yup, that’s what I was alluding to, while it may not still be the case for transistors, they did manage to take 50 odd years to get there, push that trend line from the figure 50 years heh (not saying you should, 5 seems much more conservative)
Take a look at Nvidias pace wrt Moore’s law (of FLOPS) netrouting.com/nvidia-surpassing-moores-law-gpu-i…