I’ve read a bunch of posts in the last few weeks that say ‘Moore’s Law is over’, not as their key point but as an axiom from which they make further claims. The problem is: this isn’t really true. A bunch of things have changed since Moore’s paper, but the law still roughly holds.

Moore’s law claims that the number of transistors that you can put on a chip (implicitly, for a fixed cost: you could always put more transistors in a chip by paying more) doubles roughly every 18 months. This isn’t quite true anymore, but it was never precisely true and it remains a good rule of thumb. But a load of related things have changed.

First, a load of the free lunches were eaten. Moore’s paper was written in 1965. Even 20 years later, modern processors had limited arithmetic. The early RISC chips didn’t do (integer) divide (sometimes even multiply) in hardware because you could these with a short sequence of add and shift operations in a loop (some CISC chips had instructions for these but implemented them in microcode). Once transistor costs dropped below a certain point, of course you would do them in hardware. Until the mid ‘90s, most consumer CPUs didn’t have floating-point hardware. They had to emulate floating point arithmetic in software. Again, with more transistors, adding these things is a no brainer: they make things faster because they are providing hardware for things that people were already doing.

This started to end in the late ‘90s. Superscalar out-of-order designs existed because just running a sequence of instructions faster was no longer something you got for free. Doubling the performance of something like an 8086 was easy. It wasn’t even able to execute one instruction per cycle and a lot of things were multi-instruction sequences that could become single instructions if you had more transistors, Once you get above one instruction per cycle with hardware integer multiply and divide and hardware floating point, doubling is much harder.

Next, around 2007, Dennard Scaling ended. Prior to this, smaller feature sizes meant lower leakage. This meant that you got faster clocks in the same power budget. The 100 MHz Pentium shipped in 1994. The 1 GHz Pentium 3 in 2000. Six years after that, Intel shipped a 3.2 GHz Pentium 4, which was incredibly power hungry in comparison. Since then, we haven’t really seen an increase in clock speed.

Finally, and most important from a market perspective, demand slowed. The first computers I used were fun but you ran into hardware limitations all of the time. There was a period in the late ‘90s and early 2000s when every new generation of CPU meant you could do new things. These were things you already had requirements for, but the previous generation just wasn’t fast enough to manage. But the things people use computers for today are not that different from the things they did in 2010. Moore’s Law outpaced the growth in requirements. And the doubling in transistor count is predicated on having money from selling enough things in the previous generation. The profits from the 7 nm process funded 4 nm, which funds 2 nm, and so on.

The costs of developing new processes has also gone up but this requires more sales (or higher margins) to fund. And we’ve had that, but mostly driven by bubbles causing people to buy very-expensive GPUs and similar. The rise of smartphones was a boon because it drove a load of demand: billions of smartphones now exist and have a shorter lifespan than desktops and laptops.

Somewhere, I have an issue of BYTE magazine about the new one micron process. It confidently predicted we’d hit physical limits within a decade. That was over 30 years ago. We will eventually hit physical limits, but I suspect that we’ll hit limits of demand being sufficient to pay for new scaling first.

The slowing demand is, I believe, a big part of the reason hyperscalers push AI: they are desperate for a workload that requires the cloud. Businesses compute requirements are growing maybe 20% year on year (for successful growing companies). Moore’s law is increasing the supply per dollar by 100% every 18 months. A few iterations of that and outsourcing compute stops making sense unless you can convince them that they have some new requirements that massively increase their demand.

Moore's second law - Wikipedia

@david_chisnall Agreed. The “Moore’s Law is over” crowd are usually entirely ignorant of the fact that progress in alignment with Moore’s Law sits under an international working group that has operated for decades to ensure continued advancement in the sector. The current instantiation of that is the International Roadmap for Devices and Systems, which is hosted within the IEEE.

@david_chisnall This is a pre-competitive collaboration between all the major suppliers and institutions that sets out all the difficult challenges that must be overcome across the next 15 years, in order to continue to scale at the same rate, and which identifies technology requirements for each stage of that progression.

It is freely available, so to understand what the next 15 years looks like, one only has to go to https://irds.ieee.org and read it.

IRDS

@david_chisnall It also matters that, as "the cloud" and the approaches that work for it are better and better understood, managing on-site infrastructure with cloud approaches is becoming more accessible.

That is not to say that there aren't still things that a dedicated DC can do better than smaller on-site installs, ofc.

From what I'm told from a cloud CTO friend, the hosting business is mostly a marginal race to the bottom at this point, which obviously presents its own issues re. quality.

@ermo @david_chisnall
There are things that could make cloud attractive, even if the hardware costs are higher than on-premise: security, updating software, resiliency/backups, load balancing, network management, etc. Perhaps one day these will evolve to something that can be easily done on-premise; but nowadays cloud can take care of a lot of the administrative headaches.

> Doubling the performance of something like an 8086 was easy

This feels unfair. At the time, I don't think it was seen as easy. And at every point, there was always a sense that Moore's Law is about to fail, because known techniques were understood to be viable for the next few years, but beyond that, something new was going to be needed. And then, we always found something new. And we probably can again, except, as you've pointed out, it's getting harder and harder to find any useful use for extra power.

@david_chisnall

@BenAveling

Easy is relative. When you double the number of transistors available for a design like the 8086, there are a load of things you can do with them that will have an immediate impact on performance for most workloads. The same doubling for a modern CPU will need to be mostly spent on clever structures for trying to keep execution units busy. Doubling the number of execution units would have almost no impact on performance.

@david_chisnall I'm with you on the background but I don't think the core argument holds. Are you saying cost per transistor has continued to decline on the Moore's Law curve? I think that's been flat for over a decade unless I'm missing some major data on that. And, having been part of these decisions, choosing cloud is largely about replacing CapEx and headcount with OpEx and a more flexible balance sheet. No need for any conspiracy.
@david_chisnall excellent gathering of threads. Not sure I agree on the weave but that’s just because I need to think it through (disclaimer: employer=imec-int.com - IRDS is our bread and butter but we tend be ahead of it e.g. https://spectrum.ieee.org/cmos-2)
CMOS 2.0: The Future of Transistors and Interconnects

Imec’s Julien Ryckaert explains the next phase of Moore’s Law

IEEE Spectrum

@david_chisnall this makes sense. They gotta induce demand and they're failing at it.

The other thing that stands out is that for a given workload, a computer that can do it has been getting cheaper. You could survive just fine doing web browsing and word processing on a computer that costs $25 brand new, or used for almost free.

We will, of course, hit a limit somewhere and we're all anxious about it. But when that happens, computers become durable goods, more like a car or fridge than something you have to compulsively upgrade all the time

@fluffykittycat

I refer to this as the Oracle problem. In the early ‘90s, if you were using a database to manage things like payroll and inventory, you needed a big server. Paying for an expensive database was a good idea because you really needed to get the last bit of efficiency out of the system.

By the early 2000s, your company’s database might have doubled in size (7% annual growth), but computers were 64x faster for the same price. Now you could (and a lot of companies did, but shouldn’t) handle the same workload in Access on a moderately good desktop. Another decade later and they could buy three cheap Arm SBCs for under $100 and set up Postgres with replication and handle the same workload without noticeably spiking the CPU usage. Not only did the hardware cost drop to almost nothing, the cost of an expensive database went from a rounding error in the accounting to the vast majority of the cost.

@david_chisnall

My only quibble is with your last paragraph. I think the push for AI datacenters is less a desire to manufacture a demand for cloud services than an attempt to con a lot of sovereign wealth funds and billionaires looking for someplace to park their investment money that will give them a return.

During a lawsuit about Musk’s purchase of Twitter, discovery revealed Musk’s text messages. In them (and the Epstein files echo this) you had billionaires sounding like idiots.

AI datacenters are a meme stock.

@david_chisnall Maybe we have an overrepresentation of millenials who have "new computer every 2 years" ingrained from their teenage years. Basically your point, but maybe also just my sample of people I mostly talk to.

@wink

When I went to university, they had a three-year rolling upgrade programme. 1/3 of the computers were replaced every year and you really didn’t want to be using the ones from the oldest upgrade cycle. By the time I finished my PhD, it was a five-year cycle and no one was complaining. Apparently seven years became common a bit after that. When I left Microsoft, they’d just removed their rolling upgrade programme entirely and all new computers required a business justification (‘the old one broke’ was a legitimate justification).

@david_chisnall Interesting. I never paid attention to stuff at uni, but basically my whole working life with provided hardware (2006-2026) it's been "new machine after 3y earliest".
@david_chisnall I worked in the semiconductor industry in the 90s and early 2000s and every new technology node there were some saying this is the end of Moores law. It was always that we have reached the limit of technology or lithography. But some how the next technology step always came about