The Evolution of Compute: Revisiting the Pace of Progress

Coverage of lessw-blog

ยท PSEEDR Editorial

A retrospective analysis by lessw-blog examines how semiconductor roadmaps have defied predicted limits through architectural innovation rather than traditional scaling.

In a recent retrospective analysis, lessw-blog revisits the landscape of AI compute scaling, comparing current realities against predictions made four years ago. The original discourse emerged during a pivotal moment in AI history-post-GPT-3 but pre-ChatGPT-when the "Scaling Hypothesis" was still a contentious debate rather than an accepted industry standard. The central question then, as it is now, was whether the physical limitations of semiconductor manufacturing would cap the exponential growth required to train increasingly large models.

The post examines the often-cited claim by NVIDIA CEO Jensen Huang that "Moore's Law is dead." While the strict definition of Moore's Law (transistor density doubling every two years via lithography) faces undeniable physical hurdles, lessw-blog argues that the functional pace of progress has not stagnated. Evidence for this is drawn from the International Roadmap for Devices and Systems (IRDS), which has updated its forecasts to significantly delay several previously warned limits. In short, the "hard walls" of physics are being circumvented.

This resilience in hardware progress is attributed to a shift in engineering philosophy. The industry is moving away from a singular focus on pitch scaling-simply shrinking the size of features on a chip-toward Design Technology Co-Optimization (DTCO) and advanced packaging. The author highlights backside power routing as a specific, high-impact innovation. By moving power delivery to the back of the wafer, engineers can reduce the unit cell area substantially without needing to shrink the transistors themselves. This allows for continued density improvements and efficiency gains, essential for the power-hungry clusters running today's Vector DBs and LLMs.

For PSEEDR readers, this analysis offers a hopeful but grounded perspective on the future of AI infrastructure. It suggests that while the methods of achieving performance gains are becoming more complex (and likely more expensive per unit of design effort), the fundamental trajectory of compute scaling remains intact. The "end of progress" has been postponed yet again.

We recommend reading the full post to understand the specific technical milestones that have shifted the roadmap.

Read the full post on LessWrong

Key Takeaways

Read the original post at lessw-blog

Sources