# The End of Scaling Laws

Canonical: https://mosiah.org/articles/the-end-of-scaling-laws/
Interactive: https://mosiah.org/#Articles%2Fthe-end-of-scaling-laws

# The End of Scaling Laws

//A New Paradigm in AI Development//

//Related:// [[sources|Article Sources/the-end-of-scaling-laws]] · [[notes|Article Notes/the-end-of-scaling-laws]] · [[metadata|Article Metadata/the-end-of-scaling-laws]] · [[Published Pieces]]

As we stand at the frontier of artificial intelligence, a paradigm shift is emerging. The long-held belief in the power of ever-larger models and datasets—often referred to as "scaling laws"—is being challenged. This essay explores the factors contributing to this shift and speculates on the future of AI development.

<div class="captioned-image-container">

<figure>
<a href="https://substackcdn.com/image/fetch/$s_!MVN4!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F25c79b85-3561-4b0f-bc5c-a8fd2f0bbb74_1792x1024.webp" class="image-link image2 is-viewable-img" target="_blank"></a>
<div class="image2-inset">
<img src="https://substackcdn.com/image/fetch/$s_!MVN4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F25c79b85-3561-4b0f-bc5c-a8fd2f0bbb74_1792x1024.webp" class="sizing-normal" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/25c79b85-3561-4b0f-bc5c-a8fd2f0bbb74_1792x1024.webp&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:832,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:731918,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/webp&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" srcset="https://substackcdn.com/image/fetch/$s_!MVN4!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F25c79b85-3561-4b0f-bc5c-a8fd2f0bbb74_1792x1024.webp 424w, https://substackcdn.com/image/fetch/$s_!MVN4!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F25c79b85-3561-4b0f-bc5c-a8fd2f0bbb74_1792x1024.webp 848w, https://substackcdn.com/image/fetch/$s_!MVN4!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F25c79b85-3561-4b0f-bc5c-a8fd2f0bbb74_1792x1024.webp 1272w, https://substackcdn.com/image/fetch/$s_!MVN4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F25c79b85-3561-4b0f-bc5c-a8fd2f0bbb74_1792x1024.webp 1456w" sizes="100vw" data-fetchpriority="high" width="1456" height="832" />
<div class="image-link-expand">
<div class="pencraft pc-display-flex pc-gap-8 pc-reset">
<img src="data:image/svg+xml;base64,PHN2ZyByb2xlPSJpbWciIHdpZHRoPSIyMCIgaGVpZ2h0PSIyMCIgdmlld2JveD0iMCAwIDIwIDIwIiBmaWxsPSJub25lIiBzdHJva2Utd2lkdGg9IjEuNSIgc3Ryb2tlPSJ2YXIoLS1jb2xvci1mZy1wcmltYXJ5KSIgc3Ryb2tlLWxpbmVjYXA9InJvdW5kIiBzdHJva2UtbGluZWpvaW49InJvdW5kIiB4bWxucz0iaHR0cDovL3d3dy53My5vcmcvMjAwMC9zdmciPjxnPjx0aXRsZT48L3RpdGxlPjxwYXRoIGQ9Ik0yLjUzMDAxIDcuODE1OTVDMy40OTE3OSA0LjczOTExIDYuNDMyODEgMi41IDkuOTExNzMgMi41QzEzLjE2ODQgMi41IDE1Ljk1MzcgNC40NjIxNCAxNy4wODUyIDcuMjM2ODRMMTcuNjE3OSA4LjY3NjQ3TTE3LjYxNzkgOC42NzY0N0wxOC41MDAyIDQuMjY0NzFNMTcuNjE3OSA4LjY3NjQ3TDEzLjY0NzMgNi45MTE3Nk0xNy40OTk1IDEyLjE4NDFDMTYuNTM3OCAxNS4yNjA5IDEzLjU5NjcgMTcuNSAxMC4xMTc4IDE3LjVDNi44NjExOCAxNy41IDQuMDc1ODkgMTUuNTM3OSAyLjk0NDMyIDEyLjc2MzJMMi40MTE2NSAxMS4zMjM1TTIuNDExNjUgMTEuMzIzNUwxLjUyOTMgMTUuNzM1M00yLjQxMTY1IDExLjMyMzVMNi4zODIyNCAxMy4wODgyIiAvPjwvZz48L3N2Zz4=" />
<img src="data:image/svg+xml;base64,PHN2ZyB4bWxucz0iaHR0cDovL3d3dy53My5vcmcvMjAwMC9zdmciIHdpZHRoPSIyMCIgaGVpZ2h0PSIyMCIgdmlld2JveD0iMCAwIDI0IDI0IiBmaWxsPSJub25lIiBzdHJva2U9ImN1cnJlbnRDb2xvciIgc3Ryb2tlLXdpZHRoPSIyIiBzdHJva2UtbGluZWNhcD0icm91bmQiIHN0cm9rZS1saW5lam9pbj0icm91bmQiIGNsYXNzPSJsdWNpZGUgbHVjaWRlLW1heGltaXplMiBsdWNpZGUtbWF4aW1pemUtMiI+PHBvbHlsaW5lIHBvaW50cz0iMTUgMyAyMSAzIDIxIDkiPjwvcG9seWxpbmU+PHBvbHlsaW5lIHBvaW50cz0iOSAyMSAzIDIxIDMgMTUiPjwvcG9seWxpbmU+PGxpbmUgeDE9IjIxIiB4Mj0iMTQiIHkxPSIzIiB5Mj0iMTAiPjwvbGluZT48bGluZSB4MT0iMyIgeDI9IjEwIiB5MT0iMjEiIHkyPSIxNCI+PC9saW5lPjwvc3ZnPg==" class="lucide lucide-maximize2 lucide-maximize-2" />
</div>
</div>
</div>
</figure>

</div>

###### *illustration of the scaling laws of mammals and their lifetime heartbeats, showing the relationship between size, lifespan, and total heartbeats across species*

## The Scaling Paradigm

For years, the AI community has operated under the assumption that bigger is better. Larger models, trained on ever-increasing amounts of data, have consistently shown improved performance across a wide range of tasks. This led to the formulation of scaling laws, which suggested that model performance would continue to improve as we increased model size and training data.

## Signs of Diminishing Returns

However, recent developments suggest we may be approaching the limits of this paradigm:

1.  Scarcity of High-Quality Data: As discussed in our exploration of data quality, not all data is created equal. High-quality, human-created data—especially in domains requiring deep understanding of human experience—is becoming increasingly scarce.

2.  Limitations of Synthetic Data: While synthetic data has proven valuable for certain quantifiable tasks, particularly in STEM fields, it shows significant limitations in humanities and areas requiring nuanced understanding of human behavior and culture.

3.  Perceived Intelligence Plateau: Despite continued increases in model size and complexity, the perceived intelligence of AI models seems to be leveling off. The average person struggles to differentiate between models like GPT-3.5 and Claude 3.5 Sonnet, even though the latter may have a significantly higher estimated IQ.

## The Diplomacy Lesson

The work on AI for the game Diplomacy by Noam Brown and his team provides a crucial insight. In complex, multi-agent environments with incomplete information, purely synthetic data and self-play are insufficient to achieve optimal performance. This suggests a fundamental limitation in our ability to create truly general intelligence through scaling alone.

## The Quality Imperative

As we approach the limits of available high-quality data, the focus must shift from quantity to quality. This shift presents several challenges:

1.  Data Curation: Identifying and collecting high-value data becomes increasingly critical and increasingly difficult.

2.  Domain-Specific Challenges: While STEM fields may continue to benefit from synthetic data and formal problem-solving approaches, humanities and social sciences present unique challenges that resist synthetic solutions.

3.  Ethical Considerations: As we scrape the bottom of the barrel for quality data, we encounter increasing ethical concerns about data usage, privacy, and the potential for bias in our models.

## Beyond Scaling: New Frontiers in AI Development

As we reach the end of simple scaling laws, new avenues for advancement emerge:

1.  Architectural Innovations: Instead of just making models bigger, we may need to fundamentally rethink how they're structured.

2.  Multimodal Integration: Combining different types of data and model architectures may yield synergistic improvements that pure scaling cannot achieve.

3.  Interactive Learning: Developing AI systems that can learn and adapt through interaction, much like humans do, may be key to overcoming the limitations of static datasets.

4.  Focused Specialization: Rather than striving for general intelligence, we may see a trend towards highly specialized AI systems that excel in specific domains.

5.  Human-AI Collaboration: The future may lie not in AI that replaces human intelligence, but in systems that augment and collaborate with human experts.

## Conclusion: The New Paradigm

The end of scaling laws doesn't mean the end of progress in AI. Rather, it signals a shift towards more nuanced, quality-focused approaches. The next breakthrough in AI may not come from a model with quadrillions of parameters, but from one that can truly understand and interact with the complexities of human experience.

As we move forward, the challenge will be to develop AI systems that can not only process vast amounts of information but also grasp the subtle, contextual, and often ambiguous nature of human knowledge and interaction. This new paradigm will require interdisciplinary collaboration, ethical consideration, and a deep appreciation for the complexity of intelligence itself.

The future of AI lies not in mimicking human intelligence through brute force, but in developing systems that can truly complement and enhance human capabilities in ways we have yet to imagine.

---

//Originally published on Choir Substack: [[https://choir.substack.com/p/the-end-of-scaling-laws|https://choir.substack.com/p/the-end-of-scaling-laws]].//
