80/20 - AI's Predictable Unpredictability
Aurora Consurgens, late 14th century. An alchemical treatise shows the stages of the magnum opus inside a single “philosophical egg”. Just as the dragon and raven survive the alchemist’s furnace to spark new life, disruptive technologies follow a power‑law purge where roughly 20 % persist through the flames and are reborn as the next dominant force.
Looking back on my first interaction with ChatGPT in late 2022, I remember thinking "it's incredible about 80% of the time." The step change from prior answer engines and assistants was startling, so the occasional flap wasn't much of a concern. Even the full-blown hallucinations were not so concerning, considering how often it was remarkably competent.
I started working on SNDOUT, my first major generative AI project in early 2023. It was inspired by the realization that LLMs can be treated as a latent knowledge graph waiting to be interrogated. The discovery came from testing the depth of LLM’s training data — a deep dive into obscure West African music knowledge, a music lineage I've invested significant time exploring and knew little was documented on the web. There was a curious mixture of unearthed new knowledge and completely fabricated, but highly believable, history. I quickly learned that LLMs often hallucinate in sparse data spaces — as evident in the depth of obscure 1960s Igbo Highlife music lore.
In SNDOUT, I could reliably extract broad musical influences and genre classifications — the system excelled at mapping well-documented connections between artists. But ask it to trace the specific migration of a particular drumming technique from Lagos to Accra in 1968, and it would weave together plausible but fictional narratives, complete with invented recording studio names and fabricated session musician credits.
This pattern of selective reliability wasn't unique to obscure music history, and the lessons informed my work building AI products and services in other domains. My use of generative AI expanded to include code writing assistance, synthetic data generation, and even some creative assistance. The phenomenon holds true: great performance in well-trodden domains, creative fabrication in sparse territories.
Two years ago I assumed by now the capabilities of LLMs would exceed most humans for many tasks, such as coding and managing my schedule. Around that time I started following Gary Marcus, AI researcher and critic who argued LLMs were beginning to plateau, which reoriented my expectations a bit. But over the past two years one thing has remained remarkably stable: a consistent 80/20 threshold.
Before I jump into what this means in the context of AI, let's look at the history of the 80/20 principle. Originally called the "Pareto Principle," it came from Italian economist Vilfredo Pareto in the late 19th century. In 1896, while studying wealth distribution in Italy, Pareto noticed that approximately 80% of the land was owned by just 20% of the population. This pattern of unequal distribution intrigued him, and he found similar ratios in other countries.
In 1997, Richard Koch wrote a popular business consulting book called "The 80/20 Principle," where he credits the Pareto Principle for identifying a recurring mathematical pattern foundational to how the world works. While the key insight is that 80% of effects concentrate in 20% of causes, there's a more general rule that reflects 80% displacement and 20% persistence. This pattern shows up in nature, notably in fractal geometry and metabolic scaling. In other words, it reflects something fundamental about how complex systems naturally organize for optimal function.
This pattern of 80% displacement echoes most major technological shifts in history. The printing press didn't just replace scribes — it eliminated 80% of hand-copying while creating entirely new roles: editors, publishers, typesetters. Steam power didn't just replace horses — it automated 80% of manual transportation while creating locomotive engineers, railway planners, and industrial logistics specialists.
This technical pattern shows up in enterprise adoption too: roughly 80% of enterprise AI initiatives fail to reach production or meet their objectives. But here's what's interesting — this isn't because AI doesn't work. It's because organizations underestimate that persistent 20% of human judgment, context, and adaptation that remains fundamentally complex.
At the technical level, AI performance reveals another manifestation of this threshold — but here it presents as a complexity floor rather than a concentration effect. Anyone who regularly uses AI for coding will recognize this failure rate — it's surprisingly consistent with "vibe coding." While AI researchers continually address these failings, the addition of complexity and organic growth keep the asymptotic boundary consistently pegged at 20%.
I realize this might sound like pattern-matching run amok — finding 80/20 splits everywhere because I'm looking for them. The persistence across technological revolutions suggests this isn't coincidence but reveals something fundamental about the boundary between what can be systematized and what requires human judgment — a mathematical constant hidden in the nature of complexity itself. Maybe it's like overlaying the golden ratio on natural forms — a mix of genuine mathematical harmony and our pattern-seeking minds finding meaningful order.
The 20% persistence pattern gives me confidence not just in my own relevance, but in human potential itself. My prediction holds strong: the AI-powered future belongs to the creative, empathetic, and imaginative problem solvers who can think outside the box. We've always been the 20%, but we now get to impact 80% of the value.
Posted
Jul 29, 2025

