In June 2024, global news outlets reported that McDonald's was ending a two-year partnership with IBM for its AI drive-thru system. This decision came after over a year of viral TikTok videos documenting ridiculous system failures—260 chicken nuggets randomly ordered, the addition of bacon to ice cream, and many more! Accuracy rates were in the low-to-mid 80% range, far below the 95% threshold required for practical deployment.1
While McDonald's may have made the headlines, they aren’t alone. In 2024, 42% of companies reported abandoning their AI initiatives. The average organization is estimated to shelve 46% of AI proofs-of-concept before production. 2 According to Stanford's AI Index Report 2024, training costs for state-of-the-art AI models have reached unprecedented levels, with flagship models like Google's Gemini Ultra costing $191 million to train — a dramatic escalation from the $4.3 million to train GPT-3 in 2020.
The industry has responded by deploying smaller, specialized models, and Gartner predicts that organizations will be thrice as likely to use task-specific models rather than general-purpose systems by 2027. This does make sense as smaller models are cheaper to train, faster to deploy, and easier to manage than massive general-purpose systems that deliver capabilities an organization may never use. However, even these smaller systems miss the fundamental issue; centralized systems at smaller scales are still “centralized” and training-dependent.
To understand why these aspects really pose an issue, let’s examine the deeper structural problems associated with them. No matter how much data one collects or how sophisticated the training is, when faced with real-world variability, training-dependent systems hit a performance ceiling. McDonald's Automated Order Taker couldn't handle varying human communication patterns like new accents, dialects, or non-standard orders.
Training-dependent AI suffers from three fundamental limitations:
The Performance Ceiling—accuracy barriers when confronting real-world variability
The Degradation Spiral—deterioration as conditions drift from training scenarios
The Brittleness Factor—catastrophic failure in the face of unexpected inputs
Would "additional training" fix the issues? Not really. Customer behaviors evolve, market conditions shift, and varying interactions create far too many edge cases. Retraining for every new scenario is unrealistic and unsustainable. Yet, enterprises continue to pour resources into this short-sighted approach.
Given these systemic problems with large, general-purpose models, the industry's pivot to smaller, specialized models does represent logical progress. Small language models have high market impact due to their efficiency 3; Microsoft's Phi-3-mini outperforms models twice its size while running on smartphones. However, even these models replicate the fundamental flaws of centralized AI, although on a smaller scale.
For instance, if an enterprise deploys specialized models for customer service, fraud detection, and inventory management, each model will operate independently. However, when the customer service model encounters a query requiring fraud analysis or inventory data, it will fail or necessitate expensive integration. When market conditions change, each model will need individual retraining cycles. This essentially means that organizations are compounding the problem, catering to multiple centralized systems instead of one distributed intelligence network, which is what the alternative offers.
The closest analogy to understanding swarm intelligence is an ant colony. While no single ant understands the entire operation, together, they build complex cities, find optimal food routes, and instantly adapt to threats. When one discovers danger, the information spreads throughout the colony, even without a central command.
Similarly, a swarm approach would deploy multiple simple agents to share insights across customer service, fraud detection, inventory management, and many other functions. When a customer inquiry reveals a potential fraud pattern, the network adapts collectively rather than requiring separate model updates. This would transform AI from a depreciating asset that degrades over time into an appreciating one that becomes more valuable with every interaction.
When individual components encounter problems, the network organically maintains operations and develops new solutions. This coordination difference explains why swarm intelligence delivers breakthrough performance while specialized models still hit the same training dependency walls.
Or consider cybersecurity applications. Instead of training one sophisticated fraud detection model that requires constant updates as attack patterns evolve, organizations deploy multiple simple agents monitoring different transaction signals. When new fraud patterns emerge, successful detection strategies spread automatically across the network without requiring retraining. Similarly, swarm intelligence can coordinate multiple agents handling demand forecasting, inventory optimization, and supplier relationships in supply chain management.
The advantages extend far beyond saving on retraining:
Emergent Problem-Solving: Solutions arise from agent coordination that weren't explicitly programmed; think Netflix discovering unexpected viewing patterns by combining millions of simple user preference signals.
Scalable Intelligence: Adding more agents increases capability without exponential cost growth, similar to how ride-sharing networks become more efficient as more drivers join without requiring system retraining.
Real-Time Market Response: Systems adapt to changing conditions faster than competitors can retrain, much like Uber's pricing algorithms adjusting to demand spikes within minutes.
Distributed Risk: No single point of failure can bring down entire operations. For instance, Amazon's network automatically reroutes orders when individual warehouses face disruptions.
Cost-Effective Scaling: Intelligence grows through coordination rather than expensive model sophistication; think Wikipedia being far more accurate than centralized encyclopedias thanks to editor collaboration.
The patterns plaguing today's AI aren't unprecedented. Remember the dot-com crash? Between March 2000 and October 2002, the NASDAQ crashed 78%, obliterating $5 trillion in market value. 4
Just like those dot-com companies that poured billions into "get big fast" strategies, today's AI companies are pouring resources into ever-larger training runs. By 2028, 33% of enterprise software applications are expected to incorporate agentic AI, up from less than 1% in 2024.5 However, many organizations use the same tools or develop similar capabilities, so they're not creating much competitive advantage.
Survivors of the dotcom crash—notably eBay and Google—won because they had built fundamentally different systems from day one. Instead of trying to scale through massive, centralized infrastructure, they built distributed systems that could grow organically and adapt to changing demands. Organizations implementing AI today must choose to see the signs and change gears or become a cautionary tale about following the crowd over the cliff.
References
Rogers, K. (2024, June 17). McDonald’s to end AI drive-thru test with IBM. Retrieved August 7, 2025, from https://www.cnbc.com/2024/06/17/mcdonalds-to-end-ibm-ai-drive-thru-test.html
S&P Global (2025, May 30). Generative AI experiences rapid adoption, but with mixed outcomes – Highlights from VotE: AI & Machine Learning. Retrieved August 7, 2025, from https://www.spglobal.com/market-intelligence/en/news-insights/research/ai-experiences-rapid-adoption-but-with-mixed-outcomes-highlights-from-vote-ai-machine-learning
Beatty, S. (2024, April 23). Tiny but mighty: The Phi-3 small language models with big potential. Retrieved August 7, 2025, from https://news.microsoft.com/source/features/ai/the-phi-3-small-language-models-with-big-potential/
Mukherjee, V. (2025, March 11). Nasdaq tumbles: A look back at the biggest stock market crashes. Retrieved August 7, 2025, from https://www.business-standard.com/markets/news/nasdaq-plunge-wall-street-dalal-street-stock-market-crashes-india-125031100713_1.html
(n.d.). Enterprise AI Market Size & Share Analysis - Growth Trends & Forecasts (2025 - 2030). Mordor Intelligence. Retrieved August 7, 2025, from https://www.mordorintelligence.com/industry-reports/enterprise-ai-market