“Why fear about one thing that isn’t going to occur?”
KGB Chairman Charkov’s query to inorganic chemist Valery Legasov in HBO’s “Chernobyl” miniseries makes a superb epitaph for the tons of of software development, modernization, and operational failures I’ve lined for IEEE Spectrum since my first contribution, to its September 2005 special issue on studying—or relatively, not studying—from software program failures. I famous then, and it’s nonetheless true twenty years later: Software program failures are universally unbiased. They occur in each nation, to giant corporations and small. They occur in business, nonprofit, and governmental organizations, no matter standing or repute.
International IT spending has greater than tripled in fixed 2025 {dollars} since 2005, from US $1.7 trillion to $5.6 trillion, and continues to rise. Regardless of further spending, software program success charges haven’t markedly improved previously twenty years. The result’s that the enterprise and societal prices of failure proceed to develop as software program proliferates, permeating and interconnecting each facet of our lives.
For these hoping AI software program instruments and coding copilots will shortly make large-scale IT software program initiatives profitable, overlook about it. For the foreseeable future, there are arduous limits on what AI can convey to the desk in controlling and managing the myriad intersections and trade-offs amongst systems engineering, venture, monetary, and enterprise administration, and particularly the organizational politics concerned in any large-scale software program venture. Few IT projects are shows of rational decision-making from which AI can or ought to study. As software program practitioners know, IT initiatives endure from sufficient administration hallucinations and delusions with out AI including to them.
As I famous 20 years in the past, the drivers of software failure regularly are failures of human creativeness, unrealistic or unarticulated venture targets, the lack to deal with the venture’s complexity, or unmanaged dangers, to call just a few that right this moment nonetheless frequently trigger IT failures. Numerous others return a long time, similar to these recognized by Stephen Andriole, the chair of enterprise expertise at Villanova University’s Faculty of Enterprise, within the diagram under first printed in Forbes in 2021. Uncovering a software program system failure that has gone off the rails in a novel, beforehand undocumented method can be shocking as a result of the overwhelming majority of software-related failures contain avoidable, recognized failure-inducing elements documented in tons of of after-action reviews, educational research, and technical and administration books for many years. Failure déjà vu dominates the literature.
The query is, why haven’t we utilized what we have now repeatedly been compelled to study?
The Phoenix That By no means Rose
Most of the IT developments and operational failures I’ve analyzed over the past 20 years have every had their very own Chernobyl-like meltdowns, spreading reputational radiation all over the place and contaminating the lives of these affected for years. Every usually has a narrative that strains belief. A main instance is the Canadian authorities’s CA $310 million Phoenix payroll system, which went dwell in April 2016 and shortly after went supercritical.
Phoenix venture executives believed they might deliver a modernized payment system, customizing PeopleSoft’s off-the-shelf payroll bundle to observe 80,000 pay guidelines spanning 105 collective agreements with federal public-service unions. It additionally was trying to implement 34 human-resource system interfaces throughout 101 authorities businesses and departments required for sharing worker information. Additional, the federal government’s developer crew thought they might accomplish this for lower than 60 percent of the seller’s proposed funds. They’d save by eradicating or deferring crucial payroll capabilities, decreasing system and integration testing, lowering the variety of contractors and authorities employees engaged on the venture, and forgoing very important pilot testing, together with a host of other overly optimistic proposals.
Phoenix’s payroll meltdown was preordained. In consequence, over the previous 9 years, round 70 % of the 430,000 present and former Canadian federal authorities workers paid via Phoenix have endured paycheck errors. At the same time as not too long ago as fiscal 12 months 2023–2024, a 3rd of all workers experienced paycheck mistakes. The continuing monetary stress and anxieties for hundreds of workers and their households have been immeasurable. Not solely are recurring paycheck troubles sapping worker morale, however in at the very least one documented case, a coroner blamed an worker’s suicide on the insufferable monetary and emotional pressure she suffered.
By the top of March 2025, when the Canadian government had promised that the backlog of Phoenix errors would lastly be cleared, over 349,000 were still unresolved, with 53 % pending for greater than a 12 months. In June, the Canadian government as soon as once more committed to considerably decreasing the backlog, this time by June 2026. Given earlier guarantees, skepticism is warranted.
The query is, why haven’t we utilized what we have now repeatedly been compelled to study?
What proportion of software program initiatives fail, and what failure means, has been an ongoing debate inside the IT group stretching back decades. With out diving into the talk, it’s clear that software program growth stays one of many riskiest technological endeavors to undertake. Certainly, in keeping with Bent Flyvbjerg, professor emeritus on the College of Oxford’s Saїd Enterprise Faculty, complete information reveals that not solely are IT initiatives dangerous, they’re the riskiest from a value perspective.
The CISQ report estimates that organizations in the US spend greater than $520 billion yearly supporting legacy software program methods, with 70 to 75 % of organizational IT budgets dedicated to legacy upkeep. A 2024 report by companies firm NTT DATA discovered that 80 % of organizations concede that “insufficient or outdated expertise is holding again organizational progress and innovation efforts.” Moreover, the report says that just about all C-level executives consider legacy infrastructure thwarts their skill to reply to the market. Even so, provided that the price of changing legacy methods is usually many multiples of the price of supporting them, enterprise executives hesitate to replace them till it’s now not operationally possible or cost-effective. The opposite purpose is a well-founded fear that changing them will flip right into a debacle like Phoenix or others.
Nonetheless, there have been ongoing makes an attempt to enhance software program growth and sustainment processes. For instance, we have now seen growing adoption of iterative and incremental methods to develop and maintain software program methods via Agile approaches, DevOps methods, and different associated practices.
The objective is to ship usable, reliable, and reasonably priced software program to finish customers within the shortest possible time. DevOps strives to perform this repeatedly all through the complete software program life cycle. Whereas Agile and DevOps have proved profitable for a lot of organizations, in addition they have their share of controversy and pushback. Provocative reviews declare Agile initiatives have a failure rate of up to 65 percent, whereas others declare as much as 90 percent of DevOps initiatives fail to meet organizational expectations.
It’s best to be cautious of those claims whereas additionally acknowledging that efficiently implementing Agile or DevOps strategies takes constant management, organizational self-discipline, persistence, funding in coaching, and tradition change. Nevertheless, the identical necessities have at all times been true when introducing any new software program platform. Given the historic lack of organizational resolve to instill confirmed practices, it’s not shocking that novel approaches for creating and sustaining ever extra advanced software program methods, irrespective of how efficient they might be, can even regularly fall brief.
Persisting in Silly Errors
The irritating and perpetual query is why fundamental IT project-management and governance errors throughout software program growth and operations proceed to happen so usually, given the near-total societal reliance on dependable software program and an extensively documented historical past of failures to study from? Subsequent to electrical infrastructure, with which IT is more and more merging right into a mutually codependent relationship, the failure of our computing methods is an existential risk to fashionable society.
Frustratingly, the IT group stubbornly fails to learn from prior failures. IT project managers routinely claim that their venture is someway totally different or distinctive and, thus, classes from earlier failures are irrelevant. That’s the excuse of the smug, although often not the ignorant. In Phoenix’s case, for instance, it was the federal government’s second payroll-system replacement attempt, the primary effort ending in failure in 1995. Phoenix venture managers ignored the well-documented causes for the primary failure as a result of they claimed its classes weren’t relevant, which did nothing to maintain the managers from repeating them. Because it’s been stated, we study extra from failure than from success, however repeated failures are rattling costly.
Not all software program growth failures are unhealthy; some failures are even desired. When pushing the bounds of creating new sorts of software program merchandise, applied sciences, or practices, as is occurring with AI-related efforts, potential failure is an accepted risk. With failure, expertise will increase, new insights are gained, fixes are made, constraints are higher understood, and technological innovation and progress proceed. Nevertheless, most IT failures right this moment aren’t associated to pushing the progressive frontiers of the computing artwork, however the edges of the mundane. They don’t symbolize Austrian economist Joseph Schumpeter’s “gales of creative destruction.” They’re extra like gales of economic destruction. Simply what number of extra enterprise resource planning (ERP) project failures are wanted earlier than success turns into routine? Such failures needs to be referred to as IT blunders, as studying something new from them is doubtful at finest.
Was Phoenix a failure or a blunder? I argue strongly for the latter, however on the very least, Phoenix serves as a grasp class in IT project mismanagement. The query is whether or not the Canadian authorities discovered from this expertise any greater than it did from 1995’s payroll-project fiasco? The government maintains it will learn, which is likely to be true, given the Phoenix failure’s excessive political profile. However will Phoenix’s classes lengthen to the thousands of outdated Canadian government IT systems needing alternative or modernization? Hopefully, however hope isn’t a strategy, and purposeful motion shall be crucial.
The IT group has striven mightily for many years to make the incomprehensible routine.
Repeatedly making the identical errors and anticipating a distinct end result isn’t studying. It’s a farcical absurdity. Paraphrasing Henry Petroski in his e-book To Engineer Is Human: The Role of Failure in Successful Design (Classic, 1992), we might have discovered the best way to calculate the software program failure resulting from danger, however we have now not discovered the best way to calculate to get rid of the failure of the thoughts. There are a plethora of examples of initiatives like Phoenix that failed partly resulting from bumbling administration, but this can be very tough to seek out software program initiatives managed professionally that also failed. Discovering examples of what may very well be termed “IT heroic failures” is like Diogenes searching for one trustworthy man.
The implications of not studying from blunders shall be a lot better and extra insidious as society grapples with the rising results of artificial intelligence, or extra precisely, “clever” algorithms embedded into software program methods. Hints of what would possibly occur if previous classes go unheeded are discovered within the spectacular early automated decision-making failure of Michigan’s MiDAS unemployment and Australia’s Centrelink “Robodebt” welfare systems. Each used questionable algorithms to establish misleading cost claims with out human oversight. State officers used MiDAS to accuse tens of hundreds of Michiganders of unemployment fraud, whereas Centrelink officers falsely accused tons of of hundreds of Australians of being welfare cheats. Untold numbers of lives won’t ever be the identical due to what occurred. Authorities officers in Michigan and Australia positioned far an excessive amount of belief in these algorithms. They needed to be dragged, kicking and screaming, to acknowledge that one thing was amiss, even after it was clearly demonstrated that the software program was untrustworthy. Even then, officers tried to downplay the errors’ impression on folks, then fought towards paying compensation to these adversely affected by the errors. Whereas such habits is legally termed “maladministration,” administrative evil is nearer to actuality.
So, we’re left with solely an expert and private obligation to reemphasize the apparent: Ask what you do know, what you must know, and the way large the hole is between them earlier than embarking on creating an IT system. If nobody else has ever efficiently constructed your system with the schedule, funds, and performance you requested for, please clarify why your group thinks it might. Software program is inherently fragile; constructing advanced, safe, and resilient software program methods is tough, detailed, and time-consuming. Small errors have outsize results, every with an nearly infinite variety of methods they’ll manifest, from inflicting a minor practical error to a system outage to permitting a cybersecurity risk to penetrate the system. The extra advanced and interconnected the system, the extra alternatives for errors and their exploitation. A pleasant begin can be for senior administration who management the purse strings to lastly deal with software program and systems development, operations, and sustainment efforts with the respect they deserve. This not solely means offering the personnel, monetary assets, and management assist and dedication, but additionally the skilled and private accountability they demand.
It’s well-known that honesty, skepticism, and ethics are important to reaching venture success, but they’re usually absent. Solely senior administration can demand they exist. As an illustration, honesty begins with the forthright accounting of the myriad of dangers concerned in any IT endeavor, not their rationalization. It’s a widespread “secret” that it’s far simpler to get funding to repair a troubled software program growth effort than to ask for what’s required up entrance to handle the dangers concerned. Vendor puffery may be authorized, however which means the IT buyer wants a healthy skepticism of the usually too-good-to-be-true guarantees distributors make. As soon as the contract is signed, it’s too late. Moreover, computing’s malleability, complexity, pace, low value, and talent to breed and retailer data combine to create moral conditions that require deep reflection about computing’s penalties on people and society. Alas, moral concerns have routinely lagged when technological progress and earnings are to be made. This observe should change, particularly as AI is routinely injected into automated methods.
Within the AI group, there was a motion towards the concept of human-centered AI, that means AI methods that prioritize human wants, values, and well-being. This implies making an attempt to anticipate the place and when AI can go incorrect, transfer to get rid of these conditions, and construct in methods to mitigate the results in the event that they do occur. This idea requires software to each IT system’s effort, not simply AI.
Given the historic lack of organizational resolve to instill confirmed practices…novel approaches for creating and sustaining ever extra advanced software program methods…can even regularly fall brief.
Lastly, venture cost-benefit justifications of software program developments not often take into account the monetary and emotional misery positioned on finish customers of IT systems when one thing goes incorrect. These embody the long-term failure after-effects. If these prices needed to be taken absolutely into consideration, similar to within the instances of Phoenix, MiDAS, and Centrelink, maybe there may very well be extra realism in what’s required managerially, financially, technologically, and experientially to create a profitable software system. It might be a forlorn request, however certainly it’s time the IT group stops repeatedly making the identical ridiculous errors it has made since at the very least 1968, when the time period “software crisis” was coined. Make new ones, rattling it. As Roman orator Cicero stated in Philippic 12, “Anybody could make a mistake, however solely an fool persists in his error.”
Particular due to Steve Andriole, Hal Berghel, Matt Eisler, John L. King, Roger Van Scoy, and Lee Vinsel for his or her invaluable critiques and insights.
From Your Website Articles
Associated Articles Across the Internet

