Tuesday, March 17, 2026

Cybersecurity's Original Sin: Why the Industry Has Failed to Fix Its Own Known Problems — and Who Benefits from the Failure

 

Executive Summary

In 2001, the SANS Institute published a landmark list of the ten most common cybersecurity mistakes made by individuals and organizations. Twenty-five years later, nine of those ten problems remain largely unsolved. This is not an accident, a technical mystery, or a resource problem. It is the predictable outcome of three interlocking failures: a broken economic incentive structure that shifts the cost of insecurity onto the victims; a legal framework built around End User License Agreements (EULAs) that systematically shield vendors from liability for the consequences of those failures; and an industry that has found it more profitable to treat symptoms than to cure diseases.

Here I examine each of the SANS Top 10 mistakes in light of these structural forces, traces the deliberate legal strategy that the software industry adopted in the late 1990s and early 2000s to disclaim responsibility for insecure products, and argues that the only path to genuine improvement runs through liability reform, regulatory mandate, and a fundamental shift in who pays when software fails.

 

1. The “SANS Top 10 Mistakes Individuals Make”: A 25-Year Audit

The table below maps each of the original SANS 2001 mistakes to its current status and to the specific EULA or contractual defense that the software industry deployed to insulate itself from accountability for that category of failure.

 


 

Of the ten systemic problems identified a quarter-century ago by one of the most respected cybersecurity education bodies in the world, none has been comprehensively fixed, only two show partial improvement, and the partial improvements both have caveats. Default credential hardening was driven not by the industry’s conscience but by regulatory pressure in specific verticals and by the reputational damage caused by IoT botnets that made headlines. Backup adoption improved primarily because cloud vendors found a profitable subscription model to attach to the solution. In both cases, the fix arrived when it became economically advantageous, not before.

 

2. The EULA as a Liability Shield: A Legal Strategy, Not a Legal Inevitability

To understand why these problems persisted, let’s examine the legal strategy that the software industry consciously built in the 1990s. As software became infrastructure i.e., running hospitals, banks, power grids, and government systems, the industry faced a fork in the road. It could accept that infrastructure-grade software carried infrastructure-grade liability, as manufacturers of aircraft, pharmaceuticals, and automobiles do or it could construct a legal argument that software was categorically different and immune from ordinary product liability doctrine.

It chose the latter, and it did so deliberately.

 “As-Is” — The Core Legal Fiction

The phrase that appears in virtually every EULA, in some form, is: "THE SOFTWARE IS PROVIDED 'AS IS', WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED." This single clause, replicated across hundreds of thousands of software products, effectively disclaims liability for every item on the SANS Top 10. The vendor ships unpatched code: no warranty. The vendor enables unnecessary services by default: no warranty. The vendor’s product is exploited via a known vulnerability: no warranty. The user clicks a malicious attachment that the vendor’s own email client failed to flag: user error.


2.1 How the Strategy Worked in Practice

The legal argument rested on several pillars, each of which was constructed and defended by industry lobbying and sustained by courts that were reluctant to impose product liability on an industry they did not fully understand:

       Pillar 1: Software is speech, not a product  therefore product liability doctrine should not apply.

       Pillar 2: Software is infinitely complex; defects are inherent and unknowable therefore no warranty of fitness is commercially reasonable.

       Pillar 3: Users accept the EULA before use therefore any risk that is not patched after installation is assumed by the user.

       Pillar 4: Vendor patches are advisory, not mandatory therefore a breach caused by an unpatched system is the customer’s fault, not the vendors.

 

The combined effect was a complete inversion of the normal manufacturer-consumer liability relationship. A car manufacturer who knowingly ships a car with faulty brakes faces recall obligations, regulatory sanction, and civil liability. A software vendor who knowingly ships code with buffer overflow vulnerabilities, a class of defect that has been fully understood since the 1970s, faces no equivalent obligation. The EULA said so. Courts largely agreed with this between 1995 to 2020.

2.2 Mapping the EULA Defense to the SANS 10

The following analysis shows how each of the SANS 10 mistakes was directly addressed by the software industry’s legal strategy, not by fixing the underlying problem, but by contractually assigning the consequence to someone else.

Mistakes 1 & 5: Default Passwords and Excessive Privilege

Vendors shipped products with default credentials (admin/admin, root/root, guest/guest) and with services running at maximum privilege levels because it reduced support calls and accelerated deployment. When those defaults were exploited, the EULA defense was clean: the vendor had provided a working product; the customer had failed to harden it. The customer accepted the risk by accepting the license. This argument was so embedded in industry culture that regulatory language mandating unique default credentials did not begin appearing until the mid-2010s in specific sectors, and did not reach general software until the UK’s Product Security and Telecommunications Infrastructure Act in 2022, more than two decades after SANS identified the problem.

Mistake 2: Unpatched Software

The unpatched software problem is where the EULA defense is most nakedly visible. Vendors issue security advisories, which are legally non-binding recommendations. The EULA disclaims any liability for harm caused by vulnerabilities that are disclosed but not yet patched by the customer. This creates a remarkable legal structure: the vendor acknowledges a defect exists, declines to make the patching mandatory, and successfully argues in court that the resulting breach is the customer’s fault for not acting on the advisory quickly enough. The Equifax breach of 2017, which exposed 147 million people’s sensitive data, was caused by an unpatched Apache Struts vulnerability for which a patch had been available for two months. Equifax paid a settlement, but no software vendor faced any liability whatsoever.

Mistakes 3 & 4: Open Ports and Lack of Auditing

Vendors ship operating systems and applications with services enabled by default, listening on open ports, because doing so is more convenient for the general user. The cost of a dramatically expanded attack surface is borne by the customer. Similarly, robust logging and auditing capabilities are routinely sold as premium features or enterprise add-ons rather than included as baseline security controls. In both cases, the EULA assigns the risk of misconfiguration and of inadequate visibility entirely to the buyer. The vendor has no obligation to ship a secure default configuration, only a functional one.

Mistake 7: Email Attachments and Social Engineering

This is perhaps the most cynical application of the EULA defense. Vendors classify social engineering attacks such as phishing, malicious attachments, fraudulent links as “user error” rather than as failures of the products that failed to detect or block them. This categorization is enormously convenient for vendors of email clients, antivirus software, and endpoint protection platforms, because it means that a breach caused by a sophisticated phishing email is never the vendor’s liability, even if the vendor’s product was specifically marketed and sold to prevent exactly that attack. The product’s failure becomes the user’s fault by contractual definition.

Mistakes 8, 9 & 10: Policy, Backups, Physical Security

These three items were explicitly and entirely relegated to customer responsibility via EULA. Vendors had no obligation to provide policy templates, backup functionality, or guidance on physical access controls. That these items appear on a list of mistakes individuals make rather than on a list of mistakes vendors make by not providing them reflects how completely the industry had succeeded in framing security as a user behavior problem rather than a product design problem.

 

3. Economic Insecurity Model

The EULA shield did not merely protect vendors from lawsuits. It allowed an entire secondary industry to be built on top of the resulting insecurity. Understanding this economy is essential to understanding why the SANS 10 problems have been so durable.

3.1 The Insecurity Industrial Complex

By the mid-2000s, the cybersecurity industry had grown into a multi-billion-dollar ecosystem whose revenue was structurally dependent on the persistence of the problems identified in the SANS 2001 list. Every unpatched system is a market for a patch management platform. Every misconfigured firewall is a market for a configuration auditing tool. Every phishing attack is a market for security awareness training. Every breach is a market for incident response consulting, forensic analysis, cyber insurance, and remediation services.

The Core Perverse Incentive

The cybersecurity industry is more analogous to the pharmaceutical industry which profits from managing chronic conditions than to the sanitation industry, which profits from eliminating disease vectors. A software vendor that ships memory-safe code by default eliminates the market for a class of endpoint detection tools. A platform that enforces unique passwords at the system level eliminates the market for enterprise password management solutions. The industry therefore has a structural incentive to treat symptoms, not causes.


3.2 Security as a Premium Feature

This incentive model expanded the practice of selling security capabilities as premium add-ons to an insecure base product. For example, advanced logging, privileged access management, multi-factor authentication enforcement, and threat detection have historically been available only in enterprise-tier licenses, priced beyond the reach of small and mid-size organizations. This means that the organizations least able to absorb a breach are precisely the ones denied access to the tools that could prevent it not because those tools are technically complex to provide, but because they are more profitable when segmented as premium features.

The SANS mistake of lacking auditing and logging (Mistake 4) persists not because logging is technically difficult, but because comprehensive logging has been deliberately withheld from base-tier products and sold separately. Microsoft’s decision in 2023 to expand default logging access following criticism from CISA and Congress after the Storm-0558 breach exposed that basic log data was unavailable to affected organizations, showed only external pressure forced the feature out of the premium tier.

3.3 The Compliance Illusion

The rise of compliance frameworks, PCI-DSS, HIPAA, SOX, ISO 27001, SOC 2, etc.,  created what appeared to be accountability but in practice created a substitute for security. Organizations could satisfy a compliance audit by demonstrating the presence of certain controls without those controls being effective. A web application firewall counts for PCI compliance whether or not the underlying application has SQL injection vulnerabilities. An annual security awareness training tick-box satisfies HIPAA whether or not employees’ behavior changes.

Compliance became the goal, and security became incidental. Vendors benefited from this substitution because compliance requirements created mandatory procurement events where organizations had to buy something to satisfy the auditor without requiring that the purchased products actually solve the underlying problem. The SANS 10 mistakes survived twenty-five years in part because compliance frameworks never required that they be fixed, only that a compensating control be installed.

 

 

4. The One Problem That Did Get Fixed — and Why

The partial progress on default credentials (SANS Mistake 1) is instructive precisely because it is the exception. Understanding why this one item moved while the others did not, reveals the mechanism that would be required to fix the rest of the items in the list.

Default credential exploitation became impossible for the industry to disclaim for three reasons. First, IoT botnets with the Mirai botnet in 2016 being the canonical example, caused widespread, visible, and publicly attributable damage that could not be narratively reassigned to user error. The victims were not sophisticated enterprises who had “failed to patch”. They were home users who had plugged in a device and did exactly what the vendor intended. The EULA defense collapsed under the weight of that narrative. Second, regulatory action followed: California’s SB-327 (2018), the UK PSTI Act, and NIST guidance created binding obligations that the EULA could not override. Third, the reputational damage to specific vendor categories namely IP camera manufacturers, home router vendors,  was direct and sustained.

The lesson is clear: the EULA shield holds until the cost of the insecurity can no longer be externalized. When breaches become politically and reputationally undeniable, when regulatory liability attaches, and when the victim class is sympathetic and visible, the industry fixes the problem. The SANS 10 items that remain unfixed are those where the cost is still successfully shifted onto enterprises, SMBs, and individuals who lack the political leverage to force accountability.

 

5. When Software Kills: Operational Technology and the End of the EULA Shield

Every argument surveyed so far, the EULA shield, the compliance illusion, the perverse economics of the insecurity industry, rests on a single enabling condition: that the harm caused by insecure software can be successfully shifted to the user. Data breaches are expensive and embarrassing, but their victims are diffuse, their damages are monetized at pennies per record, and their consequences fall on consumers and enterprises rather than on vendors. The entire legal and economic architecture of the past twenty-five years was built to exploit this condition.

Operational Technology (OT) changes that condition fundamentally. When insecure software controls a water treatment plant, a power grid, a hospital's life support systems, or an oil pipeline, the harm it enables is no longer a data record on a dark web marketplace. It is a city without power, a patient who cannot receive surgery, a chemical plant that explodes, or drinking water laced with caustic lye. These are physical damage to people, not just breach notification letters. Physical harm is very difficult to disclaim in a EULA.

5.1 What OT Is and Why It Is Different

Operational Technology refers to the hardware and software systems that monitor and control physical processes in the real world: industrial control systems (ICS), supervisory control and data acquisition systems (SCADA), programmable logic controllers (PLCs), distributed control systems (DCS), and the human-machine interfaces (HMIs) that operators use to manage them. These systems run power generation and distribution, water and wastewater treatment, oil and gas pipelines, chemical manufacturing, railway signaling, hospital equipment, building management, and nuclear facilities.

OT systems differ from conventional IT in ways that make the standard cybersecurity toolkit of patching quickly, isolating the compromised host, restoring from backup either impractical or dangerous. IT systems have a lifetime measured in years. OT systems have a lifetime measured in decades.  Many OT systems run continuously; a patch window that requires a twenty-minute shutdown may mean a production halt that costs millions. Legacy PLCs and SCADA components may run unpatched operating systems not because operators are negligent but because the vendor discontinued support and replacement requires re-engineering an entire physical production line. Some OT systems are safety-critical: patching incorrectly, or at the wrong moment in a process cycle, can itself cause the kind of physical accident the system was designed to prevent. And critically, the consequence of a failure is not measured in lost records but in physical damage, environmental harm, and human lives.

 

5.2 The Attack Record: From Proof of Concept to Mass Casualty Risk

The history of OT cyberattacks is a compressed, accelerating timeline from theoretical concern to near-catastrophe. The following cases are not hypotheticals from a threat model; they are documented incidents that reveal exactly what the EULA shield has permitted to persist in systems where the stakes are human life.

Stuxnet (Discovered 2010): The Proof That Physics Could Be Hacked

Stuxnet was the event that forced the world to take OT security seriously. A sophisticated worm, widely attributed to US and Israeli intelligence services, was introduced into Iran's Natanz uranium enrichment facility via infected USB drives, exploiting SANS Mistake 7 (unknown media/attachments), in an environment that was physically air-gapped from the internet. Once inside, it targeted Siemens S7-315 and S7-417 PLCs controlling centrifuge motor speeds, causing approximately 900 centrifuges to spin at destructive speeds while simultaneously reporting normal operation to the operators monitoring the facility.

The physical consequence was unambiguous: roughly 20% of Iran's uranium enrichment capacity was destroyed, not metaphorically, but as shattered physical machinery. Stuxnet demonstrated three things that the industry could no longer deny: that software running OT systems could be weaponized to cause deliberate physical destruction, that air gaps were not the security guarantee they appeared to be and that the attacker did not need to be present or even known to the victim to cause large-scale physical harm.

Ukraine Power Grid Attacks (2015 & 2016): Lights Out as a Cyberweapon

In December 2015, malware subsequently named BlackEnergy was used to attack three Ukrainian regional electricity distribution companies, cutting power to approximately 230,000 customers in the middle of a Ukrainian winter. The attackers had spent months conducting reconnaissance inside the utilities' IT networks  facilitated by spear-phishing emails exploiting SANS Mistake 7, before pivoting to OT networks and issuing commands to open breakers at multiple substations simultaneously. They then overwrote firmware on serial-to-Ethernet converters to prevent remote recovery, and launched a telephony denial-of-service attack against the utilities' customer service lines to prevent customers from reporting the outage.

A year later, in December 2016, a second attack using more sophisticated malware named Industroyer targeted Kiev's transmission substation. This attack was designed not just to cut power but to permanently damage physical substation equipment by causing rapid, repeated switching cycles attempting long-term physical destruction, not just temporary disruption. Had it succeeded more fully, restoration would have required physical replacement of high-voltage equipment with months-long lead times.

TRITON / TRISIS (2017): Targeting the Last Line of Defense

The most alarming OT attack on record was not one that caused a disaster — it was one that tried to cause a disaster by disabling the systems specifically designed to prevent one. In 2017, attackers subsequently attributed to a Russian state research institute, compromised the Safety Instrumented System (SIS) at a petrochemical facility in Saudi Arabia. SIS controllers are the last line of defense in industrial facilities: they monitor for dangerous process conditions and automatically shut down or trip equipment when readings exceed safe thresholds. They exist specifically to prevent explosions, chemical releases, and other catastrophic physical events.

TRITON was designed to reprogram the Triconex SIS controllers to either disable safety shutdowns entirely or cause them to fail in dangerous ways. The attack was only discovered because a programming error in the malware accidentally triggered a safety shutdown, alerting operators to the intrusion. Had the malware functioned as intended, subsequent process failures would have had no automatic safety backstop. The intended consequence was a large-scale explosion or toxic chemical release at an operational industrial facility resulting in a potential mass casualty event, deliberately engineered through software compromise.

No EULA clause covers this scenario. No compliance framework tick-box adequately addresses it. The 'software is provided as-is' disclaimer is not a legal defense against the deliberate engineering of an explosion at a petrochemical plant.

Colonial Pipeline (2021): Infrastructure Held Hostage

The Colonial Pipeline ransomware attack is the case study that brought OT cybersecurity into mainstream political consciousness. On 7 May 2021, the DarkSide ransomware group compromised Colonial Pipeline's IT network using a leaked employee password found on the dark web, SANS Mistake 1 (weak/reused credentials) combined with Mistake 3 (an unmonitored legacy VPN account that had not been decommissioned). The company shut down the pipeline which supplies approximately 45% of fuel consumed on the US East Coast, not because the OT systems were themselves compromised, but as a precaution to prevent the ransomware spreading from IT to OT.

The physical and economic consequences were immediate and national in scale. Fuel shortages spread across the Southeast. Panic buying caused station outages in multiple states. The US government declared a state of emergency. The price of gasoline spiked. Colonial Pipeline paid a $4.4 million ransom. President Biden addressed the situation publicly and called for new cybersecurity standards for critical infrastructure. The TSA subsequently issued binding security directives for pipeline operators becoming the first such mandatory regulatory requirements in the sector's history.

The Colonial Pipeline attack is the canonical illustration of the central thesis of this paper: the EULA shield held right up until the consequences became impossible to externalize. When a data breach at an insurance company exposes medical records, the victims are dispersed and the narrative can be managed. When half the US East Coast cannot buy petrol and the President holds a press conference, the narrative management is over.

Oldsmar Water Treatment Facility (2021): Poison in the Tap

On 5 February 2021, an operator at the Oldsmar, Florida water treatment plant watched his mouse cursor move across the screen without his input and navigate to the controls for sodium hydroxide (lye) concentration in the water supply raising the setpoint from 100 parts per million to 11,100 parts per million, a factor of 111 times the normal level. The operator immediately reversed the change and notified supervisors. Subsequent investigation found that the facility's SCADA systems were running an end-of-life operating system, that all operator computers shared a single password for remote access, and that the facility had no firewall protecting its internet-connected control systems.

Every single one of these conditions, the unpatched operating system, the shared weak password, the absent firewall, the unmonitored remote access appears on the SANS Top 10 from 2001. The facility that nearly poisoned 15,000 people's drinking water was running on the exact configuration that the cybersecurity industry had documented as dangerous a full two decades earlier. The investigation raised questions about whether this was a genuine external attack or an internal error. However, either conclusion is damning. If it was an attack, the facility's defenses were wholly inadequate. If it was an insider error, the facility had no controls preventing an individual from making a catastrophic configuration change undetected.

 

5.3 The OT Attack Escalation Curve

These cases are not historical curiosities. Since 2020, researchers have recorded nearly 100 publicly known cyber incidents with direct physical consequences to OT operations. Sadly, this figure represents only the incidents that became public. The threat landscape has broadened from sophisticated state-sponsored operations to ransomware-as-a-service operators who breach OT environments as a side effect of IT network compromises, and to hacktivist groups that specifically target industrial infrastructure as a form of political disruption.

The IT/OT convergence that accelerated through the 2010s has created a structural vulnerability: OT systems designed for isolation are now reachable via the same IT network pathways that carry email, remote desktop access, and VPN connections. Attackers do not need to understand industrial control protocols to cause physical damage. They need only to compromise the IT network and then shut down, encrypt, or corrupt the systems that operators rely on to monitor and control physical processes.

5.4 OT as the Forcing Function: Why This Changes Everything

The reason OT attacks matter so profoundly to the question of software liability is not primarily technical. It is legal and political. Every structural argument that allowed the software industry to avoid liability for twenty-five years breaks down when applied to OT environments.

       The 'no physical harm' defense collapses. EULA clauses disclaim liability for 'loss of data' and 'business interruption’. Courts are increasingly finding that they cannot disclaim liability for bodily injury, death, or large-scale physical destruction caused by foreseeable product failures.

       The 'user error' reclassification fails. When a water utility operator follows the vendor's documented procedure for remote access and an attacker exploits a default credential that the vendor chose not to change, the vendor cannot credibly characterize the resulting harm as user error. The user did what the product was designed to let them do.

       The victim class becomes sympathetic and visible. The legal and political economy of liability reform requires sympathetic, identifiable victims. 15,000 people potentially poisoned by their drinking water, or 230,000 people losing power in winter, or fuel shortages causing national panic — these are victim classes that generate congressional hearings, presidential statements, and binding regulatory directives in ways that dispersed data breach victims do not.

       The causal chain becomes legally traceable. In a data breach, establishing that a specific vendor's product defect caused a specific plaintiff's specific harm is difficult. In a SCADA attack, the causal chain from unpatched PLC firmware to physical damage is often direct, documented, and incontrovertible.

 

5.5 The Regulatory Response: OT as the Catalyst for Software-as-Product

The policy response to OT attacks is already beginning to do what twenty-five years of data breach advocacy could not: force software to be treated as a product with attendant safety obligations rather than as a service provided 'as-is' with no warranty.

The EU's Cyber Resilience Act, which entered into force on 10 December 2024 and whose main obligations apply from 11 December 2027, is the most significant development in this space. For the first time, it imposes mandatory cybersecurity requirements on manufacturers of products with digital elements as a condition of market access, not as a voluntary framework or a compliance checkbox, but as a legal prerequisite to selling in the world's largest single market. Manufacturers must ship products in secure-by-default configurations, maintain vulnerability handling processes, provide security updates throughout the product lifecycle, and cannot disclaim these obligations via EULA. Penalties reach up to €15 million or 2.5% of global annual turnover.

Critically, the CRA explicitly acknowledges that the pre-existing legal framework — which covered physical product safety but not cybersecurity — was inadequate precisely because software vulnerabilities in connected products create physical safety risks. The regulation treats software security as a product safety issue, not a customer configuration problem. This is the legal foundation that the industry argued for twenty-five years did not exist.

In the United States, the Colonial Pipeline attack triggered the first binding TSA security directives for pipeline operators, ending a decade of voluntary guidelines that the industry had largely ignored. The Biden Administration's National Cybersecurity Strategy explicitly called for shifting liability to software vendors. CISA's Secure by Design initiative, while not yet backed by legislation, signals the direction of federal regulatory thinking. The Oldsmar incident directly accelerated EPA and White House action on water sector cybersecurity mandates.

The Trajectory Is Clear

OT attacks are doing in two to three years what twenty-five years of IT data breaches could not: creating the political and legal conditions for software product liability. When software failures kill people or shut down national infrastructure, the 'as-is, no warranty' defense becomes politically untenable and legally vulnerable. The CRA is the first binding operationalization of this shift. It will not be the last.


5.6 What OT Liability Means for the Software Industry's Business Model

If software vendors are required to stand behind the security of their products as a condition of market access as the CRA now mandates for the EU market, the implications for the industry’s existing business model are profound and largely irreversible. It is tempting to treat this shift as narrowly applicable to OT: to pipeline controllers, SCADA systems, and industrial PLCs where the physical consequences of failure are undeniable. That reading would be a serious mistake. The liability principles that OT attacks have forced into the open do not stop at the OT boundary. They apply, with equal logical force, to every category of software and IT system (enterprise applications, cloud platforms, operating systems, consumer software) and the IT networks that connect them all. OT did not create a new legal theory; it destroyed the industry’s last credible argument for why the existing theory should not apply.

Consider what the “no physical harm” carve-out actually required the industry to argue: that software controlling a hospital’s billing system, an airline’s reservation platform, a bank’s transaction processing engine, or a government’s benefits distribution network occupies a categorically different liability universe than software controlling a centrifuge or a pipeline valve. This was always a legally and morally incoherent position. A ransomware attack that shuts down a hospital’s IT network by delaying surgeries, disabling medication dispensing systems, and forcing emergency patients to be diverted, causes patient deaths just as surely as a compromised OT safety system. The 2020 ransomware attack on Düsseldorf University Hospital, which forced the diversion of an emergency patient who subsequently died, illustrates the point with fatal clarity. The harm was not caused by compromised OT. It was caused by insecure IT which the SANS Institute documented in 2001.

The implications for the software industry’s business model therefore extend across its entire product portfolio, not only its industrial control offerings. First, ‘secure by default’ becomes a universal legal obligation, not a feature tier or a marketing differentiator. A vendor cannot ship any software whether a SCADA platform, an enterprise resource planning system, a cloud-hosted collaboration tool, or a consumer application with default credentials, unnecessary open services, and no patching mechanism, and then disclaim liability when it is exploited. The SANS  mistakes 1, 2, 3, and 5 become potential bases for regulatory action and civil liability across all product categories, not merely in industrial environments.

Second, the security premium tier model becomes legally precarious across the entire software market. If comprehensive logging (SANS Mistake 4) is a safety-critical feature in an OT environment, it is equally a safety-critical feature in a hospital’s electronic health record system, a financial institution’s trading platform, and a government agency’s identity management infrastructure. The practice of selling visibility, access controls, and audit capabilities as enterprise-tier add-ons while deploying base-tier products in environments where a breach causes public harm is not a business model that can survive serious product liability scrutiny in OT or IT. For example, Microsoft’s grudging expansion of default logging access in 2023, after Congressional pressure following the Storm-0558 breach, demonstrates that this argument is already winning in the IT domain. OT simply made the stakes undeniable.

Third, the lifecycle support obligation expands dramatically and universally. Under the CRA, manufacturers must provide security updates for the expected lifetime of any product with digital elements not only industrial equipment. For OT systems that may be in service for fifteen to thirty years, this is a well-understood engineering challenge. But the same principle applies to the enterprise software that runs on servers for a decade past its vendor’s stated support window, the embedded software in medical devices certified for a twenty-year deployment cycle, and the IT infrastructure components that organizations cannot replace on the vendor’s preferred upgrade cadence. “End of support” cannot remain a vendor’s mechanism for transferring accumulated security debt to its customers. That practice created the legacy vulnerability crisis in both OT and IT environments, and it is no longer defensible in either.

Fourth, and most consequentially, the ‘software is speech, not a product’ legal fiction becomes impossible to sustain across the entire software industry once it has collapsed at the OT boundary. The intellectual argument that software should be exempt from product liability doctrine was always weakest where the connection between software state and real-world consequence was most direct. OT made that connection undeniable. But the same connection exists, one or two causal steps further removed, in virtually every category of software deployed in critical systems. The ransomware that encrypts a hospital’s patient records does not directly control a ventilator but it prevents the staff who operate that ventilator from accessing the patient’s medication history, allergy records, and treatment protocols. The distinction between “software that causes physical harm directly” and “software that causes physical harm indirectly through IT failure” is not a distinction that courts, regulators, or juries are likely to treat as dispositive once they have accepted the underlying liability principle. OT broke the seal. IT will not remain exempt.

The software industry will resist these changes with the same lobbying intensity it deployed to construct the EULA shield in the first place. It will argue that OT is a special case, that IT systems are too complex for liability to attach, that innovation will be chilled, and that the market should be allowed to self-correct. These are the same arguments it has made for thirty years while the SANS Top 10 remained unfixed. The outcome is not predetermined. But the direction of the argument has shifted and the shift was driven not by the cybersecurity community’s advocacy, but in the OT world by centrifuges spinning themselves to destruction, power grids going dark in winter, and pipelines shutting down the Eastern Seaboard. The lesson for IT is not that its problems are different. It is that IT has not yet produced a catastrophe visible enough to force the same reckoning. When it does, the liability framework that OT built will be waiting.

6. The Path Forward: Making the Broken Economics Whole

If these problems persist because the economic and legal incentive structure makes persistence more profitable than resolution, then the solution set is clear, even if politically difficult.

6.1 Software Product Liability Reform

The most structurally significant change would be to make software vendors liable, in law, for foreseeable harms caused by known vulnerability classes that were not mitigated by design. This does not require perfection; it requires that vendors be held to the same standard of care as manufacturers in other safety-critical industries. A vendor who ships a product with SQL injection vulnerabilities in 2026 is not making an unknowable mistake; they are shipping a product with a defect class that has been fully documented, freely preventable by design, and actively exploited for thirty years. The “software is too complex” defense is no longer credible.

The Biden Administration’s National Cybersecurity Strategy (2023) explicitly endorsed shifting liability to software vendors, stating that “liability should be placed on the entities that fail to take reasonable precautions to secure their software.” The EU’s Cyber Resilience Act (2024) begins to operationalize this for products sold in the European market. The direction of travel is clear; the pace is the variable.

6.2 Secure-by-Default as a Regulatory Mandate

For items on the SANS 10 that relate to default configurations such as open ports, excessive privilege, default credentials, regulatory mandates for secure-by-default design are the direct remedy. This approach has precedent: automobile safety standards, pharmaceutical manufacturing standards, and aviation design standards all impose mandatory baseline safety requirements on products regardless of what liability waivers the manufacturer might prefer. There is no principled reason software should be exempt.

6.3 Ending the Security Premium Tier

Baseline security capabilities i.e., comprehensive logging, multi-factor authentication, privilege management, audit trails, should be included in base-tier products for any software deployed in critical infrastructure, healthcare, financial services, or government. The practice of selling visibility and access controls as premium features in sectors where a breach causes public harm is not a business model that should be publicly subsidized by the victims of that harm.

6.4 Mandatory Breach Disclosure with Root Cause Attribution

Current breach disclosure requirements, where they exist, do not typically require vendors to disclose whether a breach resulted from a known, unpatched vulnerability class in their own product. Requiring root-cause attribution in mandatory disclosures would create market pressure on vendors whose products repeatedly appear as the enabling condition for breach incidents. Transparency alone does not fix incentive structures, but it creates the informational precondition for market and regulatory pressure to do so. Sites like https://privacyrights.org categorize and list data breaches so publishing information about data breaches is not unusual.

7. Conclusion

The SANS Top 10 list of 2001 is not a monument to the difficulty of cybersecurity. It is a monument to the success of a legal and economic strategy that made insecurity profitable and consequence-free for the parties best positioned to fix it. The EULA disclaimer of liability was not an inevitable feature of the software landscape; it was a deliberate choice, made by an industry that understood exactly what it was disclaiming and why.

Twenty-five years of harm in the form of breaches, ransomware, identity theft, critical infrastructure disruption have been made possible not primarily by sophisticated attackers, but by the legal architecture that ensured no one with the power to fix the root causes had to bear the cost of leaving them unfixed. The attackers in our memes are clever. But the most consequential exploits of the past twenty-five years were not executed in terminal windows. They were drafted by lawyers, ratified by courts, and embedded in the license agreements that nobody reads.

 

Sources & Further Reading

This blog was written with the assistance of claude.ai

(2001). Mistakes People Make that Lead to Security Breaches. https://sans.org/security-resources/mistakes

(2023). Top 10 Cybersecurity Misconfigurations. https://cisa.gov/news-events/cybersecurity-advisories/aa23-278a

(2023). National Cybersecurity Strategy. https://bidenwhitehouse.archives.gov/wp-content/uploads/2023/03/National-Cybersecurity-Strategy-2023.pdf

(2023). Human Error Remains the Top Security Issue. https://www.techtarget.com/searchsecurity/news/252522226/SANS-Institute-Human-error-remains-the-top-security-issue 

(2025). Since Stuxnet: A History of Critical Infrastructure Attacks. https://www.forescout.com/blog/cybersecurity-threat-evolution-of-otics-and-iot-devices/

(2025). OT Cybersecurity: The Top 10 Attacks Since 2020. https://waterfall-security.com/ot-insights-center/ot-cybersecurity-insights-center/ot-cybersecurity-the-top-10-attacks-since-2020/

(2024). Behind the Breach: Analyzing Critical ICS/OT Cyberattacks. https://www.opswat.com/blog/behind-the-breach-analyzing-critical-ics-ot-cyberattacks

(2021). Recommendations Following the Oldsmar Water Treatment Facility Cyber Attack. https://www.dragos.com/blog/recommendations-following-the-oldsmar-water-treatment-facility-cyber-attack/

(2023). Cybersecuring the Pipeline. https://houstonlawreview.org/article/73666-cybersecuring-the-pipeline

(2024). Cyber Resilience Act — Regulation (EU) 2024/2847. https://eur-lex.europa.eu/eli/reg/2024/2847/oj/eng

(2018) “Click Here to Kill Everybody”, Bruce Schneier, https://www.schneier.com/books/click-here/

Friday, February 17, 2023

4 Tips for Prioritizing Metrics



There are three phases to setting up a security metrics program. First, you must collect the data, analyze it and finally report your findings. The Collection phase involves installing sensors (Zeek, Commercial IDS/IPS, vulnerability scanners, etc.). The Analysis phase uses tools such as ServiceNow, Remedy, Crystal Reports, etc. The Reporting phase is the process of creating workbooks (weekly, monthly, yearly) and a set of Operational, Incident and Compliance reports. Your target audiences include your boss, your boss’ boss, IT manager peers, security team, your Board, internal audit, CFO/COO and units involved with regulatory compliance. Matt Tolbert gave a talk in 2007 on Security metrics that resonated with me. Here are some notes I took from his talk for prioritizing security metrics. A good metrics reporting package should include:

  1. Operational. Examples of these reports include helpdesk tickets completed, security project status, # of security scans completed and their results, inventory of hardware or software connected to your network. The target audience includes your boss, IT manager peers, and your security team.

  2. Incident. These are the number of reported security incidents and their status such as success/failure, financial and reputational impact, “after action reports”, legal status. The target audience is your boss, your boss’ boss and your Board of Directors/Trustees.

  3. Compliance. These metrics show how effective your security controls, services and training are in complying with whatever security or data standards your organization has to be in compliance.  The target audience includes your boss, your boss’ boss, internal audit and units involved with regulatory compliance. 

  4. Executive. These metrics are similar to the Compliance metric but they also show the value of the security controls, services and training you’ve installed. They should also show areas that need improvement as well as showing progress to meeting the organization’s business goals.

Tolbert suggested 4 benefits from building a security metrics program. These metrics can help:

  1. Provide budget and staffing justification for expansion

  2. See what risks your organization really faces

  3. See what risks your organization will face long and short tem

  4. Gauge the effectiveness of your team(s) effectiveness in meeting your security control requirements.

There are a number of good resources on building a security metrics program. Matt Tolbert’s 2007 “Effective Security Metrics” presentation is a great summary.  Andrew Jaquith’s book “Security Metrics: Replacing Fear, Security and Doubt” is another good resource. It’s one of my “bibles” of security metrics. The Educause “Effective Security Metrics: A Guide to Effective Security Metrics” ( https://www.educause.edu/focus-areas-and-initiatives/policy-and-security/cybersecurity-program/resources/information-security-guide/toolkits/effective-security-metrics ) has good high level points to setting up a security metric framework regardless of your industry sector.


Sunday, April 3, 2022

Motorcycle Riding and Being A CISO

 I was checking out some YouTube videos and ran across this one with Michael Jordan, Charles Barkley and Oprah Winfrey. Towards the end of the video, Michael talks about being a defensive driver if you ride a motorcycle. "You have to be really focused and see the traffic ahead", says Jordan.  He then takes a dig at Charles.  Check out https://www.youtube.com/watch?v=t_Q1k2r2yao.

 

I've ridden bicycles almost all of my life and motorcycles for the last third of my life. When I'm on the bike (either type), I am looking ahead to see what traffic patterns are there and trying to anticipate how I can maneuver through those patterns safely and efficiently. My nephew and I used to play a game when he was younger. We'd be in the mall and the challenge was to walk through a crowd from point A to B without missing a step or stopping because someone stepped in front of you. You had to watch the traffic flow and make your best guess on  where and when an opening would occur.

 

This is one of the things a CISO or security architect should practice. You want to look at threat intel, network traffic or attack patterns and chart a course of action based on your past knowledge as well as your ability to guess what will happen next. Sure, sometimes you guess wrong but you use that knowledge to improve your prediction capability. Sound like machine learning? Probably.

 

Next time you ride a bicycle or motorcycle, see if you ride defensively by looking ahead and anticipating the next action that can happen. Take that skill and apply it to designing your security architecture.

Sunday, November 28, 2021

Is Protecting Admin Privs on Endpoints Still Relevant?

 

The post-pandemic WFH (Work From Home)  model should force us to reevaluate the effectiveness of our security architectures. The most common reason for wanting administrative privileges on a device is that the local IT support can't install needed software when it's required by the business. I ask my SANS students how long it takes to install a software package for a business unit. The answers range from 1-2 weeks to 6 or more months because of a software review process. 

Admin privileges on endpoints

I want to emphasize that I'm NOT talking about administrative privileges on Active Directory or some other central management (Kaseya, Solarwinds, etc.)  domain accounts. I'm talking about local accounts and accounts on standalone computers. 

Is the "User having (local/standalone) admin privileges on a computer" as bad a security risk as people say it is? I emphasize the term "local/standalone" admin accounts. I think it is not.  Why? 

  

1) in the old days, having admin privileges on a multi-user system was a big deal. If you were in administrator/root mode and your account got owned, the consequence of that breach would impact ALL of the users on that system.  For large multiuser systems, that could be hundreds to thousands of users.  I understand why there was concern about the administrative/root accounts being secure. For servers that provide a service to multiple remote (to the server) users, it makes sense to restrict admin privileges on the server(s).

  

 2) In today's BYOD world, users are admin/root and general users simultaneously. There usually is only user per device. The impact of an admin/root failure is limited to the individual.   Phishing, ransomware attacks are just as damaging regardless of the entity that triggered the attack being a general user or admin level account.  Smartphones, tablets, etc. have merged the admin and general privilege levels into a single account so it makes no sense to "restrict" admin privileges on those devices today. You can't enforce that.  

  

What about high risk data exposure? Such data exposures can happen in either admin/root or general user mode. For most of the hits I've seen over the years, the damage was done regardless of the privilege level of the account involved.

  

It comes down to training. I've said in a previous blog entry that a poorly trained sysadmin is one of the greatest threats to an organization's data and infrastructure. Organizations should require a minimum amount of training for employees who want administrative privileges on a device. 


Thursday, April 15, 2021

Time to Train -

 "Excuse me, sir. How do I get to Carnegie Hall?"

"Practice, Practice, Practice."


I've always said that a poorly trained sysadmin is one of the greatest threats to any organization's infrastructure. The military training module may seem archaic and cumbersome but it is effective. There is a significant amount of investment in creating an effective training program. I believe the correct technical description is "it ain't cheap".  Organizations that fail to train their technical and general user staff in basic or advanced IT security practices are doomed to suffer multiple failures. 

I'm not going to dive into pedagogy (can't help but giggle everytime I hear that word) or the merits of a good training program. Too much has been said on those topics. Instead, I'm going to present my idea of a training roadmap here:




 Here we have 3 main training tracks:

  • Technical track - the target audiences are system administrators, developers, IT Security analysts/architects. These training programs are designed to enhance your staff's technical knowledge.
  • Awareness track - the target audiences are your general staff, management. These training programs are designed to make your workforce aware of the laws, regulations, best practices for handling your organization's sensitive data. In addition, these programs show your staff the different types of physical and cyber attacks they may see and how to respond to these threats.
  • User (How-to) track - this training program teaches your staff how to use the day to day tools of your business. It covers things like how to:
    • use Microsoft Office, Adobe Acrobat tools
    • use graphical design tools
    • use collaboration tools
    • use in-house tools
    • use external software or hardware products.
There needs to be a blend of externally developed training materials (SANS Secure the Human, Skillsetsonline, LinkedIn Learning, etc.) and "local" training for in-house applications.

Take a look at the above roadmap and I would like to hear your suggestions on how to improve or implement the roadmap.


Saturday, January 23, 2021

Resilience Is the Key to a Successful Defensive Strategy


The main mission of any CISO is not to prevent breaches of their infrastructure, rather, it's to safeguard you organizations' sensitive data and identity. I've said many time in the past that there are no device breach notifications but there are plenty of data breach notification laws. There are many ways to protect data and identity like encryption, monitoring outbound traffic, increasing user awareness, multi-factor authentication. These are important things but they are a means to achieve a goal. Resilience is the key to a sound defensive strategy. Here are some thoughts.

  1. We play defense not offense. 95% of companies hire cybersecurity people to defend their company from cyberattacks. They don't hire them to attack other sites. That's what the remaining 5% do. However, to play good defense, one must know how to play good offense. In other words, a Blue Team should have strong Red Team skills.
  2. One must accept the fact that a breach will happen regardless of whatever controls are in place. The old defensive strategy of building a "wall" to keep the bad guys out has failed. While there are many variants of the now popular Zero Trust Network philosophy, there are 2 key points that must be in place:
    1. The network is hostile.
    2. Data and identity are the new borders
  3. The key to a successful defensive strategy is resilience not prevention.  A sound resilience strategy is key to recovery.

Resilience 

I could give the Webster's dictionary definition of resilience but let me give you an example.

Ransomware is one of the destructive attacks that has affected a large number of organizations and people recently.  It's been around since 1989 but what made it popular was the introduction of cryptocurrency as the payment mechanism. For example, the Virginia State prescription monitoring database was hit with a ransomware attack in 2009 and the attackers demanded a $10M ransom. The state didn't pay and restored from backups. There was a disruption of service, some loss of data but the service recovered. Collecting the $10M in small bills requires a bunch of duffel bags and every LEO in the planet watching those bags to see who collects them. 

This incident convinced me that the best defense against ransomware attacks is not "prevention", rather, it is "recovery". Take the time to carefully align file permissions with need-to-access requirements of the business. This is a difficult step. It may limit ransomware damage by limiting the files the malware can access.
 
A good backup strategy is the best defense in this case. A system gets hit with ransomware, you wipe it, patch it, restore your data from good backups and then move on with your business. A good resilience strategy should include these steps:

  • find your sensitive data. Consolidate it into something like a data lake. 
  • Map where your sensitive data goes within your network borders as well as outside your borders. 
  • Backup this data lake by taking snapshots, doing old school incremental backups and store the backups offline in a read-only mode. For example, NetApp devices allow the creation of a read-only snapshot.
  • Test your recovery processes frequently.

The old RFC 1244 "Site Security Handbook" describes two defensive strategies: "Protect and Proceed" and "Pursue and Prosecute".  It set the following conditions for each of these approaches:
 
Protect and Proceed
      1. If assets are not well protected.
      2. If continued penetration could result in great
         financial risk.
      3. If the possibility or willingness to prosecute
         is not present.
      4. If user base is unknown.
      5. If users are unsophisticated and their work is
         vulnerable.
      6. If the site is vulnerable to lawsuits from users, e.g.,
         if their resources are undermined.
   Pursue and Prosecute
      1. If assets and systems are well protected.
      2. If good backups are available.
      3. If the risk to the assets is outweighed by the
         disruption caused by the present and possibly future
         penetrations.
      4. If this is a concentrated attack occurring with great
         frequency and intensity.
      5. If the site has a natural attraction to intruders, and
         consequently regularly attracts intruders.
      6. If the site is willing to incur the financial (or other)
         risk to assets by allowing the penetrator continue.
      7. If intruder access can be controlled.
      8. If the monitoring tools are sufficiently well-developed
         to make the pursuit worthwhile.
      9. If the support staff is sufficiently clever and knowledgable
         about the operating system, related utilities, and systems
         to make the pursuit worthwhile.
      10. If there is willingness on the part of management to
          prosecute.

                                             Figure 1. Protect and Proceed vs Pursue and Prosecute

My ransomware scenario's recovery process is an implementation of requirement listed in RFC 1244 - " "Attempts will be made to actively interfere with the intruder's processes, prevent further access and begin immediate damage assessment and recovery." 

This is an example of resilience. Andy Greenberg's book "Sandworm" has a chapter dedicated to resilience. Dan Geer's essay "A Rubicon" is another example of the importance of resilience. Creating an "parallel" network universe addresses interdependency issues and allows for a quick recovery.

We should certainly spend funds on detection tools but the bulk of present-day defenses should be focused on how we recover from an attack. Resilience processes such as backups,  monitoring and disrupting outbound traffic to questionable sites are examples of a good resilience strategy. 

You're going to get breached at some point in time. How fast you recover can limit the damage done to your business processes. 

References


https://assets.documentcloud.org/documents/4366740/Geer-Webready-Updated.pdf
https://tools.ietf.org/html/rfc1244
https://mysupport.netapp.com/NOW/public/eseries/sam_archive1150/index.html#page/GUID-8538272A-B802-49D9-9EA2-96C82DAD26A2/GUID-F6C0C512-F196-4008-97AE-EA06EE4D32F6.html