Why Vulnerability Management Breaks Down in Industrial Systems
The patch-prioritize-verify cycle of enterprise vulnerability management rests on assumptions that collapse in OT environments. A mature OT vulnerability program looks fundamentally different — and must be built from scratch.
Part of the Phase I — Observation series
By Michael E. Ruiz
Vulnerability management, in the way most security programs practice it, is built on an assumption that breaks immediately in operational technology environments: that the remediation actions available in response to a discovered vulnerability are roughly equivalent to the risk of leaving the vulnerability unpatched. Scan, prioritize by CVSS score, patch, verify, repeat. The model works tolerably well in enterprise IT because the cost of patching, typically a maintenance window, a reboot, and a configuration test, is generally low and predictable. In OT, neither the cost nor the availability of remediation is predictable, and this difference invalidates most of the logic underlying traditional vulnerability management programs.
Start with the patch question. In an enterprise environment, a vendor releasing a security patch for an operating system is an event that triggers a defined process: test in staging, deploy in waves, verify. In an OT environment, a vendor releasing a patch for a PLC firmware or a DCS component is an event that triggers a question: has the vendor tested this patch against our specific configuration of this system, running this version of the application logic, communicating with these specific field devices? If the answer is no, and it frequently is, applying the patch is not a straightforward risk reduction action. It is a change to a production system whose behavior after the change is not fully characterized. Some manufacturers recommend against patching outside of planned turnaround windows, not because they are indifferent to security but because the stability of a running process has its own risk calculus.
Then there is the legacy problem. Industrial environments routinely contain systems running operating systems that have been end-of-life for years. Windows XP embedded on HMI workstations. Windows 2003 on historian servers. These systems exist because replacing them requires replacing or re-qualifying the application that runs on them, which may in turn require process re-validation, regulatory approval, or equipment replacement. The cost of modernization is not a license fee. It is a capital project. Telling an operations team that they need to address an unpatched vulnerability on a system that cannot be patched without a multi-million dollar project is accurate information that produces no useful action.
CVSS scores make this worse. A CVSS 9.8 critical vulnerability on a network-isolated historian connected only to a data diode is a different risk than the same CVSS 9.8 vulnerability on an engineering workstation with RDP access from the corporate network. The score is the same. The actual risk is not.
Vulnerability management systems that sort remediation priority by severity score without accounting for reachability, exploitability in context, and consequence produce priority queues that consistently send engineers after the wrong problems. Contextual vulnerability management, which accounts for network topology, asset criticality, and process consequence in the prioritization model, is not an advanced capability in OT. It is the minimum viable approach.
There is a third dimension that gets less attention: the cost of assessing vulnerabilities in the first place. Running a credentialed scan against an OT asset requires credentials, which requires an account management process that many OT environments do not have. It may require agents, which raises the same concerns about resource consumption discussed in the monitoring context. It may require a maintenance window because the scan itself, not just the patch, poses operational risk. Each of these friction points reduces scan coverage, and reduced scan coverage means the vulnerability database is built on a sample of the environment, not the whole. Organizations that report vulnerability counts for OT environments should be asked how that number was derived before anyone interprets it.
What a mature OT vulnerability program actually looks like is considerably different from its IT counterpart. It starts with a prioritized asset inventory, because not all assets are equal and the ones that matter most are in the control path for safety-critical or high-consequence processes. It accounts for compensating controls: an unpatched PLC behind a properly configured industrial DMZ, with no remote access and process-level anomaly detection, carries a different residual risk than an identical PLC with a direct connection to the corporate WAN. It defines remediation options beyond patching, including network segmentation, application whitelisting, and disabling unused communication ports, as legitimate and trackable risk reduction actions. And it measures progress not by patch rate alone but by risk reduction across the portfolio of assets that actually matter.
None of this is an argument for ignoring vulnerabilities in OT environments. It is an argument for treating the problem as fundamentally different from enterprise vulnerability management, with different tools, different processes, different success metrics, and different conversations with leadership. The organizations that get this right are the ones that stopped trying to apply the IT playbook and built something that reflects how industrial systems actually work.
These ideas are available as keynote presentations and executive briefings. Explore speaking topics →