All problems in computer science can be solved by another level of abstraction.

What Asset Context Changes About Vulnerability Findings

A vulnerability scanner, operating against a competent target environment, will produce several hundred findings per thousand hosts on any given pass. At enterprise scale this becomes a torrent. At mid-market scale — a hundred to a few thousand hosts — it becomes the thing that breaks the security team, because the team is small enough that every finding lands on the same handful of people and large enough that the aggregate cannot be processed by anyone running on human attention alone.

The conventional response is to sort by CVSS and work top-down. This produces a report. It does not produce a decision. CVSS is calibrated against a hypothetical environment — the vulnerability as it would behave in a lab, against a generic deployment, with default configuration, in isolation. The environment in front of you is none of those things. A CVSS 9.8 that affects a host with no exposed ports, no business-sensitive data, and no privileged reachability to anything important is less urgent than a CVSS 7.2 on the system that runs order entry. Anyone who has worked a triage queue for a month knows this; most of the industry keeps pretending otherwise because the alternative has been operationally infeasible.

The alternative is context-aware scanning. Not a new scanner; a layer above the scanner that interprets findings against the environment they actually live in. That layer requires two substrates — an authoritative asset inventory with business context attached, and a synthesis function that can combine finding data with asset data and produce ranked output at human scale. Both have been available in principle for years. The combination has only become operationally tractable in the last eighteen months.

The shape of the output

Start from what the output looks like when the context is there. Here is an illustrative excerpt from a real engagement (details abstracted, numbers representative):

Finding: CVE-2024-XXXX, CVSS 9.1, authentication bypass in embedded web UI, no vendor patch available.

Asset: cam-odessa-03, location: Odessa retail store, function: security camera management UI, business role: loss prevention supporting $N/day in POS volume at that location, dependents: loss-prevention reporting pipeline, network adjacency: flat with POS terminals (no segmentation).

Impact assessment: High. The authentication bypass plus the flat-network adjacency means an attacker on the camera UI can reach the POS terminals. The affected software is deployed at 4 of 22 store locations.

Similar findings: 3 other assets at Big Lake, Midland, and Crane locations running the same software version. Prioritize collectively.

Remediation path: Vendor offers no patch. Compensating control: interim network segmentation between camera VLAN and POS VLAN at affected locations. Long-term: replace camera management software. Estimated compensating-control deployment: 2 hours per site.

Compare that to what a scanner emits alone:

CVE-2024-XXXX, CVSS 9.1, host 172.102.38.106, “Authentication bypass in embedded Apache 2.4.16.”

The first is a decision. The second is a data point. The gap between them is the work, and the work is what the context layer produces.

Why this is now tractable

The two substrates each have their own reasons for being more available than they used to be.

The asset inventory substrate has been slowly improving for a decade. The adoption of CMDB tooling — NetBox in the operator-friendly end of the market, ServiceNow in the enterprise end, a proliferation of smaller tools in between — has made it realistic to maintain authoritative asset data at organizations that would have considered that a capability too expensive to justify five years ago. The tooling is still underused; the capability exists.

What the CMDB provides, when maintained, is the static half of context: what each asset is, what it runs, what it depends on, who owns it, what function it serves. The dynamic half — what’s happening to it right now, who’s reaching it, what vulnerabilities are present — comes from the operational tooling (scanners, flow collectors, log pipelines). The connection between the static and dynamic halves is the interpretation layer.

Four years ago the interpretation layer was human effort. An analyst looked at the finding, manually queried the CMDB, manually walked the dependencies, manually formulated an impact assessment, manually wrote the ranked narrative. The time cost was roughly fifteen minutes per nontrivial finding, which at a thousand findings per pass meant the interpretation was rationed to the worst few dozen and the rest got the CVSS sort. Organizations that tried to fully interpret their scan output — a small number did — staffed a team around it.

Today, the interpretation layer can be a synthesis function driven by a language model with strong context handling. Given structured input — the scanner finding, the asset record, the dependency graph, the organization’s business function map — the model produces the ranked narrative in seconds. The output quality, when the inputs are clean, is comparable to what a trained analyst produces. When the inputs are dirty, it’s worse, but the failure modes are the same as the human’s and are debuggable.

This is not “AI replaces analysts.” The failure rate is high enough — and the failures are subtle enough — that the output is a draft, not a deliverable. A human with the context pre-loaded reviews, corrects, and approves. But the human is now working at ten to twenty times the throughput they were working at when they had to construct the context themselves. That throughput change is what makes context-aware scanning tractable at the small-senior scale where most mid-market security teams operate.

What the capability actually requires

If the operational output is tractable, what’s keeping more organizations from producing it? Five things, ranked roughly by severity:

1. A CMDB that isn’t fiction.

The largest gap. Most asset inventories in the wild are one of three things: a spreadsheet updated annually, a network scan dump rebranded as an inventory, or a CMDB that was populated once and has drifted so far from reality that nobody trusts it. None of these supports context-aware synthesis. The minimum viable CMDB for this work has authoritative asset records, updated as infrastructure changes, with business function and dependency relationships maintained alongside the technical metadata.

This is the hard part. The tooling is available; the discipline is rare. Most of the engagement time on a context-aware scanning build-out goes here, not to the synthesis layer.

2. Structured scanner output.

If the scanner produces PDFs, the context layer cannot consume them. What’s needed is structured machine-readable output — API access, JSON or XML exports, or a dedicated findings platform (DefectDojo, Kenna, Nucleus) that normalizes output from multiple scanners into a consistent schema. Most enterprise-grade scanners have this; some mid-market tools still emit PDF-first and require extraction work.

3. A synthesis layer the organization can actually run.

Two shapes exist. The first is a self-hosted synthesis running against a local language model — operationally complex but keeps all data in-house. The second is an API-driven synthesis against a hosted model — operationally simple but requires sending scanner findings and asset metadata to a third-party API, which has data handling implications that the organization has to think through. Most teams land on the hosted path with a specific data minimization discipline (send the finding and the relevant asset fields, never the whole CMDB record, never credentials, never customer-identifying data).

The synthesis prompt engineering is less hard than people expect and less easy than it looks. The important thing is that the prompt returns structured output — ranked severity, impact narrative, remediation path — not free-form prose. Structured output can be validated. Prose can’t.

4. A review loop calibrated to the team.

Context-aware scanning produces draft output. Drafts need review. The review loop has to be fast enough that it doesn’t re-introduce the bottleneck the synthesis was meant to remove. In practice this means the reviewer sees the finding, the asset, the draft narrative, the confidence score, and can approve-or-edit-and-approve in under a minute. The interface matters. A bad review interface turns a 15x throughput gain into a 2x throughput gain, and operators notice.

5. Feedback to the synthesis layer.

Reviewer edits carry signal. An organization that captures the delta between the draft and the approved version builds a private training corpus over time. This is where the capability becomes genuinely differentiated — not because the synthesis is better than the generic baseline on day one, but because it gets measurably better on week twelve when it’s been trained against this environment’s particular patterns.

Organizations that skip this step operate indefinitely on the baseline capability. Organizations that maintain it pull ahead, and the gap is not trivially replicable by a competitor without equivalent deployment time in a similar environment.

What this changes about the practice

The economic implication for a security practice offering this capability is significant. A small-senior team (three to five humans) can cover vulnerability management for an order of magnitude more assets than they could without the context layer. The ceiling on what a single senior analyst can triage in a day moves from roughly fifty nontrivial findings to roughly five hundred, assuming the infrastructure is clean. That moves the economics of vulnerability-management-as-a-service from a capability that only the large organizations can afford to run in-house to one that mid-sized practices can deliver to their clients profitably.

The strategic implication is more interesting. The practices that build this capability well build a compounding advantage over the ones that don’t. The CMDB discipline, the structured findings workflow, the review loop, the private corpus — all of it takes 12 to 24 months to reach real operational maturity, and none of it is the kind of capability that can be dropped in wholesale from a vendor. The practices that start building now are building the moat for the decade.

And the practices that don’t are going to find themselves in an awkward position two years out, when their competitors are offering vulnerability management at a materially different price point with materially better output. CVSS sort isn’t going to hold against that for long.

What the output is not

Two things to be honest about.

The synthesis layer does not replace judgment. It produces a first draft that’s good enough for a reviewer to approve in a minute per finding. It does not produce a deliverable that can go directly to a client without human review. The distinction matters because the failure cases are not random — they’re systematic, often in ways that the synthesis model has no way to detect. An asset misclassified in the CMDB produces a ranked narrative that’s subtly wrong in every pass until the CMDB is corrected. A new vulnerability class the model hasn’t seen produces a narrative with the wrong shape. These are catchable in review; they’re not catchable automatically.

The synthesis layer does not secure anything. It ranks findings. The ranked findings still have to be acted on. Organizations that build the capability and don’t attach it to a remediation pipeline end up with very good reports and the same vulnerability posture they had before. The scanning work is an instrument; the instrument doesn’t play itself.

Both of those caveats are obvious to practitioners and routinely missed by leaders outside the discipline. Part of building this capability is managing the expectations around it from the start, so that the ranking output gets used correctly and the organization doesn’t conclude that “we have context-aware scanning now” means “we have security.”

Nobody ever has security. We have practices.

← index