Select Page

Gopal Dommety

|
originally published on Feb 16, 2026
Share

Software delivery is a cross-system chain: Git → CI/CD → artifacts → Kubernetes/runtime → cloud controls → incidents → remediation. The tools are excellent in isolation, but during a production incident or a cloud posture failure, teams still waste time reconstructing context across systems.

A context graph changes that. It connects:

  • Entities (repos, PRs, builds, images, services, clusters, cloud resources, policies, findings, incidents, owners)
  • Relationships (produces, deployed_to, depends_on, owned_by, violates, approved_by)
  • Decision traces (gates, exceptions, approvals, rollouts, remediation outcomes)

This is what enables fast diagnosis and safe remediation—especially for DevOps diagnostics and CSPM remediation, where “what changed?” and “what’s impacted?” are the critical questions.

Below are 12 practical use-cases, grouped by persona.

SRE Use-Cases (1–3): Faster incident triage and lower MTTR

1) “What changed recently for this service/environment?”

Question: What changed in the last N minutes/hours that could explain the incident?
Context graph pulls: deploy traces, config changes, feature flags, dependency upgrades, owner/on-call
Outcome: A ranked list of change candidates linked to owners.

2) “Is there a correlated deploy/config/flag change?”

Question: Did a deployment or config change occur right before the incident started?
Context graph pulls: incident start time, rollout stages, diffs, canary progression
Outcome: Strong correlation signal + the safest rollback target.

3) “What’s the blast radius?”

Question: What other services/customers/environments could be impacted?
Context graph pulls: dependency graph, shared infra components, common base images, shared cloud resources
Outcome: Targeted containment and faster stakeholder comms.

DevOps Diagnostics Use-Cases (4–6): Systematic debugging across pipelines + runtime

4) “Why did this deployment fail?”

Question: What is the root cause chain from pipeline → artifact → runtime?
Context graph pulls: CI stage results, artifact provenance, deployment events, cluster events, policy gates
Outcome: One trace instead of jumping across 6 tools.

5) “Infra drift and configuration diagnostics”

Question: Is this an application issue, or drift in infra/config (Helm values, env vars, cluster policies)?
Context graph pulls: config/version diffs, drift signals, admission policy changes, runtime events
Outcome: Clear diagnosis path + recommended next checks.

6) “Cross-env mismatch diagnostics”

Question: Why does staging work but production fails?
Context graph pulls: environment diffs (configs, secrets references, policies, dependencies, cloud resources)
Outcome: Pinpoints the minimal delta that explains the behavior.

AppSec Use-Cases (7–8): Risk-in-context and supply chain linkage

7) “Which vulnerabilities are actually deployed in production?”

Question: What findings exist in running workloads vs just in repos?
Context graph pulls: CVE → SBOM/package → image digest → deployment → prod env
Outcome: Less noise, better prioritization.

8) “Where did the risky component come from?”

Question: Which repo/PR/build introduced this vulnerable dependency or misconfig?
Context graph pulls: provenance + SBOM mapping, PR history, reviewer/approval trail
Outcome: Targeted fix + opportunity to add preventive guardrails.

CSPM Remediation Use-Cases (9–10): Fixing cloud posture safely (without breaking apps)

9) “CSPM finding → real impact mapping”

Question: Does this CSPM finding matter, and what will it break if we fix it?
Context graph pulls: cloud resource → workload/service → owners → traffic/dependency context → environment criticality
Outcome: You can prioritize the CSPM backlog based on application impact.

Example: “S3 bucket public” becomes:

  • which app uses it,
  • whether it serves production traffic,
  • whether it contains regulated data,
  • whether a policy fix breaks downstream processing.

10) “Auto-remediate CSPM issues with guardrails”

Question: Can we fix this posture issue automatically, and under what constraints?
Context graph pulls: environment, criticality, change window, compensating controls, ownership approvals
Outcome: Safe auto-remediation for low-risk items; approval-gated remediation for high-risk items.

Common patterns

  • auto-close overly permissive security groups in dev
  • require approval for prod internet exposure
  • attach evidence lineage for compliance (who changed what, when, why)

AI Coding Agent Use-Cases (11–12): Governed code changes and safe fixes

11) “Generate a compliant fix PR”

Question: Can an agent propose a fix that matches repo rules and release constraints?
Context graph pulls: repo policies, required tests, secure patterns, banned deps, service ownership, rollout rules
Outcome: A PR that is more likely to pass review and succeed in deployment.

12) “Choose auto vs approve vs manual (contextual trust)”

Question: Should this remediation be auto-executed, suggested, or routed for approval?
Context graph pulls: prod/dev, regulated scope, blast radius, policy, change freeze windows, prior incidents
Outcome: Safe automation that doesn’t create operational risk.

Why these use-cases work: the “change story” backbone

All of these use-cases rely on reconstructing one chain:

PR/commit → build → artifact/image → deploy → runtime → cloud posture → policy decisions → incidents → remediation outcomes

That chain is what context graphs make queryable.

How to get started

You don’t need every tool integrated to unlock DevOps diagnostics and CSPM remediation value.

Start with:

  • Git + CI/CD (change lineage)
  • Kubernetes/runtime inventory (what’s running where)
  • CSPM source (cloud posture findings)
  • Incident tool (PagerDuty/Jira) — optional but powerful early

Then expand into:

  • SBOM/attestations, service catalog ownership, feature flags, IAM relationships.

Key takeaways

  • DevOps diagnostics improves when pipeline events connect to runtime reality.
  • CSPM remediation becomes safer when cloud findings map to application impact and owners.
  • Contextual trust enables auto-remediation without chaos.
  • A minimal context graph can deliver value quickly; expand iteratively.
Tags : security

Gopal Dommety, Ph.D. is the Chief Executive Officer of OpsMx, a company advancing the automation and security of software delivery for the modern enterprise. Under his leadership, OpsMx is redefining how organizations build, secure, and release software, enabling developers to deliver innovation with speed, safety, and confidence. A technologist and inventor, Dr. Dommety holds over 70 patents and is the principal author of several Internet Protocols (RFCs) that power today’s global networking infrastructure. His work has shaped critical areas of large-scale distributed systems, algorithmic design, and secure automation. He has also authored more than 20 peer-reviewed papers, book chapters, and journal publications, and previously led the Mind-Map Project, an AI research initiative focused on modeling behavioral and personality traits from user-generated data. Before founding OpsMx, he was a General Partner at Neem Capital, a technology-focused investment firm, and held senior leadership roles in product management, research, and engineering at major technology companies and startups. Rooted in humble beginnings from a remote village in India, Gopal’’s career is guided by the principles of simplicity, first-principles thinking, and purpose-driven innovation—values that continue to shape his vision for building secure, intelligent, and resilient technology systems that move the world forward.

0 Comments

Submit a Comment

Your email address will not be published.

This site uses Akismet to reduce spam. Learn how your comment data is processed.