From Dashboards to Decision-Makers
For the past decade, the cloud operations industry has pursued a vision of intelligent infrastructure management under the banner of AIOps — using machine learning to analyze logs, detect anomalies, and recommend remediation actions. The results have been mixed. Most AIOps tools have been good at identifying problems but poor at fixing them, generating alert fatigue rather than operational improvement. The human operator remained firmly in the loop, interpreting AI recommendations and manually executing fixes.
In 2026, the industry is attempting a much more ambitious leap. Agentic AI — autonomous AI systems capable of taking independent action within defined boundaries — is being applied to cloud infrastructure operations. Instead of AI that alerts you to a problem and suggests a fix, the new generation of tools promises AI that detects the problem, diagnoses the root cause, implements the remediation, validates the fix, and documents the incident — all without human intervention.
Microsoft’s launch of agentic capabilities within Azure Copilot marked a watershed moment for the category. The system deploys six specialized agents — migration, deployment, optimization, observability, resiliency, and troubleshooting — that can analyze infrastructure telemetry across Azure services, identify degraded performance or impending failures, correlate signals across multiple data sources, and execute remediation playbooks autonomously. Microsoft’s vision for 2026 is to make agentic AI a standard part of how Azure customers build and operate applications rather than a niche capability.
Amazon Web Services followed in December 2025 with the public preview of AWS DevOps Agent, positioning it as an autonomous, always-on on-call engineer. The agent builds a topology map of application resources and relationships, then correlates telemetry from CloudWatch, Datadog, New Relic, and Splunk alongside deployment history from GitHub and GitLab CI/CD pipelines. When alerts trigger, it automatically investigates by analyzing logs, traces, and code changes to surface root causes and recommend mitigation steps.
Ericsson’s launch of its Agentic rApp as a Service on AWS at MWC 2026 targets telecommunications infrastructure with similar ambitions: AI agents that can manage the complexity of modern telecom networks. The system features specialized AI agents coordinated by a supervisor agent, integrated with Ericsson’s Intelligent Automation Platform. Ericsson’s AI solutions for network optimization already handle more than 100 million AI inferences daily across 11 million cells serving over 2 billion subscribers, and field testing of the new agentic platform is underway with leading CSPs including Vivo Brazil.
The claims are dramatic. Vendors report mean time to resolution (MTTR) reductions of 40% to 80% depending on incident type. AWS says its DevOps Agent reduces MTTR from hours to minutes. Google Cloud’s SRE teams are using Gemini CLI across the entire incident lifecycle — paging, mitigation, root cause analysis, and postmortem — to keep mean time to mitigation low. But the reality behind these headline numbers is more nuanced, and the path from demo to production is longer and harder than the marketing suggests.
What “Agentic” Actually Means for Infrastructure
The term “agentic AI” gets thrown around loosely in the technology industry, so it’s worth being precise about what it means in the context of cloud operations.
An agentic AI system for infrastructure management has four key capabilities that distinguish it from traditional AIOps:
Autonomous reasoning. The agent can analyze a novel situation — one it hasn’t been explicitly programmed to handle — and develop a diagnostic and remediation plan. This goes beyond pattern matching on known incident types. A truly agentic system can reason about unfamiliar failure modes by combining its understanding of system architecture, dependency relationships, and operational principles.
Tool use and execution. The agent doesn’t just recommend actions — it can execute them. It has authenticated access to infrastructure APIs, deployment pipelines, configuration management systems, and monitoring platforms. It can scale Kubernetes clusters, modify load balancer configurations, trigger database failovers, restart services, and update DNS records.
Multi-step planning. Complex infrastructure incidents typically require a sequence of coordinated actions. The agent can plan multi-step remediation — first isolating the affected component, then diagnosing the root cause, then implementing a fix, then validating the fix, then gradually restoring traffic. Each step’s outcome informs the next step’s actions.
Learning and adaptation. The agent improves over time, incorporating the outcomes of its interventions into its knowledge base. When a particular remediation approach fails, the agent learns not to repeat it. When a new type of incident occurs, the agent’s handling of it becomes a template for future similar events.
These capabilities, implemented well, could transform cloud operations from a reactive, labor-intensive discipline into a proactive, largely automated one. The key phrase is “implemented well.”
Real-World Results: The Good and the Complicated
Organizations that have deployed agentic AI for cloud operations report genuine improvements, but the results require careful interpretation.
The MTTR reduction numbers that appear in vendor materials typically refer to specific incident categories where the agentic system excels: resource exhaustion (CPU, memory, disk), certificate expirations, known misconfiguration patterns, and routine scaling events. For these “known-known” incidents — problems with well-understood symptoms and proven remediation — agentic AI is genuinely transformative. A system that can detect a filling disk, identify the responsible process, clean up temporary files or expand storage, and close the incident in under a minute delivers enormous value. Industry data shows organizations implementing intelligent automation resolve 47% of routine incidents without human intervention, reducing MTTR by 68% for those specific incident types.
But infrastructure incidents aren’t all known-knowns. The incidents that consume the most human engineering time are typically complex, multi-system failures involving cascading effects, race conditions, or subtle configuration interactions. For these incidents, agentic AI systems perform more like very fast triage assistants than autonomous operators. They can gather relevant telemetry, identify correlations, and narrow the diagnostic space, but they often lack the contextual understanding to determine the right remediation for novel failure modes.
Several organizations report a pattern they call “confident wrong actions” — situations where the agentic system, operating with the authority to make changes, implements a remediation that is technically valid but contextually inappropriate. Scaling up a service that is failing due to a dependency issue, for example, or restarting a stateful service that requires careful coordination. These incidents, while not catastrophic when proper guardrails are in place, erode the trust needed for expanded autonomy.
The organizations reporting the best results share common characteristics: they have well-instrumented infrastructure with comprehensive telemetry, they have mature runbook documentation that agents can be trained on, they deploy agents incrementally starting with low-risk actions, and they maintain human oversight with graduated escalation policies.
Advertisement
The Maturity Gap: Fast Growth, Stubborn Barriers
The Dynatrace Pulse of Agentic AI 2026 report, surveying 919 senior leaders directly involved in agentic AI implementation, provides the clearest picture of the adoption landscape. The numbers reveal rapid acceleration alongside persistent barriers.
On the adoption side, 50% of organizations now have agentic AI projects in production for limited use cases, 44% have projects in broad adoption across select departments, and 23% have reached mature, enterprise-wide integration. ITOps and DevOps lead adoption at 72%, followed by software engineering at 56% and customer support at 51%. Gartner predicts that 40% of enterprise applications will feature task-specific AI agents by the end of 2026, up from less than 5% in 2025.
But roughly half of all agentic AI projects remain stuck in proof-of-concept or pilot stages, and security concerns are a primary reason. The barriers to broader production deployment remain formidable.
Trust and governance. Giving an AI agent the authority to modify production infrastructure is a significant governance decision. Most organizations require extensive testing, approval processes, and risk assessments before granting autonomous execution authority. ServiceNow’s 2025 Enterprise AI Maturity Index found that fewer than 1% of surveyed organizations scored above 50 out of 100 on AI maturity, and the overall high score actually fell 12 points year-over-year. The regulatory environment for certain industries — financial services, healthcare, government — adds additional layers of approval.
Observability prerequisites. Agentic AI systems are only as good as the telemetry they can access. Organizations with fragmented monitoring, incomplete logging, or siloed observability tools cannot provide agents with the comprehensive view needed for accurate diagnosis. Many organizations discover that their observability infrastructure needs significant upgrades before agentic AI can be deployed effectively. The Dynatrace report shows observability adoption is highest during implementation (69%), followed by operationalization (57%) and development (54%).
Integration complexity. A useful agentic system needs to interact with dozens of tools: cloud provider APIs, container orchestration platforms, CI/CD pipelines, configuration management systems, incident management platforms, communication tools, and more. Building and maintaining these integrations is a significant engineering effort, and the heterogeneity of most enterprise environments makes standardization difficult.
Skill requirements. Ironically, deploying agentic AI for operations requires deep operational expertise. Someone needs to define the boundaries of agent autonomy, design the escalation policies, validate the agent’s reasoning, and intervene when things go wrong. The people best qualified for this work are experienced SREs and platform engineers — the same people the technology is supposed to augment.
Cost. The large language models and specialized AI models that power agentic cloud operations are not cheap to run. The inference costs for processing high-volume telemetry data in real-time, reasoning about complex system states, and generating remediation plans can be substantial. That said, 74% of organizations surveyed expect their agentic AI budgets to increase in the next 12 months, often by an additional $2 to $5 million or more.
The Architecture of Autonomous Operations
Organizations successfully deploying agentic AI for cloud operations are converging on a common architectural pattern that balances autonomy with safety.
Tiered autonomy. Rather than granting agents blanket authority, organizations define tiers of actions based on risk. Tier 1 actions — read-only operations like gathering telemetry, querying logs, checking configurations — are fully autonomous. Tier 2 actions — low-risk changes like scaling, restarting non-critical services, updating routing weights — are autonomous with logging and review. Tier 3 actions — high-risk changes like database failovers, configuration changes to core services, or actions affecting production data — require human approval.
Guardrails and constraints. Agents operate within defined boundaries: they cannot make changes during maintenance windows, cannot modify infrastructure tagged as critical without approval, cannot execute actions that would reduce redundancy below defined thresholds, and cannot spend above defined cost limits. These guardrails prevent the most damaging potential failures.
Feedback loops. Every agent action generates feedback data: was the remediation successful? Did it cause any side effects? How did it compare to what a human would have done? This feedback is used to continuously refine the agent’s reasoning and expand (or contract) its autonomy boundaries.
Shadow mode. Many organizations deploy agents in “shadow mode” first — the agent analyzes incidents and proposes actions without executing them, allowing humans to evaluate its judgment before granting execution authority. Shadow mode periods of weeks or months are common before agents are promoted to autonomous operation.
The Observability Ecosystem Is Going Agentic
The shift to agentic operations isn’t limited to cloud providers. The entire observability and incident management ecosystem is integrating autonomous capabilities.
Dynatrace introduced a new agentic AI foundation at its Perform 2026 conference, with agents that can continuously detect changes, assess impact, and automatically respond — advancing toward auto-remediation, auto-prevention, and auto-optimization. Dynatrace and ServiceNow have deepened their strategic collaboration, combining Dynatrace’s real-time causal intelligence with ServiceNow’s automated closed-loop workflows for detecting, diagnosing, and remediating incidents autonomously.
Datadog launched Bits AI, a collection of agents designed to act as digital teammates: an AI SRE for on-call duties, a Dev Agent for coding, and a Security Analyst for incident response. When an alert fires, Bits AI begins investigating on its own — gathering telemetry, reading runbooks, and testing multiple hypotheses — aiming to have a root cause hypothesis ready before an engineer even checks in.
This ecosystem convergence matters because enterprise environments rarely run on a single platform. A production incident might span AWS infrastructure monitored by Datadog, trigger an alert in PagerDuty, require investigation of a deployment logged in GitHub Actions, and ultimately need a Kubernetes configuration change managed through Argo CD. Agentic systems that can operate across these tool boundaries deliver the most value.
What Needs to Happen for Enterprise Adoption
Closing the gap between the 50% that have limited production deployments and full enterprise-wide integration requires progress on several fronts.
Standardized agent interfaces. The cloud operations ecosystem needs standardized APIs and protocols for agent interaction with infrastructure tools. The proliferation of proprietary agent frameworks from different vendors creates fragmentation and integration overhead. Industry initiatives around open agent standards are emerging but not yet mature.
Better evaluation frameworks. Organizations need systematic ways to evaluate agent performance before and during production deployment. Chaos engineering practices — intentionally injecting failures to test resilience — are being adapted for agent evaluation, but standardized benchmarks for agentic operational AI do not yet exist.
Graduated trust mechanisms. The industry needs better patterns for incrementally expanding agent autonomy based on demonstrated competence. Binary trust decisions — the agent either has authority or it doesn’t — are too coarse. Fine-grained, dynamically adjustable trust levels that expand or contract based on agent performance would enable faster, safer adoption.
Cost optimization. The inference costs of running large AI models for operational reasoning need to decrease significantly for the economics to work across a broad range of organizations. Smaller, specialized models fine-tuned for operational tasks — rather than general-purpose LLMs — may provide the necessary cost reduction.
The trajectory is clear: agentic AI will increasingly manage cloud infrastructure, starting with routine operations and gradually expanding to more complex scenarios. The question is not whether autonomous infrastructure operations will arrive, but how quickly organizations can build the trust, governance, and technical foundations to adopt them safely.
Advertisement
🧭 Decision Radar (Algeria Lens)
| Dimension | Assessment |
|---|---|
| Relevance for Algeria | Medium — Algeria’s growing cloud adoption (data center market projected at $447M by 2035) means agentic operations will become relevant as organizations scale beyond manual management, especially for Algeria Telecom’s new 400G backbone infrastructure |
| Infrastructure Ready? | Partial — Algeria’s cloud infrastructure is still maturing. The 2025 Oran AI data center initiative and Huawei partnership for 400G optical networks are positive, but limited data center density and slow internet speeds (ranked fourth-slowest globally) constrain agentic AI deployment |
| Skills Available? | No — SRE and advanced DevOps expertise is scarce in Algeria. Agentic AI requires deep platform engineering knowledge to configure autonomy boundaries, design escalation policies, and validate agent reasoning. Huawei’s ICT Competition 2025-2026 and university programs are building pipeline but not production-ready talent |
| Action Timeline | 12-24 months — Monitor developments now, begin proof-of-concept pilots as Algeria’s cloud infrastructure matures and regional cloud providers expand offerings |
| Key Stakeholders | Algeria Telecom operations teams, Sonatrach IT infrastructure, government digital transformation agencies, cloud-first startups, university computer science departments |
| Decision Type | Educational — Track the technology evolution, invest in DevOps/SRE training programs, and prepare observability infrastructure for eventual agentic adoption |
Quick Take: Agentic cloud operations will matter for Algeria as the country’s digital infrastructure scales, but the prerequisites — mature observability, comprehensive telemetry, and SRE expertise — are still being built. Algerian organizations should focus on foundational observability and DevOps practices now, which will both improve current operations and prepare the ground for agentic automation when the infrastructure supports it.
Sources & Further Reading
- Agentic Cloud Operations: A New Way to Run the Cloud — Microsoft Azure Blog
- AWS DevOps Agent: Accelerate Incident Response and Improve System Reliability — AWS Blog
- Ericsson Launches Agentic rApp as a Service on AWS — Ericsson Press Release
- The Pulse of Agentic AI in 2026 — Dynatrace Global Report
- Dynatrace and ServiceNow Deepen Strategic Collaboration for Autonomous IT Operations — ServiceNow Newsroom
- From Paging to Postmortem: Google Cloud SREs on Using Gemini CLI for Outage Response — InfoQ
- Gartner Predicts 40% of Enterprise Apps Will Feature AI Agents by 2026 — Gartner





Advertisement