Operations | Monitoring | ITSM | DevOps | Cloud

How Autonomous Are Your IT Operations, Really?

This post introduces a six-level maturity model that defines what true autonomy looks like in IT operations, from basic AI chat interfaces to fully coordinated agent ecosystems. ITOps teams have more automation tooling than ever, and yet incident response still depends heavily on human judgment to hold it together. Alerts fire, engineers dig through dashboards, context gets assembled by hand, and someone at the end of the workflow makes the final call.

What is Agentic Observability?

Agentic observability is the instrumentation and correlation needed to explain and control agent behavior across multi-step workflows. Legacy observability focuses on runtime health and service behavior. You monitor metrics like CPU usage, memory, latency, and error rates to confirm that applications and infrastructure are functioning as expected. When a workflow degrades, the proximate cause is often a crash, timeout, permission error, or resource constraint.

GPU Fragmentation Is Killing AI Economics

By 2026, the GPU shortage isn’t a supply-chain hiccup anymore. It’s baked into the system. Even after pouring billions into CapEx, most enterprises still want 40% more GPU capacity than they actually have. And it’s not because they’re chasing moonshots. Technology companies are training foundation models while serving inference for millions of users on the same clusters. AI labs are juggling fine-tuning, evaluation, and real-time experimentation side by side.

How AI-Powered ATS Systems Are Transforming Modern Recruitment

Recruitment has changed dramatically over the past decade. Companies are no longer relying on manual CV screening and gut-feel interviews. Instead, AI-powered Applicant Tracking Systems (ATS) are reshaping how organizations hire - faster, smarter, and with less bias.

AI-ready sovereignty playbook 2026: how to run gen-AI workloads (ethically) in the EU

Sovereignty is a concept that can have shown nuances in the way it is currently used by states and industry to describe some services. The term “strategic autonomy” has also been used, as to describe the need for governments to ensure that they have a hand on the full value chain (or at least know the gaps and accept the risks) and can apply their rules while it seats in its jurisdiction (autonomy derives from the greek autos (self) nomos (rule).

What Is LLMjacking? The New AI Cybercrime Stealing Cloud AI Compute

LLMjacking is a new cybercrime where attackers steal access to cloud-hosted AI models and use them for free — while the victim pays the bill. In this video, we break down what LLMjacking is, how attackers exploit compromised credentials and exposed APIs, and why security teams should treat AI infrastructure as a high-value attack target. Discovered by the Sysdig Threat Research Team, LLMjacking is quickly becoming the AI-era equivalent of cryptojacking — except instead of mining cryptocurrency, attackers run expensive large language models (LLMs) at scale.

Meet the new Bits AI SRE: Deeper reasoning, twice as fast

When we announced Bits AI SRE at DASH 2025, we introduced an autonomous SRE agent that investigates alerts the moment they trigger. Bits AI SRE reads the same telemetry data as your team, understands your architecture, and follows your runbooks to identify likely root causes before you even open your laptop. It’s your AI teammate that’s always on call.

How AI lets you talk to your company's data and get answers instantly

In this conversation recorded at Elastic’s New York office, three product leaders discuss how AI agents are transforming enterprise software. The discussion features Steve Kearns (general manager, Search solutions at Elastic), Mike Nichols (general manager, Security solutions at Elastic), and Baha Azarmi (general manager, Observability at Elastic). They explain how Elastic Agent Builder allows teams to interact with their data using natural language instead of complex queries.

How LLMs can help boost productivity

Learn how large language models (LLMs) are transforming productivity in business, coding, research, and daily workflows. Discover practical ways to use AI tools to automate tasks and improve efficiency. Additional Resources: About Elastic Elastic, the Search AI Company, enables everyone to find the answers they need in real time, using all their data, at scale. Elastic’s solutions for search, observability, and security are built on the Elastic Search AI Platform — the development platform used by thousands of companies, including more than 50% of the Fortune 500.

Your Questions About AI-Assisted Development Answered

We recently hosted a webinar on AI-assisted development with DORA, and the audience had a lot of questions—far more than we could get to in an hour. I picked out six that get at the stuff people are wrestling with day to day. These aren't the easy questions, and I don't think there are necessarily easy answers, but I've spent the past year building and shipping with AI coding tools and observing (literally) what happens when that code hits production. Here's what I have.