Operations | Monitoring | ITSM | DevOps | Cloud

Debug PostgreSQL query latency faster with EXPLAIN ANALYZE in Datadog Database Monitoring

In PostgreSQL, the EXPLAIN ANALYZE statement gives you a detailed report of what actually happens when you execute a query. This kind of information is important for troubleshooting slow queries, but using EXPLAIN ANALYZE to collect this data is often challenging in a production environment. Datadog Database Monitoring now supports automatic collection of EXPLAIN ANALYZE plans for PostgreSQL, enabling you to easily capture execution details that help you troubleshoot slow queries.

Datadog acquires Propolis

Generative AI enables teams to write and ship code faster than ever. But current methods for testing and quality assurance have not evolved to match the new pace and scale of deployments. Manual and deterministic testing paths quickly become obsolete when new features are released, and they fundamentally can’t test AI outputs, leaving a massive untested surface area. To keep up, teams need new testing methods that can define what goals users have, and ensure that their outcomes match.

Unify and correlate frontend and backend data with retention filters

Teams can use Datadog Real User Monitoring (RUM) and RUM without Limits to get full visibility into the frontend health of their applications while retaining only the sessions that contain critical problems that affect the end-user experience. But application errors or slowness often result from backend issues, such as database bottlenecks. To diagnose these issues, you need to correlate the frontend data from RUM with the backend data from Datadog Application Performance Monitoring (APM).

Easily Map Logs to OCSF with Datadog Observability Pipelines

Normalizing security logs into the Open Cybersecurity Schema Framework (OCSF) is often complex, manual, and time-consuming. With Datadog Observability Pipelines, you can easily transform logs into OCSF format—right in your own environment—before routing them to destinations like Splunk, CrowdStrike, and AWS Security Lake. This video show how Security teams can use Observability Pipelines to: Collect, process, and transform logs into OCSF format automatically.

Monitor Arista VeloCloud SD-WAN performance with Datadog

As organizations grow their cloud environments and branch office networks, maintaining reliable connectivity and application performance becomes more complex. VeloCloud SD-WAN provides dynamic, policy-based routing to help ensure that your connectivity is dependable and cost-efficient, and that your applications perform consistently.

Building reliable dashboard agents with Datadog LLM Observability

This article is part of our series on how Datadog’s engineering teams use LLM Observability to iterate, evaluate, and ship AI-powered agents. In this first story, the Graphing AI team shares how they instrumented their widget- and dashboard-generation agents with LLM Observability to detect regressions and debug failures faster. Visibility into how large language model (LLM) applications behave in real time is essential for building reliable AI-driven systems at Datadog.

How we built an AI SRE agent that investigates like a team of engineers

We built Bits AI SRE to help engineers investigate and solve production incidents, one of the most difficult aspects of operating distributed systems today. As environments grow more dynamic and complex, resolving issues becomes more challenging. Failures now span more services, involve noisier signals, and encompass larger volumes of telemetry data, making it hard for on-call engineers to find root causes quickly. Today, Bits AI SRE is already helping teams decrease time to resolution by up to 95%.

Automate flaky test fixes with the Bits AI Dev Agent and Test Optimization

Flaky tests are a significant source of inefficiency that impacts many engineering teams. Along with failing your build, they interrupt your entire development flow, generate excessive CI/CD noise, and, critically, compromise developer trust in the test suite itself. Datadog Test Optimization enables you to manage test suites at scale by pinpointing the flakiest tests, analyzing their history across hundreds of runs, and automatically surfacing the root cause.

Datadog integrations 2025 recap: Observability for AI, security, and hybrid cloud

The year 2025 marked a major milestone in the Datadog integrations ecosystem as we surpassed 1,000 integrations. Along the way, we also added over 110 new technology partners and expanded coverage across the fastest growing software categories, including AI, distributed security, hybrid infrastructure, and data intelligence. This recap highlights the most impactful integrations we released this year and how they connect to these broader technology trends.

Bring faster visibility into AWS Lambda functions with remote instrumentation

Comprehensive observability is critical for running performant, reliable, and secure serverless workloads. However, configuring and maintaining that visibility across hundreds or thousands of serverless functions can be difficult to scale and sustain. Developers across teams often manage serverless functions using different infrastructure as code (IaC) frameworks, as well as different review, deployment, and update processes.

Build custom apps in seconds with conversational AI in App Builder

Using a drag-and-drop interface, engineering teams can create apps that support troubleshooting, improve day-to-day operations, and offer self-service access without leaving Datadog. With the new conversational AI feature, teams can turn an idea into a working app in seconds. Watch the video to see how it works..

Implement dbt data quality checks with dbt-expectations

dbt is one of the most popular solutions for data transformations and modeling. Many commercial data pipelines rely on dozens, or even hundreds, of individual dbt jobs. Data engineers, data platform engineers, and analytics engineers who own these pipelines need to maintain a testing framework to prevent mistakes in data processing that can compromise analysis.

Troubleshoot faster with the GitLab Source Code integration in Datadog

Developers and SREs who rely on GitLab to develop their services often face significant friction when troubleshooting errors or fixing issues that degrade code quality. To understand the context of a problem, they resort to tab-hopping between observability tools and GitLab, connecting stack traces, spans, and profiles back to the right files and commits.

Check out features we announced at AWS re:Invent in the latest episode of This Month in Datadog

Tune in for spotlights of Bits AI SRE, now generally available, and Datadog’s MCP Server, which connects AI agents to our platform by ingesting prompts and mapping them to Datadog resources and data. Plus, we cover how to: Search logs at petabyte scale in your own infrastructure with CloudPrem Break down costs drivers at the prefix level with Storage Management Create workflows that adapt to real-world complexity with Agent Builder Detect and block credential leaks with Secret Scanning.