Operations | Monitoring | ITSM | DevOps | Cloud

How Diffusion Transformer Models Power Hyper-Realistic AI Avatar Videos

The AI avatar videos from a year ago still had a tell. The mouth movement was a little off, the facial expressions were a bit stiff. It was a quality that made it obvious that you were looking at a digital human and not a real one. The uncanny valley issue was not a small aesthetic problem, it was the only thing that stopped the practical adoption of anything other than novelty use cases.

Run Local LLMs on Mac to Cut Claude Costs

Part of the motivation for this post is how cloud API economics are shifting: Anthropic is moving large enterprise customers toward per-token, usage-based billing (unbundled from flat seat fees), which makes “always call the API” a moving cost line for teams at scale. A hybrid or local layer is one way to keep spend bounded while you still use premium models where they matter.

Rootly's Dan Sadler: why AI coding tools are driving more incidents + why reliability is the product

Cortex co-founder and CTO Ganesh Datta sits down with Dan Sadler, VP of Engineering at Rootly. Dan explains how Rootly treats reliability as a product feature rather than just a technical metric, and why culture might be the most impactful element of building reliable systems.

When agents orchestrate agents, who's watching?

You used to monitor services. Then you started monitoring AI calls inside services. Now your AI agent is spinning up other AI agents to complete tasks. Your old monitoring instincts need to evolve. This isn't hypothetical. Agentic architectures are already in production. Coding agents are calling search agents; orchestrators are spawning specialized sub-agents for retrieval, planning, and execution. Teams are shipping these systems faster than they're figuring out how to watch them.

What does using AI for post-mortems actually mean?

Everyone is using AI to help with post-mortems now. The pitch is obvious: post-mortems are time-consuming, the blank page is brutal, and AI is very good at producing structured, confident-sounding documents quickly. We're not here to push back on that. We've built AI into our own post-mortem experience, pulling your Slack thread, timeline, PRs, and custom fields together and giving your team a meaningful starting point in seconds. We think that's genuinely valuable, and the teams using it agree.

How it feels to run an incident with AI SRE

We've been building the broader incident.io platform for several years now, and one thing we've learned is that UX matters more here than almost anywhere else. When an incident fires, there's no room for poorly designed interfaces or fumbling through features you haven't touched in a while. The product has to be ergonomic: easy to pick up, easy to navigate, with the right things at your fingertips at exactly the right moment. We've put a lot of effort into this over the last 5 years.

AI for Incident Response: Should You Build or Buy?

SREs and platform teams are overwhelmed by the effort of manually troubleshooting ever-more complex cloud-native environments. This pain is driving a breakneck adoption of AI SRE solutions that promise to automate core reliability practices, from root cause analysis to capacity planning. For teams with strong engineering talent, creating a DIY AI SRE seems like a straightforward challenge.