Operations | Monitoring | ITSM | DevOps | Cloud

Latest News

Grafana Loki: Open Source Log Aggregation Inspired by Prometheus

Logging solutions are a must-have for any company with software systems. They are necessary to monitor your software solution’s health, prevent issues before they happen, and troubleshoot existing problems. The market has many solutions which all focus on different aspects of the logging problem. These solutions include both open source and proprietary software and tools built into cloud provider platforms, and give a variety of different features to meet your specific needs.

Log Management & Managed Open Distro ELK Platform, Logit.io Launch New Teams & Users UI

Log management and managed Open Distro ELK provider Logit.io announced today that they've launched an entirely new redesign of their teams and users pages to improve the user experience for users that wish to add additional members to teams and create new teams easier.

How Cloudflare Logs Provide Traffic, Performance, and Security Insights with Coralogix

Cloudflare secures and ensures the reliability of your external-facing resources such as websites, APIs, and applications. It protects your internal resources such as behind-the-firewall applications, teams, and devices. This post will show you how Coralogix can provide analytics and insights for your Cloudflare log data – including traffic, performance, and security insights.

Is "Vendor-Owned" Open Source an Oxymoron?

Open source is eating the world. Companies have realized and embraced that, and ever more companies today are built around a successful open source project. But there’s also a disturbing counter-movement: vendors relicensing popular open source projects to restrict usage. Last week it was Grafana Labs which announced relicensing Grafana, Loki and Tempo, its popular open source monitoring tools, from Apache2.0 to the more restrictive GNU AGPLv3 license.

Security Log Management Done Right: Collect the Right Data

Nearly all security experts agree that event log data gives you visibility into and documentation over threats facing your environment. Even knowing this, many security professionals don’t have the time to collect, manage, and correlate log data because they don’t have the right solution. The key to security log management is to collect the correct data so your security team can get better alerts to detect, investigate, and respond to threats faster.

Announcing the LogDNA and Sysdig Alert Integration

LogDNA Alerts are an important vehicle for relaying critical real-time pieces of log data within developer and SRE workflows. From Slack to PagerDuty, these Alert integrations help users understand if something unexpected is happening or simply if their logs need attention. This allows for shorter MTTD (mean time to detection) and improved productivity.

How to do network traffic analysis with VPC Flow Logs on Google Cloud

Network traffic analysis is one of the core ways an organization can understand how workloads are performing, optimize network behavior and costs, and conduct troubleshooting—a must when running mission-critical applications in production. VPC Flow Logs is one such enterprise-grade network traffic analysis tool, providing information about TCP and UDP traffic flow to and from VM instances on Google Cloud, including the instances used as Google Kubernetes Engine (GKE) nodes.

Cyber Defense Magazine Names ChaosSearch "Cutting Edge" in Cybersecurity Analytics

Exciting news — ChaosSearch won the 2021 InfoSec “Cutting Edge in Cybersecurity Analytics” award from Cyber Defense Magazine! We’re honored to be recognized for our innovation in delivering security insights at scale. The InfoSec panel of judges is made up of certified security pros who understand what SecOps teams care about and how log analytics should be applied to keep data secure.

Monitoring Model Drift in ITSI

I’m sure many of you will have tried out the predictive features in ITSI, and you may even have a model or two running in production to predict potential outages before they occur. While we present a lot of useful metrics about the models’ performance at the time of training, how can you make sure that it is still generating accurate predictions? Inaccuracy in models as the underlying data or systems change over time is natural.