Operations | Monitoring | ITSM | DevOps | Cloud

The latest News and Information on Log Management, Log Analytics and related technologies.

CriblCon 2023 Keynote Session

On July 17th, 2023, more than 400 Cribl users came together at The Mirage in Las Vegas to celebrate each other and the power of learning at CriblCon. The theme of our conference, “Do Different,” resonated throughout the day, emphasizing our commitment to innovation and highlighting the distinctive approach our customers and employees bring to every aspect of their work.

How to use Splunk Universal Forwarders With BindPlane OP

A tutorial on how to start collecting data from your splunk universal forwarders using BindPlane as an aggregator, giving you the ability to start sending telemetry data to multiple destinations. About ObservIQ: observIQ brings clarity and control to our customer's existing observability chaos. How? Through an observability pipeline: a fast, powerful and intuitive orchestration engine built for the modern observability team. Our product is designed to help teams significantly reduce cost, simplify collection, and standardize their observability data.

What Is Adaptive Thresholding?

Adaptive thresholding is a term used in computer science and — more specifically — across IT Service Intelligence (ITSI), for analyzing historical data to determine key performance indicators (KPIs) in your IT environment. Among other things, it’s used to govern KPI outliers in an effort to foster more meaningful and trusted performance monitoring alerts.

Your First 100 Days With Cribl: Why Having an Onboarding Process Matters

The process of adding new data to operations and security analytics tools is familiar to admins. New data onboarding can be a tiresome process that takes up too much time and delays getting value from the new data. The process typically begins with the admin engaging the data source owner, getting the wrong data sample, and then having to try again.

How to Remove Fields with Empty Values From Your Logs

Much of the log data we handle doesn’t offer substantial insight and can be conveniently removed from your logs, helping us reduce costs. What may seem like a small adjustment, like deleting an attribute, can have significant implications when scaled up. A typical case involves fields in your logs presenting empty values or housing data considered irrelevant. Below we’ll take a look at a few examples of what this looks like and how you can take action in BindPlane OP.

Continuous Observability: Shedding Light on CI/CD Pipelines

DevOps is not just about operating software in production, but also releasing that software to production. Well-functioning continuous integration/continuous delivery (CI/CD) pipelines are critical for the business, and this calls for quality observability to ensure that Lead Time for Changes is kept short and that broken and flaky pipelines are quickly identified and remediated.

Democratizing Data Through Secure Self-Service Concierge Access of Cribl Stream

Ah, the age-old question of how to manage screen time for kids – it’s like trying to navigate a minefield of Peppa Pig, Paw Patrol, and PJ Masks! I mean, who knew Octonauts and Bubble Guppies would become household names? As a dad of two young kids, managing screen time is a balancing act, especially keeping our 5-year-old happy with access to her shows.

Cribl Stream Projects

The increasing demand for Cribl Stream as an internal service is a testament to its effectiveness in improving operations and enhancing security measures. With the rise of ITOps, SecOps, SRE, DevOps, and other teams embracing Cribl Stream, we are excited to offer Cribl Stream Projects, which enables the secure expansion of Stream usage to more users within organizations. This enhances collaboration and provides deeper insights, resulting in a more personalized user experience. With Stream Projects, Cribl is the first product in the industry enabling organizations to allow teams to manage their own data without needing to understand the infrastructure or service being used to collect and route it.

Getting Started with GROK Patterns

If you’re new to logging, you might be tempted to collect all the data you possibly can. More information means more insights; at least, those NBC “the more you know” public services announcements told you it would help. Unfortunately, you can create new problems if you do too much logging. To streamline your log collection, you can apply some filtering of messages directly from the log source. However, to parse the data, you may need to use a Grok pattern.