Operations | Monitoring | ITSM | DevOps | Cloud

SigNoz

The Ultimate Guide to API Monitoring in 2024 - Metrics, Tools, and Proven Practices

According to Akamai, 83% of web traffic is through APIs. Microservices, servers, and clients constantly communicate to exchange information. Even the Google search you made to reach this article involved your browser client calling Google APIs. Given APIs govern the internet, businesses rely on them heavily. API health is directly proportional to business prosperity. This article covers everything about API monitoring, so your API infrastructure’s health is always in check ✅.

Grafana vs Splunk - Key Features and Differences

Grafana and Splunk are both used as monitoring tools. But while Grafana is majorly used as a data visualization tool, Splunk is an enterprise security and observability platform. Monitoring tools are essential for any business that wants to have visibility into its IT infrastructure. They provide real-time data that can be used to identify and troubleshoot problems. Grafana and Splunk are two of the most popular monitoring tools on the market. So, which one is better for your business?

Launch Week, Upgrades to Metrics & Query Builder & Access Token Management - SigNal 34

Welcome to the 34th edition of our monthly product newsletter - SigNal 34! Last month was full of action. We did our first launch week, and we were thrilled to see the response. We have shipped some amazing features recently. Let’s see what humans of SigNoz were up to in the month of February 2024.

SigNoz Launch Week - Day 5 - Access Token Mgmt & Onboarding

Welcome to last day of SigNoz Launch Week! We will start with a chat with Ankit (CTO, SigNoz) about the evolution of SigNoz, our focus on OpenTelemetry and belief in open-source. Post that we will deep dive into recently released Access Token management feature which will help users programmatically access data stored in SigNoz and unlock powerful use cases. Then we will showcase our new SaaS onboarding flow which makes getting started with SigNoz much easier. We will also discuss the process we follow for improving our documentation & onboarding flow.

SigNoz Launch Week - Day 4 - Logs Pipeline

For day 4, we will showcase the recent work we have done in Logs Pipeline. With Log Pipelines, you can transform logs to suit your querying and aggregation needs before they get stored in the database. Pipelines provide a way to modify the structure and content of log data without needing to change application code or redeploy components. By extracting relevant attributes from logs, pipelines enable more efficient analysis.

SigNoz Launch Week - Day 3 - Frontend Monitoring

Welcome to SigNoz Launch week! In day 3 we will focus on monitoring frontend with SigNoz. We will take examples on how to monitor web vitals of your application and monitoring errors in client applications. This will be followed by a discussion with our maintainers on nuances of building performant frontend application for data dense products like SigNoz. Do tune in!

SigNoz Launch Week - Day 1 - Logs Explorer

Welcome to SigNoz Launch Week 1.0! This is our first launch week, and we’re excited to introduce you to some cool new features in SigNoz. We ship fast but often miss sharing the story behind these features with our community. Launch week for us is an opportunity to share the behind-the-scenes of new features that we have built in the recent past. Our open-source maintainers will share the story on the whats, whys, and hows of new upgrades to SigNoz!

Monitoring Kafka with OpenTelemetry including client side monitoring

In this video, you will see a demo of how to monitor Kafka with OpenTelemetry. We will instrument a NodeJS application using Kafka and get client side metrics like delay between producer emitting a message to consumer receiving it via distributed tracing. We will also get Kafka server metrics like consumer lag and plot it dashboards.

Latest Top 11 Log Monitoring Tools [Includes Open-Source]

For any software company, a log monitoring tool is a must for collecting, storing, and providing a centralized view of all logs from different applications and hosts for faster anomaly detection, incident resolution, and troubleshooting. They can also help detect security threats and provide audit trails. They are effective in capacity planning, decision-making, and ensuring optimized performance.