Operations | Monitoring | ITSM | DevOps | Cloud

November 2024

Data Pipelining with InfluxDB

In this blog post, we’ll explore how to build a data pipeline using Kafka, Faust, and InfluxDB to effectively ingest, transform, and store data. We’ll start with an overview of Kafka, a high-performance messaging platform, and Faust, a Python library designed for stream processing, now maintained by the community as Faust-streaming.

Optimizing Queries in InfluxDB 3.0 Using Progressive Evaluation

In a previous post, we described the technique that makes the ”most recent values” queries hundreds of times faster and has benefited many of our customers. The idea behind this technique is to progressively evaluate time-organized files until we reach the most recent values.

Part Two: InfluxDB 3.0 Under the Hood

In the first blog in this series, Setting Up InfluxDB and Visualizing Data: Part 1, we built a data collection and visualization platform for time series data using InfluxDB Cloud Serverless. Inspired by the CSTR with PID controllers use case, the project showcased how to ingest real-time data and visualize it using InfluxDB and Grafana. This follow-up post focuses on InfluxDB’s 3.0 architecture, giving an in-depth look at the platform’s inner workings.

Product Update: Introducing User Groups for InfluxDB Cloud Dedicated

We are excited to announce the launch of User Groups, a major update that facilitates enhanced security through access control in InfluxDB Cloud Dedicated. This new feature allows for more granular access management by limiting limited access accounts. Giving customers more access control helps them implement PoLP (“Principle of Least Privilege”) for improved security.

LLM Monitoring and Observability

The demand for LLM is rapidly increasing—it’s estimated that there will be 750 million apps using LLMs by 2025. As a result, the need for LLM observability and monitoring tools is also rising. In this blog, we’ll dive into what LLM monitoring and observability are, why they’re both crucial and how we can track various metrics to ensure our model isn’t just working but thriving.