Operations | Monitoring | ITSM | DevOps | Cloud

Analytics

Server Monitoring with Logz.io and the ELK Stack

In a previous article, we explained the importance of monitoring the performance of your servers. Keeping tabs on metrics such as CPU, memory, disk usage, uptime, network traffic and swap usage will help you gauge the general health of your environment as well as provide the context you need to troubleshoot and solve production issues.

How to Send Akamai Logs to LogDNA

Akamai provides the Content Delivery Network (CDN) which is a highly-distributed platform of servers optimized to deliver contents including web and media applications. These networks enable applications to easily serve content from closer to their end users. Centralizing Akamai logs increases the ability to observe the end to end application and service delivery. LogDNA is proud to enable integration with Akamai to provide better observability and a unified view for our customers.

Search Relevance - Solr & Elasticsearch Similarities

Lucene has a lot of options for configuring similarity. By extension, Solr and Elasticsearch have the same options. Similarity makes the base of your relevancy score: how similar is this document (actually, this field in this document) to the query? I’m saying the base of the score because, on top of this score, you can apply per-field boosts, function scoring (e.g. boost more recent documents) and re-ranking (e.g. Learning to Rank).

4 Business Disasters That Could Have Been Avoided With Real-Time Anomaly Detection

Digital, network-connected systems are transforming every aspect of business — from your mission-critical workloads to your most rarely used applications. But the increases in scalability and cost efficiency come at a cost. Because every system is so reliant on network connectivity, unplanned downtime is becoming increasingly expensive.

Logging Levels 101

If you’re a software developer, then you understand how vital application logging is in software development and a critical part of logging is something called logging levels. Log entries generally contain essential information—such as a timestamp, a message, and sometimes additional stuff like an exception’s stack trace. Those pieces of information are useful because they allow someone reading the log entry to understand how the application behaved in production.

Kafka Logging with the ELK Stack

Kafka and the ELK Stack — usually these two are part of the same architectural solution, Kafka acting as a buffer in front of Logstash to ensure resiliency. This article explores a different combination — using the ELK Stack to collect and analyze Kafka logs. As explained in a previous post, Kafka plays a key role in our architecture. As such, we’ve constructed a monitoring system to ensure data is flowing through the pipelines as expected.