The latest News and Information on Observabilty for complex systems and related technologies.
A Harrowing Landscape The increasing complexity of modern services is forcing IT Ops teams to employ a growing landscape of disparate tools to monitor the health of their IT Stack. In fact, the number of tools has grown so much in the last few years, that one wonders how IT Ops teams are even able to effectively configure, maintain, ingest, and process all the events that these tools create.
This post is the first in a three-part series on how to effectively monitor the hosts and systems in your ecosystem, and we're starting with the one you use most: your personal computer. Metrics are a key part of observability, providing insight into the usage of your systems, allowing you to optimize for efficiency and plan for growth. Let's take a look at the different metrics you should be monitoring.
Moving to a scalable, distributed microservice architecture poses a great deal of challenges for any organization. It gets harder to understand the system and pinpoint where errors originate. Logs get much messier, and stitching together a coherent picture of a particular request can be time-consuming or downright impossible. Distributed tracing can help with all of that.
I’ve tackled this question before: how much should my observability stack cost? While the things in that post are true now as ever, I did end on one somewhat vague conclusion. When it came to figuring out exactly what you need in your stack by drawing a straight line from the business case to the money you spend, my conclusion was that “it depends.” That’s how we approached pricing at Honeycomb: it depends on your needs, so we should give you many different options.
Logging is an essential method to understanding what’s happening in your environment. Logs help developers and system administrators understand where and when things have gone wrong. Ideally, logs on their own would suffice as indicators of what’s happening. However, there’s far too many log messages being produced in today’s world and most don’t contain the information we actually need.
Considering the scale of log data that modern cloud environments generate, it’s oftentimes prohibitively expensive to index all of it. For monitoring and logging, cost management is just as important as in other parts of the business. Whether sudden spikes of log data overwhelm databases or good business generates more activity in your environment, teams should anticipate and mitigate the steep costs that result from high log volumes.