6 Linux server performance metrics to look out for
Linux is one of the best known and most used open source operating systems. The key advantages of using Linux servers are that they are stable, secure, compatible, and customizable.
Linux is one of the best known and most used open source operating systems. The key advantages of using Linux servers are that they are stable, secure, compatible, and customizable.
Following the release of Azure Monitor for VMs in January 2020, some changes have been made to how its performance data is stored in Log Analytics. The most significant of these changes is that this data is now stored in the InsightsMetrics table rather than in the Perf table, as we discussed in our ‘What is Azure VM Insights?’ article.
There are different tools and mechanisms for autoscaling in Kubernetes at both the application and infrastructure layers to help users manage their cluster resources. In this article, we’ll explore two infrastructure autoscaling tools for Kubernetes — Ocean by Spot and open source Cluster Autoscaler.
Version 2.1 of the HAProxy Data Plane API expands support to all available request and response actions, adds Lua actions, and improves file handling. A year ago, we introduced version 1.0 of the HAProxy Data Plane API, enabling you to configure your HAProxy load balancers remotely through a modern RESTful HTTP API. That first version of the API focused on the essential behaviors for creating frontend proxies, backend server pools, ACLs and traffic switching rules.
Talk about performance monitoring to any system admin or IT manager and one of the first questions they will ask is whether the monitoring is agent-based or agentless. The moment you hear that question, you know that they are interested in an agentless monitoring solution. Such is the fear of having agents on critical servers in the infrastructure! In this article, we will discuss.
Edge computing – with its ability to help unleash the potential of smart cities and autonomous vehicles, is something often thought to exist “around the corner” as opposed to an industry already in motion.
This is a review of the last three years that we spent stabilizing Marathon. Marathon is the central workload scheduler in DC/OS. Most of the time when you launch an app or a service on DC/OS, it is Marathon that starts it on top of Apache Mesos. Mesos manages the compute and storage resources and Marathon orchestrates the workload. We sometimes dub it the “init.d of DC/OS”. Being such an integral part of DC/OS, we must ensure that it keeps functioning.
This is part of a series of articles discussing strategies to implement serverless architectural design patterns. We continue to follow this literature review. Although we use AWS serverless services to illustrate concepts, they can be applied in different cloud providers. In the previous article (Part 1) we covered the Aggregator and Data Lake patterns. In today’s article, we’ll continue in the Orchestration & Aggregation category covering the Fan-in/Fan-out and Queue-based load leveling.