Operations | Monitoring | ITSM | DevOps | Cloud

The latest News and Information on Monitoring for Websites, Applications, APIs, Infrastructure, and other technologies.

Solr Monitoring Made Easy with Sematext

As shown in Part 1 Solr Key Metrics to Monitor, the setup, tuning, and operations of Solr require deep insights into the performance metrics such as request rate and latency, JVM memory utilization, garbage collector work time and count and many more. Sematext provides an excellent alternative to other Solr monitoring tools.

Solr Open Source Monitoring Tools

Open source software adoption continues to grow. Tools like Kafka and Solr are widely used in small startups, ones that are using cloud ready tools from the start, but also in large enterprises, where legacy software is getting faster by incorporating new tools. In this second part of our Solr monitoring series (see the first part discussing Solr metrics to monitor), we will explore some of the open source tools available to monitor Solr nodes and clusters.

My acquihire adventure with a large tech company

During the spring and summer of 2018 I was negotiating a possible acquihire of Checkly / moi with Datadog. I'm only writing about it right now because Datadog recently launched their Synthetics product — the product I would be involved with as some form of product manager. This post details how this all came to be, what steps we took in the process and some reflections on the whole thing.

Handling Multiple WordPress Websites Made Easier With ManageWP

Nowadays, everyone uses WordPress on a regular basis to write articles, make notes and for just about everything. If you are working on multiple WordPress-related websites, a time will definitely come when you will find it difficult to manage everything at the same time. This is where the role of ManageWP or Manage WordPress becomes important.

Distributed Machine Learning With PySpark

Spark is known as a fast general-purpose cluster-computing framework for processing big data. In this post, we’re going to cover how Spark works under the hood and the things you need to know to be able to effectively perform distributing machine learning using PySpark. The post assumes basic familiarity with Python and the concepts of machine learning like regression, gradient descent, etc.