Operations | Monitoring | ITSM | DevOps | Cloud

Latest News

The Great Irony of Serverless Computing

Working with Serverless computing is like riding an electric bike. You get speed, flexibility, automatic assistance to scale with ease. Development is usually hassle-free because you can focus on code and only pay for usage of the service. Except when your users hit an error. Debugging that issue feels like your bike’s battery just died while climbing a steep hill.

How to Debug Slow Lambda Response Times

When you build your application on top of Lambda, AWS automatically scales the number of “workers” (think containers) running your code based on traffic. And by default, your functions are deployed to three Availability Zones (AZs). This gives you a lot of scalability and redundancy out of the box. When it comes to API functions, every user request is processed by a separate worker. So the API-level concurrency is now handled by the platform.

Serverless vs. Containers: Key Differences Explained

The “as a service” business model continues to grow rapidly, largely thanks to the rise of cloud computing. “As a service” offerings deliver IT products and technologies such as software, hardware, and data storage to consumers via the Internet, rather than having to install or manage them themselves. Serverless and containers are two such “as a service” technologies that have seen increasing adoption in recent years.

The Ultimate Guide to Monitoring Serverless Applications

Serverless applications, more often than not, have logic distributed over multiple functions and services, which with growth and agents and wrappers attached, can get more complex and costly. This is where Serverless monitoring comes in to help. But what is Serverless monitoring? Serverless monitoring allows developers to gain important insight on what happens during each execution and event, errors become more easily visible and measuring resource consumption for each invocation is possible.

10 amazing benefits of serverless technology

If you’ve heard of the term “serverless architecture” but you have no idea what it means, don’t feel bad. This buzzword (buzz-phrase?) has been on the rise for the past few years, and it doesn’t look like it’s going to stop. More and more cloud companies are trying to promote serverless architectures for businesses. There are plenty of benefits to using a serverless architecture for your business.

A day in the life of a Software Engineer: A developer's perspective on working with serverless

Want to know what actually goes on under the platform and behind the screens at Dashbird? We recently sat down for a Q&A with our CTO, Marek Tihkan, on leading and managing an engineering team. Today, we speak to Alex, one of the engineers on the Development team and the brain and elbow grease behind our newly launched Dashbird Atlas real-time 3D map of your entire serverless environment.

What alerts should you have for serverless applications?

A key metric for measuring how well you handle system outages is the Mean Time To Recovery or MTTR. It’s basically the time it takes you to restore the system to working conditions. The shorter the MTTR, the faster problems are resolved and the less impact your users would experience and hopefully the more likely they will continue to use your product! And the first step to resolve any problem is to know that you have a problem.

A techy bedtime story: The tale of wizards, caves and how Serverless came to be

We’ve noticed interesting threads floating around the internet asking how would you explain serverless to a toddler. As it happens, we just turned three years old and to mark the occasion, we decided to take up the challenge and wrote a bedtime story, for serverless enthusiasts of all ages, about how serverless came to be, from the very beginning. It is a magical tale of the ingenious, life-changing journey that got us to a new universe called Serverless.