All Posts

Performance Engineering: Series II

In this series, we are going to look at throughput optimization. Just to recap, throuhput is measured as the number of requests an app can handle in a second without compromising on the latency sla. Before we go deep into how to improve the performance let’s look at the flow in a typical web service. Most of the webservices, pull data from a data store, then processes it and serves it as part of the response. Serialization is involved in reading the request and serving the resopnse back. To understand better, let’s take a look at how a single request flows through.

Performance Engineering: Series I

Software is often delivered in different ways to different set of users. Some get it via the web, others access it through the cloud, and some get it on specific device like a smartphone or personal computer. All users have different expectations of how they use the software and it has impact on how they perceive performance of the software. From a product angle, a software meets the performance expectations of the user if it delivers the function/feature promised within the agreed time. From an engineering perspective, a software is seen as performant if it makes the most use of the resources (cpu/memory/nodes/network).