Latency metrics

Latency measures are crucial for checking how well your apps and services perform. Latency means the total time it takes for a piece of data to go from where it starts to where it ends up, usually on a network. When we talk about latency, we’re mostly talking about how fast things move in a network. It’s one of the main things we look at to see if a service is good or not. We usually measure it in milliseconds. The lower the latency, the better the user’s experience.

Read More

Caching

Introduction

Caches take advantage of the locality of reference principle: recently requested data is likely to be requested again. A cache is like short-term memory: it has a limited amount of space, but is typically faster than the original data source and contains the most recently accessed items. Caches can exist at all levels in architecture, but are often found at the level nearest to the front end where they are implemented to return data quickly without taxing downstream levels.

Read More