Last week I had an interesting conversation on how to set goals for http request latency:
1) Measure network latency client side (browser or app) rather than at the edge.
2) Set a P75 goal rather than P95 or P99
e.g. P75 <= 250 ms
3) Set a P99/P75 goal
e.g. P99/P75 <= 6
Let's look into why:
1) Measure network latency client side (browser or app) rather than at the edge.
2) Set a P75 goal rather than P95 or P99
e.g. P75 <= 250 ms
3) Set a P99/P75 goal
e.g. P99/P75 <= 6
Let's look into why:
Comments
But it is still server to server
There's nothing like rum 🤗
The real indicator of a regression is success rate dropping. Slow is bad - abandoned is worse
Hence the choice of a P75 goal.
So looking for a high P99/P75 ratio helps surface endpoints that scale poorly, e.g. endpoints that do N+1 queries or aren't paginated and return large lists.