site that my customers need to visit. The business understands that we need to serve customers with at least a basic level of QoS: no page loads over 4s
are useful. We use them all over the place... they are cheap. But, ideally, you want to load the page closer to the way a user does (all assets, javascript, etc.) Enter phantomjs
Omniture, etc. Statsd (out-of-the-box) and Metrics are mediocre approach. If we have a lot of observable data N, N̅ isn’t so useful, , |N|, q(0.5), q(0.95), q(0.99), q(0), q(1), add a lot.
8 statistical aggregates Let’s look at API latencies... say we do 1000/s, that’s 60k/minute. Over a minute of time, 60k points to 8 represents... a lot of information loss.
up... Performance like that should only happen if our binary tree is badly imbalanced (replace with countless other pathologically bad precipitates of failure); I wonder if it is... It’s almost like some requests are super slow; I wonder if they are. STOP WONDERING.
stores data... happens to store all data visualized in Circonus. We have to get data into the system. We have to get data out of the system. I don’t wonder... here’s why.