Statistics for Engineers
— White Paper —

Statistics for Engineers

by Heinrich Hartmann

Statistics is the art of turning data into information. The statistical methods you may have learned about in school are valuable tools for a wide range of analysis. Clinical researchers use statistics to help make sense of data and to draw conclusions about their area of study. Geneticists use them to predict the likelihood of hereditarily-linked diseases. And of course, software engineers use statistics to improve the quality of service they deliver to external and internal customers by analyzing operations data.

Despite the wide use of statistics across many fields, there are stark differences in how these methods are put to use. Many statistical methods assume the normal distribution of data — this assumption doesn’t necessarily hold true for operations data.

Circonus Chief Data Scientist, Heinrich Hartmann, offers best practices for software engineers trying to analyze operations data.

In this paper, you’ll learn:

  • Preferred ways to visualize data with comparisons of rug plots, histograms, and scatter and line plots
  • Why high dynamic range histograms are a good choice for large-volume operations data with a wide range of values
  • How heat maps show the change of data distributions over time
  • When to use standard deviations, quantiles, and outliers
  • How to formulate and measure Service Level Agreements



Modern IT systems collect an increasing wealth of data from network gear, operating systems, applications, and other components. This data needs to be analyzed to derive vital information about the user experience and the business performance. For instance, faults need to be detected, service quality needs to be measured and resource usage of the next days and month needs to be forecast.

Rule #1: Spend more time working on code that analyzes the meaning of metrics, than code that collects, moves, stores and displays metrics.

— Adrian Cockcroft1

Statistics is the art of extracting information from data, and hence becomes an essential tool for operating modern IT systems. Despite a rising awareness of this fact within the community (see the quote above), resources for learning the relevant statistical methods for this domain are hard to find.

The statistics courses offered in universities usually depend on their students having prior knowledge of probability, measure, and set theory, which is a high barrier of entry. Even worse, these courses often focus on parametric methods, such as t-tests, that are inadequate for this kind of analysis since they rely on strong assumptions on the distribution of data (i.e. normality) that are not met by operations data.

This lack of relevance of classical, parametric statistics can be explained by history. The origins of statistics reach back to the 17th century, when computation was expensive and data was a sparse resource, leading mathematicians to spend a lot of effort to avoid calculations.

Today the stage has changed radically and allows different approaches to statistical problems. Consider this example from a textbook2 used in a university statistics class:

A fruit merchant gets a delivery of 10,000 oranges. He wants to know how many of those are rotten. To find out he takes a sample of 50 oranges and counts the number of rotten ones. Which deductions can he make about the total number of rotten oranges?

The chapter goes on to explain various inference methods. The example translated to the IT domain could go as follows:

A DB admin wants to know how many requests took longer than one second to complete. He measures the duration of all requests and counts the number of those that took longer than one second. Done.

The abundance of computing resources has completely eliminated the need for elaborate estimations.

Therefore, this article takes a different approach to statistics. Instead of presenting textbook material on inference statistics, we will walk through four sections with descriptive statistical methods that are accessible and relevant to the case in point. We will discuss several visualization methods (section 1), gain a precise understanding of how to summarize data with histograms (section 2), visit classical summary statistics (section 3) and see how to replace them with robust, quantile-based alternatives (section 4). I have tried to keep prerequisite mathematical knowledge to a minimum (e.g., by providing source code examples along with the formulas wherever feasible. (Disclaimer: The source code is deliberately inefficient and serves only as an illustration of the mathematical calculation. Use it at your own risk!)

Visualizing Data

Visualization is the most essential data-analysis method. The human brain can process geometric information much more rapidly than numbers or language. When presented with a suitable visualization, one can almost instantly capture relevant properties such as typical values and outliers.

This section runs through the basic visualization plotting methods and discusses their properties. The Python tool chain (IPython8, matplotlib12, and Seaborn17) is used here to produce the plots. The section does not demonstrate how to use these tools. Many alternative plotting tools (R, MATLAB) with accompanying tutorials are available online. Source code and datasets can be found on GitHub.4

Rug plots

The most basic visualization method for a one-dimensional dataset X = [x1, …, xn] is the rug plot (figure 1). It consists of a single axis on which little lines, called rugs, are drawn for each sample.

A Rug Plot of web-request rates
Figure 1: A Rug Plot of web-request rates

Rug plots are suitable for all questions where the temporal ordering of the samples is not relevant, such as common values or outliers. Problems occur if there are multiple samples with the same sample value in the dataset. Those samples will be indistinguishable in the rug plot. This problem can be addressed by adding a small random displacement (jitter) to the samples.

Despite its simple and honest character, the rug plot is not commonly used. Histograms or line plots are used instead, even if a rug plot would be more suitable.


The histogram is a popular visualization method for one-dimensional data. Instead of drawing rugs on an axis, the axis is divided into bins, and bars of a certain height are drawn on top of them, so that the number of samples within a bin is proportional to the area of the bar (figure 2).

Figure 2: Histogram

The use of a second dimension often makes a histogram easier to comprehend than a rug plot. In particular, questions such as “Which ratio of the samples lies below y?” can be effectively estimated by comparing areas. This convenience comes at the expense of an extra dimension used and additional choices that have to be made about value ranges and bin sizes.

Histograms are addressed in more detail in the next section.

Scatter plots

The scatter plot is the most basic visualization of a two-dimensional dataset. For each pair of values x,y a point is drawn on a canvas that has coordinates (x,y) in a Cartesian coordinate system.

Scatter plots of request rates of two database nodes
Figure 3: Scatter plots of request rates of two database nodes

The scatter plot is a great tool to compare two metrics. Figure 3 plots the request rates of two different database nodes in a scatter plot. In the plot shown on top the points are mainly concentrated on a diagonal line, which means that if one node serves many requests, then the other is doing so as well. In the bottom plot the points are scattered all over the canvas, which represents a highly irregular load distribution, and might indicate a problem with the db configuration.

In addition to the fault-detection scenario outlined above, scatter plots are also an indispensable tool for capacity planning and scalability analysis.3,15

Line plots

The line plot is by far the most popular visualization method seen in practice. It is a special case of a scatter plot, where time stamps are plotted on the x-axis. In addition, a line is drawn between consecutive points. Figure 4 shows an example of a line plot.

A line plot of web-request rates
Figure 4: A line plot of web-request rates

The addition of the line provides the impression of a continuous transition between the individual samples. This assumption should always be challenged and taken with caution (e.g., just because the CPU was idle at 1:00 pm and 1:01 pm, this does not mean it did not do any work in between).

Sometimes the actual data points are omitted from the visualization altogether and only the line is shown. This is a bad practice and should be avoided.

The line plot is a great tool to surface time-dependent patterns such as periods or trends. For time- independent questions—typical values, for example—other methods such as rug plots might be better suited.

Which one to use?

Choosing a suitable visualization method depends on the question to be answered. Is time dependence important? Then a line plot is likely a good choice. If not, then rug plots or histograms are likely better tools. Do you want to compare different metrics with each other? Then consider using a scatter plot.

Producing these plots should become routine. Your monitoring tool might be able to provide you with some of these methods already. To get the others, figure out how to export the relevant data and import it into the software tool of your choice (Python, R, or Excel). Play around with these visualizations and see how your machine data looks.

To discover more visualization methods, check out the Seaborn gallery.18


Histograms in IT operations have two different roles: visualization method and aggregation method.

To gain a complete understanding of histograms, let’s start by building one for the web request-rate data already presented in the section on visualization. The listing in figure 5 contains a complete implementation, discussed step by step here.

  1. The first step in building a histogram is to choose a range of values that should be covered. To make this choice you need some prior knowledge about the dataset. Minimum and maximum values are popular choices in practice. In this example the value range is [500, 2200].
  2. Next the value range is partitioned into bins. Bins are often of equal size, but there is no need to follow this convention. The bin partition is represented here by a sequence of bin boundaries (line 4).
  3. Count how many samples of the given dataset are contained in each bin (lines 6-13). A value that lies on the boundary between two bins will be assigned to the higher bin.
  4. Finally, produce a bar chart, where each bar is based on one bin, and the bar height is equal to the sample count divided by the bin width (lines 14-16). The division by bin width is an important normalization, since otherwise the bar area is not proportional to the sample count. Figure 5 shows the resulting histogram.

 1| from matplotlib import pyplot as plt
 2| import numpy as np
 3| X = np.genfromtxt("DataSets/RequestRates.csv", delimiter=",")[:,1]
 4| bins = [500, 700, 800, 900, 1000, 1500, 1800, 2000, 2200]
 5| bin_count = len(bins) - 1
 6| sample_counts = [0] * bin_count
 7| for x in X:
 8|  for i in range(bin_count):
 9|       if (bins[i] <= x) and (x < bins[i + 1]):
10|           sample_counts[i] += 1
11| bin_widths = [ float(bins[i] - bins[i-1]) for i in range(1,  bin_count) ]
12| bin_heights = [ count/width for count, width in zip(sample_counts, bin_widths) ]
13|[:bin_count-1], width=bin_widths, height=bin_heights);

Result of a manual histogram implementation
Figure 5: Result of a manual histogram implementation

Different choices in selecting the range and bin boundaries of a histogram can affect its appearance considerably. Figure 6 shows a histogram with 100 bins for the same data. Note that it closely resembles a rug plot. On the other extreme, choosing a single bin would result in a histogram with a single bar with a height equal to the sample density.

Histogram plot with value range (500, 2200) and 100 equally sized bins
Figure 6: Histogram plot with value range (500, 2200) and 100 equally sized bins

Software products make default choices for the value range and bin width. Typically the value range is taken to be the range of the data, and equally spaced bins are used. Several formulas exist for selecting the number of bins that yield “ideal” results under certain assumptions—in particular, n1/2 (Excel) and 3.5 σ/n1/3 (Scott’s rule7). In practice, these choices do not yield satisfying results when applied to operations data, such as request latencies, that contain many outliers.

For readers familiar with probability theory, note that the histogram is an estimator for the probability density function.9

Histogram as aggregation method

When measuring high-frequency data such as I/O latencies, which can arrive at rates of more than 1,000 samples per second, storing all individual samples is no longer feasible. If you are willing to forget about ordering and sacrifice some accuracy, you can save massive amounts of space by using histogram data structures.

The essential idea is, instead of storing the individual samples as a list, to use the vector of bin counts that occurs as an intermediate result in the histogram computation. The example in figure 5 arrived at the following values:

sample_count = [0, 10, 8, 4, 25, 23, 4, 2]

The precise memory representation used for storing histograms varies. The important point is that the sample count of each bin is available.

Histograms allow approximate computation of various summary statistics, such as mean values and quantiles, that will be covered in the following sections. The precision depends on the bin sizes.

Also, histograms can be aggregated easily. If request latencies are available for each node of a database cluster in histograms with the same bin choices, then you can derive the latency distribution of the whole cluster by adding the sample counts for each bin. The aggregated histogram can be used to calculate mean values and quantiles over the whole cluster. This is in contrast to the situation in which mean values or quantiles are computed for the nodes individually. It is not possible to derive, for example, the 99th-percentile of the whole cluster from the 99th-percentiles of the individual nodes.14

High dynamic range histograms

An HDR (high dynamic range) histogram provides a pragmatic choice for bin width that allows a memory-efficient representation suitable for capturing data on a very wide range that is common to machine-generated data such as I/O latencies. At the same time HDR histograms tend to produce acceptable visual representations in practice.

An HDR histogram changes the bin width dynamically over the value range. A typical HDR histogram has a bin size of 0.1 between 1 and 10, with bin boundaries: 1, 1.1, 1.2,..., 9.9, 10. Similarly, between 10 and 100 the bin size is 1, with boundaries 10, 11, 12, ..., 100. This pattern is repeated for all powers of 10, so that there are 90 bins between 10k and 10k+1. The typical range of k is -127...128; often an extra bin for the 0 value is added.

The general definition is a little bit more complex and lengthy, so it is not provided here, but interested readers can refer to for more details and a memory-efficient implementation.

From the previous description it should be apparent that HDR histograms span an extremely large range of values ([10-127, 10128]). Bin sizes are similar to float-number precisions: the larger the value, the less precision is available. In addition, bin boundaries are independent of the dataset. Hence, the aggregation technique described earlier applies to HDR histograms.

Histograms as heat maps

Observing the change of data distributions over time requires an additional dimension on the histogram plot. A convenient method of doing so is to represent the sample densities as a heat map instead of a bar chart. Figure 7 shows the request-rate data visualized in such a way. Light colors mean low sample density, dark colors signal high sample density.

Request rate histogram (50 bins) presented as heatmap
Figure 7: Request rate histogram (50 bins) presented as heatmap

Multiple histogram heat maps that were captured over time can be combined into a single two-dimensional heat map.

Figure 8 shows a particularly interesting example of such a visualization for a sequence of HDR histograms of web-request latencies. Note that the distribution of the data is bimodal, with one mode constant around ~5 ms and another more diffuse mode ascending from ~10 ms to ~50 ms. In this particular case the second mode was caused by a bug in a session handler, which kept adding new entries to a binary tree. This tree had to be traversed for each incoming request, causing extended delays. Even the logarithmic growth of the average traversal time can be spotted if you look carefully.

Request latency heat map over time in Circonus
Figure 8: Request latency heat map over time in Circonus

Classical Summary Statistics

The aim of summary statistics is to provide a summary of the essential features of a dataset. It is the numeric equivalent of an “elevator pitch” in a business context: just the essential information without all the details.

Good summary statistics should be able to answer questions such as “What are typical values?” or “How much variation is in the data?” A desirable property is robustness against outliers. A single faulty measurement should not change a rough description of the dataset.

This section looks at the classical summary statistics: mean values and standard deviations. In the next section we will meet their quantile-based counterparts, which are more robust against outliers.

Mean value

The mean value or average of a dataset X = [x1,..., xn] is defined as:

mean(x1,...,xn) = 1/n ni=1xi

or when expressed as Python code:

def mean(X): return sum(X) / len(X)

The mean value has the physical interpretation of the center of mass if weights of equal weight are placed on the points xi on a (mass-less) axis. When the values of xi are close together, the mean value is a good representation of a typical sample. In contrast, when the samples are concentrated at several centers, or outliers are present, the mean value can be far from each individual data point (figure 9).

Mean values are abundant in IT operations. One common application of mean values is data rollup. When multiple samples arrive during a sampling period of one minute, for example, the mean value is calculated as a “one-minute rollup” and stored instead of the original samples. Similarly, if data is available for every minute, but you are interested only in hour intervals, you can “roll up the data by the hour” by taking mean values.

Rug plot of a two-modal dataset
Figure 9: Rug plot of a two-modal dataset

Despite their abundance, mean values lead to a variety of problems when measuring performance of services. To quote Dogan Ugurlu from “Why? Because looking at your average response time is like measuring the average temperature of a hospital. What you really care about is a patient’s temperature, and in particular, the patients who need the most help.”16 In the next section we will meet median values and quantiles, which are better suited for this kind of performance analysis.

Spike erosion

Viewing metrics as line plots in a monitoring system often reveals a phenomenon called spike erosion.5 To reproduce this phenomenon, select a metric (e.g., ping latencies) that experiences spikes at discrete points in time and zoom in on one of those spikes and read the height of the spike at the y-axis. Now zoom out of the graph and read the height of the same spike again. Are they equal?

Figure 10 shows such a graph. The spike height has decreased from 0.8 to 0.35.

How is that possible? The result is an artifact of a rollup procedure that is commonly used when displaying graphs over long time ranges. The amount of data gathered over the period of one month (more than 40,000 minutes) is larger than the amount of pixels available for the plot. Therefore, the data has to be rolled up to larger time periods before it can be plotted. When the mean value is used for the rollups, the single spike is averaged with an increasing number of “normal” samples and hence decreases in height.

Ping latency spike on a view range of 6h vs. 48h
Figure 10: Ping latency spike on a view range of 6h vs. 48h

How to do better? The immediate way of addressing this problem is to choose an alternative rollup method, such as max values, but this sacrifices information about typical values. Another more elegant solution is to roll up values as histograms and display a two-dimensional heat map instead of a line plot for larger view ranges. Both methods are illustrated in the graph on the far right in figure 10.

Deviation measures

Once the mean value μ of a dataset has been established, the next natural step is to measure the deviation of the individual samples from the mean value. The following three deviation measures are often found in practice.

The maximal deviation is defined as

maxdev(x1,...,xn)= max{|xi −μ||i=1,...,n}

and gives an upper bound for the distance to the mean in the dataset.

The mean absolute deviation is defined as

mad(x1,...,xn) = 1/n ni=1|xi − μ|

and is the most direct mathematical translation of a typical deviation from the mean.

The standard deviation is defined as

stddev(x1,...,xn) = √( 1/n ni=1(xi − μ)2 )

While the intuition behind this definition is not obvious, this deviation measure is very popular for its nice mathematical properties (as being derived from a quadratic form). In fact, all three of these deviation measures fit into a continuous family of p-deviations,11 which features the standard deviation in a “central” position.

Figure 11 shows the mean value and all three deviation measures for a request latency dataset. You can immediately observe the following inequalities:

mad(x1, ... , xn) ≤ stddev(x1, ... , xn) ≤ maxdev(x1, ... , xn)

This relation can be shown to hold true in general. The presence of outliers affects all three deviation measures significantly. The maximal deviation is even larger than the mean value and thus represents possible deviations of the dataset into the negative value range, which is impossible for request latencies.

A request latency dataset
Figure 11: A request latency dataset

Caution with the standard deviation

Many of us remember the following rule of thumb from school:

  • 68 percent of all samples lie within one standard deviation of the mean.
  • 95 percent of all samples lie within two standard deviations of the mean.
  • 99.7 percent of all samples lie within three standard deviations of the mean.

These assertions rely on the crucial assumption that the data is normally distributed. For operations data this is almost never the case, and the rule fails quite drastically: in the previous example more than 0.97 percent lie within one standard deviation of the mean value.

The following war story can be found in P. K. Janert’s book10:

An SLA (service level agreement) for a database defined a latency outlier as a value outside of three standard deviations. The programmer who implemented the SLA check remembered the above rule naively and computed the latency of the slowest 0.3 percent of the queries instead.

This rule has little to do with the original definition in practice. In particular, this rule labels 0.3 percent of each dataset blindly as outliers. Moreover, it turned out that the reported value captured long- running batch jobs that were on the order of hours. Finally, the programmer hard-coded a seemingly reasonable threshold value of ~50 seconds, and that was reported as the “three standard deviations,” regardless of the actual input.

The actual SLA was never changed.

Quantiles And Outliers

The classical summary statistics introduced in the previous section are well suited for describing homogeneous distributions but are easily affected by outliers. Moreover, they do not contain much information about the tails of the distribution.

A quantile is a flexible tool that offers an alternative to the classical summary statistics which is less susceptible to outliers.

Before introducing quantiles, we need to recall the following concept. The (empirical) cumulative distribution function CDF(y) for dataset X, at a value y, is the ratio of samples that are lower than the value y:


Or expressed in Python code:

def CDF(X,y): 
lower_count = 0
for x in X: if x <= y: lower_count += 1
return float(lower_count) / len(X)

Figure 12 shows an example for a dataset of request rates. Note CDF(X,y) takes values between 0 and 1 and is monotonically increasing as a function of y.

The cumulative distribution function for a data set of request rates 1.0
Figure 12: The cumulative distribution function for a data set of request rates 1.0

Now we turn to the definition of quantiles. Fix a number q between 0 and 1 and a dataset X of size n.

Roughly speaking, a q-quantile is a number y that divides X into two sides, with a ratio of q samples lying below y and the remaining ratio of 1 - q samples lying above y.

More formally, a q-quantile for X is a value y so that:

  • at least q · n samples are less than or equal to y and
  • at least (1 - q) · n samples are greater than or equal to y

Familiar examples are the minimum, which is a 0-quantile; the maximum, which is a 1-quantile; and the median, which is a 0.5-quantile. Common names for special quantiles include percentiles for k/100-quantiles and quartiles for k/4-quantiles.

Note that quantiles are not unique. There are ways of making them unique, but those involve a choice that is not obvious. Wikipedia lists nine different choices that are found in common software products.13 Therefore, if people talk about the q-quantile or the median, one should always be careful and question which choice was made.

As a simple example of how quantiles are non-unique, take a dataset with two values X = [10,20]. Which values are medians, 0-quantiles, 0.25-quantiles? Try to figure it out yourself.

The good news is that q-quantiles always exist and are easy to compute. Indeed, let S be a sorted copy of the dataset X such that the smallest element X is equal to S[0] and the largest element of X is equal to S[n - 1]. If d = floor(q · (n – 1)), then S[d] will have d+1 samples S[0],...,S[d], which are less than or equal to S[d], and nd + 1 samples S[d],..., S[n], which are greater than or equal to S[d]. It follows that S[d] = y is a q-quantile. The same argument holds true for d = ceil(q ·(n–1)).

The following listing is a Python implementation of the above construction:

def quantile_range(q,X):
	S = sorted(X)
	n = len(X)
	return (
		S[int(math.floor(q * (n - 1)))],
		S[int(math.ceil(q * (n - 1)))] 

It is not hard to see that the above construction consists of the minimal and maximal possible q-quantiles. The notation Qmin (X, q) is the minimal q-quantile. The minimal quantile has the property Qmin (X, q) y if and only if at least n · q samples of X are less than or equal to y. A similar statement holds true for the maximal quantile when checking ratios of samples that are greater than y.

Quantiles are closely related to the cumulative distribution functions discussed in the previous section. Those concepts are inverse to each other in the following sense: If CDF(X, y) = q, then y is a q-quantile for X. Because of this property, cumulative distribution function values are also referred to as inverse quantiles.

Applications to service level monitoring

Quantiles and CDFs provide a powerful method to measure service levels. To see how this works, consider the following SLA that is still commonly seen in practice: “The mean response time of the service shall not exceed three milliseconds when measured each minute over the course of one hour.”

This SLA does not do a good job of capturing the service experience of consumers. First, the requirement can be violated by a single request that takes more than 90ms to complete. Also, a long period where low overall load causes the measured request to finish within 0.1 ms can compensate for a short period where lots of external requests are serviced with unacceptable response times of 100 ms or more.

An SLA that captures the quality of service as experienced by the customers looks like this: 80 percent of all requests served by the API within one hour should complete within 3ms.

Not only is this SLA easier to formulate, but it also avoids the above problems. A single long-running request does not violate the SLA, and a busy period with long response times will violate the SLA if more than 20 percent of all queries are affected.

To check the SLA, here are two equivalent formulations in terms of quantiles and CFDs:

  • The minimal 0.8-quantile is at most 3 ms: Qmin (X1h, 0.8) 3ms.
  • The 3-ms inverse quantile is larger than 0.8: CDF(X1h, 3ms) 0.8.

Here X1h denotes the samples that lie within a one-hour window. Both formulations can be used to monitor service levels effectively. Figure 13 shows Qmin (X1h, 0.8) as a line plot. Note how on June 24 the quantile rises above 3 ms, indicating a violation of the SLA. Figure 14 shows a plot of the inverse quantile CDF(3ms, X1h), which takes values on the right axis between 0 percent and 100 percent. The SLA violation manifests as the inverse quantile dropping below 80 percent.

Hence, quantiles and inverse quantiles give complementary views of the current service level.

Histogram metric with quantile Qmin(0.8) over 1h windows
Figure 13: Histogram metric with quantile Qmin(0.8) over 1h windows
Histogram metric with inverse quantile CFD(3ms) over 1h windows
Figure 14: Histogram metric with inverse quantile CFD(3ms) over 1h windows


This article has presented an overview of some statistical techniques that find applications in IT operations. We discussed several visualization methods, their qualities and relations to each other. Histograms were shown to be an effective tool for capturing data and visualizing sample distributions. Finally, the concepts of percentiles and inverse percentiles were shown to have application to monitoring service levels.


  1. Cockcroft, A. 2014. Monitorama - Please, no more Minutes, Milliseconds, Monoliths or Monitoring Tools;
  2. Georgii, H.-O. 2002. Stochastik. DeGruyter.
  3. Gunther, N. J. 2007. Guerrilla Capacity Planning. Berlin: Springer-Verlag.
  4. Hartmann, H. 2015. Statistics for Engineers;
  5. Hartmann, H. 2016. Show Me the Data;
  6. HDR Histogram: a high dynamic range histogram;
  7. Histogram;
  8. IPython;
  9. Izenman, A. J. 2008. Modern Multivariate Statistical Techniques. New York: Springer-Verlag.
  10. Janert, P. K. 2010. Data Analysis with Open Source Tools. O’Reilly.
  11. Lp space;
  12. Matplotlib;
  13. Quantile;
  14. Schlossnagle, T. 2015. The problem with math: why your monitoring solution is wrong;
  15. Schwarz, B. 2015. Practical Scalability Analysis with the Universal Scalability Law;
  16. Ugurlu, D. 2013. The Most Misleading Measure of Response Time: Average;
  17. Waskom, M. 2015. Seaborn: statistical data visualization;
  18. Waskom, M. 2015. Seaborn example gallery; index.html

See where Circonus can take you.

Learn how monitoring without limits can drive new levels of value and impact in your business and operations.

Free Trial Circonus Sandbox