An example of using algebraic transformations for code optimization.

Created by Peter Kankowski
Last changed
Contributors: Orib, Greg Wiszkowitz, James Mayfield, and Ace
Filed under Low-level code optimization

Share on social sitesReddit Digg Delicious Buzz Facebook Twitter

Calculating standard deviation in one pass

Standard deviation is a statistic parameter that helps to estimate the dispersion of data series. It's usually calculated in two passes: first, you find a mean, and second, you calculate a square deviation of values from the mean:

double std_dev1(double a[], int n) {
    if(n == 0)
        return 0.0;
    double sum = 0;
    for(int i = 0; i < n; ++i)
       sum += a[i];
    double mean = sum / n;
    double sq_diff_sum = 0;
    for(int i = 0; i < n; ++i) {
       double diff = a[i] - mean;
       sq_diff_sum += diff * diff;
    double variance = sq_diff_sum / n;
    return sqrt(variance);

But you can do the same thing in one pass. Rewrite the formula in the following way:

s_x^2 = 1/N * sum(x_i - x_mean)^2 = 1/N * sum(x_i^2 - 2*x_i*x_mean + x_mean^2) = 1/N * sum(x_i^2) - 2*x_mean/N * sum(x_i) + N*x_mean^2/N = 1/N * sum(x_i^2) - 2*x_mean^2 + x_mean^2 = 1/N * sum(x_i^2) - x_mean^2
double std_dev2(double a[], int n) {
    if(n == 0)
        return 0.0;
    double sum = 0;
    double sq_sum = 0;
    for(int i = 0; i < n; ++i) {
       sum += a[i];
       sq_sum += a[i] * a[i];
    double mean = sum / n;
    double variance = sq_sum / n - mean * mean;
    return sqrt(variance);

Unfortunately, the result will be inaccurate when the array contains large numbers (see the comments below).

This trick is mentioned in an old Soviet book about programmable calculators (here is the reference for Russian readers: Финк Л. Папа, мама, я и микрокалькулятор. — М.: Радио и связь, 1988).

Peter Kankowski
Peter Kankowski

About the author

Peter is the developer of Aba Search and Replace, a tool for replacing text in multiple files. He likes to program in C with a bit of C++, also in x86 assembly language, Python, and PHP. He can be reached at kankowski@gmail.com.

Created by Peter Kankowski
Last changed
Contributors: Orib, Greg Wiszkowitz, James Mayfield, and Ace


Ten recent comments are shown below. Show all comments


James, thank you, really very good point. Mark Hoemmen's paper has a perfect example to illustrate the danger: take only three numbers 10000, 10001 and 10002, use floats and get the wrong result.

(It's probably not accidental that M. Hoemmen is PhD student on Berkeley, where Prof. W. Kahan is. And I wouldn't be surprised if James is there too. :) )

That simple example shows once again why Prof. W. Kahan designed the x87 FPU with the possibility to use 80 bits for partial results. But Microsoft intentionally designed compilers to ignore that feature (and more).

Well that's a good moment to visit again Prof. Kahan's web page


and also read, for example

Marketing versus Mathematics


"Old Kernighan-Ritchie C works better than ANSI C or Java!"

"In 1980 we went to Microsoft to solicit language support for the 8087, for

which a socket was built into the then imminent IBM PC. Bill Gates attended

our meeting for a while and then prophesied that almost none of those sockets

would ever be filled! He departed, leaving a dark cloud over the discussions.

Microsoft’s languages still lack proper support for Intel’s floating-point."



"In particular Bill Gates Jr., Microsoft’s language expert, disparaged the extra-wide format in 1982 with consequences that persist today in Microsoft’s languages for the PC. Sun’s Bill Joy did likewise."


Floating-Point Arithmetic Besieged by “Business Decisions”


The first paper is from 2000 and I don't know if Java now has the extended precision floating point for partial results. Modern CPU instructions "for multimedia" of course don't. Which makes them potentially dangerous when used on formulas like the starting one.

Finally, from:


"Routine use of far more precision than deemed necessary by clever but numerically naive programmers, provided it does not run too slowly, is the best way available"

Arne Hormann,

It does not require two divisions - only one.

// for sample variance
double std_dev(double a[], int n) {
    if(n == 0)
        return 0.0;
    int i = 0;
    double meanSum = a[0];
    double stdDevSum = 0.0;
    for(i = 1; i < n; ++i) {
        double stepSum = a[i] - meanSum;
        double stepMean = ((i - 1) * stepSum) / i;
        meanSum += stepMean;
        stdDevSum += stepMean * stepSum;
    // for poulation variance: return sqrt(stdDevSum / elements);
    return sqrt(stdDevSum / (elements - 1));

Arne, thank you for your algorithm. I've already seen similar algorithms but I don't know how they are derived. I'd greatly appreciate if you would point me to some source which explains derivation of this and/or similar algorithms. According to some experiments I made for some similar calculations the algorithms like the one you presented here give significantly better results compared to the "one pass, full squares" (which can produce even completely wrong results for "inconvenient" input) but they can still be noticeably less accurate than the real two pass algorithm. Do you know of any source where such error analysis is shown? Thank you.

David Jones,

As documented in previous comments, the error can be large for the one-pass method.

But sometimes the input to sqrt can be negative, resulting in... well, who knows what, given that you're coding in C.

Try a list with 3 copies of 1.4592859018312442e+63 for example.

Peter Kankowski,

Thank you for the example. There seems to be a problem with very large values and little or no difference between them. std_dev1 works fine, while std_dev2 gives incorrect results.

Thanks for your stimulating input!,
typedef struct statistics_s {
    size_t k; // sample count
    double Mk; // mean
    double Qk; // standard variance
} statistics_t;

static void statistics_record (statistics_t *stats, double x) {
    if (1 == stats->k) {
        stats->Mk = x;
        stats->Qk = 0;
    } else {
        double d = x - stats->Mk; // is actually xk - M_{k-1},
                                  // as Mk was not yet updated
        stats->Qk += (stats->k-1)*d*d/stats->k;
        stats->Mk += d/stats->k;

static double statistics_stdev (statistics_t *stats)
    return sqrt (stats->Qk/stats->k);

static size_t statistics_nsamples (statistics_t *stats)
    return stats->k;

static double statistics_variance (statistics_t *stats)
    return stats->Qk / (stats->k - 1);

static double statistics_stdvar (statistics_t *stats)
    return stats->Qk / stats->k;

static double statistics_init (statistics_t *stats)
    stats->k = 0;

Knuth actually has one pass algorithm like this

for N=1
_M = x
_C = 0.0;
for N>1
delta = x - _M;
_M += delta / N;
_C += delta*(x-_M)

_M gets you mean, _C / (N-1) is variance


Thanks, Ildar. There's so many gems in Knuth's toolkit.

@Ace, if you're still around, walk thru the algebra carefully, paying attention to the fact that n times the average is definitionally the sum.



above PDF link is broken. Below link can be used.


Peter Kankowski,
Thank you very much for the updated link.

Your name:


Please ignore this field: