Geometric Mean and Geometric Standard Deviation
A friend is writing her master’s thesis in a subfield where data is typically summarized using geometric statistics: geometric means (GMs) and geometric standard deviations (GSDs), and sometimes even geometric standard errors (GSEs), whatever those are. Oh and occasionally also ‘geometric confidence intervals’ and ‘geometric interquartile ranges’.
…Most of which are (a) not something anyone really has intuition for and (b) surprisingly hard to find references for online, compared to regular ‘arithmetic’ statistics.
I was trying to help her understand these, but it took a lot of work to find easilyreadable references online, so I wanted to write down what I figured out.
(later edit: I wish I had saved the sources though.)
fyi: I think there is something in this that is defined incorrectly (a couple people emailed me about it) but I haven't had a chance to go back and fix it because I need to dig up some sources again. It's probably still useful for intuition but definitely doublecheck the facts.
1. What’s the point of using a Geometric Mean?
The geometric mean of a dataset \(\b{x} = \{x_i\}\) is given by:
\[\GM[\b{x}] = (\prod x_i)^{\frac{1}{n}}\]Though it is easier to understand through this equation^{1}
\[\GM[\b{x}] = e^{\AM[\log \b{x}]}\]For example:
\[\GM[1,2,3,4,5] = \sqrt[5]{1\cdot 2 \cdot 3 \cdot 4 \cdot 5} = e^{\frac{1}{n}(\log 1 + \log 2 + \log 3 + \log 4 + \log 5)} \approx 2.61\]A more meaningful example: if something increases by \(x_i\) percents for \(i = (1 \ldots 10)\), then its total increase is by \(\GM(x_i)^{10}\). Multiplying by 10 different numbers gives the same result as multiplying by their geometric mean ten times.
(The base of the exponent and logarithm can be anything. It cancels out – if you use \(\log_b\), you raise \(b\) to the power afterwards: \(\GM[\b{x}] = b^{\AM[\log_b \b{x}]}\).)
(By the way: using \(\GM[]\) and \(\AM[]\) as notations for these things is definitely not conventional, but I think it makes it easier to read in settings where you’re using both.)
The exponentialoflogarithms formula means that you are computing “the average of the logarithms of your data”, and then rescaling that back so the units work out. The same process will be used to construct the geometric standard deviation / confidence intervals / interquartile ranges / whatever: (a) calculate the regular statistic on the log data, like \(\text{SD}[\log x]\), then rescale back: \(\text{GSD}[x] = e^{SD[\log x]}\). We might call this “logtransformed” statistics; I don’t know where the word “geometric” came from.
The simplest case where logtransformed statistics are more useful is if we are dealing with data whose values range over many orders of magnitudes. This is true for things that we already use logarithmic scales for, like brightness of stars, loudness of sounds, acidity/basicity of solutions, frequency of sounds, or power of earthquakes.
In all these cases, notice that you don’t really know what the original measurements were. Like, presumably Richter measures the energy of the earthquake, but who knows what the values actually look like. Is a 7.5 Richter earthquake \(10^{15} J\) of energy? \(10^{20} J\)? Who knows?^{2} If we did statistics on the nontransformed values themselves, a single huge value like \(10^{20}\) would become the mean of any data set, so our summary statistics would not summarize much of anything.
Basically, each of these scales is measuring the logarithms of something, but for reporting data we don’t care as much what the ‘something’ is. If that’s the case, I think you may as well just logtransform all your data and then be done with it. Don’t worry about geometric means, just logarithm everything and take arithmetic means. But that’s my opinion.^{3}
The geometric mean and related statistics are for when you logtransform your data for analysis, and then want to transform it back. For instance, if you wanted to report the power of an earthquake in Joules, not in Richter, maybe because you’re comparing it to other numbers reporting in Joules. Or, say, if the original data came in very sensible units like parts per million
or growth % yearoveryear
, and you don’t want to report a value that’s been logtransformed to be unrecognizable.
1b. But actually: when do you use it?
It’s turning out to be surprisingly tricky to get a straight answer. Some people think the answer is ‘almost always’. But it appears to be pretty subjective.
Here’s the basic reason, though:
 You want to use a geometric mean if it makes more sense for your data to be multiplied together, rather than added together.
(In fact, it may as well just be called a ‘multiplicative mean’, and then let’s call the arithmetic mean the ‘additive mean’ at the same time.)
Here are some signs that this might be true:
 the factors that cause your data to vary at all apply ‘multiplicatively’, instead of ‘additively’
 that is, a factor’s effect is proportional to the size of your data. This would be anything like an ‘increase in efficiency’ or an ‘increased rate of occurrence’
 in practice most things, apparently, work this way. Growth rates, heights, densities, power outputs, dollar amounts, disease rates, …
 but generally you may not be sure, or able to tell, if this is true, so here are some more signs:
 corollary: the logarithm of the data looks more like a normal distribution than the data itself (ie your data is closer to lognormal)
 the data definitely cannot go below 0, by definition of what you’re measuring
 and effectively never goes to 0, either, since that would make the geometric mean 0 also (though see note, below)
 the data do not have a constant translation factor (ie, if you could just as easily have measured \(x + 50\) instead of \(x\), then the logarithm is going to be meaningless.)
 so don’t take the GM of temperature scales, unless you translate them to Kelvin first
 the data do have an arbitrary multiplicative factor (if you could have just as easily measured \(50x\) instead of \(x\), by changing units)
 the data are normalized against some constant value (so, if it is a % of anything)
 or, generally, the data are ratios, such as concentrations of a substance, rates of occurrence of an event, or changes in a value per some unit of time.
People online seem to think that many – maybe even most – quantities turn out to vary multiplicatively, more or less. Probably many more than are being currently summarized using geometric means!
I’d say a general rule of thumb is: if you feel, for your type of data and some value of \(N\), that a reasonable average of \(10^N\) and \(10^{N}\) is \(1\) rather than \(\frac{10^N}{2}\), you should use a geometric mean.
Examples:
 Company growth? Doubling revenue and halving revenue should average to \(1\), not \(1.25\). Geometric.
 Disease incidence? 1in100 and 1in10000 should probably average to 1in1000, not 1in200. Geometric.
 Microbe concentration? 1 partpermillion and 10000 partspermillion should probably be 100 partspermillion. Geometric. One bad lake shouldn’t mess up them all.
 Heights of kids in a classroom? a bunch of 5.somethings and some 6.somethings — it doesn’t really matter, they’ll be really close.
 Distances of fire stations to homes in a city? Tricky. I’d say arithmetic; you want your mean to be sensitive to that small cluster of houses that is weirdly 10x as far from service as everyone else.
 Split times on a 10km race? Arithmetic, you almost certainly want the mean to be one tenth of the total time.
 Speeds on each leg of a road trip, like \(60\) mph and \(45\) mph. Trick question, neither, you want the harmonic mean, which is pretty much only for speeds; see the end of this article.
Put differently: if massive values that are many orders of magnitude higher than others are expected and should not basically delete all your other data, then use a geometric mean.
1c. Important note about 0
If your data includes \(0\), then the geometric mean is 0.
But wait. What if you measure, like, bacteria concentration in a lake (a definitely orderofmagnitudebased value), and get 0 because you don’t detect anything? Or you detect ‘no earthquakes’ on some particular day?
And indeed, it turns out lots of scientists have data with 0s in it, and they’ve just been … working around it?
Okay. Maybe it is possibly reasonable to either:
 delete 0 values from your data, or
 replace 0s with ‘very small numbers that aren’t 0’, such as \(\e = 10^{k}\) for some \(k\) that makes them smaller than all your nonzero values
…in order to keep your geometric mean from being literally 0. Why: maybe your instruments aren’t sensitive to detect values that are very near 0 and just report 0 instead. Or maybe your data is a ratio like ‘counts of a molecule per million liters’, and it is assumed that, no matter what, if you sampled enough, the thing you’re counting would show up at least once, if you kept looking, so it can’t ‘really’ be 0.
But, and this is important: if there are enough zero data points that your choice of \(\e\) for rounding is changing your geometric mean significantly, you are probably doing something wrong.
And if you decide instead of delete the 0 values, you better report the result as “the average X when X was present”, rather than just “the average X”, or you’re just “lying with statistics”.
It’s pretty confusing, though. You aren’t wrong for being unsure. People use geometric means all the time with data that can be zero, maybe doing one of the above workarounds, and I really doubt they’re all handling it correctly or reporting it correctly afterwards.
Oh, and by the way: if you are geometricmeaning a bunch of, say, annual company growth percentages, and getting 0 – say, for example, the data \(x = (20\%, 0\%, 50\%, 120\%)\) – you’re doing it wrong. Those growth rates are percentages and need to be changed to actual multiplicative factors \(x = (0.8, 1.0, 1.5, 2.2)\), which gets rid of the zeroes. The only \(0\%\) growth is your company going out of business.
2. Equations for Geometric Statistics
Once again, the geometric mean is the logtransformed arithmetic mean:
\[\GM[x] = e^{\AM[\log x]} = \sqrt[n]{\prod_i x_i} = \prod_i \sqrt[n]{x_i}\]By the AMGM inequality, which is often just referred to as AMGM, the geometric mean is always less than the arithmetic mean (if the inputs are all positive. Otherwise all bets are off. Don’t use GM on negative numbers!):
\[\AM[x] \geq \GM[x]\]Another important relationship, which explains why GM works so well for ratios:
\[\GM \Big[ \frac{x_i}{y_i} \Big] = \frac{\GM[x_i]}{\GM[y_i]}\]By the way, don’t program the formula \(\GM[x] = \sqrt[n]{\prod x_i}\) into a computer (if you do not know what you are doing). Use a builtin library for it, or use one of the other two formulas (\(= e^{\AM[\log x]}\) or \(= \prod \sqrt[n]{x_i}\)). Multiplying arbitrarily long lists of numbers together can overflow the data values in some programming languages, causing them to wrap around to negative numbers, rendering all your math utterly wrong. In particularly unfortunate cases, the resulting value will, by chance, be close and so seem reasonable, yet have been computed an entirely wrong way.
The geometric standard deviation (GSD) is the same transformation, applied to the regular standard deviation.
\[\text{GSD}[x] = e^{\text{SD}[\log x]}\]This is going to be useful if and only it was a good idea to use a geometric mean on your data, and particularly, IMO, if your data is positively skewed. That’s cause: using a Geometric Standard Deviation, the phrase “68.2% of values fall within one standard deviation of the mean” means something different:
The GSD, instead of giving an equal range on either side of the mean, gives an equal factor:
\[\begin{aligned} e^{AM \pm SD} &= e^{AM}e^{\pm SD} \\ &= \GM[x] \text{GSD}[x]^{\pm 1} \\ &= \GM[x] \; {}_{\div}^{\times} \; \text{GSD}[x] \\ & \stackrel{!}{\neq} \GM[x] \pm \text{GSD}[x] \end{aligned}\]Note that \(\GM[x] \text{GSD}[x]^{\pm 1}\) means the two values are \((\GM[x] \text{GSD}[x], \frac{\GM[x]}{\text{GSD}[x]})\). Clearly the GSD should be not too different from \(1\), such that 68.2% of values fall in the range of \(\text{GSD}[x]^{\pm 1}\).
Analogously, if your data is welldescribed by a GM + GSD, it’s probably not welldescribed by an AM + SD, because it should be positively skewed, while your SD would suggest that the data is spread evenly around the mean. (Consider how weird it would be to say that “X% of the data falls within \(\frac{\mu}{2}\) and \(2 \mu\)” for nonskewed data like a normal distribution.)
If reporting confidence intervals for a given \(P\)value, using geometric statistics, they will also not be the same distance from the geometric mean. For \(P=0.95\), with zscore \(z = 1.96\) (ie laughably permissive, false positives everywhere), the confidence intervals are:
\[\text{CI}_{0.95} = \GM[x] e^{\pm 1.96 \text{SD}[x]} = \GM[x] GSD^{\pm 1.96}\]For interquartile ranges, it’s:
\[\text{IQR} = \GM[x] e^{\pm 0.67 \text{SD}[x]} = \GM[x] GSD^{\pm 0.67}\]What about geometric standard error? This would say “how far is the sample geometric mean \(\GM[x_i]\) from the true geometric mean of the data \(\GM[\b{x}]\)?”
Reminder, since this one is a little less common: the standard error (or more precisely, the “standard error of the mean”) of a set of \(N\) samples drawn from a normal distribution \(\cal{N}(\mu, \sigma^2)\) is the “standard deviation of the mean of \(N\) values from the true mean of the distribution” and is given by \(\text{SE}[x,N] = \frac{\sigma_x}{\sqrt{N}}\). Intuitively, as we sample more values from the normal distribution, our computed mean is also a normally distributed random variable, but has a smaller standard deviation, by a factor of \(\frac{1}{\sqrt{N}}\).
This measurement is actually reported in papers sometimes (apparently), but hard to find a good equation for online. It turns out that it’s given by:
\[\text{GSE}[x, N] = \frac{\mu_G}{\sqrt{N}} \sigma_{\log x}\]Where \(\sigma_{\log x}\) is the standard deviation of \(\log x\), not the ‘geometric standard deviation’, and \(\mu_G\) is the true geometric mean of \(x\) (or at least, one that you have way more confidence in, like from a separate and much larger study).
Since you don’t normally know \(\mu_G\), you can use the sample mean \(\GM[x]\), but then, like with the regular standard deviation and error formulas, you have to change \(N \ra N1\):
\[\text{GSE}[x, N] = \frac{\GM[x]}{\sqrt{N1}} \text{SD}[\log x]\]The geometric standard error tells us that “our calculation of \(\GM[x]\) from \(N\) samples” has a probability distribution:
\[\GM_N[x] \sim \cal{N}(\mu_G, \frac{\mu_G}{\sqrt{N}} \sigma_{\log x}) = \cal{N}(\mu_G, GSE[x,N])\]And the previous formula is our best estimate of it, given a sample of \(N\) values.
This is weird to compute. Until now we did not care about \(\sigma_{\log x}\). (And I assume, but am actually not sure, that \(\sigma_{\log x}\) has another factor of \(\frac{1}{N1}\) in it, since it’s also computed from the sample?) But it should, for sufficiently high \(N\), give the correct numerical difference between \(\GM[x]\) and the true \(\mu_x\).
I’m not sure it’s a good idea to use the GSE, because I don’t think people will know how to think about it. It is a bit unintuitive: it is the only statistic we’ve talked about that is additive (so it’s talking about a difference \(\GM_N[x]  \mu_G\), rather than a ratio \(\frac{GM_N[x]}{\mu_G}\)). It’s based on the idea that, even for lognormal data, the sample geometric mean \(\GM_N[x]\) is normally distributed. (I did not say lognormal, though it may also be that! Even lognormal distribution samples are normally distributed for highenough \(N\) – they are just also massively skewed.) Maybe some subfield thinks this makes perfect sense as a summary statistic – I don’t know. I would avoid it.
Trivia: Some Other Means
By the way, there are other ways to summarize data than AM and GM. You can create all sorts of means using the same formula:
\[\E_f [ X ] = f^{1} \E [ f(X) ]\]For instance:
 AM: \(f(x) = x\)
 GM: \(f(x) = \log x\)
 Harmonic Mean: \(f(x) = x^{1}\)
 Good for averaging rates of things, like the rates of a car over different legs of a trip. If you travel \(60\) mph one way and \(40\) mph the other way over the same distance (maybe there was some traffic), then your average speed is \(\frac{2}{\frac{1}{60} + \frac{1}{40}} \approx 48 \text{ mph}\), which is useful: the total travel time is equal if you travel \(48\) mph the whole time.
 Root Mean Square: \(f(x) = x^2\)
 Used in electrical engineering, especially to measure the average strength of electrical signals. When \(\AM[x]=0\) (such as for an alternating current), then \(\text{RMS}(x) = \sigma_x\), the regular standard deviation.
 Generalized means: screw it, \(f(x) = x^p\) for any \(p\).
 \(\infty\)mean: set \(p=\infty\) and you get, weirdly, \(\E_{\infty}[x] = \max(x)\).
 \(\infty\)mean: it’s just \(\min(x)\)
 basically turn \(p\) up to get more contributions from higher values and turn it down to get more contributions from lower ones.
 just kidding, don’t actually use anything except the first four, no one will know what you’re talking about if you use this.
Generally speaking if you find yourself with data of the form \(y = f(x)\), and you know it’s of the form \(y =f(x)\), it probably will be a good idea to summarize it with statistics on \(f^{1} (y)\). These are just specific implementations of that idea. The geometric mean turns out to be the safest option, in general, because \(x \mapsto \log x\) has the nice property of smoothing out any kind of polynomial \(\ra p \log x\) and exponential \(e^x \mapsto x\).
[disclaimer: I am not a mathematician. What I wrote here is my best understanding at the time. Let me know if you think something is wrong or incomplete.]

I’m using \(\log x\equiv \log_e x \equiv \ln x\) throughout this page. This is normal in some fields and totally weird in others. Sorry if it’s weird to you. ↩

Actually I looked this up and Richter is totally different: it measures the amplitude of the local tremor, as recorded on a seismograph (the displacement of a needle, or whatever) which is then normalized based on the distance from the measurement to the epicenter of the earthquake. It turns out there are tons of other scales, and some of them do a better job of measuring the actual energy of the earthquake. ↩

Although you can introduce artifacts in your data by doing this. If a bunch of values are uniformly distributed in \([1, 1000]\) (so \(AM \sim 500\)) and you take all their logarithms, they’re not uniformly distributed anymore – their average is going to be higher than \(\log 500\). This had better not affect your statistics’ conclusions. ↩