# Bimodal distribution

{{#invoke:Hatnote|hatnote}}Template:Main other

Figure 1. A simple bimodal distribution, in this case a mixture of two normal distributions with the same variance but different means. The figure shows the probability density function (p.d.f.), which is an average of the bell-shaped p.d.f.s of the two normal distributions.
Figure 2. Histogram of body lengths of 300 weaver ant workers.[1]
Figure 3. A bivariate, multimodal distribution.

In statistics, a bimodal distribution is a continuous probability distribution with two different modes. These appear as distinct peaks (local maxima) in the probability density function, as shown in Figure 1.

More generally, a multimodal distribution is a continuous probability distribution with two or more modes, as illustrated in Figure 3.

## Terminology

When the two modes are unequal the larger mode is known as the major mode and the other as the minor mode. The least frequent value between the modes is known as the antimode. The difference between the major and minor modes is known as the amplitude. In time series the major mode is called the acrophase and the antimode the batiphase.

## Gatlung's classification

Gatling introduced a classification system (AJUS) for distributions[2]

• A: unimodal distribution - peak in the middle
• J: unimodal - peak at either end
• U: bimodal - peaks at both ends
• S: bimodal or multimodal - multiple peaks

This classification has since been modified slightly:

• J (modified) - peak on right
• L: unimodal - peak on left
• F: no peak (flat)

Under this classification bimodal distributions are classified as type S or U.

## Examples

Bimodal distributions occur both in mathematics and in the natural sciences.

### Probability distributions

Important bimodal distributions include the arcsine distribution and the beta distribution. Others include the U-quadratic distribution.

The ratio of two normal distributions is also bimodally distributed. Let

${\displaystyle R={\frac {a+x}{b+y}}}$

where a and b are constant and x and y are distributed as normal variables with a mean of 0 and a standard deviation of 1. R has a known density that can be expressed as a confluent hypergeometric function.[3]

The distribution of the reciprocal of a t distributed random variable is bimodal when the degrees of freedom are more than one. Similarly the reciprocal of a normally distributed variable is also bimodally distributed.

### Occurrences in nature

Examples of variables with bimodal distributions include the time between eruptions of certain geysers, the color of galaxies, the size of worker weaver ants, the age of incidence of Hodgkin's lymphoma, the speed of inactivation of the drug isoniazid in US adults, the absolute magnitude of novae, and the circadian activity patterns of those crepuscular animals that are active both in morning and evening twilight. In fishery science multimodal length distributions reflect the different year classes and can thus be used for age distribution- and growth estimates of the fish population[4] Sediments are usually distributed in a bimodal fashion.

## Origins

{{#invoke:main|main}}

### Mathematical

A bimodal distribution most commonly arises as a mixture of two different unimodal distributions (i.e. distributions having only one mode). In other words, the bimodally distributed random variable X is defined as ${\displaystyle Y}$ with probability ${\displaystyle \alpha }$ or ${\displaystyle Z}$ with probability ${\displaystyle (1-\alpha ),}$ where Y and Z are unimodal random variables and ${\displaystyle 0<\alpha <1}$ is a mixture coefficient.

Mixtures with two distinct components need not be bimodal and two component mixtures of unimodal component densities can have more than two modes. There is no immediate connection between the number of components in a mixture and the number of modes of the resulting density.

### Biology

In biology four factors are known to contribute to bimodal distributions of population sizes:

• the initial distribution of individual sizes
• the distribution of growth rates among the individuals
• the size and time dependence of the growth rate of each individual
• mortality rates that may affect each size class differently.

The bimodal distribution of sizes of weaver ant workers shown in Figure 2 arises due to existence of two distinct classes of workers, namely major workers and minor workers.[1] In this case, Y would be the size of a random major worker, Z the size of a random minor worker, and α the proportion of worker weaver ants that are major workers.

The distribution of fitness effects of mutations for both whole genomes[5][6] and individual genes[7] is also frequently found to be bimodal with most mutations being either neutral or lethal with relatively few having intermediate effect.

## General properties

A mixture of two unimodal distributions with differing means is not necessarily bimodal. The combined distribution of heights of men and women is sometimes used as an example of a bimodal distribution, but in fact the difference in mean heights of men and women is too small relative to their standard deviations to produce bimodality.[8]

Bimodal distributions have the peculiar property that - unlike the unimodal distributions - the mean may be a more robust sample estimator than the median.[9] This is clearly the case when the distribution is U shaped like the arcsine distribution. It may not be true when the distribution has one or more long tails.

### Moments of mixtures

Let

${\displaystyle f(x)=pg_{1}(x)+(1-p)g_{2}(x)}$

where gi is a probability distribution and p is the mixing parameter.

The moments of f(x) are[10]

${\displaystyle \mu =p\mu _{1}+(1-p)\mu _{2}}$
${\displaystyle \nu _{2}=p[\sigma _{1}^{2}+\delta _{1}^{2}]+(1-p)[\sigma _{2}^{2}+\delta _{2}^{2}]}$
${\displaystyle \nu _{3}=p[S_{1}\sigma _{1}^{3}+3\delta _{1}\sigma _{1}^{2}+\delta _{1}^{3}]+(1-p)[S_{2}\sigma _{2}^{3}+3\delta _{2}\sigma _{2}^{2}+\delta _{2}^{3}]}$
${\displaystyle \nu _{4}=p[K_{1}\sigma _{1}^{4}+4S_{1}\delta _{1}\sigma _{1}^{3}+6\delta _{1}^{2}\sigma _{1}^{2}+\delta _{1}^{4}]+(1-p)[K_{2}\sigma _{2}^{4}+4S_{2}\delta _{2}\sigma _{2}^{3}+6\delta _{2}^{2}\sigma _{2}^{2}+\delta _{2}^{4}]}$

where

${\displaystyle \mu =\int {xf(x)dx}}$
${\displaystyle \delta _{i}=\mu _{i}-\mu }$
${\displaystyle \nu _{r}=\int {(x-\mu )^{r}f(x)dx}}$

and Si and Ki are the skewness and kurtosis of the ith distribution.

## Mixture of two normal distributions

It is not uncommon to encounter situations where an investigator believes that the data comes from a mixture of two normal distributions. Because of this, this mixture has been studied in some detail.[11]

A mixture of two normal distributions has five parameters to estimate: the two means, the two variances and the mixing parameter. A mixture of two normal distributions with equal standard deviations is bimodal only if their means differ by at least twice the common standard deviation.[8] Estimates of the parameters is simplified if the variances can be assumed to be equal (the homoscedastic case).

It is obvious that if the means of the two normal distributions are equal that the combined distribution is unimodal. Conditions for unimodality of the combined distribution were derived by Eisenberger.[12] Necessary and sufficient conditions for a mixture of normal distributions to be bimodal have been identified by Ray and Lindsay.[13]

A mixture of two approximately equal mass normal distributions have a negative kurtosis since the two modes on either side of the center of mass effectively flatten out the distribution.

A mixture of two normal distributions with highly unequal mass have a positive kurtosis since the smaller distribution lengthens the tail of the more dominant normal distribution.

Mixtures of other distributions require additional parameters to be estimated.

### Mixture of two normal distributions with equal variances

If the case of equal variance the mixture is unimodal if and only if[14]

${\displaystyle d\leq 1}$

or

${\displaystyle \left\vert \log(1-p)-\log(p)\right\vert \geq 2\log(d-{\sqrt {d^{2}-1}})+2d{\sqrt {d^{2}-1}}}$

where p is the mixing parameter and d is

${\displaystyle d={\frac {\mu _{1}-\mu _{2}}{2{\sqrt {\sigma _{1}\sigma _{2}}}}}}$

where μ1 and μ2 are the means of the two normal distributions and σ1 and σ2 are their standard deviations.

## Summary statistics

Bimodal distributions are a commonly used example of how summary statistics such as the mean, median, and standard deviation can be deceptive when used on an arbitrary distribution. For example, in the distribution in Figure 1, the mean and median would be about zero, even though zero is not a typical value. The standard deviation is also larger than deviation of each normal distribution.

Although several have been suggested, there is no presently generally agreed summary statistic (or set of statistics) to quantify the parameters of a general bimodal distribution. For a mixture of two normal distributions the mean and standard deviation along with the mixing parameter (a weighing system for the combination) are usually used - a total of five parameters.

### Ashman's D

A statistic that may be useful is Ashman's D:[15]

${\displaystyle D=2^{\frac {1}{2}}{\frac {\left|\mu _{1}-\mu _{2}\right|}{\sqrt {(\sigma _{1}^{2}+\sigma _{2}^{2})}}}}$

where μ1, μ2 are the means and σ1 σ2 are the standard deviations.

For a mixture of two normal distributions D > 2 is required for a clean separation of the distributions.

### Bimodality index

The bimodality index assumes that the distribution is a sum of two normal distributions with equal variances but differing means.[16] It is defined as follow:

${\displaystyle \delta ={\frac {|\mu _{1}-\mu _{2}|}{\sigma }}}$

where μ1, μ2 are the means and σ is the common standard deviation.

${\displaystyle BI=\delta {\sqrt {p(1-p)}}}$

where p is the mixing parameter.

### Bimodal separation

This index assumes that the distribution is a mixture of two normal distributions with means (μ1 and μ2) and standard deviations (σ1 and σ2):[17]

${\displaystyle S={\frac {\mu _{1}-\mu _{2}}{2(\sigma _{1}+\sigma _{2})}}}$

### Bimodality coefficient

Sarle's bimodality coefficient b is[18]

${\displaystyle \beta ={\frac {\gamma ^{2}+1}{\kappa }}}$

where γ is the skewness and κ is the kurtosis. The kurtosis is here defined to be the standardised fourth moment around the mean. The value of b lies between 0 and 1.[19] The logic behind this coefficient is that a bimodal distribution will have very low kurtosis, an asymmetric character, or both - all of which increase this coefficient.

The formula for a finite sample is[20]

${\displaystyle b={\frac {g^{2}+1}{k+{\frac {3(n-1)^{2}}{(n-2)(n-3)}}}}}$

where n is the number of items in the sample, g is the sample skewness and k is the sample excess kurtosis.

The value of b for the uniform distribution is 5/9. This is also its value for the exponential distribution. Values greater than 5/9 may indicate a bimodal or multimodal distribution. The maximum value (1.0) is reached only by a Bernoulli distribution with only two distinct values or the sum of two different Dirac delta functions.

The distribution of this statistic is unknown. It is related to a statistic proposed earlier by Pearson - the difference between the kurtosis and the square of the skewness (vide infra).

### Bimodality amplitude

This is defined as[17]

${\displaystyle A_{B}={\frac {A_{1}-A_{an}}{A_{1}}}}$

where A1 is the amplitude of the smaller peak and Aan is the amplitude of the antinode.

AB is always < 1. Larger values indicate more distinct peaks.

### Bimodal ratio

This is the ratio of the left and right peaks.[17] Mathematically

${\displaystyle R={\frac {A_{r}}{A_{l}}}}$

where Al and Ar are the amplitudes of the left and right peaks respectively.

### Bimodality parameter

This parameter (B) is due to Wilcock.[21]

${\displaystyle B={\frac {A_{r}}{A_{l}}}\sum P_{i}}$

where Al and Ar are the amplitudes of the left and right peaks respectively and Pi is the logarithm taken to the base 2 of the proportion of the distribution in the ith interval. The maximal value of B is 1.

## Statistical tests

A number of tests are available to determine if a data set is distributed in a bimodal (or multimodal) fashion.

### Graphical methods

In the study of sediments particle size is frequently bimodal. Empirically it has been found useful to plot the frequency against the log( size ) of the particles.[22][23] This usually gives a clear separation of the particles into a bimodal distribution. In geological applications the logarithm is normally taken to the base 2. The log transformed values are referred to as phi (Φ) units. This system is known as the Krumbein (or phi) scale.

An alternative method is to plot the log of the particle size against the cumulative frequency. This graph will usually consist two reasonably straight lines with a connecting line corresponding to the antimode.

Statistics

Approximate values for several statistics can be derived from the graphic plots.[22]

${\displaystyle {\mathit {Mean}}={\frac {\phi _{16}+\phi _{50}+\phi _{84}}{3}}}$
${\displaystyle {\mathit {StdDev}}={\frac {\phi _{84}-\phi _{16}}{4}}+{\frac {\phi _{95}-\phi _{5}}{6.6}}}$
${\displaystyle {\mathit {Skew}}={\frac {\phi _{84}+\phi _{16}-2\phi _{50}}{2(\phi _{84}-\phi _{16})}}+{\frac {\phi _{95}+\phi _{5}-2\phi _{50}}{2(\phi _{95}-\phi _{5})}}}$
${\displaystyle {\mathit {Kurt}}={\frac {\phi _{95}-\phi _{5}}{2.44(\phi _{75}-\phi _{25})}}}$

where Mean is the mean, StdDev is the standard deviation, Skew is the skewness, Kurt is the kurtosis and φx is the value of the variate φ at the xth percentage of the distribution.

### Unimodal vs. bimodal distribution

A necessary but not sufficient condition for a symmetrical distribution to be bimodal is that the kurtosis be less than three.[24][25] Here the kurtosis is defined to be the standardised fourth moment around the mean. The reference given prefers to use the excess kurtosis - the kurtosis less 3.

Pearson in 1894 was the first to devise a procedure to test whether a distribution could be resolved into two normal distributions.[26] This method required the solution of a ninth order polynomial. In a subsequent paper Pearson reported that for any distribution skewness2 + 1 < kurtosis.[19] Later Pearson showed that[27]

${\displaystyle b_{2}-b_{1}\geq 1}$

where b2 is the kurtosis and b1 is the square of the skewness. Equality holds only for the two point Bernoulli distribution or the sum of two different Dirac delta functions. These are the most extreme cases of bimodality possible. The kurtosis in both these cases is 1. Since they are both symmetrical their skewness is 0 and the difference is 1.

Baker proposed a transformation to convert a bimodal to a unimodal distribution.[28]

Several tests of unimodality versus bimodality have been proposed: Haldane suggested one based on second central differences.[29] Larkin later introduced a test based on the F test;[30] Benett created one based on the G test.[31] Tokeshi has proposed fourth test.[32][33] A test based on a likelihood ratio has been proposed by Holzmann and Vollmer.[14]

A method based on the score and Wald tests has been proposed.[34] This method can distinguish between unimodal and bimodal distributions when the underlying distributions are known.

### Antimode tests

Statistical tests for the antimode are known.[35]

Otsu's method

Otsu's method is commonly employed in computer graphics to determine the optimal separation between two distributions.

### General tests

To test if a distribution is other than unimodal, several additional tests have been devised: the bandwidth test,[36] the dip test,[37] the excess mass test,[38] the MAP test,[39] the mode existence test,[40] the runt test,[41][42] the span test,[43] and the saddle test.

The dip test is available for use in R.[1] The values for the dip statistic values range between 0 to 1. Values less than 0.05 indicating significant bimodality and values greater than 0.05 but less than 0.10 suggesting bimodality with marginal significance.

### Silverman's test

Silverman introduced a bootstrap method for the number of modes.[36] The test uses a fixed bandwidth which reduces the power of the test and its interpretability. Under smoothed densities may have an excessive number of modes whose count during bootstrapping is unstable.

### Special cases

Additional tests are available for a number of special cases

Mixture of two normal distributions

A study of a mixture density of two normal distributions data found that separation into the two normal distributions was difficult unless the means were separated by 4-6 standard deviations.[44]

In astronomy the Kernel Mean Matching algorithm is used to decide if a data set belongs to a single normal distribution or to a mixture of two normal distributions.

Beta-normal distribution

This distribution is bimodal for certain values of is parameters. A test for these values has been described.[45]

## Parameter estimation and fitting curves

Assuming that the distribution is known to be bimodal or has been shown to be bimodal by one or more of the tests above, it is frequently desirable to fit a curve to the data. This may be difficult.

Bayesian methods may be useful in difficult cases.

### Software

Two normal distributions

A package for R is available for testing for bimodality.[2] This package assumes that the data are distributed as a sum of two normal distributions. If this assumption is not correct the results may not be reliable. It also includes functions for fitting a sum of two normal distributions to the data.

Assuming that the distribution is a mixture of two normal distributions then the expectation-maximization algorithm may be used to determine the parameters. Several programmes are available for this including Cluster.[46]

Other distributions

The mixtools package also available for R can test for and estimate the parameters of a number of different distributions.[47]

Another package for a mixture of two right tailed gamma distributions is available.[48]

Several other packages for R are available to fit mixture models: these include flexmix,[49] mcclust,[50] mixdist[51]

The programme SWRC ﬁt can fit a number of bimodal distributions.[3]

The statistical programme SAS can also fit a variety of mixed distributions with the command PROCFREQ.

## References

1. {{#invoke:Citation/CS1|citation |CitationClass=journal }}
2. Galtung J (1969) Theory and methods of social research. Universitetsforlaget, Oslo ISBN 0043000177
3. Fieller E (1932) The distribution of the index in a noraml bivariate population. Biometrika (24):428-440
4. Introduction to tropical fish stock assessment
5. {{#invoke:Citation/CS1|citation |CitationClass=journal }}
6. {{#invoke:Citation/CS1|citation |CitationClass=journal }}
7. {{#invoke:Citation/CS1|citation |CitationClass=journal }}
8. {{#invoke:Citation/CS1|citation |CitationClass=journal }}
9. Mosteller F, Tukey JW (1977) Data analysis and regression: a second course in statistics. Reading, Mass, Addison-Wesley Pub Co
10. Robertson CA, Fryer JG (1969) Some descriptive properties of normal mixtures. Skandinavisk Aktuarietidskrift 69: 137–146
11. Eisenberger I (1964) Genesis of bimodal distributions. Technometrics 6 (4) 357-363
12. Ray S, Lindsay BG (2005) The topography of multivariate normal mixtures. Ann Stat 33 (5) 2042-2065
13. Holzmann H,Vollmer S (2008) A likelihood ratio test for bimodality in two-component mixtures – with application to regional income distribution in the EU. AStA 2(1)57-69 Cite error: Invalid <ref> tag; name "Holzmann2008" defined multiple times with different content
14. Ashman KM, Bird CM, Zepf SE(1994) Astronomical J 108: 2348
15. Wang J, Wen S, Symmans WF, Pusztai L, Coombes KR (2009) The bimodality index: a criterion for discovering and ranking bimodal signatures from cancer gene expression profiling data. Cancer Inform 7:199-216
16. Zhang C, Mapes BE, Soden BJ (2003) Bimodality in tropical water vapour. Q J R. Meteorol Soc 129: 2847–2866 doi: 10.1256/qj.02.16
17. Ellision AM (1987) Effect of seed dimorphism on the density-dependent dynamics of experimental populations of Atriplex triangularis (Chenopodiaceae). Am J Botany 74(8): 1280-1288
18. Pearson K (1916) Mathematical contributions to the theory of evolution, XIX: Second supplement to a memoir on skew variation. Phil Trans Roy Soc London. Series A 216 (538–548): 429–457. Bibcode 1916RSPTA.216..429P. doi:10.1098/rsta.1916.0009. JSTOR 91092
19. SAS Institute Inc. (2012). SAS/STAT 12.1 user’s guide. Cary, NC: Author.
20. Wilcock PR (1993) The critical shear stress of natural sediments. J Hydraul Engrg ASCE 119:491-505
21. Folk RL, Ward WC (1957) Brazos River bar: a study in the significance of grain size parameters. J Sedim Petrol 27: 3–26
22. Dyer KR (1970) Grain-size parameters for sandy gravels. J Sedim Petrol 40 (2) 616-620
23. Gneddin OY(2010) Quantifying Bimodality.
24. Muratov AL, Gnedin OY (2010) Modeling the metallicity distribution of globular clusters. Ap J (submitted) arXiv:1002.1325
25. Pearson K (1894) Contributions to the mathematical theory of evolution: On the dissection of asymmetrical frequency-curves. Phil Trans Roy Soc Series A, Part 1, 185: 71-90
26. Pearson K (1929) Editorial note. Biometrika 21: 370-375
27. Baker GA (1930) Transformations of bimodal distributions. Ann Math Stat 1 (4) 334-344
28. Haldane JBS (1951) Simple tests for bimodality and bitangentiality. Ann Eugenics 16 (1) 359–364 DOI: 10.1111/j.1469-1809.1951.tb02488.x
29. Larkin RP (1979) An algorithm for assessing bimodality vs. unimodality in a univariate distribution. Behavior Research Methods 11 (4) 467-468 DOI: 10.3758/BF03205709
30. Bennett SC (1992) Sexual dimorphism of Pteranodon and other pterosaurs, with comments on cranial crests. J Vert Paleont 12 (4) 422-434
31. Tokeshi M (1992) Dynamics and distribution in animal communities; theory and analysis. Researches in Population Ecology 34:249–273
32. Barreto S, Borges PAV, Guo Q (2003) A typing error in Tokeshi’s test of bimodality. Global Ecology & Biogeography 12: 173–174
33. Carolan AM, Rayner JCW (2001) One sample tests for the location of modes of nonnormal data. J Applied Mathematics and Decision Science 5(1) 1–19
34. Hartigan JA (2000) Testing for antimodes. Studies in Classification, Data Analysis, and Knowledge Organization 169-181
35. Silverman BW (1981) Using kernel density estimates to investigate multimodality. J Roy Statist Soc Ser B 43:97-99 Cite error: Invalid <ref> tag; name "Silverman1981" defined multiple times with different content
36. Hartigan JA, Hartigan PM (1985) The dip test of unimodality. Ann Statist 13 (1) 70-84
37. Mueller DW, Sawitzki G (1991) Excess mass estimates and tests for multimodality. JASA 86, 738 -746
38. Rozál GPM Hartigan JA (1994) The MAP test for multimodality. J Classification 11 (1) 5-36 DOI: 10.1007/BF01201021
39. Minnotte MC (1997) Nonparametric testing of the existence of modes. Ann Statist 25 (4) 1646-1660
40. Hartigan JA, Mohanty S (1992) The RUNT test for multimodality. J Classifcation 9: 63-70
41. Andrushkiw RI, Klyushin DD, Petunin YI (2008) Theory Stoch Processes 14 (1) 1-6
42. Hartigan JA (1988) The span test of multimodality
43. Jackson PR, Tucker GT, Woods HF (1989) Testing for bimodality in frequency distributions of data suggesting polymorphisms of drug metabolism--hypothesis testing. Br J Clin Pharmacol 28(6) 655–662
44. http://www.amstat.org/sections/srms/Proceedings/y2002/Files/JSM2002-000150.pdf
45. https://engineering.purdue.edu/~bouman/software/cluster/
46. http://cran.r-project.org/web/packages/mixtools/index.html
47. http://cran.r-project.org/web/packages/discrimARTs/discrimARTs.pdf
48. http://cran.r-project.org/web/packages/flexmix/index.html
49. http://cran.r-project.org/web/packages/mclust/index.html
50. http://cran.r-project.org/web/packages/mixdist/index.html
1. REDIRECT Template:Probability distributions