On the Estimation of Confidence Intervals for Binomial Population Proportions in Astronomy: The Simplicity and Superiority of the Bayesian Approach
Ewan CameronA Department of Physics, Swiss Federal Institute of Technology (ETH Zurich), CH-8093 Zurich, Switzerland.
B Email: cameron@phys.ethz.ch
Publications of the Astronomical Society of Australia 28(2) 128-139 https://doi.org/10.1071/AS10046
Submitted: 03 December 10 Accepted: 01 March 2011 Published: 16 June 2011
Journal Compilation © Astronomical Society of Australia 2011
Abstract
I present a critical review of techniques for estimating confidence intervals on binomial population proportions inferred from success counts in small to intermediate samples. Population proportions arise frequently as quantities of interest in astronomical research; for instance, in studies aiming to constrain the bar fraction, active galactic nucleus fraction, supermassive black hole fraction, merger fraction, or red sequence fraction from counts of galaxies exhibiting distinct morphological features or stellar populations. However, two of the most widely-used techniques for estimating binomial confidence intervals — the ‘normal approximation’ and the Clopper & Pearson approach — are liable to misrepresent the degree of statistical uncertainty present under sampling conditions routinely encountered in astronomical surveys, leading to an ineffective use of the experimental data (and, worse, an inefficient use of the resources expended in obtaining that data). Hence, I provide here an overview of the fundamentals of binomial statistics with two principal aims: (i) to reveal the ease with which (Bayesian) binomial confidence intervals with more satisfactory behaviour may be estimated from the quantiles of the beta distribution using modern mathematical software packages (e.g. r, matlab, mathematica, idl, python); and (ii) to demonstrate convincingly the major flaws of both the ‘normal approximation’ and the Clopper & Pearson approach for error estimation.
Keywords: methods: data analysis — methods: statistical
1 Introduction
One problem frequently encountered in astronomical research is that of estimating a confidence interval (CI) on the value of an unknown population proportion based on the observed number of success counts in a given sample. The unknown population proportion may be, for instance, the intrinsic fraction of barred disk galaxies at a specific epoch to be inferred from the observed number of barred disks in a volume-limited sample (e.g., Elmegreen et al. 1990; van den Bergh 2002; Cameron et al. 2010; Nair & Abraham 2010), with the corresponding binomial CI used to evaluate the hypothesis that the bar fraction changes with redshift relative to a local benchmark (e.g., Cameron et al. 2010). Experiments to investigate the role of mass and environment in quenching star-formation via measurement of the galaxy red sequence fraction (e.g., Baldry et al. 2006; Hester et al. 2010; Ilbert et al. 2010), or to investigate whether or not major mergers were more frequent at high redshift via measurement of the close-pair/asymmetric fraction (e.g., De Propris et al. 2005; Conselice et al. 2008; López-Sanjuan et al. 2010), also routinely present this class of problem.
However, the two most commonly used methods for estimating CIs on binomial population proportions — the ‘normal approximation’ and the Clopper & Pearson (1934) approach — exhibit significant flaws under routine sampling conditions (cf. Vollset 1993; Santner 1998; Brown et al. 2001, 2002). In particular, the ‘normal approximation’ (also called the ‘Poisson error’) may systematically underestimate the CI width necessary to provide coverage at the desired level, especially for small samples, but even for rather large samples when the true population proportion is either very low or very high. If used naïvely the ‘normal approximation’ has the potential to mislead one into over-stating the significance of one’s inferences concerning the physical system under study formulated on the basis of the observed data.
Astronomers aware of these flaws in the ‘normal approximation’ often adopt the alternative Clopper & Pearson (1934) approach to CI estimation by way of reference to the CI tables in Gehrels (1986). Unfortunately, the Clopper & Pearson (1934) approach suffers from the opposite problem to that of the ‘normal approximation’ — namely, a systematic over-estimation of the CI width required to provide the desired coverage (Clopper & Pearson 1934; Neyman 1935; Gehrels 1986; Agresti & Coull 1998). In scientific research this over-estimation of the statistical measurement uncertainties may mislead one into placing insufficient confidence in the experimental outcomes, resulting in an inefficient use of the measured data (and hence the resources expended in obtaining that data). Indeed, it has been well argued by Agresti & Coull (1998) that in many practical applications even the ‘normal approximation’, despite its flaws, is preferable to the Clopper & Pearson (1934) approach.
Fortunately, there exist a multitude of alternative methods for generating CIs on binomial population proportions, many of which exhibit far more satisfactory behaviour than either the ‘normal approximation’ or the Clopper & Pearson (1934) approach — see Agresti & Coull (1998) and Brown et al. (2001) for various examples. Here I review both the theory and application of one of these methods — use of the beta distribution quantiles — deriving from a simple Bayesian analysis in which a uniform (‘non-informative’) prior is adopted for the true population proportion (e.g., Gelman et al. 2003). As I will demonstrate, the beta distribution generator for binomial CIs is both theoretically well motivated and easily applied in practice using widely available mathematical software packages (e.g., r, matlab, mathematica, idl, python). Ultimately, I advocate strongly that this strategy for estimating binomial CIs be adopted in future studies aiming to constrain fundamental population proportions in astronomical research (e.g., the galaxy bar fraction, red sequence fraction, or merger fraction) — especially for samples intrinsically of small to intermediate size, or when the subdivision of larger samples for analytical purposes produces sparsely populated data bins.
2 The Binomial Distribution
In probability theory any experiment for which there are only two possible random outcomes — success, occurring with probability p, or failure, occurring with probability q = (1 – p) — is referred to as a Bernoulli trial. Examples of Bernoulli trials in astronomical research may include asking whether or not a randomly sampled galaxy is barred, red-sequence, or merging. The probability, P, of observing a particular number of successes, k, in a series of n independent Bernoulli trials (with common success probability p) is governed by the binomial probability function: 1
where 0 ≤ k ≤ n, k ∈ ℤ (an integer), and
(see, for example, Quirin 1978). Note that the probabilities given by the n + 1 possible values of k correspond to the n + 1 terms of the binomial expansion of (p + q) n . The number of barred systems counted in a given sample of disk galaxies is a classic example of a binomially distributed variable in astronomy. The corresponding expectation value for the number of successes is with a variance of . Moreover, the expectation value for the fraction of successes, k/n, is equal to the Bernoulli trial success probability (also referred to as the ‘underlying population proportion’), p, and its variance is pq/n.
2.1 An Intermission: Just What Is a Confidence Interval?
As explained eloquently by both Kraft et al. (1991) and Ross (2003), there is a fundamental difference between the ‘classical’ and ‘Bayesian’ definitions of the term ‘confidence interval’. In classical statistical theory a binomial CI is defined as a pair of random variables, Pl and Pu , (with each random variable necessarily a finite, real-valued, measurable function; cf. Rao & Swift 2006) operating on the set of all possible experimental outcomes, , such that if the experiment were to be repeated by a sufficiently large number of independent observers then the fraction of observers for whom the true value of the underlying population proportion is covered by their realisation of these random variables — i.e., for whom P l (θ i ) = p l < p < p u = P u (θ i ) — is guaranteed to converge to (at least) a specific value, c, termed ‘the confidence level’. In the Bayesian paradigm, on the other hand, the underlying population proportion is treated as an unknown model parameter and the binomial CI defined as an interval, (p l , p u ), to which the experimenter believes may be assigned a probability, c, of containing the true value of p, based upon consideration of the likelihood function for p given the experimental data and the strength of any a priori beliefs or expectations regarding the system under study. (Indeed, acknowledging the significant conceptual differences between these alternative approaches to the binomial CI, the term ‘credible interval’ is often used instead in Bayesian analysis to avoid confusion with the classical nomenclature.) Importantly, as noted by Kraft et al. (1991), regardless of one’s philosophical position regarding these two statistical systems, ‘the Bayesian definition of confidence intervals reflects common astronomical usage better than the classical definition’.
Of the three binomial CI generators discussed in this review, only that attributed to Clopper & Pearson (1934) is consistent with the classical definition for all possible values of the underlying population proportion and sample size. However, I will argue that (at least) in the case of the binomial distribution, Bayesian CIs provide generally more satisfactory behaviour for astronomical purposes than their classical counterparts, even when evaluated against a performance diagnostic based on the classical definition — namely, the coverage fraction (or ‘effective coverage’) at given p and n.
3 The Beta Distribution Generator for Binomial Confidence Intervals
In astronomical data analysis it is standard practice to adopt the measured success fraction (also referred to as the ‘observed population proportion’), , as one’s ‘best guess’ of the underlying population proportion. In statistical terms, is employed as a point estimator for p. The likelihood of observing the result, , for a given value of p is, of course, proportional to pk q n–k . Normalisation of this likelihood function over 0 < p < 1 defines a ‘beta distribution’ with integer parameters a = k + 1 and b = n – k + 1:
where q = 1 – p (e.g., Gelman et al. 2003; Ross 2003). Differentiation of this likelihood function reveals that our best guess, , is in fact the maximum likelihood estimator of p. 2 The characteristic shape of the (beta distribution) likelihood function for p is illustrated in Figure 1 at a variety of ‘measured’ success fractions for samples of sizes n = 6 (left panel) and n = 36 (right panel). At small n, the likelihood function for p is markedly asymmetric (except where ), but at intermediate n it is visibly converging towards a narrow, symmetric, (pseudo-)normal distribution — the motivation behind the ‘normal approximation’ discussed in Section 4.
Given no a priori knowledge to inform one’s expectations regarding the experimental outcome, one may suppose that all values of p are equally ‘probable’. Formally, this condition is characterised via the Bayes–Laplace uniform prior, for which P prior(p) = 1 over 0 < p < 1. Application of Bayes’ theorem under this assumption allows one to treat the normalised likelihood function for p as a posterior probability distribution. Thus, the quantiles of the beta distribution from Eqn 2 may be used directly to estimate (Bayesian) confidence intervals on the underlying population proportion given the observed data. 3 Specifically, the lower and upper bounds, pl and pu , defining an ‘equal-tailed’ (or ‘central’) interval for p at a nominal confidence level of c = 1–α are given by the quantiles:
Note that the bounds of this ‘equal-tailed’ interval (which partition the probability of p greater than pu equal to that of p less than pl ) will be necessarily asymmetric about the maximum likelihood value, , (except at ) owing to the asymmetric nature of the (beta distribution) likelihood function for p (shown in Figure 1). As I will demonstrate below, binomial CIs generated in this manner have one rather desirable property, not shared by either the ‘normal approximation’ or the Clopper & Pearson (1934) approach — namely, their mean effective coverage is consistently very close to the nominal confidence level, even in the small sample size regime.
In the upper panel of Figure 2 , I examine first the effective coverage, ce , of ‘equal-tailed’ binomial CIs defined via the beta distribution for a range of population proportions and sample sizes (0.025 ≤ p ≤ 0.975 and 1 ≤ n ≤ 100) at a nominal level of cn ≈ 0.683 (1σ) — with the effective coverage defined as the fraction of samples drawn from the binomial probability function with given p and n for which the corresponding realisation of the CI under investigation encompasses the true population proportion. Thus, the effective coverage fractions, ce , presented here are computed as the sum of all binomial probabilities P(k, n, p) over {k : 0 ≤ k ≤ n, k ∈ ℤ}, for which the triad {k, n, p} produces a confidence interval, (pl , pu ), containing (covering) p.
One of the most striking features of this plot is the remarkable sensitivity of the effective coverage to the true underlying population proportion and sample size. This so-called ‘oscillation signature’ is an inherent property of all deterministic (i.e., non-randomising) generators for binomial CIs, arising from the discreteness of the binomial distribution. 4 Despite these oscillations it is clear that the beta distribution CIs do achieve an effective coverage close to (or slightly greater than) the desired confidence level over the vast majority of the parameter space explored here. Indeed, even at the extremes of p 1/6 and p 5/6, where the oscillations are initially rather large, there is evidently a rapid increase in coverage stability with increasing sample size, such that the ‘oscillation signature’ is vastly suppressed by n 40, and effectively eliminated (at least for 0.025 ≤ p ≤ 0.975) by n 80 (unlike in the case of the ‘normal approximation’ examined in Section 4).
In the lower panel of Figure 2, I examine the corresponding mean effective coverage (averaged uniformly over 0.025 ≤ p ≤ 0.975) as a function of sample size. Whereas the effective coverage at given p and n shown in the upper panel is consistent with the classical notion of confidence interval performance, the mean effective coverage may be considered a ‘Bayesian’ CI performance diagnostic — i.e., if one really does hold all p values equally ‘probable’ a priori, then one’s favoured CI generator should be at least expected to provide coverage consistent with the nominal level in the long-term average of all equivalent experiments. Inspection of the lower panel of Figure 2 confirms a very close agreement between the mean effective coverage of the beta distribution CI generator and the nominal confidence level, independent of n.
Most modern mathematical software packages provide robust, easy-to-use library functions for computing beta distribution quantiles (e.g., the qbeta routine in r; the quantile and betadistribution commands in mathematica; the betaincinv function in matlab; the ibeta function in idl; or the dist.beta.ppf function in python). Explicit code fragments demonstrating the implementation of these commands are provided in the Appendix to this paper, and I advocate strongly the use of these recipes for the computation of confidence intervals on binomial population proportions in future astronomical studies. In Tables 1 and 2 in the Appendix, I present compilations of ‘equal-tailed’ CIs generated in this manner at nominal confidence levels of 1σ and 3σ, respectively, for all possible observed success counts in sample sizes up to n = 20. These tables are intended both as a convenient reference for use directly in studies involving samples of 20 objects or less, and as a benchmark against which to confirm the correct implementation of the beta distribution CI generator for users newly adopting this technique.
A note on the above The (non-informative) Bayes–Laplace uniform prior may in fact be viewed as the special case of P prior(p) = B(1, 1) = 1 within a wider family of possible conjugate priors for the binomial population proportion based on the beta distribution. Another popular non-informative prior for p is the Jeffreys prior of P prior(p) = B(1/2, 1/2) 5 (cf. Brown et al. 2001; Gelman et al. 2003), which is, by design, proportional to the square root of the Fisher information. Application of the Jeffreys prior returns a posterior probability distribution for p of B(k + 1/2, n – k + 1/2). The performance of binomial CIs generated via beta distribution quantiles based on the Jeffreys prior differ insignificantly from those based on the uniform prior over 0.025 ≤ p ≤ 0.975 when n 2 — consistent with the description of both these priors as ‘non-informative’ (i.e., that even for small sample sizes the shape of the posterior probability distribution in both cases is strongly governed by the likelihood function of the observed data). (See the recent review by Cousins, Hymes & Tucker 2009 for a thorough evaluation of the performance of Bayesian CIs constructed with the Jeffreys prior.) Hence, while the specific results presented in this paper are computed exclusively using the uniform prior, for the purposes of our general discussion regarding the superiority of the beta distribution quantile technique over the ‘normal approximation’ and the Clopper & Pearson (1934) approach, these two non-informative priors may be considered interchangable.
4 The ‘Normal Approximation’
For a system with an underlying binomial population proportion, p, neither very close to 0 or 1, one may suppose (with reference to the central limit theorem) that the distribution of the statistic in a series of independent samples of a fixed ‘large’ size will follow approximately a normal distribution. Under the assumptions of this ‘normal approximation’ (also called the ‘Poisson error’) one may employ the standard Wald test criterion, established by Wald & Wolfowitz (1939), to construct a two-sided confidence interval for p. Specifically, at a confidence level of c = 1 – α one may expect that the true value of p lies within the interval
where , and z 1–α/2 is defined with reference to the standard normal distribution:
Values of z 1–α/2 for particular confidence levels may be obtained from reference tables in statistical textbooks (e.g., Quirin 1978) or computed within one’s favourite mathematical software package (e.g., the qnorm function in r). Of course, the most commonly used formula for constructing error bars on measured galaxy bar fractions, (e.g., Elmegreen et al. 1990), is simply the application of Eqn 4 at z 1–α/2 = 1, corresponding to a 1σ confidence level of c ≈ 0.683. The cases of z 1–α/2 = 2 and 3 (i.e., 2σ and 3σ errors) correspond to higher confidence levels of c ≈ 0.954 and 0.997, respectively.
As noted above, the key assumption behind this approach to binomial CI estimation — that the distribution of may be approximated via a normal distribution with mean p and variance pq/n — is reasonable only under the conditions of a ‘large’ sample size and p neither very close to 0 or 1. In Figure 3 I compare the distribution of the statistic (computed directly from the binomial probability function) against the shape of the corresponding ‘normal approximation’ for three different values of the underlying population proportion (p = 1/3, 1/2, and 5/6) and two different sample sizes (n = 6 and 36). In the small sample size example (n = 6) the ‘normal approximation’ provides a reasonable representation of the distribution at p = 1/3 and p = 1/2, but performs poorly at p = 5/6 (i.e., p close to 1). However, in the intermediate sample size example (n = 36) there is now a clear convergence towards a normal distribution in even at p = 5/6. These examples, presented in Figure 3, serve to illustrate the nature of deviations from ‘normality’ in the distribution of at small n and/or extreme p values. The impact of these deviations on the performance of the ‘normal approximation’ as a binomial CI generator is examined below.
In Figure 4 I present the effective coverage of binomial CIs estimated via the ‘normal approximation’ as a function of p and n at a nominal confidence level of cn ≈ 0.683 (1σ). As in the case of the beta distribution quantile approach described above, there is a clear ‘oscillation signature’ visible in this figure, reflecting a marked sensitivity in the coverage performance to the value of the underlying population proportion and sample size. 6 However, it is also evident that the ‘normal approximation’ suffers a systematic decline in performance both for small n and towards extreme values of p near 0 or 1, generating binomial CIs with effective coverage far below the desired level. The strict symmetry of the ‘normal approximation’ CI about the observed success fraction — which at low or high may even extend (unphysically) to p ≤ 0 or p ≥ 1 — regardless of the inherent asymmetry in the likelihood distribution for p (see Figure 1) is the principal cause of these coverage failures. The poor performance of the ‘normal approximation’ at small n is further highlighted in the corresponding plot of mean effective coverage against sample size shown in the lower panel of Figure 4. For the 1σ CIs examined here (and popularly adopted in studies of the galaxy bar fraction) the mean effective coverage of the ‘normal approximation’ is far below the nominal level for n 20, and should thus be strictly avoided in this small sample size regime. Indeed, although its mean effective coverage does ultimately improve with increasing n, one may be well advised to avoid the ‘normal approximation’ altogether in light of its poor effective coverage at extreme p values and the ready availability of a superior CI generator in the form of the (Bayesian) beta distribution quantiles described in Section 3.
The flaws in the ‘normal approximation’ as a CI generator described above were a great source of concern for statisticians in the 1930s, prompting the search for alternatives able to ensure universal coverage of at least the nominal level (thereby satisfying the classical definition of the term ‘confidence interval’) while remaining readily computable given the limited aids available at the time (such as reference tables of quantiles for standard distributions). The most popular of all such proposed alternatives was the Clopper & Pearson (1934) approach (cf. Gehrels 1986), which I review below.
5 The Clopper & Pearson Approach
In their landmark 1934 paper Clopper & Pearson presented a direct method for constructing ‘classical’ confidence intervals on inferred population proportions based on quantiles of the binomial probability function (Eqn 1), guaranteed to provide a coverage probability of at least (but likely exceeding) the nominal confidence level. The ‘two-sided’ (Clopper & Pearson 1934) CI at c = 1 – α is constructed by solving the following equations for the lower and upper bounds, Pl (k) = pl and Pu (k) = pu :
and
where k is again the observed number of successes (e.g., barred galaxies) in the sample, and n the total sample size. Note that in the extreme cases of or 1, the Clopper & Pearson (1934) formulae reduce simply to:
Modern mathematical software packages, such as r and matlab, support easy-to-use library functions (e.g., binom.test in the stats package in r; or binofit in the statistics toolbox in matlab) for computation of Clopper & Pearson (1934) confidence limits, which employ robust algorithms for the solution of Eqns 5 and 6. Alternatively, there exist numerous reference tables of pre-computed binomial CIs based on the Clopper & Pearson (1934) approach — most notably those of Gehrels (1986), a popular reference for estimating uncertainties in astronomical population proportions.
In the upper panel of Figure 5 , I examine the effective coverage of CIs generated via the Clopper & Pearson (1934) approach as a function of p and n at a nominal confidence level of c ≈ 0.683 (1σ). In contrast with the results for both the beta distribution and the ‘normal approximation’ reviewed above, the Clopper & Pearson (1934) CIs provide coverage far exceeding the nominal confidence level over much of this parameter space. The Clopper & Pearson (1934) coverage excess is also clearly evident in the corresponding mean effective coverage for this CI generator, plotted as a function of sample size in the lower panel of Figure 5. Although the Clopper & Pearson (1934) CIs do eventually converge to the nominal level at very large n, in the small to intermediate sample size regime their mean effective coverage is consistently far above the desired level. This point is in fact acknowledged in Gehrels (1986), although it appears not to be widely appreciated considering the frequency with which these CIs are treated as a ‘gold standard’ in astronomical papers.
6 Mean Confidence Interval Widths
To illustrate the influence of the choice of CI generator on the estimated magnitude of the relevant uncertainties (i.e., the error bar size), I compare in Figure 6 the mean widths of c ≈ 0.683 (1σ) CIs estimated via the (‘equal-tailed’) beta distribution quantile technique, the ‘normal approximation’, and the Clopper & Pearson (1934) approach as a function of p for samples of sizes n = 6 (left panel) and n = 36 (right panel). In the small sample size regime (where the ‘normal approximation’ fails to provide sufficient coverage at p 1/6 and p 5/6; see Figure 4) the mean CI widths are markedly smaller (by as much as Δ p ∼ –0.15) than those derived using the beta distribution technique (which generally provides superior coverage at these p values; see Figure 2). (Of course, the beta distribution should not be viewed as a strict benchmark for the ideal CI width, since its coverage is indeed prone to erratic performance at certain p values — the ‘oscillation signature’ to which all non-randomising binomial CI generators are prone; although, as we have argued above, its performance may be considered the best of the three approaches examined in this study.) In the intermediate sample size regime, the mean widths of these these two CI generators are in much better agreement, except at the extremes of p 1/20 and p 19/20 where a marked under-estimation is still evident in the ‘normal approximation’ CIs. The Clopper & Pearson (1934) CIs, on the other hand, exhibit a much greater mean width than those of the beta distribution or ‘normal approximation’, regardless of p — reflecting the substantial coverage excess demonstrated for this CI generator in Section 5 (see Figure 5). These examples verify that the choice of CI generator can indeed have a substantial impact on the magnitude of the estimated uncertainties, thereby confirming this choice to be an important practical consideration for effective astronomical data analysis.
7 Conclusions
I have reviewed the performance of three alternative methods for estimating confidence intervals on binomial population proportions; namely, the beta distribution quantile technique, the ‘normal approximation’, and the Clopper & Pearson (1934) approach (cf. Gehrels 1986). Despite their current popularity in astronomical research, the latter two CI generators are demonstrated to perform poorly under sampling conditions routinely encountered in observational studies, with the ‘normal approximation’ failing to provide CIs of sufficient width to achieve coverage at the nominal confidence level, and the Clopper & Pearson (1934) approach producing CIs far wider than necessary to achieve the nominal coverage. In contrast, the (Bayesian) beta distribution quantile technique is revealed to be a well-motivated alternative, consistently providing a mean level of coverage close to the nominal level, even for small-to-intermediate sample sizes. Given that the beta distribution generator for binomial CIs may be easily implemented using modern mathematical software packages, I advocate strongly that this technique be adopted in future studies aiming to constrain the true values of astronomical propulation proportions (e.g., the galaxy bar fraction, red sequence fraction, or merger fraction).
Acknowledgments
The author would like to thank Matthew Prescott for his assistance in defining the idl code fragment, Carlos López for supplying the python code fragment, Roban Kramer for numerous helpful discussions on the role of statistics in astronomy, and the anonymous referee for a thorough reading of the paper and many insightful comments.
References
Agresti, A. and Coull, B. A., 1998, The American Statistician, 52, 119| Crossref | GoogleScholarGoogle Scholar |
Baldry, I. K., Balogh, M. L., Bower, R. G., Glazebrook, K., Nicol, R. C., Bamford, S. P. and Budavari, T., 2006, MNRAS, 373, 469
| Crossref | GoogleScholarGoogle Scholar |
Burgasser, A. J., Kirkpatrick, J. D., Reid, N. I., Brown, M. E., Miskey, C. L. and Gizis, J. E., 2003, ApJ, 586, 512
| Crossref | GoogleScholarGoogle Scholar |
Brown, L. D., Cai, T. T. and DasGupta, A., 2001, Statistical Science, 16, 101
Brown, L. D., Cai, T. T. and DasGupta, A., 2002, The Annals of Statistics, 30, 160
| Crossref | GoogleScholarGoogle Scholar |
Cameron, E. et al., 2010, MNRAS, 409, 346
| Crossref | GoogleScholarGoogle Scholar |
Clopper, C. J. and Pearson, E. S., 1934, Biometrika, 26, 404
| Crossref | GoogleScholarGoogle Scholar |
Conselice, C. J., Rajgor, S. and Myers, R., 2008, , , 909
Cousins, R. D., Hymes, K. E. and Tucker, T., 2009, NIM, 612, 388
| Crossref | GoogleScholarGoogle Scholar |
De Propris, R., Liske, J., Driver, S. P., Allen, P. D. and Cross, N. J. G., 2005, ApJ, 130, 1516
Elmegreen, D. M., Elmegreen, B. G. and Bellin, A. D., 1990, ApJ, 364, 415
| Crossref | GoogleScholarGoogle Scholar |
Gehrels, N., 1986, ApJ, 303, 336
| Crossref | GoogleScholarGoogle Scholar | 1:CAS:528:DyaL28XhvFKgsbY%3D&md5=3e027fefcff229d2c5619efab98bd7bdCAS |
Gelman A. , Carlin J. B. , Stern H. S. & Rubin D. B. , 2003, Bayesian Data Analysis, (New York: Chapman & Hall)
Hester, J. A., 2010, ApJ, 720, 191
| Crossref | GoogleScholarGoogle Scholar | 1:CAS:528:DC%2BC3cXhtF2kt7zK&md5=9de365d714c93872907a1fbc3439a5a8CAS |
Ilbert, O. et al., 2010, ApJ, 709, 644
| Crossref | GoogleScholarGoogle Scholar | 1:CAS:528:DC%2BC3cXisFCht7w%3D&md5=c8e25b9fc83ec7662b88c6a2551af477CAS |
Kraft, R. P., Burrows, D. N. and Nousek, J. A., 1991, ApJ, 374, 344
| Crossref | GoogleScholarGoogle Scholar |
Quirin W. L. , 1978, Probability and Statistics (New York: Harper & Row Publishers)
López-Sanjuan, C., Balcells, M., Pérez-González, P. G., Barro, G., Gallego, J. and Zamorano, J., 2010, A&A, 518, 20
| Crossref | GoogleScholarGoogle Scholar |
Nair, P. B. and Abraham, R. G., 2010, ApJL, 714, L260
| Crossref | GoogleScholarGoogle Scholar |
Neyman, J., 1935, The Annals of Mathematical Statistics, 6, 111
| Crossref | GoogleScholarGoogle Scholar |
Rao, M. M. and Swift, R. J., 2006, Mathematics and Its Applications, , 582
Ross, T. D., 2003, Computers in Biology and Medicine, 33, 509
| Crossref | GoogleScholarGoogle Scholar | 12878234PubMed |
Santner, T. J., 1998, Teaching Statistics, 20, 20–23
| Crossref | GoogleScholarGoogle Scholar |
van den Bergh, S., 2002, AJ, 124, 782
| Crossref | GoogleScholarGoogle Scholar |
Vollset, S. E., 1993, Statistics in Medicine, 12, 809
| Crossref | GoogleScholarGoogle Scholar | 1:STN:280:DyaK3szhvVSiug%3D%3D&md5=4a6f98f79cd241216088dc80e1310ed5CAS | 8327801PubMed |
Wald, A. and Wolfowitz, J., 1939, The Annals of Mathematical Statistics, 10, 105
| Crossref | GoogleScholarGoogle Scholar |
1 One may note that the correct terminology in a statistical context is actually ‘binomial probability mass function’, owing to the discrete nature of the binomial distribution, i.e., that there exist a finite number of possible k values (the integers from 0 to n, inclusive) to which non-zero probabilities may be assigned. (As distinct from the alternative case of a ‘probability density function’, such as the Bayesian posterior probability distribution for p considered in Section 3, for which non-zero probabilities may only be assigned to measurable intervals on the real number line, and not individual — or even countable sets of — real numbers.) Nevertheless, to avoid any confusion with the more commonly used definition of the term ‘mass function’ in astronomy I adopt the shorter expression ‘binomial probability function’ herein.
2 Technically, when (or 1) the likelihood function for p has no zero first derivative on the open interval, (0, 1), although the function itself is indeed strictly increasing as p → 0 (or 1). In this case one may choose to adopt the median (50% quantile) of the (beta distribution) likelihood function as one’s best guess for p, or else to compute a ‘one-sided’ confidence interval bounding p instead. In either case, one proceeds using similar principles.
3 Astronomers familiar with the work of Burgasser et al. (2003) on binarity in brown dwarfs may be familiar with the algorithm for recovering confidence intervals on p given in their Appendix, which is in fact equivalent to the Bayesian approach with uniform prior presented here (although Burgasser et al. 2003 make no explicit reference to either Bayes’ theorem or the beta distribution).
4 Brown et al. (2001) describe the ‘oscillation signature’ as the challenge of ‘lucky p, lucky n’ — namely that for certain (‘lucky’) combinations of underlying population proportion and sample size there exist two almost equally likely values closely straddling the true p. For instance, if p = 1/5 and one has a sample of size n = 3, the possible values are 0, 1/3, 2/3, and 1, occurring with frequencies 0.512, 0.384, 0.096, and 0.008, respectively. Tailoring a binomial CI specifically to this situation, one could define and (with ϵ an arbitrarily small constant necessary to ensure p is contained within the open interval (pl , pu ) for k = 0 and 1), returning an effective coverage of ce = 0.512 + 0.384 = 0.896. However, applying the same CI generator to a system with p = 1/3 (and again n = 3) for which the possible values occur with frequencies 0.296, 0.444, 0.222, and 0.037 (rounded to 3 decimal places), one obtains an effective coverage of only ce = 0.444! For further discussion of the impact of the ‘oscillation signature’ on binomial CIs the interested reader is referred to Agresti & Coull (1998) and Brown et al. (2001, 2002).
5 Note that the factorial functions used in the beta distribution definition of Eqn 2 must be replaced by gamma functions according to the relation (m)! = Γ(m +1) in order to handle the non-integer input in this case.
6 It is important also to note that this ‘oscillation signature’ is evident even in binomial CIs generated via the ‘normal approximation’ at very large sample sizes, as thoroughly demonstrated by Brown et al. (2001, 2002). For instance, Brown et al. (2001) describe the erratic behaviour of the ‘normal approximation’ coverage at a nominal level of cn = 0.95 for a system with p = 0.005, whereby there is a steady convergence in ce towards 0.95 for n increasing until n = 592, at which point coverage falls suddenly to ce = 0.792! Similarly, Brown et al. (2002) demonstrate that in order to ensure coverage stays at or above a nominal level of cn = 0.93 for a system with p = 0.1 using the ‘normal approximation’ one requires a sample size of at least n = 286, whereas for the Bayesian (Jeffreys non-informative prior) CI this criterion is satisfied by n = 47.
A CI Code Fragments & CI Reference Tables
Here I provide simple code fragments demonstrating the implementation of the beta distribution CI generator via standard library routines in r, matlab, mathematica, idl, and python. The correct performance of these code fragments in one’s preferred mathematical software package may be verified by comparison against the reference tables of binomial CIs presented here in Tables 1 and 2. As in the main body of this paper I denote the nominal confidence level c, the observed success count k, and the sample size n. In the following it is assumed that these variables have already been defined computationally by the user with c a real/double and k and n integers.
-
p_lower < - qbeta((1-c)/2,k+1,n-k+1)
-
p_upper <- qbeta(1-(1-c)/2,k+1,n-k+1)
-
p_lower = betaincinv((1-c)/2,k+1,n-k+1)
-
p_upper = betaincinv(1-(1-c)/2,k+1,n-k+1)
-
plower = Quantile[BetaDistribution[k+1,n-k+1],(1-c)/2]
-
pupper = Quantile[BetaDistribution[k+1,n-k+1],1-(1-c)/2]
-
p_lower = IMSL_BETACDF((1-c)/2,k+1,n-k+1,/INVERSE)
-
p_upper = IMSL_BETACDF(1-(1-c)/2,k+1,n-k+1,/INVERSE) otherwise, iteratively:
-
z = FINDGEN(10000)*0.0001
-
Beta = IBETA(k+1,n-k+1,z)
-
il = VALUE_LOCATE(Beta,(1-c)/2)
-
iu = VALUE_LOCATE(Beta,1-(1-c)/2)
-
p_lower = z[il]
-
p_upper = z[iu]
-
import scipy.stats.distributions as dist
-
p_lower = dist.beta.ppf((1-c)/2.,k+1,n-k+1)
-
p_upper = dist.beta.ppf(1-(1-c)/2.,k+1,n-k+1)