fbpx
Wikipedia

Checking whether a coin is fair

In statistics, the question of checking whether a coin is fair is one whose importance lies, firstly, in providing a simple problem on which to illustrate basic ideas of statistical inference and, secondly, in providing a simple problem that can be used to compare various competing methods of statistical inference, including decision theory. The practical problem of checking whether a coin is fair might be considered as easily solved by performing a sufficiently large number of trials, but statistics and probability theory can provide guidance on two types of question; specifically those of how many trials to undertake and of the accuracy of an estimate of the probability of turning up heads, derived from a given sample of trials.

A fair coin is an idealized randomizing device with two states (usually named "heads" and "tails") which are equally likely to occur. It is based on the coin flip used widely in sports and other situations where it is required to give two parties the same chance of winning. Either a specially designed chip or more usually a simple currency coin is used, although the latter might be slightly "unfair" due to an asymmetrical weight distribution, which might cause one state to occur more frequently than the other, giving one party an unfair advantage.[1] So it might be necessary to test experimentally whether the coin is in fact "fair" – that is, whether the probability of the coin's falling on either side when it is tossed is exactly 50%. It is of course impossible to rule out arbitrarily small deviations from fairness such as might be expected to affect only one flip in a lifetime of flipping; also it is always possible for an unfair (or "biased") coin to happen to turn up exactly 10 heads in 20 flips. Therefore, any fairness test must only establish a certain degree of confidence in a certain degree of fairness (a certain maximum bias). In more rigorous terminology, the problem is of determining the parameters of a Bernoulli process, given only a limited sample of Bernoulli trials.

Preamble edit

This article describes experimental procedures for determining whether a coin is fair or unfair. There are many statistical methods for analyzing such an experimental procedure. This article illustrates two of them.

Both methods prescribe an experiment (or trial) in which the coin is tossed many times and the result of each toss is recorded. The results can then be analysed statistically to decide whether the coin is "fair" or "probably not fair".

  • Posterior probability density function, or PDF (Bayesian approach). Initially, the true probability of obtaining a particular side when a coin is tossed is unknown, but the uncertainty is represented by the "prior distribution". The theory of Bayesian inference is used to derive the posterior distribution by combining the prior distribution and the likelihood function which represents the information obtained from the experiment. The probability that this particular coin is a "fair coin" can then be obtained by integrating the PDF of the posterior distribution over the relevant interval that represents all the probabilities that can be counted as "fair" in a practical sense.
  • Estimator of true probability (Frequentist approach). This method assumes that the experimenter can decide to toss the coin any number of times. The experimenter first decides on the level of confidence required and the tolerable margin of error. These parameters determine the minimum number of tosses that must be performed to complete the experiment.

An important difference between these two approaches is that the first approach gives some weight to one's prior experience of tossing coins, while the second does not. The question of how much weight to give to prior experience, depending on the quality (credibility) of that experience, is discussed under credibility theory.

Posterior probability density function edit

One method is to calculate the posterior probability density function of Bayesian probability theory.

A test is performed by tossing the coin N times and noting the observed numbers of heads, h, and tails, t. The symbols H and T represent more generalised variables expressing the numbers of heads and tails respectively that might have been observed in the experiment. Thus N = H + T = h + t.

Next, let r be the actual probability of obtaining heads in a single toss of the coin. This is the property of the coin which is being investigated. Using Bayes' theorem, the posterior probability density of r conditional on h and t is expressed as follows:

 

where g(r) represents the prior probability density distribution of r, which lies in the range 0 to 1.

The prior probability density distribution summarizes what is known about the distribution of r in the absence of any observation. We will assume that the prior distribution of r is uniform over the interval [0, 1]. That is, g(r) = 1. (In practice, it would be more appropriate to assume a prior distribution which is much more heavily weighted in the region around 0.5, to reflect our experience with real coins.)

The probability of obtaining h heads in N tosses of a coin with a probability of heads equal to r is given by the binomial distribution:

 

Substituting this into the previous formula:

 

This is in fact a beta distribution (the conjugate prior for the binomial distribution), whose denominator can be expressed in terms of the beta function:

 

As a uniform prior distribution has been assumed, and because h and t are integers, this can also be written in terms of factorials:

 

Example edit

For example, let N = 10, h = 7, i.e. the coin is tossed 10 times and 7 heads are obtained:

 

The graph on the right shows the probability density function of r given that 7 heads were obtained in 10 tosses. (Note: r is the probability of obtaining heads when tossing the same coin once.)

 
Plot of the probability density f(r | H = 7, T = 3) = 1320 r7 (1 − r)3 with r ranging from 0 to 1

The probability for an unbiased coin (defined for this purpose as one whose probability of coming down heads is somewhere between 45% and 55%)

 

is small when compared with the alternative hypothesis (a biased coin). However, it is not small enough to cause us to believe that the coin has a significant bias. This probability is slightly higher than our presupposition of the probability that the coin was fair corresponding to the uniform prior distribution, which was 10%. Using a prior distribution that reflects our prior knowledge of what a coin is and how it acts, the posterior distribution would not favor the hypothesis of bias. However the number of trials in this example (10 tosses) is very small, and with more trials the choice of prior distribution would be somewhat less relevant.)

With the uniform prior, the posterior probability distribution f(r | H = 7,T = 3) achieves its peak at r = h / (h + t) = 0.7; this value is called the maximum a posteriori (MAP) estimate of r. Also with the uniform prior, the expected value of r under the posterior distribution is

 


Estimator of true probability edit

The best estimator for the actual value   is the estimator  .

This estimator has a margin of error (E) where   at a particular confidence level.

Using this approach, to decide the number of times the coin should be tossed, two parameters are required:

  1. The confidence level which is denoted by confidence interval (Z)
  2. The maximum (acceptable) error (E)
  • The confidence level is denoted by Z and is given by the Z-value of a standard normal distribution. This value can be read off a standard score statistics table for the normal distribution. Some examples are:
Z value Confidence level Comment
0.6745 gives 50.000% level of confidence Half
1.0000 gives 68.269% level of confidence One std dev
1.6449 gives 90.000% level of confidence "One nine"
1.9599 gives 95.000% level of confidence 95 percent
2.0000 gives 95.450% level of confidence Two std dev
2.5759 gives 99.000% level of confidence "Two nines"
3.0000 gives 99.730% level of confidence Three std dev
3.2905 gives 99.900% level of confidence "Three nines"
3.8906 gives 99.990% level of confidence "Four nines"
4.0000 gives 99.993% level of confidence Four std dev
4.4172 gives 99.999% level of confidence "Five nines"
  • The maximum error (E) is defined by   where   is the estimated probability of obtaining heads. Note:   is the same actual probability (of obtaining heads) as   of the previous section in this article.
  • In statistics, the estimate of a proportion of a sample (denoted by p) has a standard error given by:
 

where n is the number of trials (which was denoted by N in the previous section).

This standard error   function of p has a maximum at  . Further, in the case of a coin being tossed, it is likely that p will be not far from 0.5, so it is reasonable to take p=0.5 in the following:

   

And hence the value of maximum error (E) is given by

 

Solving for the required number of coin tosses, n,

 

Examples edit

1. If a maximum error of 0.01 is desired, how many times should the coin be tossed?

 
  at 68.27% level of confidence (Z=1)
  at 95.45% level of confidence (Z=2)
  at 99.90% level of confidence (Z=3.3)

2. If the coin is tossed 10000 times, what is the maximum error of the estimator   on the value of   (the actual probability of obtaining heads in a coin toss)?

 
 
  at 68.27% level of confidence (Z=1)
  at 95.45% level of confidence (Z=2)
  at 99.90% level of confidence (Z=3.3)

3. The coin is tossed 12000 times with a result of 5961 heads (and 6039 tails). What interval does the value of   (the true probability of obtaining heads) lie within if a confidence level of 99.999% is desired?

 

Now find the value of Z corresponding to 99.999% level of confidence.

 

Now calculate E

 

The interval which contains r is thus:

 
 

Other approaches edit

Other approaches to the question of checking whether a coin is fair are available using decision theory, whose application would require the formulation of a loss function or utility function which describes the consequences of making a given decision. An approach that avoids requiring either a loss function or a prior probability (as in the Bayesian approach) is that of "acceptance sampling".[2]

Other applications edit

The above mathematical analysis for determining if a coin is fair can also be applied to other uses. For example:

  • Determining the proportion of defective items for a product subjected to a particular (but well defined) condition. Sometimes a product can be very difficult or expensive to produce. Furthermore, if testing such products will result in their destruction, a minimum number of items should be tested. Using a similar analysis, the probability density function of the product defect rate can be found.
  • Two party polling. If a small random sample poll is taken where there are only two mutually exclusive choices, then this is similar to tossing a single coin multiple times using a possibly biased coin. A similar analysis can therefore be applied to determine the confidence to be ascribed to the actual ratio of votes cast. (If people are allowed to abstain then the analysis must take account of that, and the coin-flip analogy doesn't quite hold.)
  • Determining the sex ratio in a large group of an animal species. Provided that a small random sample (i.e. small in comparison with the total population) is taken when performing the random sampling of the population, the analysis is similar to determining the probability of obtaining heads in a coin toss.

See also edit

References edit

  1. ^ However, if the coin is caught rather than allowed to bounce or spin, it is difficult to bias a coin flip's outcome. See Gelman, Andrew; Deborah Nolan (2002). "Teacher's Corner: You Can Load a Die, But You Can't Bias a Coin". American Statistician. 56 (4): 308–311. doi:10.1198/000313002605. S2CID 123597087.
  2. ^ Cox, D.R., Hinkley, D.V. (1974) Theoretical Statistics (Example 11.7), Chapman & Hall. ISBN 0-412-12420-3
  • Guttman, Wilks, and Hunter: Introductory Engineering Statistics, John Wiley & Sons, Inc. (1971) ISBN 0-471-33770-6
  • Devinder Sivia: Data Analysis, a Bayesian Tutorial, Oxford University Press (1996) ISBN 0-19-851889-7

checking, whether, coin, fair, this, article, includes, list, general, references, lacks, sufficient, corresponding, inline, citations, please, help, improve, this, article, introducing, more, precise, citations, january, 2010, learn, when, remove, this, messa. This article includes a list of general references but it lacks sufficient corresponding inline citations Please help to improve this article by introducing more precise citations January 2010 Learn how and when to remove this message In statistics the question of checking whether a coin is fair is one whose importance lies firstly in providing a simple problem on which to illustrate basic ideas of statistical inference and secondly in providing a simple problem that can be used to compare various competing methods of statistical inference including decision theory The practical problem of checking whether a coin is fair might be considered as easily solved by performing a sufficiently large number of trials but statistics and probability theory can provide guidance on two types of question specifically those of how many trials to undertake and of the accuracy of an estimate of the probability of turning up heads derived from a given sample of trials A fair coin is an idealized randomizing device with two states usually named heads and tails which are equally likely to occur It is based on the coin flip used widely in sports and other situations where it is required to give two parties the same chance of winning Either a specially designed chip or more usually a simple currency coin is used although the latter might be slightly unfair due to an asymmetrical weight distribution which might cause one state to occur more frequently than the other giving one party an unfair advantage 1 So it might be necessary to test experimentally whether the coin is in fact fair that is whether the probability of the coin s falling on either side when it is tossed is exactly 50 It is of course impossible to rule out arbitrarily small deviations from fairness such as might be expected to affect only one flip in a lifetime of flipping also it is always possible for an unfair or biased coin to happen to turn up exactly 10 heads in 20 flips Therefore any fairness test must only establish a certain degree of confidence in a certain degree of fairness a certain maximum bias In more rigorous terminology the problem is of determining the parameters of a Bernoulli process given only a limited sample of Bernoulli trials Contents 1 Preamble 2 Posterior probability density function 2 1 Example 3 Estimator of true probability 3 1 Examples 4 Other approaches 5 Other applications 6 See also 7 ReferencesPreamble editThis article describes experimental procedures for determining whether a coin is fair or unfair There are many statistical methods for analyzing such an experimental procedure This article illustrates two of them Both methods prescribe an experiment or trial in which the coin is tossed many times and the result of each toss is recorded The results can then be analysed statistically to decide whether the coin is fair or probably not fair Posterior probability density function or PDF Bayesian approach Initially the true probability of obtaining a particular side when a coin is tossed is unknown but the uncertainty is represented by the prior distribution The theory of Bayesian inference is used to derive the posterior distribution by combining the prior distribution and the likelihood function which represents the information obtained from the experiment The probability that this particular coin is a fair coin can then be obtained by integrating the PDF of the posterior distribution over the relevant interval that represents all the probabilities that can be counted as fair in a practical sense Estimator of true probability Frequentist approach This method assumes that the experimenter can decide to toss the coin any number of times The experimenter first decides on the level of confidence required and the tolerable margin of error These parameters determine the minimum number of tosses that must be performed to complete the experiment An important difference between these two approaches is that the first approach gives some weight to one s prior experience of tossing coins while the second does not The question of how much weight to give to prior experience depending on the quality credibility of that experience is discussed under credibility theory Posterior probability density function editOne method is to calculate the posterior probability density function of Bayesian probability theory A test is performed by tossing the coin N times and noting the observed numbers of heads h and tails t The symbols H and T represent more generalised variables expressing the numbers of heads and tails respectively that might have been observed in the experiment Thus N H T h t Next let r be the actual probability of obtaining heads in a single toss of the coin This is the property of the coin which is being investigated Using Bayes theorem the posterior probability density of r conditional on h and t is expressed as follows f r H h T t Pr H h r N h t g r 0 1 Pr H h p N h t g p d p displaystyle f r mid H h T t frac Pr H h mid r N h t g r int 0 1 Pr H h mid p N h t g p dp nbsp where g r represents the prior probability density distribution of r which lies in the range 0 to 1 The prior probability density distribution summarizes what is known about the distribution of r in the absence of any observation We will assume that the prior distribution of r is uniform over the interval 0 1 That is g r 1 In practice it would be more appropriate to assume a prior distribution which is much more heavily weighted in the region around 0 5 to reflect our experience with real coins The probability of obtaining h heads in N tosses of a coin with a probability of heads equal to r is given by the binomial distribution Pr H h r N h t N h r h 1 r t displaystyle Pr H h mid r N h t N choose h r h 1 r t nbsp Substituting this into the previous formula f r H h T t N h r h 1 r t 0 1 N h p h 1 p t d p r h 1 r t 0 1 p h 1 p t d p displaystyle f r mid H h T t frac N choose h r h 1 r t int 0 1 N choose h p h 1 p t dp frac r h 1 r t int 0 1 p h 1 p t dp nbsp This is in fact a beta distribution the conjugate prior for the binomial distribution whose denominator can be expressed in terms of the beta function f r H h T t 1 B h 1 t 1 r h 1 r t displaystyle f r mid H h T t frac 1 mathrm B h 1 t 1 r h 1 r t nbsp As a uniform prior distribution has been assumed and because h and t are integers this can also be written in terms of factorials f r H h T t h t 1 h t r h 1 r t displaystyle f r mid H h T t frac h t 1 h t r h 1 r t nbsp Example edit For example let N 10 h 7 i e the coin is tossed 10 times and 7 heads are obtained f r H 7 T 3 10 1 7 3 r 7 1 r 3 1320 r 7 1 r 3 displaystyle f r mid H 7 T 3 frac 10 1 7 3 r 7 1 r 3 1320 r 7 1 r 3 nbsp The graph on the right shows the probability density function of r given that 7 heads were obtained in 10 tosses Note r is the probability of obtaining heads when tossing the same coin once nbsp Plot of the probability density f r H 7 T 3 1320 r7 1 r 3 with r ranging from 0 to 1 The probability for an unbiased coin defined for this purpose as one whose probability of coming down heads is somewhere between 45 and 55 Pr 0 45 lt r lt 0 55 0 45 0 55 f p H 7 T 3 d p 13 displaystyle Pr 0 45 lt r lt 0 55 int 0 45 0 55 f p mid H 7 T 3 dp approx 13 nbsp is small when compared with the alternative hypothesis a biased coin However it is not small enough to cause us to believe that the coin has a significant bias This probability is slightly higher than our presupposition of the probability that the coin was fair corresponding to the uniform prior distribution which was 10 Using a prior distribution that reflects our prior knowledge of what a coin is and how it acts the posterior distribution would not favor the hypothesis of bias However the number of trials in this example 10 tosses is very small and with more trials the choice of prior distribution would be somewhat less relevant With the uniform prior the posterior probability distribution f r H 7 T 3 achieves its peak at r h h t 0 7 this value is called the maximum a posteriori MAP estimate of r Also with the uniform prior the expected value of r under the posterior distribution is E r 0 1 r f r H 7 T 3 d r h 1 h t 2 2 3 displaystyle operatorname E r int 0 1 r cdot f r mid H 7 T 3 mathrm d r frac h 1 h t 2 frac 2 3 nbsp Estimator of true probability editThe best estimator for the actual value r displaystyle r nbsp is the estimator p h h t displaystyle p frac h h t nbsp This estimator has a margin of error E where p r lt E displaystyle p r lt E nbsp at a particular confidence level Using this approach to decide the number of times the coin should be tossed two parameters are required The confidence level which is denoted by confidence interval Z The maximum acceptable error E The confidence level is denoted by Z and is given by the Z value of a standard normal distribution This value can be read off a standard score statistics table for the normal distribution Some examples are Z value Confidence level Comment 0 6745 gives 50 000 level of confidence Half 1 0000 gives 68 269 level of confidence One std dev 1 6449 gives 90 000 level of confidence One nine 1 9599 gives 95 000 level of confidence 95 percent 2 0000 gives 95 450 level of confidence Two std dev 2 5759 gives 99 000 level of confidence Two nines 3 0000 gives 99 730 level of confidence Three std dev 3 2905 gives 99 900 level of confidence Three nines 3 8906 gives 99 990 level of confidence Four nines 4 0000 gives 99 993 level of confidence Four std dev 4 4172 gives 99 999 level of confidence Five nines The maximum error E is defined by p r lt E displaystyle p r lt E nbsp where p displaystyle p nbsp is the estimated probability of obtaining heads Note r displaystyle r nbsp is the same actual probability of obtaining heads as r displaystyle r nbsp of the previous section in this article In statistics the estimate of a proportion of a sample denoted by p has a standard error given by s p p 1 p n displaystyle s p sqrt frac p 1 p n nbsp where n is the number of trials which was denoted by N in the previous section This standard error s p displaystyle s p nbsp function of p has a maximum at p 1 p 0 5 displaystyle p 1 p 0 5 nbsp Further in the case of a coin being tossed it is likely that p will be not far from 0 5 so it is reasonable to take p 0 5 in the following s p displaystyle s p nbsp p 1 p n 0 5 0 5 n 1 2 n displaystyle sqrt frac p 1 p n leq sqrt frac 0 5 times 0 5 n frac 1 2 sqrt n nbsp And hence the value of maximum error E is given by E Z s p Z 2 n displaystyle E Z s p frac Z 2 sqrt n nbsp Solving for the required number of coin tosses n n Z 2 4 E 2 displaystyle n frac Z 2 4 E 2 nbsp Examples edit 1 If a maximum error of 0 01 is desired how many times should the coin be tossed n Z 2 4 E 2 Z 2 4 0 01 2 2500 Z 2 displaystyle n frac Z 2 4 E 2 frac Z 2 4 times 0 01 2 2500 Z 2 nbsp n 2500 displaystyle n 2500 nbsp at 68 27 level of confidence Z 1 n 10000 displaystyle n 10000 nbsp at 95 45 level of confidence Z 2 n 27225 displaystyle n 27225 nbsp at 99 90 level of confidence Z 3 3 2 If the coin is tossed 10000 times what is the maximum error of the estimator p displaystyle p nbsp on the value of r displaystyle r nbsp the actual probability of obtaining heads in a coin toss E Z 2 n displaystyle E frac Z 2 sqrt n nbsp E Z 2 10000 Z 200 displaystyle E frac Z 2 sqrt 10000 frac Z 200 nbsp E 0 0050 displaystyle E 0 0050 nbsp at 68 27 level of confidence Z 1 E 0 0100 displaystyle E 0 0100 nbsp at 95 45 level of confidence Z 2 E 0 0165 displaystyle E 0 0165 nbsp at 99 90 level of confidence Z 3 3 3 The coin is tossed 12000 times with a result of 5961 heads and 6039 tails What interval does the value of r displaystyle r nbsp the true probability of obtaining heads lie within if a confidence level of 99 999 is desired p h h t 5961 12000 0 4968 displaystyle p frac h h t frac 5961 12000 0 4968 nbsp Now find the value of Z corresponding to 99 999 level of confidence Z 4 4172 displaystyle Z 4 4172 nbsp Now calculate E E Z 2 n 4 4172 2 12000 0 0202 displaystyle E frac Z 2 sqrt n frac 4 4172 2 sqrt 12000 0 0202 nbsp The interval which contains r is thus p E lt r lt p E displaystyle p E lt r lt p E nbsp 0 4766 lt r lt 0 5170 displaystyle 0 4766 lt r lt 0 5170 nbsp Other approaches editOther approaches to the question of checking whether a coin is fair are available using decision theory whose application would require the formulation of a loss function or utility function which describes the consequences of making a given decision An approach that avoids requiring either a loss function or a prior probability as in the Bayesian approach is that of acceptance sampling 2 Other applications editThe above mathematical analysis for determining if a coin is fair can also be applied to other uses For example Determining the proportion of defective items for a product subjected to a particular but well defined condition Sometimes a product can be very difficult or expensive to produce Furthermore if testing such products will result in their destruction a minimum number of items should be tested Using a similar analysis the probability density function of the product defect rate can be found Two party polling If a small random sample poll is taken where there are only two mutually exclusive choices then this is similar to tossing a single coin multiple times using a possibly biased coin A similar analysis can therefore be applied to determine the confidence to be ascribed to the actual ratio of votes cast If people are allowed to abstain then the analysis must take account of that and the coin flip analogy doesn t quite hold Determining the sex ratio in a large group of an animal species Provided that a small random sample i e small in comparison with the total population is taken when performing the random sampling of the population the analysis is similar to determining the probability of obtaining heads in a coin toss See also editBinomial test Coin flipping Confidence interval Estimation theory Inferential statistics Loaded dice Margin of error Point estimation Statistical randomnessReferences edit However if the coin is caught rather than allowed to bounce or spin it is difficult to bias a coin flip s outcome See Gelman Andrew Deborah Nolan 2002 Teacher s Corner You Can Load a Die But You Can t Bias a Coin American Statistician 56 4 308 311 doi 10 1198 000313002605 S2CID 123597087 Cox D R Hinkley D V 1974 Theoretical Statistics Example 11 7 Chapman amp Hall ISBN 0 412 12420 3 Guttman Wilks and Hunter Introductory Engineering Statistics John Wiley amp Sons Inc 1971 ISBN 0 471 33770 6 Devinder Sivia Data Analysis a Bayesian Tutorial Oxford University Press 1996 ISBN 0 19 851889 7 Retrieved from https en wikipedia org w index php title Checking whether a coin is fair amp oldid 1205738026, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.