Coupon collector's problem

Graph of number of coupons, n vs the expected number of tries needed to collect them all, E (T )

In probability theory, the coupon collector's problem describes the "collect all coupons and win" contests. It asks the following question: Suppose that there is an urn of n different coupons, from which coupons are being collected, equally likely, with replacement. What is the probability that more than t sample trials are needed to collect all n coupons? An alternative statement is: Given n coupons, how many coupons do you expect you need to draw with replacement before having drawn each coupon at least once? The mathematical analysis of the problem reveals that the expected number of trials needed grows as .[1] For example, when n = 50 it takes about 225[2] trials to collect all 50 coupons.

Understanding the problem

The key to solving the problem is understanding that it takes very little time to collect the first few coupons. On the other hand, it takes a long time to collect the last few coupons. In fact, for 50 coupons, it takes on average 25 trials to collect the very last coupon after the other 49 coupons have been collected. This is why the expected time to collect all coupons is much longer than 50. The idea now is to split the total time into 50 intervals where the expected time can be calculated.

Answer

The following table (click [show] to expand) gives the expected number of tries to get sets of 1 to 100 coupons.

Solution

Calculating the expectation

Let T be the time to collect all n coupons, and let ti be the time to collect the i-th coupon after i  1 coupons have been collected. Think of T and ti as random variables. Observe that the probability of collecting a new coupon is pi = (n  (i  1))/n. Therefore, ti has geometric distribution with expectation 1/pi. By the linearity of expectations we have:

Here Hn is the n-th harmonic number. Using the asymptotics of the harmonic numbers, we obtain:

where is the Euler–Mascheroni constant.

Now one can use the Markov inequality to bound the desired probability:

Calculating the variance

Using the independence of random variables ti, we obtain:

The is a value of the Riemann zeta function (see Basel problem).

Now one can use the Chebyshev inequality to bound the desired probability:

Tail estimates

A different upper bound can be derived from the following observation. Let denote the event that the -th coupon was not picked in the first trials. Then:

Thus, for , we have .

Extensions and generalizations

Here m is fixed. When m = 1 we get the earlier formula for the expectation.

See also

Notes

  1. Here and throughout this article, "log" refers to the natural logarithm rather than a logarithm to some other base. The use of Θ here invokes big O notation.
  2. E(50) = 50(1 + 1/2 + 1/3 + ... + 1/50) = 224.9603, the expected number of trials to collect all 50 coupons. The approximation for this expected number gives in this case .
  3. Stadje, Wolfgang (1990), "The Collector's Problem with Group Drawings", Advances in Applied Probability, 22 (4): 866–882
  4. Sardy, Sylvain; Velenik, Yvan (2010), "Paninimania: sticker rarity and cost-effective strategy" (PDF), Swiss Statistical Society, Bulletin, 66: 2–6
  5. Flajolet, Philippe; Gardy, Danièle; Thimonier, Loÿs (1992), "Birthday paradox, coupon collectors, caching algorithms and self-organizing search", Discrete Applied Mathematics, 39 (3): 207–229

References

External links

This article is issued from Wikipedia - version of the 11/17/2016. The text is available under the Creative Commons Attribution/Share Alike but additional terms may apply for the media files.