The expected number of picks needed equals the sum of the probabilities that at least t picks are needed, which means that t−1 subsets left at least one value uncovered. We can use inclusion-exclusion to get the probability that at least one value is uncovered.
The probability that a particular set of k values is uncovered after t−1 subsets are chosen is
Bigg(fracn−kchoosernchooserBigg)t−1
So, by inclusion-exclusion, the probability that at least one value is uncovered is
sumnk=1nchoosek(−1)k−1Bigg(fracn−kchoosernchooserBigg)t−1
And then the expected number of subsets needed to cover everything is
sumit=1nftysumnk=1nchoosek(−1)k−1Bigg(fracn−kchoosernchooserBigg)t−1
Change the order of summation and use s=t−1:
sumnk=1nchoosek(−1)k−1sumis=0nftyBigg(fracn−kchoosernchooserBigg)s
The inner sum is a geometric series.
sumnk=1nchoosek(−1)k−1fracnchoosernchooser−n−kchooser
nchoosersumnk=1(−1)k−1fracnchooseknchooser−n−kchooser
I'm sure that should simplify further, but at least now it's a simple sum. I've checked that this agrees with the coupon collection problem for r=1.
Interestingly, Mathematica "simplifies" this sum for particular values of r, although what it returns even for the next case is too complicated to repeat, involving EulerGamma, the gamma function at half-integer values, and PolyGamma[0,1+n].
No comments:
Post a Comment