Thank you very much, Ray !

You mentioned Pearson's result. Could you post 
a pointer to a book/paper which I can use and cite. 
It is likely that I can not find the original from 1901,
but any recent discussion of his or related result would do.

Also, correct me if I'm wrong, but I think, 
if N -> infinity, then P -> 1 (for any K=2,3,4,...), given that p>q,
and P -> 0, if p<q.

Thank you again.

[EMAIL PROTECTED] (Ray Koopman) wrote in message news:<[EMAIL PROTECTED]>...
> [EMAIL PROTECTED] (Dan Esperantos) wrote in message 
> news:<[EMAIL PROTECTED]>...
> > Suppose we are given a multinomial probability distribution
> >                                         N!
> > P(X_1=N_1, X2=N_2,..., X_K=N_K) = ---------------- p^N_1 * q^(N-N_1),
> >                                   N_1! N_2!...N_K! 
> > 
> > where p + q*(K-1) = 1, and N_1 + N_2 + N_K = N.
> > 
> > How to compute an approximation of the following
> > probability, where {X_i}s are distributed as above:
> > 
> > P( X_1 = max{X_1, X_2,..., X_K} )?
> > 
> > if p > q or if p > q*(K-1)
> > 
> > For example, when K = 2 we simply have (assume N is even):
> >                                               N      N!
> > P( X_1 = max{X_1, X_2} ) = P( X_1 >= X_2 ) = Sum  -------- p^i*(1-p)^(N-i)
> >                                             i=N/2 i!(N-i)!
> > 
> > Are any approximations known for such sums?
> > What about the case of general K? 
> > In my applications N is on order of several hundreds,
> > so exact computation is not feasible.
> 
> Let Y_i = X_1 - X_i, i = 2,...,K. Then
> 
> P(X_1 >= max{X_2,...,X_K}) = P(Y_2 >= 0 & ... & Y_K >= 0)
>                            = P(min{Y_2,...,Y_K} >= 0).
> 
> The {Y_i} are exchangeable with means = N(p-q),
> 
> variances = Var(X_1) + Var(X_i) - 2 Cov(X_1,X_i)
>           = Np(1-p) + Nq(1-q) - 2(-Npq),
>           = N(p + q - (p-q)^2),
> 
> covariances = Var(X_1) - Cov(X_1,X_i) - Cov(X_1,X_j) + Cov(X_i,X_j)
>             = Np(1-p) - 2(-Npq) + (-Nq^2)
>             = N(p - (p-q)^2).
> 
> If Nq is sufficiently large -- say 10 or so? -- then the distribution
> of {X_i}, and hence of {Y_i}, is approximately multivariate normal,
> and the desired probability is approximately equal to the probability
> that K-1 exchangeable multinormal variables whose means, variances,
> and covariances are given above, are all positive (or, equivalently,
> that their minimum is positive).
> 
> K = 2: P = F(z), where F is the standard normal cdf, 
> and z = Mean(Y_i)/SD(Y_i) = N(p-q)/Sqrt(N(p + q - (p-q)^2))).
> 
> K = 3: Pearson (1901), working on a more general problem, reduced
> the obvious two-dimensional integral to a one-dimensional integral:
>                 R
> P = (F(z))^2 + Int f(z,z;r)dr, where z is as defined above,
>                 0                      
> f(x,y;r) = exp((-1/2)(x^2 + y^2 - 2rxy)/(1-r^2))/(2Pi Sqrt(1-r^2))
> 
> is the bivariate standard normal pdf with correlation r, and
> 
> R = Cov(Y_i,Y_j)/Var(Y_i) = (p-(p-q)^2)/(p+q-(p-q)^2)
> 
> is the correlation of Y_i and Y_j. The integrand simplifies a little
> when we set x = y = z, but numeric integration is still necessary.
> 
> K = 4: I vaguely remember seeing a paper about the three-dimensional
> case, but I have no idea whose or where or when.
> 
> K > 4: ???
.
.
=================================================================
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at:
.                  http://jse.stat.ncsu.edu/                    .
=================================================================

Reply via email to