Prove correlate di Bernoulli, distribuzione multivariata di Bernoulli?


14

Sto semplificando una domanda di ricerca che ho al lavoro. Immagina di avere 5 monete e di chiamare le teste un successo. Queste sono monete MOLTO distorte con probabilità di successo p = 0.1. Ora, se le monete fossero indipendenti, ottenendo quindi la probabilità di almeno 1 o più testa è molto semplice, 1(11/10)5 . Nel mio scenario, le mie prove di Bernoulli (lancio di monete) non sono indipendenti. Le uniche informazioni a cui ho accesso sono la probabilità di successi (ognuno è p = .1) e le correlazioni teoriche di Pearson tra le variabili binarie.

Esiste un modo per calcolare la probabilità di un successo o più solo con queste informazioni? Sto cercando di evitare un approccio basato sulla simulazione perché questi risultati teorici verranno utilizzati per guidare l'accuratezza di uno studio di simulazione. Ho esaminato la distribuzione multivariata di Bernoulli ma non credo di poterlo specificare completamente solo con correlazioni e probabilità marginali di successo. Un mio amico mi ha consigliato di costruire una copula gaussiana con i margini di bernoulli (usando il pacchetto R copula) e quindi di utilizzare la pMvdc()funzione su un campione di grandi dimensioni per ottenere la probabilità che desidero, ma non sono esattamente sicuro di come procedere con esso.


La distribuzione multivariata di Bernoulli è stata descritta qui: arxiv.org/abs/1206.1874
Tim

C'è un elemento temporale tra le prove o sono tutti in parallelo? Se precedente, puoi fare un'ipotesi semplificativa secondo cui dipende solo da t r i a l i - n , dove n ti dà l'ordine del tuo modello Markov? trialitrialinn
Zhubarb,

Risposte:


16

No, this is impossible whenever you have three or more coins.

The case of two coins

Let us first see why it works for two coins as this provides some intuition about what breaks down in the case of more coins.

Let X and Y denote the Bernoulli distributed variables corresponding to the two cases, XBer(p), YBer(q). First, recall that the correlation of X and Y is

corr(X,Y)=E[XY]E[X]E[Y]Var(X)Var(Y),

E[X]E[Y]Var(X)Var(Y), so by knowing the correlation, you also know E[XY]. Now, XY=1 if and only if both X=1 and Y=1, so

E[XY]=P(X=1,Y=1).

By knowing the marginals, you know p=P(X=1,Y=0)+P(X=1,Y=1), and q=P(X=0,Y=1)+P(X=1,Y=1). Since we just found that you know P(X=1,Y=1), this means that you also know P(X=1,Y=0) and P(X=0,Y=0), but now you're done, as the probability you are looking for is

P(X=1,Y=0)+P(X=0,Y=1)+P(X=1,Y=1).

Now, I personally find all of this easier to see with a picture. Let Pij=P(X=i,Y=j). Then we may picture the various probabilities as forming a square:

Here, we saw that knowing the correlations meant that you could deduce P11, marked red, and that knowing the marginals, you knew the sum for each edge (one of which are indicated with a blue rectangle).

The case of three coins

This will not go as easily for three coins; intuitively it is not hard to see why: By knowing the marginals and the correlation, you know a total of 6=3+3 parameters, but the joint distribution has 23=8 outcomes, but by knowing the probabilities for 7 of those, you can figure out the last one; now, 7>6, so it seems reasonable that one could cook up two different joint distributions whose marginals and correlations are the same, and that one could permute the probabilities until the ones you are looking for will differ.

Let X, Y, and Z be the three variables, and let

Pijk=P(X=i,Y=j,Z=k).

In this case, the picture from above becomes the following:

enter image description here

The dimensions have been bumped by one: The red vertex has become several coloured edges, and the edge covered by a blue rectangle have become an entire face. Here, the blue plane indicates that by knowing the marginal, you know the sum of the probabilities within; for the one in the picture,

P(X=0)=P000+P010+P001+P011,

and similarly for all other faces in the cube. The coloured edges indicate that by knowing the correlations, you know the sum of the two probabilities connected by the edge. For example, by knowing corr(X,Y), you know E[XY] (exactly as above), and

E[XY]=P(X=1,Y=1)=P110+P111.

So, this puts some limitations on possible joint distributions, but now we've reduced the exercise to the combinatorial exercise of putting numbers on the vertices of a cube. Without further ado, let us provide two joint distributions whose marginals and correlations are the same:

enter image description here

Here, divide all numbers by 100 to obtain a probability distribution. To see that these work and have the same marginals/correlations, simply note that the sum of probabilities on each face is 1/2 (meaning that the variables are Ber(1/2)), and that the sums for the vertices on the coloured edges agree in both cases (in this particular case, all correlations are in fact the same, but that's doesn't have to be the case in general).

Finally, the probabilities of getting at least one head, 1P000 and 1P000, are different in the two cases, which is what we wanted to prove.

For me, coming up with these examples came down to putting numbers on the cube to produce one example, and then simply modifying P111 and letting the changes propagate.

Edit: This is the point where I realized that you were actually working with fixed marginals, and that you know that each variable was Ber(1/10), but if the picture above makes sense, it is possible to tweak it until you have the desired marginals.

Four or more coins

Finally, when we have more than three coins it should not be surprising that we can cook up examples that fail, as we now have an even bigger discrepancy between the number of parameters required to describe the joint distribution and those provided to us by marginals and correlations.

Concretely, for any number of coins greater than three, you could simply consider the examples whose first three coins behave as in the two examples above and for which the outcomes of the final two coins are independent from all other coins.


3

Correlated Bernoulli trials lead to a beta-binomial distribution for the counted outcomes. It should be possible to parameterise this distribution to give a specified correlation value, and then calculate the probability you want.


Isn't a beta-binomial just a Binomial whose success probability parameter is a random variable following a Beta? How does that apply to the OP's problem?
A. G.

1
Yes, that is one characterisation of the distribution. It is also one of the solutions of correlated Bernoulli trials (see e.g., Hisakado et al 2006)
Reinstate Monica

So it is! Upvoted.
A. G.

Utilizzando il nostro sito, riconosci di aver letto e compreso le nostre Informativa sui cookie e Informativa sulla privacy.
Licensed under cc by-sa 3.0 with attribution required.