In lecture 19, we saw how to obtain the expected distance between two uniform r.v. How would we do the same, but with the Normal distribution?
Given Z1,Z2∼N(0,1), Z1 and Z2 are i.i.d., can we find E|Z1−Z2|?
Now, we could just jump in and try using 2D LOTUS, but let's take a step back and recall what we have seen earlier in Lecture 14 about the linearity of Normals, and see if we can apply what we know about MGFs.
Recall from Lecture 17 that the MGF of a sum of independent r.v. is just the product of their respective MGF.
M(X+Y)=E(et(X+Y))=E(etX)E(etY)=E(eμ1t+12σ21t2)E(eμ2t+12σ22t2)=E(e(μ1+μ2)t+12(σ21+σ22)t2)◼Returning now to our original question, we want to find E|Z1−Z2|, with Z1,Z2∼N(0,1).
Keep in mind that Z1−Z2∼N(0,2). But since this is just the standard normal scaled with √2, we can do this:
E|Z1−Z2|=E|√2Z|, where Z∼N(0,1)=√2E|Z|=√2∫∞−∞|z|1√2πe −z22dz=√2πA generalization of the binomial distribution, Mult(n,→p) has parameter n for number of items observed, with probability vector →p=p1,p2,…,pk) such that pj≥0 and ∑pi=1
So while the binomial has 2 classes, with success equating to p and failure equating to q, the multinomial has k classes, each with respective probability pi.
We have n objects which we are independently putting into k categories.
→X=Mult(n,→p), with →X=(X1,X2,…,Xk)Pj=P(category j)Xj=number of objects in category jBecause this is a joint distribution, Mult(n,→p) has a joint PMF.
P(X1=n1,X2=n2,…,Xk=nk)=n!n1!n2!…nk!pn11pn22…pnkkGiven →X∼Multk(n,→p), what is the marginal distribution of Xj?
Well, since we are interested in only the two cases of an object being in class k or not, this is binomial.
So for the marginal distribution Xj of Mult,
Xj∼Bin(n,pj)E(Xj)=npjVar(Xj)=npj(1−pj)Given a situation where there are n voters in a population sample, and there are 10 political parties, →X=(X1,X2,…,X10)∼Mult(n,(p1,p2,…,p10))...
Let us say that political parties 3 through 10 are relatively minor compared to parties 1 and 2. We can describe this case where we want to lump together parties 3 through 10 with another multinomial Y such that:
→Y=(Y1,Y2,Y3,…,10)∼Mult(n,(p1,p2,p3,…,10))Here, we gather up the counts/probabilities for parties 3 through 10, and now we have a multinomial with essentially three classes.
Say that we have a k-class →X∼Multk(n,→p), but we know X1=n1, but we don't know about the other k−1 classes...
Given X1=n1
(X2,…,Xk)∼Multk-1(n−n1,(p′2,…,p′k)) where p′j=P(in class j|not in class 1)∼Mult((n−n1),(p21−p1,…,pk1−p1)) or alternatively... ∼Mult((n−n1),(p2p2+⋯+pk,…,pkp2+⋯+pk)) or alternatively...All we are doing here is simply re-normalizing the probability vector to take into account that we have information that X1=n1. Multinomials are simple and intuitive!
The Cauchy Distribution is T=XY with X,Y i.i.d. N(0,1).
It looks simple enough and appears to be quite useful, but it does have some weird properties.
Can you find the PDF of T?
Let us try to find the CDF, and derive the PDF from that.
CDF: P(XY≤t)=P(X|Y|) following from the symmetry of the Normal=P(X≤t|Y|)=∫∞−∞∫t|y|−∞1√2πe−x221√2πe−y22dxdy=1√2π∫∞−∞e−y22∫t|y|−∞1√2πe−x22dxdy starting from x=1√2π∫∞−∞e−y22Φ(t|y|)dybut we have an even function=√2π∫∞0e−y22Φ(t|y|)dyPDF: F′(t)=√2π∫∞0ye−y221√2πe−t2y22dy=1π∫∞0ye−(1+t2)y22dyand with u=(1+t2)y22, du=(1+t2)ydy=1π∫∞0(1+t2)(1+t2)ye−(1+t2)y22dy=1π∫∞011+t2e−udu=1π(1+t2) and just integrate this to get the CDFCDF: =tan−1(t)πRecall that the PDF of N(0,1) is ϕ(y)
P(X≤t|Y|)=∫infty−∞P(X≤t|Y|∣Y=y)ϕ(y)dy=∫∞−∞Φ(t|y)ϕ(y)dy which is pretty much what we did aboveView Lecture 20: Multinomial and Cauchy | Statistics 110 on YouTube.