Multinomial Distribution

The random variable X = (X1, X2, ..., Xk) denoting the outcome of n trials, whre Xi = frequencyor number of outcome Ei with respective probability pi ( i = 1, 2, ..., k), is said to have multinomial distribution with parameters (n, p1, p2, ..., pk), if its probability mass function (p.m.f) is given by $$p(x_1, x_2, ..., x_n) \ = \ \frac{n!}{x_1! x_2! ... x_k!} \ p_1^{x_1} p_2^{x_2} ... p_k^{x_k}$$ $$where, \ \sum_{x=1}^k \ x_i \ = \ n \ and \ \sum_{i=1}^k \ p_i \ = \ 1.$$

Summary

The random variable X = (X1, X2, ..., Xk) denoting the outcome of n trials, whre Xi = frequencyor number of outcome Ei with respective probability pi ( i = 1, 2, ..., k), is said to have multinomial distribution with parameters (n, p1, p2, ..., pk), if its probability mass function (p.m.f) is given by $$p(x_1, x_2, ..., x_n) \ = \ \frac{n!}{x_1! x_2! ... x_k!} \ p_1^{x_1} p_2^{x_2} ... p_k^{x_k}$$ $$where, \ \sum_{x=1}^k \ x_i \ = \ n \ and \ \sum_{i=1}^k \ p_i \ = \ 1.$$

Things to Remember

  1. When k = 2, the mulltinomial distribution reduces to binomial distribution with parameters n and p.
  2. When k = 3, then the multinomial distribution reduces to trinomial distribution

MCQs

No MCQs found.

Subjective Questions

No subjective questions found.

Videos

No videos found.

Multinomial Distribution

Multinomial Distribution

An experiment is said to be a multinomial experiment if

  1. the experiment consists of n fixed trials.
  2. for each trail there are k ( > 2 i.e. 3 or more ) possible outcomes.
  3. the trials are independent.
  4. the probability for each outcome remains the same from trial to trial.

The examples of multinomial experiment are

  1. a number of throws of a fair die in which rach throws can result six different outcomes.
  2. the number of selection or drawings of balls ar random with replacement from a box, containing 20 balls of which 2 are white, 4 are black, 6 red and 8 green balls.
  3. a number of people, randomly selected to ask them the reaction to a new hydropower project in a location, in which the outcomes could be

(a) like the project

(b) dislike the project

(c) indifferent

Definition :

The random variable X = (X1, X2, ..., Xk) denoting the outcome of n trials, whre Xi = frequencyor number of outcome Ei with respective probability pi ( i = 1, 2, ..., k), is said to have multinomial distribution with parameters (n, p1, p2, ..., pk), if its probability mass function (p.m.f) is given by

$$p(x_1, x_2, ..., x_n) \ = \ \frac{n!}{x_1! x_2! ... x_k!} \ p_1^{x_1} p_2^{x_2} ... p_k^{x_k}$$

$$where, \ \sum_{x=1}^k \ x_i \ = \ n \ and \ \sum_{i=1}^k \ p_i \ = \ 1.$$

Derivation of multinomial distribution

Let us perform n independent trials, each of which may result any one of k (>2) possible outcomes at eah tria. Let E1, E2, ..., Ek be the k mutually exclusive and exhaustive events (outcomes) of a trial with corresponding probability p1, p2, ..., pk so thatp1 + p2 + ... + pk = 1. Again, let X1, X2, ..., Xk be the frequency or number of times of respective events E1, E2, ..., Ek occurs in the n trials so that each Xi can take the values in the set { 0, 1, 2, ..., n } and X1 + X2 + ... + Xk = n.

consider an arrangement of outcomes, in which E1 occurs x1 times, E2 occurs x2 times, ... Ek occurs xk times as E1E1 ... E1, E2E2 ... E2, ..., EkEk ... Ek.

x1 times, x2 times, ..., xk times,

Here, the number of different possible arrangements of n trials of which x1 are of event E1, x2 are of event E2, ..., xk aare of event Ek is equal to $$\frac{n!}{x_1! x_2! ... x_k!}.$$

The probability of this sequence of occurance of events is p1x1 p2x2 ... pkxk for each arrangement. Then, in n trials, the probability that E1 occurs x1 times, E2 occurs x2 times, ..., and Ek occurs xk times is given by

$$p(X_1 \ = \ x_1, X_2 \ = \ x_2, ..., X_n \ = \ x_n) \ = \ \frac{n!}{x_1! x_2! ... x_k!} \ p_1^{x_1} p_2^{x_2} ... p_k^{x_k}$$

which is the probability mass (pmf) of multinomial distribution with parameters (n, p1, p2, ..., pk).

Since the above probability mass function is the general term in the multinomial expansion of ( p1 + p2 + ... + pk)n, the distribution with its probability mass function (pmf) is called multinomial distribution. therefore it can easily be proved that

$$\sum_{x} \ p(x_1, x_2, ..., x_n) \ = \sum_{x} \frac{n!}{x_1! x_2! ... x_k!} \ p_1^{x_1} p_2^{x_2} ... p_k^{x_k}$$

= ( p1 + p2 + ... + pk)n

= 1

If a random vector X = ( X1, X2, ..., Xk ) follows multinomial distribution with parameters (n, p1, p2, ..., pk) then we denote it as

(X1, X2, ..., Xk) ~ Mk(n, p1, p2, ..., pk).

It is important to note that the multinomial distribution is a ganeralization of binomial distribution. The multinomial distribution is complicated distribution because it is a multi varite distribution involving k variables X1, X2, ..., Xk but only k-1 variables are independent, as X1 + X2 + ... + Xk = n. However, this distribution has wide applications such as in sampling with replacement of which individuals or observations are classified into more than two categories or groups.

Reduction of multinomial distribution when k = 2 and k = 3

When k = 2, then the multinomial distribution has probability mass function (pmf)

$$p(x_1, x_2) \ = \ \frac{n!}{x_1! x_2! ... x_k!} \ p_1^{x_1} p_2^{x_2}$$

since x1 + x2 = n and p1 + p2 = 1, we get x2 = n -x1 and p2 = 1 - p1. Then,

$$p(x_1, n-x_1) \ = \ \frac{n!}{x_1! (n-x_1)! ... x_k!} \ p_1^{x_1} (1-p_1)^{n-x_1}$$

if we write x1 = x and p1 = p, then

$$p(x, n-x) \ = \ \frac{n!}{x! (n-x)! ... x_k!} \ p^{x} (1-p)^{n-x}$$

$$\Rightarrow \ p(x) \= \ \frac{n!}{x! (n-x)! ... x_k!} \ p^{x} (q)^{n-x} \ \ \ ; \ x \ = \ 0, \ 1, \ 2, \ ..., \ n$$

Where,

x = Number of successes

n-x = Number of failures

p = probability of successes of an event in a single trial

q = 1 - p = probability of failures in a single trial.

This p(x) is the pmf of binomial distrinution with parameters n anf p.

Thus, when k = 2, the mulltinomial distribution reduces to binomial distribution with parameters n and p. Symbolically,

(X1, X2) ~ M2 (n, p1, p2) = B (n, p1)

$$\Rightarrow \ X_1 \ ~ \ B(n, p_1)$$

Similarly, when k = 3, then the multinomial distribution reduces to trinomial distribution with probability mass function (pmf)

$$p(x_1, x_2, x_3) \ = \ \frac{n!}{x_1! x_2! x_3!} \ p_1^{x_1} \ p_2^{x_2} \ p_3^{x_3}$$

where, x1 + x2 + x3 = n and p1+ p2 + p3 = 1.

We have x3 = n - x1 - x2 and p3 = 1 - p1 - p2. It implies that

$$p(x_1, x_2, n-x_1-x_2) \ = \ \frac{n!}{x_1! x_2! (n-x_1-x_2)!} \ p_1^{x_1} \ p_2^{x_2} \ (1-p_1-p_2)^{n-x_1-x_2}$$

which is the probability mass function of trinomial distribution of two random variables X1 and X2.

Symbolically,

(X1, X2, X3) ~ M3 (n, p1, p2, p3)

$$\Rightarrow \ (X_1, X_2) \ ~ \ M_3 \ (n, p_1, p_2).$$

Bibliography

Sukubhattu N.P. (2013). Probability & Inference - II. Asmita Books Publishers & Distributors (P) Ltd., Kathmandu.

Larson H.J. Introduction to Probability Theory and Statistical Inference. WileyInternational, New York.

Lesson

Discrete probability distribution

Subject

Statistics

Grade

Bachelor of Science

Recent Notes

No recent notes.

Related Notes

No related notes.