Unit II (1) Basic Terminology: (i) Exhaustive Events: A set of events is said to be exhaustive, if it includes all the possible events. For example, in tossing a coin there are two exhaustive cases either head or tail and there is no other possibility. (ii) Mutually exclusive events: If the occurrence if one of the events precludes the occurrence of all others, then such a set of events is said to be mutually exclusive. Just as tossing a coin, head comes up or the tail and both canβt happen at the same time, i.e. these are two mutually exclusive cases. (iii)Equally likely events: If one of the events cannot be expected to happen in preference to another then such events are said to be equally likely. For instance, in tossing a coin, the coming of the head or the tail is equally likely. Thus, when a die is thrown, the turnings up of the six different faces of the die are exhaustive, mutually exclusive and equally likely. (2) Definition of Probability: If there are n exhaustive, mutually exclusive and equally likely cases of which m are favorable to an event A, then the probability (ππ) of the happening of ππ π΄π΄ is ππ(π΄π΄) = ππ = . ππ
As there are ππ β ππ cases in which π΄π΄ will not happen (denoted by π΄π΄β² ), the chance of π΄π΄ not ππβππ ππ = 1 β = 1 β ππ. ππ. ππ. ππ(π΄π΄β² ) = 1 β ππ(π΄π΄) so that happening is ππ or ππ(π΄π΄β² ) so that ππ = β²
ππ
ππ
ππ(π΄π΄) + ππ(π΄π΄ ) = 1. i.e. if an event is certain to happen then its probability is unity, while if it is certain not to happen, its probability is zero. Note: This definition of probability fails when (i) number of outcomes is infinite (not exhaustive) and (ii) outcomes are not equally likely. (3) Statistical (or Empirical) definition of probability: If in ππ trails, an event π΄π΄ happens ππ ππ times, then the probability (ππ) of happening of π΄π΄ is given by ππ = ππ(π΄π΄) = limππββ ππ
(4) Probability and set notations: (i) Random experiment: Experiments which are performed essentially under the same conditions and whose results cannot be predicted are known as random experiments. E.g. Tossing a coin or rolling a die are random experiments. (ii) Sample Space: The set of all possible outcomes of a random experiment is called sample space for that experiment and is denoted by ππ. The elements of the sample space are called the sample points. E.g. On tossing a coin, the possible outcomes are head (π»π») and tail (ππ). Thus ππ = {π»π», ππ}. Page 1 of 12
(5)
(6)
(7)
(8)
(9)
(iii)Event: The outcomes of a random experiment is called an event. Thus every subset of a sample space ππ is an event. The null set ππ is also an event and is called an impossible event. Probability of an impossible event is zero i.e. ππ(ππ) = 0. Axioms: (i) The numerical value of the probability lies between 0 and 1. i.e for any event A of S, 0 β€ ππ(π΄π΄) β€ 1. (ii) The sum of probabilities of all sample events is unity. i.e. P(S)=1. (iii)Probability of an event made of two or more sample events is the sum of their probabilities. Notations: (i) Probability of happening of events A or B is written as ππ(π΄π΄ + π΅π΅) or ππ(π΄π΄βπ΅π΅). (ii) Probability of happening of both the events A and B is written as ππ(π΄π΄π΄π΄) or ππ(π΄π΄ β© π΅π΅). (iii)βEvent A implies (β) event Bβ is expressed as π΄π΄ β π΅π΅. (iv) βEvents A and B are mutually exclusiveβ is expressed as π΄π΄ β© π΅π΅ = ππ. (v) For any two events A and B, ππ(π΄π΄ β© π΅π΅ β² ) = ππ(π΄π΄) β ππ(π΄π΄ β© π΅π΅). Similarly, ππ(π΄π΄β² β© π΅π΅) = ππ(π΅π΅) β ππ(π΄π΄ β© π΅π΅). Addition law of probability or theorem of total probability: (i) If the probability of an event A happening as a result of a trial is P(A) and the probability of a mutually exclusive event B happening is P(B), then the probability of either of the events happening as a result of the trial is ππ(π΄π΄ + π΅π΅) ππππ ππ(π΄π΄ βͺ π΅π΅) = ππ(π΄π΄) + ππ(π΅π΅). (ii) If A and B are two events (not mutually exclusive), then ππ(π΄π΄ + π΅π΅) = ππ(π΄π΄) + ππ(π΅π΅) β ππ(π΄π΄π΄π΄). or ππ(π΄π΄ βͺ π΅π΅) = ππ(π΄π΄) + ππ(π΅π΅) β ππ(π΄π΄ β© π΅π΅). In general, for any number of mutually exclusive events π΄π΄1 , π΄π΄2 , β¦ , π΄π΄ππ , we have ππ(π΄π΄1 + π΄π΄2+β¦+π΄π΄ππ=ππ(π΄π΄1)+πππ΄π΄2+β¦+ππ(π΄π΄ππ) (iii)If A, B, C are any three events, then ππ(π΄π΄ + π΅π΅ + πΆπΆ ) = ππ(π΄π΄) + ππ(π΅π΅) + ππ(πΆπΆ ) β ππ(π΄π΄π΄π΄) β ππ(π΅π΅π΅π΅ ) β ππ(πΆπΆπΆπΆ) + ππ(π΄π΄π΄π΄π΄π΄) Independent Events: Two events are said to be independent, if happening or failure of one does not affect the happening or failure of the other .Otherwise the events are said to be dependent. For two dependent events A and B, the symbol ππ(π΅π΅/π΄π΄) denotes the probability of occurrence of B, when A has already occurred. It is known as conditional probability and read as a βprobability of B given Aβ. Multiplication law of probability or theorem of compound probability: If the probability of an event A happening as a result of trial is P(A) and after A has happened the probability of an event B happening as a result of another trial (i.e. conditional Page 2 of 12
probability of B given A) is ππ(π΅π΅/π΄π΄), then the probability of both the events A and B happening as a result of two trials is ππ(π΄π΄π΄π΄) = ππ(π΄π΄ β© π΅π΅) = ππ(π΄π΄). ππ(π΅π΅/π΄π΄). If the events A and B are independent, i.e. if the happening of B does not depend on whether A has happened or not, then ππ(π΅π΅/π΄π΄) = ππ(π΅π΅) and ππ(π΄π΄/π΅π΅) = ππ(π΄π΄).Therefore, ππ(π΄π΄π΄π΄) ππππ ππ(π΄π΄ β© π΅π΅) = ππ(π΄π΄). ππ(π΅π΅). In general, ππ(π΄π΄1 π΄π΄2 β¦ π΄π΄ππ ) = ππ(π΄π΄1 ). ππ(π΄π΄2 ) β¦ ππ(π΄π΄ππ ) Note: If ππ1 , ππ2 be the probabilities of happening of two independent events, then (i) the probability that the first event happens and the second event fails is ππ1 (1 β ππ2 ) (ii) the probability that both events fail to happen is (1 β ππ1 )(1 β ππ2 ) (iii)the probability that atleast one of the events happens is 1 β (1 β ππ1 )(1 β ππ2 )
Problems based on conditional probability:
1. Two cards are drawn in succession from a pack of 52 cards. Find the chance that the first is a king and the second a queen if the first card is (i) replaced (ii) not replaced. 2. A pair of dice is tossed twice. Find the probability of scoring 7 points (a) once, (b) atleast once (c) twice. 3. There are two groups of subjects: one of which consists of 5 science and 3 engineering subjects, and the other consists of 3 science and 5 engineering subjects. An unbiased die is cast. If the number 3 or 5 turns up, a subject is selected at random from the first group, otherwise the subject is selected at random from second group. Find the probability that an engineering subject is selected ultimately. 4. A box A contains 2 white and 4 black balls. Another box B contains 5 white and 7 black balls. A ball is transferred from the box A to the box B. Then a ball is drawn from the box B. Find the probability that it is white. 5. Two persons A and B toss an unbiased coin alternatively on the understanding that the first who gets the head wins. If A starts the game, find their respective chances of winning. 6. Two cards are selected at random from 10 cards numbered 1 to 10. Find the probability p that the sum is odd, if (i) the two cards are drawn together (ii) the two cards are drawn one after the other without replacement. (iii) the two cards are drawn one after the other with replacement. 7. Given ππ(π΄π΄) = 1/4, ππ(π΅π΅) = 1/3 and ππ(π΄π΄ βͺ π΅π΅) = 1/2 , evaluate ππ(π΄π΄/π΅π΅), ππ(π΅π΅/π΄π΄), ππ(π΄π΄ β© π΅π΅β² ) ππππππ ππ(π΄π΄/π΅π΅β² ) 8. A can hit a target 3 times in 5 shots, B 2 times in 5 shots and C 3 times in 4 shots. They fire a volley. What is the probability that (i) two shots hit, (ii) atleast two shots hit? 9. A problem in mechanics is given to three students A, B and C whose chances of solving it are 1 1
1
, and respectively. What is the probability that the problem will be solved?
2 3
4
Page 3 of 12
10. The students in a class are selected at random, one after the other, for an examination. Find the probability p that the boys and girls in the class alternate if (i) the class consists of 4 boys and 3 girls (ii) the class consists of 3 boys and 3 girls. 11. A and B throw alternatively with a pair of dice. A wins if he throws 6 before B throws 7 and B wind if he throws 7 before A throws 6. If A begins, find his chance of winning. BAYE'S THEOREM: An event A corresponds to a number of exhaustive events π΅π΅1 , π΅π΅2 , β¦ , π΅π΅ππ . If ππ(π΅π΅ππ ) and ππ(π΄π΄/π΅π΅ππ ) are given, then ππ(π΅π΅ππ /π΄π΄)
ππ (π΅π΅ππ ).ππ(π΄π΄/π΅π΅ππ ) . ππ (π΅π΅ππ ).ππ(π΄π΄/π΅π΅ππ )
=β
Note: The probabilities ππ(π΅π΅ππ ), ππ = 1,2, β¦ , ππ are called a priori probabilities because these exist before we get any information from the experiment. The probabilities ππ(π΄π΄/π΅π΅ππ ), ππ = 1,2, β¦ ππ are called posteriori probabilities, because these are found after the experiment results are known. Problems based on Baye's theorem: 1. Three machines ππ1 , ππ2 ππππππ ππ3 produce identical items. Of their respective output 5%, 4% and 3% are faulty. On a certain day ππ1 has produced 25% of the total output, ππ2 has produced 30% and ππ3 the remainder. An item selected at random is found to be faulty. What are the chances that it was produced by the machine with the highest output? 2. There are three bags: first containing 1 white, 2 red , 3 green balls; second 2 white, 3 red, 1 green balls and third 3 white, 1 red, 2 green balls. Two balls are drawn from a bag chosen at random. These are found to be one white and one red. Find the probability that the balls so drawn came from the second bag. 3. In a certain college, 4% of the boys and 1% of the girls are taller than 1.8m. Furthermore 60% of the students are girls. If a student is selected at random and is found to be taller than 1.8m. what is the probability that the student is a girl? 4. In a bolt factory, there are four machines A, B, C and D manufacturing 20%, 15%, 25% and 40% of the total output respectively. Of their outputs 5%, 4%, 3% and 2% in the same order are defective bolts. A bolt is chosen at random from the factory's production and is found defective. What is the probability that the bolt was manufactured by machine A or machine D? 5. The contents of three urns are: 1 white, 2 red, 3 green balls; 2 white, 1red, 1green balls; 4 white, 5 red, 3 green balls. Two balls are drawn from an urn chosen at random. These are found to be one white and one green. Fine the probability that the balls so drawn came from the third urn. Random Variable: If a real variable ππ be associated with the outcome of a random experiment, then since the values which ππ takes depends on chance, it is called a random variable or a stochastic variable or simply a variate. For instance, if a random experiment πΈπΈ consists of tossing a pair of dice, the sum ππ of the two numbers which turn up have the value 2,3,4, β¦ ,12 Page 4 of 12
depending on chance. Then ππ is the random variable. It is a function whose values are real numbers and depend on chance. If in a random experiment, the event corresponding to a number β²ππβ² occurs, then the corresponding random variable ππ is said to assume the values β²ππβ² and the probability of the event is denoted by ππ(ππ = ππ). Similarly, the probability of the ππ β€ ππ is written as ππ(ππ β€ ππ). If a random variable takes a finite or countable set of values, it is called a discrete random variable. On the other hand, if it assumes an infinite number of uncountable values, it is called a continuous random variable. Discrete Probability Distribution: Suppose a discrete random variable ππ is the outcome of some experiment. If the probability that ππ takes the values π₯π₯ππ is ππππ , then ππ(ππ = π₯π₯ππ ) = ππππ ππππ ππ(π₯π₯ππ ) for ππ = 1,2, .. where (i) ππ(π₯π₯ππ ) β₯ 0 for all values of ππ, (ii) β ππ(π₯π₯ππ ) = 1. The set of values π₯π₯ππ with their probabilities ππππ constitutes a discrete probability distribution of the discrete random variable ππ. For example, the discrete probability distribution of ππ, the sum of the numbers which turn on tossing a pair of dice is given by the following table: ππ = π₯π₯ππ ππ(π₯π₯ππ )
2
3
4
5
6
7
8
9
10
11
12
1/36
2/36
3/36
4/36
5/36
6/36
5/36
4/36
3/36
2/36
1/36
Distribution function: The distribution function πΉπΉ(π₯π₯) of the discrete random variable is defined by πΉπΉ (π₯π₯) = ππ(ππ β€ π₯π₯) = βπ₯π₯ππ=1 ππ(π₯π₯ππ ) where π₯π₯ is any integer. This distribution function is also sometimes called as cumulative distribution function. Problems based on discrete probability distribution: 1. A die is tossed thrice. A success is 'getting 1 or 6' on a toss. Find the mean and variance of the number of successes. 2. A random variable X has the following probability distribution X 0 1 2 3 4 5 6 7 P(X) 0 k 2k 2k 3k ππ 2 2ππ 2 7ππ 2 + ππ Find: (i) the value of βkβ; (ii)ππ(1.5 < ππ < 4.5/ππ > 2); (iii) the smallest value of ππ for which ππ(ππ β€ ππ) >
1 2
3. The probability density function of a variate X is X 0 1 2 3 4 5 6 P(X) K 3k 5k 7k 9k 11k 13k (i) Find ππ(ππ < 4), ππ(ππ β₯ 5), ππ(3 < ππ β€ 6) (ii) What will be minimum value of ππ so that ππ(ππ β€ 2) > 0.3 Continuous Probability Distribution: When a variate X takes every value in an interval, it gives rise to continuous distributions of X. The distribution defined by the variates like heights or weights are continuous distributions. Page 5 of 12
A major conceptual difference, however, exists between the discrete and continuous probabilities. When thinking in discrete terms, the probability associated with an event is meaningful. With continuous events, however, where the number of events is infinitely large, the probability that a specific event will occur is practically zero. For this reason, continuous probability statements must be worded somewhat differently from discrete ones. Instead of finding the probability that π₯π₯ equals some value, we find the probability of π₯π₯ falling in a small interval. Thus the probability distribution of a continuous variate π₯π₯ is defined by a function ππ(π₯π₯) such 1
that the probability of the variate π₯π₯ falling in the small interval π₯π₯ β ππππ to 1
1
2
1
π₯π₯ + ππ is ππ(π₯π₯)ππππ . 2
Symbolically it can be expressed as ππ οΏ½π₯π₯ β ππππ β€ π₯π₯ β€ π₯π₯ + πππποΏ½ is ππ(π₯π₯)ππππ . Then ππ(π₯π₯) is 2
2
called the probability density function and the continuous curve π¦π¦ = ππ(π₯π₯) is called the probability curve. β The density function ππ(π₯π₯) is always positive and β«ββ ππ(π₯π₯ )ππππ = 1 (i.e., the total area under the probability curve and the x-axis is unity which corresponds to the requirement that the total probability of happening of an event is unity. π₯π₯ Distribution function: If πΉπΉ (π₯π₯) = ππ(ππ β€ π₯π₯) = β«ββ ππ(π₯π₯)ππππ , then F(x) is defined as cumulative
distribution function or simply the distribution function of the continuous random variable X. It is the probability that the value of the vairate X will be β€ π₯π₯. The distribution function F(x) has the following properties: (i) πΉπΉ β² (π₯π₯) = ππ(π₯π₯) β₯ 0, so that F(x) is a non-decreasing function. (ii) πΉπΉ (ββ) = 0 (iii) πΉπΉ (β) = 1 ππ
ππ
ππ
(iv) ππ(ππ β€ ππ β€ ππ) = β«ππ ππ(π₯π₯)ππππ = β«ββ ππ(π₯π₯)ππππ β β«ββ ππ(π₯π₯)ππππ = πΉπΉ (ππ) β πΉπΉ(ππ) Problems based on discrete probability distribution: 1) (i) Is the function defined as follows a density function? βπ₯π₯ π₯π₯ > 0 ππ ππ (π₯π₯) = οΏ½ 0 π₯π₯ β€ 0 (ii) If so, determine the probability that the variate having this density will fall in the interval (1,2) (iii)Also find the cumulative probability function F(2). 1) Expectation: The mean value (ππ) of the probability distribution of a variate X is commonly known as its expectation or mathematical expectation and is denoted by E(X). Page 6 of 12
ππ = πΈπΈ (ππ) =
β§ οΏ½ π₯π₯ππ ππ(π₯π₯ππ ) βͺ ππ
if X is a discrete random variable
β
β¨ βͺ οΏ½ π₯π₯ ππ(π₯π₯)ππππ if X is a continuous random variable β©ββ In general, expectation of any function ππ(π₯π₯) is given by, πΈπΈ[ππ(π₯π₯)] =
β§ οΏ½ ππ(π₯π₯ππ ) ππ(π₯π₯ππ ) βͺ ππ
if X is a discrete random variable
β
β¨ βͺ οΏ½ ππ(π₯π₯) ππ (π₯π₯)ππππ if X is a continuous random variable β©ββ 2) Variance: The variance of a distribution is given by ππ 2 =
2 β§ οΏ½(π₯π₯ππ β ππ) ππ(π₯π₯ππ ) βͺ ππ
if X is a discrete random variable
β
β¨ 2 βͺ οΏ½(π₯π₯ β ππ) ππ(π₯π₯)ππππ if X is a continuous random variable β©ββ where ππ is the standard deviation of the distribution. th 3) r moment of a random variable X is defined as πΈπΈ(ππ ππ ) and is denoted by ππππβ² ππππβ² =
ππ β§ οΏ½ π₯π₯ππ ππ(π₯π₯ππ ) βͺ ππ
if X is a discrete random variable
β
β¨ ππ βͺ οΏ½ π₯π₯ ππ(π₯π₯)ππππ β©ββ
if X is a continuous random variable
4) rth moment about the mean or the rth central moment (denoted by ππππ ) is defined by ππππ =
ππ β§ οΏ½(π₯π₯ππ β ππ) ππ(π₯π₯ππ ) βͺ ππ β
β¨ ππ βͺ οΏ½(π₯π₯ β ππ) ππ(π₯π₯)ππππ β©ββ 5) Mean deviation from mean is given by β§ οΏ½|π₯π₯ππ β ππ|ππ(π₯π₯ππ ) βͺ ππ β
β¨ βͺ οΏ½ |π₯π₯ β ππ| ππ(π₯π₯)ππππ β©ββ
if X is a discrete random variable
if X is a continuous random variable if X is a discrete random variable
if X is a continuous random variable
Page 7 of 12
Note: (i) ππ1β² = πΈπΈ(ππ) (ii) ππ2 = π£π£π£π£π£π£π£π£π£π£π£π£π£π£π£π£ = ππ 2 = πΈπΈ (ππ 2 ) β {πΈπΈ(ππ)}2 Properties of expected values: (i) If c is a constant, then E(c)=c (ii) If ππ1 , ππ2 , β¦ , ππππ are n random variables defined on a sample space S. then πΈπΈ (ππ1 + ππ2 + β¦+ππππ=πΈπΈ(ππ1)+πΈπΈ(ππ2)+β¦+πΈπΈ(ππππ) (iii) πΈπΈ (ππππ + ππ) = ππππ (ππ) + ππ where ππ and ππ are constants. (iv) If X and Y are independent random variables defined on a sample space S then πΈπΈ (ππππ) = πΈπΈ(ππ)πΈπΈ(ππ) Properties of variance: (i) ππππππ(ππ) = 0, where ππ is a constant. (ii) ππππππ(ππ + ππ) = ππππππ(ππ) + ππππππ(ππ) if ππ and ππ are independent (iii) ππππππ(ππππ) = ππ 2 ππππππ(ππ) (iv) ππππππ(ππππ + ππ) = ππ2 ππππππ(ππ) Problems: 1. In a lottery, m tickets are drawn at a time out off n tickets numbered from 1 to n. Find the expected value of the sum of the numbers on the tickets drawn. 2. X is a continuous random variable with probability density function given by ππππ , 0 β€ π₯π₯ < 2 2ππ , 2 β€ π₯π₯ < 4 Find k and mean value of X. ππ(π₯π₯) = οΏ½ βππππ + 6ππ , 4 β€ π₯π₯ < 6 3. A variate X has the following probability distribution 9 X: -3 6 P(X=x):
1/6
1/2
1/3
Find E(X) and πΈπΈ(ππ 2 ). Hence evaluate πΈπΈ(2ππ + 1)2 4. The frequency distribution of a measurable characteristic varying between 0 and 2 is π₯π₯ 3 , 0 β€ π₯π₯ β€ 1 ππ(π₯π₯) = οΏ½ Calculate the standard deviation and also the mean 3 (2 β π₯π₯) , 1 β€ π₯π₯ β€ 2 deviation about the mean. Binomial Distribution: It is concerned with trails of a repetitive nature in which only the occurrence or nonoccurrence, success or failure, acceptance or rejection, yes or no of a particular events is of interest. Page 8 of 12
If we perform a series of independent trails such that for each trail ππ is the probability of success and ππ that of failure, then the probability of ππ successes in a series of ππ trails is given by ππ πΆπΆππ ππππ ππππβππ , where ππ takes any integral value from 0 to ππ . The probabilities of successes are therefore, given by 0,1,2, . . . ππ, . . . ππ ππππ , ππ πΆπΆ1 ππ ππππβ1 , ππ πΆπΆ2 ππ2 ππππβ2 , β¦ , ππ πΆπΆππ ππππ ππππβππ , β¦ , ππππ . The probability of the number of successes so obtained is called the binomial distribution for the simple reason that the probabilities are the successive terms in the expansion of the binomial (ππ + ππ)ππ . β΄ The sum of the probabilities = ππππ + ππ πΆπΆ1 ππ ππππβ1 + ππ πΆπΆ2 ππ2 ππππβ2 + β¦ + ππ πΆπΆππ ππππ ππππβππ + β¦ + ππππ Binomial frequency distribution: If ππ independent trails constitute on experiment and this experiment is repeated ππ times, then the frequency of r successes is ππ ππ πΆπΆππ ππππ ππππβππ . The possible number of successes together with these expected frequencies constitute the binomial frequency distribution. Problems: 1. The probability that a pen manufactured by a company will be defective is 1/10. If 12 such pens are manufactured, find the probability that (a) exactly two will be defective (b) atleast two will be defective (c) none of them will be defective 2. In 256 sets of 12 tossed of coin, in how many ways one can expect 8 heads and four tails. 3. In sampling a large number of parts manufactured by a machine, the mean number of defectives in a sample of 20 is 2. Out of 1000 such samples, how many would be expected to contain atleast 3 defective parts. 4. The following data are the number of seeds germinating out of 10 on damp filter paper for 80 sets of seeds. Fit a binomial distribution to these data: x: 0 1 2 3 4 5 6 7 8 9 10 f:
6
20
28
12
8
6
0
5. Fit a binomial distribution for the following data: x: 0 1 2 3 f:
2
14
20
34
0
0
0
0
4
5
22
8
6. Determine the binomial distribution for which mean = 2(variance) and mean + variance=3. Also find ππ(ππ β€ 3). 7. Out of 800 families with 5 children each, how many would you expect to have (a) 3 boys, (b) 5 girls (c) either 2 or 3 boys? Assume equal probabilities for boys and girls.
Page 9 of 12
Poisson distribution: It is a distribution related to the probabilities of events which are extremely rare, but which have a large number of independent opportunities for occurrence. The number of persons born blind per year in a large city and the number of deaths by horse kick in an army corps are some of the phenomena, in which this law is followed. Problems: 1. If the probability a bad reaction from a certain injection is 0.001, determine the chance that out of 2000 individuals more than two will get a bad reaction. 2. In a certain factory turning our razor blades, there is a small chance of 0.002 for any blade to be defective. The blades are supplied in packets of 10, use Poisson distribution to calculate the approximate number of packets containing no defective, one defective and two defective blades respectively in a consignment of 10,000 packets. 3. Fit a Poisson distribution to the set of observations: 4 x: 0 1 2 3 f:
122
60
15
1
2
4. Fit a Poisson distribution to the following: x:
0
1
2
3
4
f:
46
38
22
9
1
5. If a random variable has a Poisson distribution such that P(1)=P(2), find (i) mean of the distribution (ii) P(4). 6. X is a Poisson variate and it is found that the probability that X=2 is two-thirds of the probability that X=1. Find the probability that X=0 and the probability that X=3. What is the probability that X exceeds 3? 7. A certain screw making machine produces on average of 2 defective screws out of 100, and packs them in boxes of 500. Find the probability that a box contains 15 defective screws. Normal Distribution or Gaussian Distribution: Now we consider a continuous distribution of fundamental importance, namely the normal distribution. Any quantity whose variation depends on random causes is distributed according to the normal law. Its importance lies in the fact that a large number of distributions approximate to the normal distribution. Let us define a variate π§π§ =
π₯π₯βππππ βππππππ
------------------------(1)
where π₯π₯ is a binomial vairate with mean ππππ and S.D. οΏ½ππππππ so that π§π§ is a variate with mean zero and variance unity. In the limit as n tends to infinity, the distribution of π§π§ becomes a continuous distribution extending from ββ to β. Page 10 of 12
It can be shown that the limiting form of the binomial distribution (1) for large values of ππ when neither p nor ππ is very small, is the normal distribution. The normal curve is of the form
ππ(π₯π₯ ) =
1
ππ β2ππ
ππ
β(π₯π₯ βππ )2 2ππ 2
where ππ and ππ are the mean and standard deviation respectively.
Properties of Normal Distribution: I. The normal curve is bell-shaped and is symmetrical about its mean. It is unimodel with ordinates decreasing rapidly on both sides of the mean. The maximum ordinate is
ππ β2ππ
found by putting π₯π₯ = ππ. As it is symmetrical, its mean, median and mode are the same. 4
1
,
II. Mean deviation from the mean ππ β ππ 5
III. Moments about the mean ππ2ππ+1 = 0; ππ2ππ = (2ππ β 1)(2ππ β 3) β¦ 3.1 ππ 2ππ IV. The probability of x lying between π₯π₯1 and π₯π₯2 is given by the area under the normal curve from π₯π₯1 to π₯π₯2 , i.e. ππ(π₯π₯1 β€ π₯π₯ β€ π₯π₯2 ) = ππ(π§π§2 ) β ππ(π§π§1 ) where ππ(π§π§) =
π§π§ β« β2ππ 0 1
ππ
βπ§π§ 2 2
ππππ . This
integral is called the probability integral or the error function due to its use in the theory of sampling and the theory of errors. (Use normal table). (i) The area under the normal curve between the ordinates π₯π₯ = ππ β ππ and π₯π₯ = ππ + ππ is 0.6826, β 68% nearly. Thus approximately 2/3 of the values lie within these limits. (ii) The area under the normal curve between π₯π₯ = ππ β 2ππ and π₯π₯ = ππ + 2ππ is 0.9544, 1
β 95.5% nearly, which implies that about 4 % of the values lie outside these limits. 2
(iii) 99.73% of the values lie between π₯π₯ = ππ β 3ππ and π₯π₯ = ππ + 3ππ i.e. only a quarter % of the whole lies outside these limits. (iv) 95% of the values lie between π₯π₯ = ππ β 1.96ππ and π₯π₯ = ππ + 1.96ππ i.e. only 5% of the values lie outside these limits. (v) 99% of the values lie between π₯π₯ = ππ β 2.58ππ and π₯π₯ = ππ + 2.58ππ i.e. only 1% of the values lie outside these limits. (vi) 99.9% of the values lie between π₯π₯ = ππ β 3.29ππ and π₯π₯ = ππ + 3.29ππ
Probable Error: Any lot of articles manufactured to certain specifications is subject to small errors. In fact, measurement of any physical quantity shows slight error. In general, these errors of manufacture or experiment are of random nature and therefore, follow a normal distribution. While quoting a specification of an experimental result, we usually mention the probable error(ππ). It is such that the probability of an error falling within the limits ππ β ππ and ππ + ππ is Page 11 of 12
exactly equal to the chance of an error falling outside these limits, i.e. the chance of an error 1
2
lying within ππ β ππ and ππ + ππ is . It is given by Probable error β ππ 2
3
Problems: 1. X is a normal variate with mean 30 and S.D. 5, find the probabilities that (i) 26 β€ ππ β€ 40, (ii) ππ β₯ 45 and (iii) |ππ β 30| > 5 2. A certain number of articles manufactured in one batch were classified into three categories according to a particular characteristic, being less than 50, between 50 and 60 and greater than 60. If this characteristic is known to be normally distributed, determine the mean and standard deviation of this batch if 60%, 35% and 5% were found in these categories. 3. In the normal distribution, 31% of the items are under 45 and 8% are over 64. Find the mean and S.D of the distribution. 4. In the test on 2000 electric bulbs, it was found that the life of a particular make was normally distributed with an average life of 2040 hours and S.D of 60 hours. Estimate the number of bulbs likely to burn for (a) more than 2150 hours, (b) less than 1950 hours and (c) more than 1920 hours and but less than 2160 hours. 5. If the probability of committing an error of magnitude x is given by π¦π¦ =
β
βππ
ππ ββ
2 π₯π₯ 2
; compute
the probable error from the following data: ππ1 = 1.305 ; ππ2 = 1.301 ; ππ3 = 1.295 ; ππ4 = 1.286 ; ππ5 = 1.318 ; ππ6 = 1.321 ; ππ7 = 1.283 ; ππ8 = 1.289 ; ππ9 = 1.300 ; ππ10 = 1.286 6. Fit a normal curve to the following distribution: x:
2
4
6
8
10
f:
1
4
6
4
1
7. For a normally distributed variate with mean 1 and S.D.3, find the probabilities that (ππ )3.43 β€ ππ β€ 6.19 (ππππ ) β 1.43 β€ ππ β€ 6.19 8. The mean and standard deviation of the marks obtained by 1000 students in an examination are respectively 34.4 and 16.5. Assuming the normality of the distribution, find the approximate number of students expected to obtain marks between 30 and 60.
Page 12 of 12
2 TABLE A Areas of a Standard Normal Distribution (Alternate Version of Appendix I Table 4)
The table entries represent the area under the standard normal curve from 0 to the specified value of z.
0
z
z
.00
.01
.02
.03
.04
.05
.06
.07
.08
.09
0.0
.0000
.0040
.0080
.0120
.0160
.0199
.0239
.0279
.0319
.0359
0.1
.0398
.0438
.0478
.0517
.0557
.0596
.0636
.0675
.0714
.0753
0.2
.0793
.0832
.0871
.0910
.0948
.0987
.1026
.1064
.1103
.1141
0.3
.1179
.1217
.1255
.1293
.1331
.1368
.1406
.1443
.1480
.1517
0.4
.1554
.1591
.1628
.1664
.1700
.1736
.1772
.1808
.1844
.1879
0.5
.1915
.1950
.1985
.2019
.2054
.2088
.2123
.2157
.2190
.2224
0.6
.2257
.2291
.2324
.2357
.2389
.2422
.2454
.2486
.2517
.2549
0.7
.2580
.2611
.2642
.2673
.2704
.2734
.2764
.2794
.2823
.2852
0.8
.2881
.2910
.2939
.2967
.2995
.3023
.3051
.3078
.3106
.3133
0.9
.3159
.3186
.3212
.3238
.3264
.3289
.3315
.3340
.3365
.3389
1.0
.3413
.3438
.3461
.3485
.3508
.3531
.3554
.3577
.3599
.3621
1.1
.3643
.3665
.3686
.3708
.3729
.3749
.3770
.3790
.3810
.3830
1.2
.3849
.3869
.3888
.3907
.3925
.3944
.3962
.3980
.3997
.4015
1.3
.4032
.4049
.4066
.4082
.4099
.4115
.4131
.4147
.4162
.4177
1.4
.4192
.4207
.4222
.4236
.4251
.4265
.4279
.4292
.4306
.4319
1.5
.4332
.4345
.4357
.4370
.4382
.4394
.4406
.4418
.4429
.4441
1.6
.4452
.4463
.4474
.4484
.4495
.4505
.4515
.4525
.4535
.4545
1.7
.4554
.4564
.4573
.4582
.4591
.4599
.4608
.4616
.4625
.4633
1.8
.4641
.4649
.4656
.4664
.4671
.4678
.4686
.4693
.4699
.4706
1.9
.4713
.4719
.4726
.4732
.4738
.4744
.4750
.4756
.4761
.4767
2.0
.4772
.4778
.4783
.4788
.4793
.4798
.4803
.4808
.4812
.4817
2.1
.4821
.4826
.4830
.4834
.4838
.4842
.4846
.4850
.4854
.4857
2.2
.4861
.4864
.4868
.4871
.4875
.4878
.4881
.4884
.4887
.4890
2.3
.4893
.4896
.4898
.4901
.4904
.4906
.4909
.4911
.4913
.4916
2.4
.4918
.4920
.4922
.4925
.4927
.4929
.4931
.4932
.4934
.4936
2.5
.4938
.4940
.4941
.4943
.4945
.4946
.4948
.4949
.4951
.4952
2.6
.4953
.4955
.4956
.4957
.4959
.4960
.4961
.4962
.4963
.4964
2.7
.4965
.4966
.4967
.4968
.4969
.4970
.4971
.4972
.4973
.4974
2.8
.4974
.4975
.4976
.4977
.4977
.4978
.4979
.4979
.4980
.4981
2.9
.4981
.4982
.4982
.4983
.4984
.4984
.4985
.4985
.4986
.4986
3.0
.4987
.4987
.4987
.4988
.4988
.4989
.4989
.4989
.4990
.4990
3.1
.4990
.4991
.4991
.4991
.4992
.4992
.4992
.4992
.4993
.4993
3.2
.4993
.4993
.4994
.4994
.4994
.4994
.4994
.4995
.4995
.4995
3.3
.4995
.4995
.4995
.4996
.4996
.4996
.4996
.4996
.4996
.4997
3.4
.4997
.4997
.4997
.4997
.4997
.4997
.4997
.4997
.4997
.4998
3.5
.4998
.4998
.4998
.4998
.4998
.4998
.4998
.4998
.4998
.4998
3.6
.4998
.4998
.4998
.4999
.4999
.4999
.4999
.4999
.4999
.4999
For values of z greater than or equal to 3.70, use 0.4999 to approximate the shaded area under the standard normal curve.