Page 1 :
Probability, Basic Definitions, Random Experiment:, o An experiment, whose all possible outcomes are known in advance but the outcome, of any specific performance cannot predicted before the completion of the, experiment, o Eg: Tossing of a coin., Sample-space, o A set of all possible outcomes associated with same random experiment, o Denoted by ‘S’., o Eg: In the experiment of tossing a die, If we are interested in the number that shows on the top, face, then sample space would be S = {1, 2, 3, 4, 5, 6}, Experiment or Trial, o It is a series of action where the outcomes are always uncertain., o Eg: - Tossing of a coin, Selecting a card from deck of cards, throwing a dice., Event, o Subset of sample – space., o In any sample space we may be interested in the occurrence of certain events rather than in the, occurrence of a specific element in the sample space., Simple Event, o If an event is a set containing only one element of the sample-space, Compound Event, o A compound event is one that can be represented as a union of sample points, o Eg:, Event of drawing a heart from a deck of cards is the subset A = {heart} of the sample space S =, {heart, spade, club, diamond}., Therefore A is a simple event. None the event B of drawing a red card is a compound event since, B = {heart U diamond} = {heart, diamond}., Probability, o If a random experiment can result in any one of N different equally likely outcomes, and if, exactly n of these outcomes favours to A,, Then the probability of event A, P (A) = n/N i.e. favourable cases/total no. of cases., o Remarks:, ▪ If the probability of certain event is one, it doesn’t mean that event is going to happen, with certainty, ▪ It’s just predicting that, the event is most likely to occur in comparison to other events., Predictions depend upon the past information and of course also on the way of analysing, the information at hand, ▪ Similarly if the probability of certain event is zero, it doesn’t mean that, the event can, never occur!
Page 2 :
Mutually exclusive Event, o If two events are mutually exclusive they cannot occur simultaneously., Independent Events, o Events are said to be independent if the occurrence or non-occurrence of one does not affect the, occurrence or non-occurrence of other., Exhaustive Event, o A set of events is said to be exhaustive if the performance of random experiment always result, in the occurrence of at least one of them, , Conditional Probability, •, •, •, , •, •, , The conditional probability of an event B is the probability that the event will occur given the, knowledge that an event A has already occurred., This probability is written P (B|A), notation for the probability of B given A., In the case where events A and B are independent (where event A has no effect on the, probability of event B), the conditional probability of event B given event A is simply the, probability of event B, that is P(B)., If events A and B are not independent, then the probability of the intersection of A and B (the, probability that both events occur) is defined by P (A and B) = P (A) P (B|A)., If E and F are two events associated with the same sample space of a random experiment, the, conditional probability of the event E given that F has occurred,, , I.e. P (E|F) is given by, , •, , Eg:, , Properties of conditional probability, Let E and F be events of a sample space S of an experiment, then we have, , Property 1, P (S|F) = P (F|F) = 1, We know that
Page 3 :
Property 2, If A and B are any two events of a sample space S and F is an event of S such that P (F) ≠ 0, then, , In particular, if A and B are disjoint events,, , Property 3, , Multiplication Theorem on Probability, •, •, , Let E and F be two events associated with a sample space S., Conditional, probability, of, event, E, given, that, denoted by P(E|F) and is given by, , •, , From this result, we can write, , F, , has, , occurred, , is
Page 4 :
•, , The above result is known as the Multiplication rule of probability., Example:, o An urn contains 10 black and 5 white balls. Two balls are drawn from the urn one after, the other without replacement. What is the probability that both drawn balls are black?, o Solution:, ▪ Let E and F denote respectively the events that first and second ball drawn are, black., ▪ P(E) = P (black ball in first draw) = 10/15, ▪ Given, • First ball drawn is black, i.e., event E has occurred, now there, are 9 black balls and five white balls left in the urn., • Therefore, the probability that the second ball drawn is black, given that the, ball, in, the, first, draw, is, black,, is, nothing, but, the conditional probability of F given that E has occurred., • i.e. P(F|E) = 9/14, ▪ By multiplication rule of probability, we have, P (E ∩ F) = P (E) P (F|E), =10/ 15 X 9/14 = 3/7, , Note:, •, , Multiplication rule of probability for more than two events If E, F and G are three events of, sample space, we have, P (E ∩ F ∩ G) = P (E) P (F|E) P (G| (E ∩ F)) = P (E) P (F|E) P (G|EF), , •, , Similarly, the multiplication rule of probability can be extended for four or more events., , Independent Events, o If E and F are two events such that the probability of occurrence of one of them is not affected, by occurrence of the other. Such events are called independent events., o Let E and F be two events associated with the same random experiment, then E and F are said, to be independent if, P (E ∩ F) = P (E). P (F), Remarks
Page 5 :
o Two events E and F are said to be dependent if they are not independent, i.e. if, P (E ∩ F) ≠ P (E). P (F), o Sometimes there is a confusion between independent events and mutually exclusive events., ▪ Term ‘independent’ is defined in terms of ‘probability of events’ whereas mutually, exclusive is defined in term of events (subset of sample space)., ▪ Mutually exclusive events never have an outcome common, but independent events, may, have common outcome., ▪ Two independent events having nonzero probabilities of occurrence cannot be mutually, exclusive, and conversely, i.e. two mutually exclusive events having nonzero, probabilities of occurrence cannot be independent., o Two experiments are said to be independent if for every pair of events E and F, where E is, associated with the first experiment and F with the second experiment,, the probability of the simultaneous occurrence of the events E and F when the two experiments, are performed is the product of P(E) and P(F) calculated separately on the basis of two, experiments, i.e. P (E ∩ F) = P (E). P (F), o Three events A, B and C are said to be mutually independent, if, P (A ∩ B) = P (A) P (B), P (A ∩ C) = P (A) P (C), P (B ∩ C) = P (B) P(C) and, P (A ∩ B ∩ C) = P (A) P (B) P(C), If at least one of the above is not true for three given events, we say that the events are not, independent., o Example, ▪ A die is thrown. If E is the event ‘the number appearing is a multiple of 3’ and F be the, event ‘the number appearing is even’ then find whether E and F are independent ?, Solution:, W.k.t the sample space is S = {1, 2, 3, 4, 5, 6}, Now E = {3, 6}, F = {2, 4, 6} and E ∩ F = {6}, Then, • P(E) = 2/6 1/3, • P(F) = 3/6 = 1/2, • P(E ∩ F) = 1/6, Clearly P (E ∩ F) = P (E). P (F). Hence E and F are independent events., , Bayes' Theorem: Description, o Also called as inverse probability theorem, o Consider that there are two bags I and II., • Bag I contains 2 white and 3 red balls, • Bag II contains 4 white and 5 red balls., • One ball is drawn at random from one of the bags.
Page 6 :
•, •, •, , •, •, , Probability of selecting any of the bags (i.e. 1/2) or probability of drawing a ball of a, particular colour (say white) from a particular bag (say Bag I)., Probability, that, the, ball, drawn, is, of, a, particular, colour,, if, we are given the bag from which the ball is drawn., To find the probability that the ball drawn is from a particular bag (say Bag II), if the, colour of the ball drawn is given we have to find the reverse probability of Bag II to be, selected when an event occurred after it is known., Famous mathematician, John Bayes' solved the problem of finding reverse probability by, using conditional probability., Hence named as ‘Bayes theorem’ which was published posthumously in 1763., , Definitions:, Partition of a sample space, •, , •, , A set of events E1, E2, ..., En is said to represent a partition of the sample space S if, o Ei ∩ Ej = φ, i ≠ j, i, j = 1, 2, 3, ..., n, o E1 ∪ Ε2 ∪ ... ∪ En= S and, o P (Ei) > 0 for all i = 1, 2, ……,n., The events E1, E2, ..., En represent a partition of the sample space S if they are pairwise disjoint,, exhaustive and have nonzero probabilities., , Theorem of total probability, o Let {E1, E2,...,En} be a partition of the sample space S,, o Suppose that each of the events E1, E2,...,En has nonzero probability of occurrence., o Let A be any event associated with S, then, , Proof, Given that E1, E2,...,En is a partition of the sample space S. Therefore,
Page 7 :
Bayes' Theorem: Proof, If E1, E2 ,..., En are n non empty events which constitute a partition, of sample space S, i.e. E1, E2 ,..., En are pairwise disjoint and E1∪ E2∪ ... ∪ En = S and, A is any event of nonzero probability, then, , Proof:, By formula of conditional probability, we know that, , Remark, The following terminology is generally used when Bayes' theorem is applied., o, o, o, o, , The events E1, E2, ..., En are called hypotheses., The probability P(Ei) is called the priori probability of the hypothesis Ei, The conditional probability P(Ei |A) is called a posteriori probability of the hypothesis Ei., Also called the formula for the probability of "causes". Since the Ei's are a partition of the sample, space S, one and only one of the events Ei occurs (i.e. one of the events E i must occur and only, one can occur). Hence, the above formula gives us the probability of a particular Ei, given that, the event A has occurred., , Random Variables and its Probability Distributions, In most of the random experiments and Sample space , we were not only interested in the particular, outcome that occurs but rather in some number associated with that outcomes as shown in, following examples/experiments.
Page 8 :
o Experiments, o In tossing two dice, we may be interested in the sum of the numbers on the, two dice., o In tossing a coin 50 times, we may want the number of heads obtained., o In the experiment of taking out four articles (one after the other) at random, from a lot of 20 articles in which 6 are defective, we want to know the, number of defectives in the sample of four and not in the particular sequence, of defective and non-defective articles., o In all the above experiments,, o We have a rule which assigns to each outcome of the experiment a single real number., o This single real number may vary with different outcomes of the experiment. Hence, it is, a variable., o Also its value depends upon the outcome of a random experiment and, hence, is called, random variable., o A random variable is usually denoted by X., o A random variable can take any real value, therefore, its co-domain is the set of real, numbers. Hence, a random variable can be defined as follows, ▪ A random variable is a real valued function whose domain is the sample space of, a random experiment., ▪ Eg: Consider the experiment of tossing a coin two times in succession, ▪ Sample space of the experiment is S = {HH, HT, TH, TT}., ▪ If X denotes the number of heads obtained, then X is a random variable and for, each outcome, its value is as given below :, • X(HH) = 2, X (HT) = 1, X (TH) = 1, X (TT) = 0, ▪ Let Y denote the number of heads minus the number of tails for each outcome, of the above sample space S., • Y (HH) = 2, Y (HT) = 0, Y (TH) = 0, Y (TT) = – 2., ▪ Hence, X and Y are two different random variables defined on the same sample, space S., ▪ Note: More than one random variables can be defined on the same sample space., , Probability distribution of a random variable, o Description giving the values of the random variable along with the corresponding probabilities, is called the probability distribution of the random variable X., o In general, the probability distribution of a random variable X is defined as follows:, o The probability distribution of a random variable X is the system of numbers
Page 9 :
o Also for all possible values of the random variable X, all elements of the sample, space are covered. Hence, the sum of all the probabilities in a probability distribution, must be one., o If xi is one of the possible values of a random variable X, the statement, X = xi is true only at some point (s) of the sample space. Hence, the probability that X takes value, xi is always nonzero, i.e. P(X = xi) ≠ 0., , Mean of a random variable, o Mean is a measure of location or central tendency in the sense that it roughly locates a, middle or average value of the random variable., o Let X be a random variable whose possible values x1, x2, x3, ..., xn occur with probabilities p1, p2,, p3,..., pn, respectively. The mean of X, denoted by μ, is the number, i.e. the mean of X is, the weighted average of the possible values of X, each value being weighted by its probability, with which it occurs., o The mean of a random variable X is also called the expectation of X, denoted by E(X), Thus, o The mean or expectation of a random variable X is the sum of the products of all possible values, of X by their respective probabilities., , Variance of a random variable, o The mean of a random variable does not give us information about the variability in the, values of the random variable., o If the variance is small, then the values of the random variable are close to the mean. Also, random variables with different probability distributions can have equal means, as shown in the, following distributions of X and Y, X, P(X), , 1, 1, 8, , 2, 2, 8, , 3, 3, 8, , 4, 4, 8, , The variables X and Y are different, however their means are same., The diagrammatic representation of these distributions are shown below:
Page 10 :
o Let X be a random variable whose possible values x1, x2,...,xn occur with probabilities p(x1),, p(x2),..., p(xn) respectively., , The non- negative number is called the standard deviation of the random variable
Page 11 :
Bernoulli Trials and Binomial Distribution, Bernoulli trials, The outcome of any trial is independent of the outcome of any other trial. In each of such trials,, the probability of success or failure remains constant. Such independent trials which have only, two outcomes usually referred as ‘success’ or ‘failure’ are called Bernoulli trials., o Trials of a random experiment are called Bernoulli trials, if they satisfy, the following conditions :, o There should be a finite number of trials., o The trials should be independent., o Each trial has exactly two outcomes: success or failure., o The probability of success remains the same in each trial., o Example: 30 Six balls are drawn successively from an urn containing 7 red and 9 black, balls. Tell whether or not the trials of drawing balls are Bernoulli trials when after each, draw the ball drawn is (i) replaced (ii) not replaced in the urn., Solution, (i) The number of trials is finite. When the drawing is done with replacement, the, probability of success (say, red ball) is p = 7/16 which is same for all six trials, (draws). Hence, the drawing of balls with replacements are Bernoulli trials., (ii)When the drawing is done without replacement, the probability of success (i.e., red, ball) in first trial is 7/16, in 2nd trial is 6/15 if the first ball drawn is red or 7/15 if the first, ball drawn is black and so on. Clearly, the probability of success is not same for all trials,, hence the trials are not Bernoulli trials, o, , Binomial distribution, The probability distribution of number of successes in an experiment consisting of n Bernoulli trials, may be obtained by the binomial expansion of (q + p)n. Hence, this distribution of number of, successes X can be written as, , The above probability distribution is known as binomial distribution with parameters n and p,, because for given values of n and p, we can find the complete probability distribution., The probability of x successes P(X = x) is also denoted by P(x) and is given by, , This P(x) is called the probability function of the binomial distribution. A binomial distribution with, n-Bernoulli trials and probability of success in each trial as p, is denoted by B (n, p).