STA301_LEC24

Page 1

Virtual University of Pakistan Lecture No. 24 of the course on Statistics and Probability by Miss Saleha Naghmi Habibullah


a

F(a) = p( x < a) =

−∞

f ( x ) dx


IN THE LAST LECTURE, YOU LEARNT • Graphical Representation of the Distribution Function of a Discrete Random Variable • Mathematical Expectation • Mean, Variance and Moments of a Discrete Probability Distribution • Properties of Expected Values


TOPICS FOR TODAY •Chebychev’s Inequality • Concept of Continuous Probability Distribution • Mathematical Expectation, Variance & Moments of a Continuous Probability Distribution


We begin with the discussion of the concept of the Chebychev’s Inequality in the case of a discrete probability distribution:


Chebychev’s Inequality If X is a random variable having mean µ and variance σ2 > 0, and k is any positive constant, then the probability that a value of X falls within k standard deviations of the mean is at least That is:

1   1 − 2  .  k 

1 P ( µ − kσ < X < µ + kσ ) ≥ 1 − 2 , k


Alternatively, we may state Chebychev’s theorem as follow:


Given the probability distribution of the random variable X with mean µ and standard deviation σ, the probability of the observing a value of X that differs the µ by k or more standard deviations cannot exceed 1/k2.


As indicated earlier, this inequality is due to the Russian mathematician P.L. Chebychev (1821-1894), and it provides a means of understanding how the

standard deviation measures variability about the mean of a random variable.

It holds for all probability distributions having finite mean and variance.


Let us apply this concept to the example of the number of petals on the flowers of a particular species that we considered earlier: EXAMPLE: If a biologist is interested in the number of petals on a particular flower, this number may take the values 3, 4, 5, 6, 7, 8, 9, and each one of these numbers will have its own probability.


The probability distribution of the random variable X is: No. of Petals X x1 = 3 x2 = 4 x3 = 5 x4 = 6 x5 = 7 x6 = 8 x7 = 9

P(x) 0.05 0.10 0.20 0.30 0.25 0.075 0.025 1


The mean of this distribution is: µ = E(X) = ∑XP(X) = 5.925 ≅ 5.9 And the standard deviation of this distribution is:

σ = S.D.( X ) = 36.925 − ( 5.925)

2

= 36.925 − 35.106 = 1.819 = 1.3


According to the Chebychev’s inequality, the probability is at least 1 - 1/22 = 1 - 1/4 = 3/4 = 0.75 that X will lie between µ - 2σ and µ + 2σ i.e. between 5.9 - 2(1.3) and 5.9 + 2(1.3) i.e. between 3.3 and 8.5


Let us have another look at the probability distribution:

No. of Petals X x1 = 3 x2 = 4 x3 = 5 x4 = 6 x5 = 7 x6 = 8 x7 = 9

P(x) 0.05 0.10 0.20 0.30 0.25 0.075 0.025 1


According to this distribution, the probability that X lies between 3.3 and 8.5 is 0.10 + 0.20 + 0.30 + 0.25 + 0.075 = 0.925 which is greater than 0.75 (AS indicated by the Chebychev’s inequality).


Finally, and most importantly, we will use the concepts in Chebychev's Rule and the Empirical Rule to build the foundation for statistical inference-making. The method is illustrated in next example.


EXAMPLE Suppose you invest a fixed sum of money in each of five business ventures. Assume you know that 70% of such ventures are successful, the outcomes of the ventures are independent of one another, and the probability distribution for the number, x, of successful ventures out of five is:


x P(x)

0 1 2 3 4 5 .002 .029 .132 .309 .360 .168

a)

Find µ = E(X). Interpret the result.

b)

Find

[

σ = E ( X − µ) Interpret the result.

2

].


c)

Graph P(x).

d) Locate µ and the interval µ + 2σ on the graph. Use either Chebychev’s Rule or the Empirical Rule to approximate the probability that x falls in this interval. Compare this result with the actual probability. e) Would you expect to observe fewer than two successful ventures out of five?


SOLUTION a)

Applying the formula, µ = E(X) = ∑x P(x) = 0(.002) + 1(.029) + 2(.132) + 3(.309) + 4.(.360) + 5(.168) = 3.50


INTERPRETATION:

On average, the number of successful ventures out of five will equal 3.5. (It should be remembered that this expected value has meaning only when the experiment – investing in five business ventures – is repeated a large number of times.)


b) Now we calculate the variance of X:


We know that σ2

= E[(X - µ)2] = ∑(x - µ)2 P(x)


Hence, we will need to construct a column of x - µ:

x 0 1 2 3 4 5

P(x) .002 .029 .132 .309 .360 .168

x-µ –3.5 –2.5 –1.5 –0.5 +0.5 +1.5

2

(x-µ) 12.25 6.25 2.25 0.25 0.25 2.25 Total

2

(x-µ) P(x) 0.02 0.18 0.30 0.08 0.09 0.38 1.05


Thus, the variance is σ2 = 1.05 and the standard deviation is

2

σ = σ = 1.05 = 1.02


This value measures the spread of the probability distribution of X, the number of successful ventures out of five.


c) The graph of P(x) is shown in the following figure with the mean µ and the interval µ + 2σ

= 3.50 + 2(1.02) = 3.50 + 2.04 = (1.46, 5.54)

shown on the graph.


p(x) 0.4 0.3 0.2 0.1 x

0 0

1

2

µ + 2σ (1.46)

3

4

µ

5

µ + 2σ (5.54)


Note particularly that Âľ = 3.5 locates the centre of the probability distribution.


Since this distribution is a theoretical relative frequency distribution that is moderately mound-shaped, we expect (from Chebychev’s Rule) at least 75% and, more likely (from the Empirical Rule), approximately 95% of observed x values to fall in the interval µ + 2σ ------ that is, between 1.46 and 5.54.


It can be seen from the above figure that the actual probability that X falls in the interval Âľ + 2Ďƒ includes the sum of P(x) for the values X = 2, X = 3, X = 4, and X = 5.


p(x) 0.4 0.3 0.2 0.1 x

0 0

1

2

µ + 2σ (1.46)

3

4

µ

5

µ + 2σ (5.54)


This probability is P(2) + P(3) + P(4) + P(5) = .132 +.309 + .360 + .168 = .969. Therefore, 96.9% of the probability distribution lies within 2 standard deviations of the mean.


This

percentage is CONSISTENT with both the Chebychev’s rule and the Empirical Rule.


d) Fewer than two successful ventures out of five implies that x = 0 or x = 1. Since both these values of x lie outside the interval Âľ + 2Ďƒ, we know from the Empirical Rule that such a result is unlikely (with approximate probability of only .05).


The exact probability, P(x < 1), is P(0) + P(1) = .002 + .029 = .031. Consequently, in a single experiment where we invest in five business ventures, we would not expect to observe fewer than two successful ones.


The key question: What is the significance of the Chebychev’s Inequality and the Empirical Rule? The answer to this question is that both these rules assist us in having a certain IDEA regarding amount of data lying between the mean minus a certain number of standard deviations and mean plus that same number of standard deviations.


Given any data-set, the moment we compute the mean and standard deviation, we HAVE an idea regarding the two points (i.e. mean minus two standard deviations, and mean plus two standard deviations) between which the BULK of our data lies.


If our data-set is humpshaped, we obtain this idea through the Empirical Rule, and if we don’t have any reason to believe that our data-set is hump-shaped, then we obtain this idea through the Chebychev’s Rule.


Next, we begin the discussion of CONTINUOUS RANDOM VARIABLES.


In this regard, the first point to be noted is that uptil now we have discussed discrete random variables – quantities that are countable.


We now begin the discussion of CONTINUOUS RANDOM VARIABLES – quantities that are measurable.


As stated in the very first lecture, continuous variables result from measurement, and can therefore take any value within a certain range. For example, the height of a normal Pakistani adult male may take any value between 5 feet 4 inches and 6 feet.


The temperature at a place, the amount of rainfall, time to failure for an electronic system, etc. are all examples of continuous random variable. Formally speaking, a continuous random variable can be defined as follows:


CONTINUOUS RANDOM VARIABLE A random variable X is defined to be continuous if it can assume every possible value in an interval [a, b], a < b, where a and b may be –∞ and +∞ respectively.


The function f(x) is called the probability density function, abbreviated to p.d.f., or simply density function of the random variable X.


A continuous probability distribution looks something like this:


f(x)

X


A p.d.f. has the following properties:

i)

f(x) > 0, for all x ∞

ii)

∫ f ( x ) dx =1

−∞


iii) The probability that X takes on a value in the interval [c, d], c < d is given by: P(c < x < d) =

d

âˆŤ

f ( x ) dx

c

which is the area under the curve y = f(x) between X = c and X = d, as shown in the following figure:


f(x) P(c < x < d)

c d The TOTAL area under the curve is 1.


In other words: 1) f(x) a non-negative function, 2) the integration takes place over all possible values of the random variable X between the specified limits, and 3) the probabilities are given by appropriate areas under the curve.


Since

k

P( X = k ) = âˆŤ f ( x ) dx = 0,

it should therefore be noted kthat the probability of a continuous random variable X taking any particular value k is always zero. That is why probability for a continuous random variable is measurable only over a given interval.


Further, since for a continuous random variable X, P(X = x) = 0 for every x, the following four probabilities are regarded as the same: P(c < X < d), P(c < X < d), P(c < X < d) and P(c < X d). They may be different for a discrete random variable.


The values (expressed as intervals) of a continuous random variable and their associated probabilities can be expressed by means of a formula.


We

now

discuss

the distribution function of a continuous random variable:


CONTINUOUS RANDOM VARIABLE A random variable X may also be defined as continuous if its distribution function F(x) is continuous and is differentiable everywhere except at isolated points in the given range.


In contrast with the graph of the distribution function of a discrete variable, the graph of F(x) in the case of a continuous variable has no jumps or steps but is a continuous function for all x-values, as shown in the following figure:


1

F(x)

F(a) 0

F(b) X


Since F(x) is a nondecreasing function of x, we have i)

f(x) > 0, x

ii)

F(x ) = ∫ f (x ) dx , for all x. −∞


The relationship between f(x) and F(x) is as follows:


f(x) is obtained by finding the derivative of F(x), i.e.

d F( x ) = f ( x) dx


Let us now explain the above concepts with the help of an example:


EXAMPLE a) Find the value of k so that the function f(x) defined as follows, may be a density function f(x) = kx, 0 < x < 2 = 0, elsewhere b)

Compute P(X = 1).

c)

Compute P(X > 1).

d)

Compute the distribution function F(x).

e)

P X < 1/2

(

)

1/ 3 < X < 2 / 3


SOLUTION a)

The function f(x) will be a density function, if

i)

f(x) > 0 for every x, and

ii)

∫ f ( x ) dx = 1

−∞


The first condition is satisfied when k > 0. The second condition will be satisfied, if

∫ f ( x ) dx = 1,

−∞


0

2

−∞ 0

0

2

2

−∞

0

2

i.e. if 1 = ∫ f ( x ) dx +∫ f ( x ) dx + ∫ f ( x ) dx i.e. if 1 = ∫ 0 dx +∫ kx dx + ∫ 0 dx

2  x2   +0 = 2k i.e. if 1 == 0 + k 2     0 This gives k = 1/2


We had f(x) = kx, 0 < x < 2 = 0, elsewhere and since we have obtained k = 1/2, hence:

x 2 ,

f ( x) =   0,

for 0 ≤ x ≤ 2 elsewhere


b) Since f(x) is continuous probability function, therefore P(X = 1) = 0.


c) P(X > 1) is obtained by computing the area under the curve (in this case, a straight line) between X=1 and X=2:

f(x) 1 f(x) = x|2

0

1

2

X


This area is obtained as follows:

P(X > 1)

= area of shaded region 2

= ∫ f ( x ) dx 1

2

=∫

1

2

x 2

dx =

[ ] 2

x 4

1

3 = 4


d) To compute the distribution function, we need to find:

x

F(x) = P(X < x) = ∫ f ( x ) dx −∞


We do so step as shown below:

by step,


For any x such that -∞ < x < 0, x

= 0 dx 0, ∫ F(x) = −∞


If 0 < x < 2, we have 0

x

x F( x ) = ∫ 0 dx + ∫   dx = −∞ 0 2 

x

[ ] x2 4

0

x = , 4

and, finally, for x > 2 we have 0

2x

2

2

F( x ) = ∫ 0 dx + ∫ dx + ∫ 0 dx = 1 −∞ 02 0


Hence F(x) = 0, for x < 0 2

x , for 0 < x < 2 = 4 =1,

for x > 2.


We will discuss the computation of the conditional probability

(

P X < 1/2

)

1/ 3 < X < 2 / 3

in the next lecture.


IN TODAY’S LECTURE, YOU LEARNT •Chebychev’s Inequality • Concept of Continuous Probability Distribution •Distribution Function of a Continuous Probability Distribution


IN THE NEXT LECTURE, YOU WILL LEARN • Mathematical Expectation, Variance & Moments of a Continuous Probability Distribution •BIVARIATE Probability Distribution


Turn static files into dynamic content formats.

Create a flipbook
Issuu converts static files into: digital portfolios, online yearbooks, online catalogs, digital photo albums and more. Sign up and create your flipbook.