Can Normal Distribution Be Skewed: Detailed Facts, Examples And FAQs

Mazhar Image1 300x135 1

Normal distribution is skewed with zero skewness, so the answer to the most common confusion can normal distribution be skewed is normal distribution is not skewed distribution as the curve of the normal distribution is symmetric without tail whose skewness is zero. The normal distribution curve is bell shaped with symmetry on the curve.

Since the skewness is lack of symmetry in the curve so if the symmetry is present in the curve there is lack of skewness.

How do you tell if the data is normally distributed?

For the data to check whether normally distributed or not just try to sketch the histogram and from the curve of the curve if the symmetry is present in the curve then the data is normally distributed, from the curve of data itself the question can normal distribution be skewed or not cleared if the concept of skewness is clear. Sketching the histogram or curve in each case is tedious or time consuming so instead of that their are number of statistical tests like Anderson-Darling statistic (AD) which are more useful to tell whether data is normally distributed or not.

The data which follows normal distribution have zero skewness in the curve and the characteristics of the curve of the skewed distribution is different without symmetry, this we will understand with the following example:

Example: Find the percent of score lies between 70 to 80 if the score of mathematics of university students are normally distributed with the mean 67 and standard deviation 9?

Can Normal Distribution Be Skewed
symmetry in the normal distribution or can normal distribution be skewed

Solution:

To find the percent of score we follow the probability for the normal distribution discussed earlier in normal distribution, so to do so first we will convert into normal variate and follow the table discussed in normal distribution to find the probability using the conversion

Z=(X-μ)/σ

we want to find the score percent between 70 and 80 so we use random variable values 70 and 80 with the given mean 67 and standard deviation 9 this gives

Z=70-67/9 = 0.333

and

Z=80-67/9 = 1.444

This we can sketch as

image 126

the above shaded area shows the region between z=0.333 and z=1.444 from the table of standard normal variate the probabilities are

P(z > 0.333)=0.3707
and
P(z > 1.444)=0.0749
so
p(0.333 < z0.333)-P(z > 1.444)=0.3707-0.0749=0.2958

so 29.58% students will score between 70 to 80 .

In the above example the skewness of the curve is zero and the curve is symmetric, to check the data is normally distributed or not we have to perform the hypothesis tests.

How do you tell if a distribution is skewed left or right?

The distribution is known to be skewed if it is right tailed or left tailed in the curve so the depending on the nature of the curve we can judge whether the distribution is positive skewed or negative skewed. The concept of skewness is discussed in detail in the articles positively and negatively skewed distribution. If the symmetry in the left side lacks the distribution is skewed left and if the symmetry lacks in the right side the distribution is skewed right. The best way to check the distribution is skewed is to check the variation in the central tendencies that is if mean<median<mode then the distribution is left skewed and if mean>median>mode then the distribution is right skewed. The geometrical representation is as follows

image 127
left skewed distribution
image 128
right skewed distribution

The measures to calculate the skewness left or right for the information given in detail in the article of skewness.

What is an acceptable skewness?

Since the skewness as earlier discussed is lack of symmetry so what range is acceptable that must be clear. The question can normal distribution is skewed arise to check whether in the normal distribution is acceptable or not and the answer of the acceptable skewness is in normal distribution because in normal distribution the skewness is zero and the distribution in which skewness is near to zero is more acceptable. So after the testing for skewness if the skewness is nearer to zero then the skewness is acceptable depending on the requirement and range for the client.

In brief the acceptable skewness is the skewness which is nearer to zero as per the requirement.

How skewed is too skewed?

The skewness is the statistical measurement to check the symmetry present in the curve of the distribution and the information and all the measures to check skewness is present or not, depending on that we can find if the distribution is far from zero then too skewed or symmetry is zero then we can say the distribution is too skewed.

How do you determine normal distribution?

To determine the distribution is normal or not we have to look the distribution have the symmetry or not if the symmetry is present and the skewness is zero then the distribution is normal distribution, the detail methods and techniques were already discussed in detail in normal distribution

Do outliers skew data?

In the distribution data if any data follow unusual way and very far or away from the usual data that is known as outlier and in most of the cases the outliers are responsible for the skewness of the distribution and because of the unusual nature of outliers the distribution have skewness, so we can say that in the distribution the outliers skew data. The outliers in all cases will not skew data they skewed data only if they also follow the systematic sequence in continuous distribution to give left or right tailed curve.

In the previous articles the detail discussion of normal distribution and skewed distribution discussed.

Negatively Skewed Distribution: 9 Facts You Should Know

image 11 300x172 1

 Skewed Distribution | skewed distribution definition

    The distribution in which symmetry is not present and the curve of the distribution shows tail either left or right side is known as skewed distribution, so skewness is the asymmetry present in the curve or histogram apart from the symmetric or normal curve.

depending on the measure of central tendencies the nature of the distribution whether skewed or not can be evaluated there is special relations between mean, mode and median in left-tailed or right-tailed skewed distribution.

image 11
left-skewed distribution
image 12
right-skewed distribution

normal distribution vs skewed | normal vs skewed distribution

Normal distributionskewed distribution
In Normal distribution the curve is symmetricIn skewed distribution the curve is not symmetric
The measure of central tendencies mean, mode and median are equalThe measure of central tendencies mean, mode and median are not equal
mean=median =modemean>median>mode or mean<median<mode
Normal distribution vs skewed distribution

skewed distribution examples in real life

skewed distribution occurs in number of real life situation like the ticket sale of the particular show or movies in different months, record of athletes performance in competition, stock market returns, real estate rates fluctuation, life cycle of specific species, income variation, exam score and many more competitive outcomes. The distribution curve which shows asymmetry occurs frequently in applications.

difference between symmetrical and skewed distribution | symmetrical and skewed distribution

The main difference between the symmetrical distributions and skewed distribution is the differences between the central tendencies mean median and mode and in addition as the name suggest in the symmetrical distribution the curve of distribution is symmetric while in the skewed distribution the curve is not symmetric but have the skewness and it may be right-tailed or left tailed or may be both tailed also, the different distribution differs only on the nature of the skewness and symmetry so all the probability distributions can be classified into these two main categories.

To find the nature of distribution whether symmetric or skewed we must have to either draw the curve of the distribution or the coefficient of skewness with the help of absolute or relative measures.

highly skewed distribution

The modal or highest value of the distribution if differs from mean and median that gives the skewed distribution, if the highest value coincides with mean and median and equal then the distribution is symmetric distribution, the highly skewed distribution may be positive or negative. The skewed distribution modal value can be find out using the coefficient of skewness.

Negatively skewed distribution| which is a negatively skewed distribution

Negatively Skewed Distribution
Negatively skewed distribution

Any distribution in which the measure of central tendencies follows the order mean<median<mode and the coefficient of skewness in negative in the negatively skewed distribution, the negatively skewed distribution is also known as left skewed distribution because in negatively skewed distribution the tail of graph or plot of information is left.

image 66
negatively skewed distribution

The coefficient of skewness for the negatively skewed distribution can easily find out with the usual methods of finding the coefficients of skewness.

negatively skewed distribution example

If 150 students in an examination performed as given below then find the nature of skewness of the distribution

marks0-1010-2020-3030-4040-5050-6060-7070-80
freq124018012421412

Solution: To find the nature of skewness of distribution we have to calculate the coefficient of skewness for which we require mean, mode, median and standard deviation for the given information so for this we will calculate these with the help of the following table

class intervalfmid value
x
c.f.d’=(x-35)/10f*d’f*d’2
0-1012512-3-36108
10-20401552-2-80160
20-30182570-1-1818
30-4003570000
40-5012458211212
50-604255124284168
60-701465138342126
70-801275150448192
total=52total=784

so the measures will be

\\begin{array}{l} Median =\\mathrm{L}+\\frac{\\left(\\frac{\\mathrm{N}}{2}-\\mathrm{C}\\right)}{\\mathrm{f}} \\times \\mathrm{h}=40+\\frac{75-70}{10} \\times 10=45 \\\\Mean (\\overline{\\mathrm{x}})=\\mathrm{A}+\\frac{\\sum_{\\mathrm{i}=1}^{\\mathrm{k}} \\mathrm{fd}^{\\prime}}{\\mathrm{N}} \\times \\mathrm{h}=35+\\frac{52}{150} \\times 10=39.16 \\end{array}

and

\\begin{aligned}   Standard Deviation }(\\sigma) &=\\mathrm{h} \\times \\sqrt{\\frac{\\sum \\mathrm{fd}^{\\prime 2}}{\\mathrm{~N}}-\\left(\\frac{\\sum \\mathrm{fd}}{\\mathrm{N}}\\right)^{2}} \\\\ &=10 \\times \\sqrt{\\frac{784}{150}-\\left(\\frac{52}{150}\\right)^{2}} \\\\&=10 \\times \\sqrt{5.10}=22.38 \\end{aligned}

hence the coefficient of skewness for the distribution is

S_k=\\frac{3(Mean-Median)}{\\sigma} \\\\=\\frac{3(39.16-45)}{22.38}=-0.782

negatively skewed distribution mean median mode

In the negatively skewed distribution mean median mode is in ascending order which represents the tail on the left side of the curve of distribution, the measure of central tendencies mean median and mode for the negatively skewed distribution follows exactly the reverse pattern of positively skewed distribution. The curve of the negatively skewed distribution is also an inverse image of the positively skewed distribution. so Mean<median<mode in negatively skewed distribution.

negatively skewed distribution curve

The nature of the curve for the negatively skewed distribution curve is left-skewed without symmetry either in a histogram or continuous curve.

image 85
negatively skewed distribution curve
image 86
left-skewed

As symmetry is the measure to calculate the asymmetry present in the distribution, so the distribution curve of negatively skewed distribution shows the asymmetry present on the left side.

positively skewed normal distribution

The continuous distribution which is following the normal distribution curve including the asymmetry by gathering the information to the right tail shows the right-skewed curve asymmetric about the median following descending order in the central tendencies mean median and mode.

FAQs

Why chi square distribution is positively skewed

The chi-square distribution gives the values from zero to infinity and the curve of the distribution gathers the information in the right tail so it shows the right-skewed curve hence the chi-square distribution is a positively skewed distribution.

Is Poisson distribution positively skewed

Yes, Poisson distribution is a positively skewed distribution as the information scattered near the right tail so the nature of the plot is positively skewed

Why does negative binomial distribution always positively skew

The negative binomial distribution is always positively skewed because negative binomial distribution is the generalization of pascal distribution which is always positively skewed so is the negative binomial distribution.

Does skewness have any impact on linear regression models My dependent variable and my interaction variable is positively skewed

The impact on linear regression of the model having my dependent variable and my interaction skewed does not mean the regression error is also skewed and vice versa as the error is skewed does not mean the variables are skewed.

Skewness: 7 Important Facts You Should Know

image 53 1

Content

 Skewness

    The curve which is the plotted observations represents the skewness if the shape of the curve is not symmetric, of the given set. In other words the lack of symmetry in the graph of the given information represents the skewness of the given set. Depending on the tail in the right or left the skewness is known as positively skewed or negatively skewed. The distribution depending on this skewness is known as positively skewed distribution or negatively skewed distribution

image 53
positively skewed curve
image 54
Negatively skewed curve

The mean, mode and median shows the nature of distribution so if the nature or shape of the curve is symmetric these measure of central tendencies are equal and for the skewed distributions these measure of central tendencies varies as either mean>median>mode or mean<median<mode.

Variance and Skewness

VarianceSkewness
Amount of variability can be obtained using varianceDirection of variability can be obtained using skewness
Application of measure of variation is in Business and economicsApplication of measure of Skewness is in medical and life sciences
variance and skewness

Measure of Skewness

To find the degree and the direction of the frequency distribution whether positive or negative the measure of skewness is very helpful even with the help of the graph we know the positive or negative nature of the skewness but the magnitude will not be exact in graphs hence these statistical measures gives the magnitude of lack of symmetry.

To be specific the measure of skewness must have

  1. Unit free so that the different distributions can be comparable if the units are same or different.
  2. Value of measure for symmetric distribution zero and positive or negative for positive or negative distributions accordingly.
  3. The value of measure should vary if we move from negative skewness to positive skewness.

There are two types of measure of skewness

  1. Absolute Measure of skewness
  2. Relative Measure of skewness

Absolute Measure of skewness

In the symmetrical distribution the mean, mode and median are same so in absolute measure of skewness the difference of these central tendencies gives the extent of symmetry in the distribution and the nature as positive or negative skewed distribution but the absolute measure for different units is not useful while comparing two sets of information.

The Absolute skewness can be obtained using

  1. Skewness(Sk)=Mean-Median
  2. Skewness(Sk)=Mean-Mode
  3. Skewness(Sk)=(Q3-Q2)-(Q2-Q1)

Relative Measure of skewness

Relative measure of skewness is used to compare the skewness in two or more distributions by eliminating the influence of variation, relative measure of skewness is known as coefficient of skewness, the following are the important relative measure of skewness.

  1. Karl Pearson’s Coefficient of Skewness

This method is used most often to calculate skewness

S_k=\\frac{Mean-Mode}{\\sigma}

this coefficient of skewness is positive for positive distribution, negative for negative distribution and zero for the symmetric distribution. This Karl Pearson’s coefficient usually lies between +1 and -1. If Mode is not defined then to calculate the Karl Pearson’s coefficient we use the formula as

S_k=\\frac{3(Mean-Mode)}{\\sigma}

If we use this relation then Karl Pearson’s coefficient lies between +3 and -3.

2. Bowleys’s Coefficient of Skewness|Quartile measure of skew ness

In Bowleys’s coefficient of skewness the quartile deviations were used to find the skewness so it is also known as quartile measure of skewness

S_k=\\frac{(Q_3-Q_2)-(Q_2-Q_1)}{(Q_3-Q_1)} \\\\=\\frac{(Q_3-2Q_2+Q_1)}{(Q_3-Q_1)}

or we can write it as

S_k=\\frac{(Q_3-M)-(M-Q_1)}{(Q_3-Q_1)} \\\\=\\frac{(Q_3-2M+Q_1)}{(Q_3-Q_1)}

this value of coefficient is zero if the distribution is symmetric and the value for positive distribution is positive, for negative distribution is negative. The value of Sk lies between -1 and +1.

3. Kelly’s Coefficient of Skewness

In this measure of skewness the percentiles and deciles are used to calculate the skewness, the coefficient is

S_k=\\frac{(P_{90}-P_{50})-(P_{50}-P_{10})}{(P_{90}-P_{10})} \\\\=\\frac{(P_{90}-2P_{50}+P_{10})}{(P_{90}-P_{10})}

where these skewness involves the 90, 50 and 10 percentiles and using deciles we can write it as

S_k=\\frac{(D_9-D_5)-(D_5-D_1)}{(D_9-D_1)} \\\\=\\frac{(D_9-2D_5+D_1)}{(D_9-D_1)}

in which 9,5 and 1 deciles were used.

4. β and γ Coefficient of Skewness| Measure of skew ness based on moments.

Using the central moments the measure of skewness the β coefficient of skewness can be define as

\\beta_1=\\frac{{\\mu_3}^2}{{\\mu_2}^3}

this coefficient of skewness gives value zero for the symmetric distribution but this coefficient not tells specifically for the direction either positive or negative, so this drawback can be removed by taking square root of beta as

\\gamma_1=\\pm \\sqrt{\\beta_1}=\\frac{\\mu_3}{{\\mu_2}^{3/2}}=\\frac{\\mu_3}{\\sigma^3}

this value gives the positive and negative value for the positive and negative distributions respectively.

Examples of skewness

  1.  Using the following information find the coefficient of skewness
Wages0-1010-2020-3030-4040-5050-6060-7070-80
No. of people121835425045208

Solution: To find the coefficient of skewness we will use karl Pearson’s coefficient

frequencymid-value(x)fxfx2
0-1012560300
10-2018152704050
20-30352587521875
30-404235147051450
40-5050452250101250
50-6045552475136125
60-702065130084500
70-8087560045000
2309300444550

the karl pearson coefficient of skewness is

\\begin{array}{l} \\text { Karl person's coefficient of skewness }=J=\\frac{\\text { Mean }-\\text { Mode }}{S . D .}\\\\ \\begin{array}{l} \\text { Mean, } \\quad \\bar{x}=\\frac{1}{N} \\sum_{i} f_{i} x_{i}, \\quad \\text { Mode }=l+\\frac{c\\left(f_{1}-f_{0}\\right)}{\\left(f_{1}-f_{0}\\right)+\\left(f_{1}-f_{2}\\right)} \\\\ \\text { Standard deviation }=\\sqrt{\\frac{1}{N} \\sum_{i} f_{i} x_{i}^{2}-\\bar{x}^{2}} \\end{array} \\end{array}

\\begin{array}{c} \\text { Mean }=\\frac{9300}{230}=40.43 \\\\ \\text { S.D. }=\\sqrt{\\frac{1}{N} \\sum_{i} f_{i} x_{i}^{2}-\\bar{x}^{2}}=\\sqrt{\\frac{1}{230}(444550)-\\left[\\frac{9300}{230}\\right]^{2}}=17.27 . \\end{array}

the modal class is maximum frequent class 40-50 and the respective frequencies are

f_{0}=42, f_{1}=50,f_{2}=45

thus

\\text { Hence, Mode }=40+\\frac{10(50-42)}{(50-42)+(50-45)}=46.15

so the coefficient of skewness will be

=\\frac{40.43-46.15}{17.27}=-0.3312

which shows the negative skewness.

2. Find the coefficient of skewness of the frequency distributed marks of 150 students in certain examination

marks0-1010-2020-3030-4040-5050-6060-7070-80
freq104020010401614

Solution: To calculate the coefficient of skewness we require mean, mode, median and standard deviation for the given information so for calculating these we form the following table

class intervalfmid value
x
c.f.d’=(x-35)/10f*d’f*d’2
0-1010510-3-3090
10-20401550-2-80160
20-30202570-1-2020
30-4003570000
40-5010458011010
50-604055120280160
60-701665136348144
70-801475150456244
total=64total=828

now the measures will be

\\begin{array}{l} Median =\\mathrm{L}+\\frac{\\left(\\frac{\\mathrm{N}}{2}-\\mathrm{C}\\right)}{\\mathrm{f}} \\times \\mathrm{h}=40+\\frac{75-70}{10} \\times 10=45 \\\\Mean (\\overline{\\mathrm{x}})=\\mathrm{A}+\\frac{\\sum_{\\mathrm{i}=1}^{\\mathrm{k}} \\mathrm{fd}^{\\prime}}{\\mathrm{N}} \\times \\mathrm{h}=35+\\frac{64}{150} \\times 10=39.27 \\end{array}

and

\\begin{aligned}   Standard Deviation }(\\sigma) &=\\mathrm{h} \\times \\sqrt{\\frac{\\sum \\mathrm{fd}^{\\prime 2}}{\\mathrm{~N}}-\\left(\\frac{\\sum \\mathrm{fd}}{\\mathrm{N}}\\right)^{2}} \\\\ &=10 \\times \\sqrt{\\frac{828}{150}-\\left(\\frac{64}{150}\\right)^{2}} \\\\&=10 \\times \\sqrt{5.33}=23.1 \\end{aligned}

hence the coefficient of skewness for the distribution is

S_k=\\frac{3(Mean-Median)}{\\sigma} \\\\=\\frac{3(39.27-45}{23.1}=-0.744

3. Find the mean, variance and coefficient of skewness of distribution whose first four moments about 5 are 2,20,40 and 50.

Solution: since the first four moments are given so

\\begin{array}{c} \\mu_{1}^{\\prime}(5)=\\frac{1}{N} \\sum_{i=1}^{k} f_{i}\\left(x_{i}-5\\right)=2 ; \\mu_{2}^{\\prime}(5)=\\frac{1}{N} \\sum_{i=1}^{k} f_{i}\\left(x_{i}-5\\right)^{2}=20 ; \\\\ \\mu_{3}^{\\prime}(5)=\\frac{1}{N} \\sum_{i=1}^{k} f_{i}\\left(x_{i}-5\\right)^{3}=40 \\quad \\text { and } \\quad \\mu_{4}^{\\prime}(5)=\\frac{1}{N} \\sum_{i=1}^{k} f_{i}\\left(x_{i}-5\\right)^{4}=50 . \\\\ \\mu_{1}^{\\prime}(5)=\\frac{1}{N} \\sum_{i=1}^{k} f_{i} x_{i}-5=2 \\\\ \\Rightarrow \\bar{x}=2+5=7 \\end{array}

so we can write it

\\begin{array}{l} \\mu_{r}=\\mu_{r}^{\\prime}(A)-{ }^{r} C_{1} \\mu_{r-1}^{\\prime}(A) \\mu_{1}^{\\prime}(A)+{ }^{r} C_{2} \\mu_{r-2}^{\\prime}(A)\\left[\\dot{\\mu}_{1}^{\\prime}(A)\\right]^{2}-\\ldots .+(-1)^{r}\\left[\\mu_{1}^{\\prime}(A)\\right]^{r} \\\\ \\text { Hence } \\mu_{2}=\\mu_{2}^{\\prime}(5)-\\left[\\mu_{1}^{\\prime}(5)\\right]^{2}=20-4=16 \\\\ \\mu_{3}=\\mu_{3}^{\\prime}(5)-3 \\mu_{2}^{\\prime}(5) \\mu_{1}^{\\prime}(5)+2\\left[\\mu_{1}^{\\prime}(5)\\right]^{3} \\\\ 40-3 \\times 20 \\times 2+2 \\times 2^{3}=-64 \\end{array}

so the coefficient of skewness is

\\beta_{1}=\\frac{\\mu_{3}^{2}}{\\mu_{2}^{3}}=\\frac{(-64)^{2}}{(16)^{3}}=-1

Positively skewed distribution definition|Right skewed distribution meaning

Any distribution in which the measure of central tendencies i.e mean, mode and median having positive values and the information in the distribution lacks the symmetry.

In other words the positively skewed distribution is the distribution in which the measure of central tendencies follows as mean>median>mode in the right side of the curve of the distribution.

If we sketch the information of the distribution the curve will be right tailed because of which positively skewed distribution is also known as right skewed distribution.

positively skewed distribution or right skewed distribution
positively/ right skewed distribution

from above curve it is clear that the mode is the smallest measure in positively or right skewed distribution and the mean is the largest measure of central tendencies.

positively skewed distribution example|example of right skewed distribution

  1. For a positively skewed or right skewed distribution if the coefficient of skewness is 0.64, find the mode and median of the distribution if mean and standard deviations are 59.2 and 13 respectively.

Solution: The given values are mean=59.2, sk=0.64 and  σ=13 so using the relation

S_k=\\frac{mean-mode}{\\sigma} \\\\0.64=\\frac{59.2-\\text { Mode }}{13} \\\\Mode =59.20-8.32=50.88 \\\\Mode =3 Median -2 Mean \\\\50.88=3 Median -2(59.2) \\\\Median =\\frac{50.88+118.4}{3}=\\frac{169.28}{3}=56.42

2. Find the standard deviation of the positively skewed distribution whose coefficient of skewness is 1.28 with mean 164 and mode 100?

Solution: In the same way using the given information and the formula for the coefficient of positively skewed distribution

S_k=\\frac{mean-mode}{\\sigma} \\\\1.28=\\frac{164-100}{\\sigma} \\\\\\sigma=\\frac{64}{1.28}=50

so the standard deviation will be 50.

3. In the quarterlies deviations if the addition of first and third quarterlies is 200 with median 76 find the value of third quartile of the frequency distribution which is positively skewed with coefficient of skewness 1.2?

Solution: To find the third quartile we have to use the relation of coefficient of skewness and quarterlies, since the given information is

S_k=1.2 \\\\Q_1+Q_3=200 \\\\Q_2=76[ \\\\S_{k}=\\frac{\\left(Q_{3}+Q_{1}-2 Q_{2}\\right)}{\\left(Q_{3}-Q_{1}\\right)} \\\\1.2=\\frac{(200-2 \\times 76)}{\\left(Q_{3}-Q_{1}\\right)} \\\\Q_{3}-Q_{1}=\\frac{48}{1.2}=40 \\\\Q_{3}-Q_{1}=40

from the given relation we have

Q_1+Q_3=200 \\\\Q_1=200-Q_3

from these two equations we can write

Q_{3}-Q_{1}=40 \\\\ Q_{3}-(200-Q_3)=40 \\\\2Q_3=240 \\\\Q_3=120

so the value of the third quartile is 120.

4. Find the coefficient of skewness for the following information

x93-9798-102103-107108-112113-117118-122123-127128-132
f25121714631

Solution: here we will use Bowley’s measure of skewness using quartiles

classfrequencycumulative frequency
92.5-97.522
97.5-102.557
102.5-107.51219
107.5-112.51736
112.5-117.51450
117.5-122.5656
122.5-127.5359
127.5-132.5160
N=60

As Nth/4=15th observation of class is 102.5-107.5 , Nth/2=30th observation of class is 107.5-112.5 and 3Nth/4=45th observation of class is 112.5-117.5 so

Q_{1}=l_{1}+\\frac{\\left(\\frac{N}{4}-m_{1}\\right) c_{1}}{f_{1}}=102.5+\\frac{\\left(\\frac{60}{4}-7\\right) 5}{12}=105.83

and

Q_{3}=l_{3}+\\frac{\\left(\\frac{3 N}{4}-m_{3}\\right) c_{3}}{f_{3}}=112.5+\\frac{\\left(\\frac{3 \\times 60}{4}-36\\right) 5}{14}=115.714

and median is

Q_{2}=l_{2}+\\frac{\\left(\\frac{N}{2}-m_{2}\\right) c_{2}}{f_{2}}=107.5+\\frac{\\left(\\frac{60}{2}-19\\right) 5}{17}=110.735

thus

Q=\\frac{Q_{3}+Q_{1}-2 M}{Q_{3}-Q_{1}}=\\frac{115.714+105.83-2 \\times 110.735}{115.714-105.83}=0.0075

which is positively skewed distribution.

where is the mean in a positively skewed distribution

We know that the positively skewed distribution is right skewed distribution so the curve is right tailed the meaning of this most of the information will be nearer to the tail so the mean in a positively skewed distribution is nearer to the tail and since in positively or right skewed distribution mean>median>mode so mean will be after the median.

Right skewed distribution mean median mode|relationship between mean median and mode in positively skewed distribution

In the positively skewed or right skewed distribution the measure of central tendencies mean, median and mode are in the order mean>median>mode, as mode is the smallest one then median and the largest central tendency is the mean which for the right tailed curve is nearer to the tail of the curve for the information.

so the relationship between mean median and mode in positively skewed distribution is in the increasing order and with the help of the difference of these two central tendencies the coefficient of skewness can be calculated, so mean, median and mode gives the nature of skewness also.

positively skewed distribution graph|positively skewed distribution curve

The graph either in the form of smooth curve or in the form of histogram for the discrete information, the nature is right tailed as the mean of the information gather around the tail of the curve as skewness of distribution discusses the shape of the distribution. Since the large amount of data is in left of the curve and tail of the curve onto the right is longer.

some of the graphs of positively distributed information are as follows

image
image 1
image 2

image 3
image 4

from the above graphs it is clear that the curve has lacking the symmetry in any aspects .

positively skewed score distribution

In any distribution if the scores are in the positively skewed that is the score following the positively skewed distribution as mean>median>mode and the curve of the distribution score having right tailed curve in which score is affected by the large value.

This type of distribution is known as positively skewed score distribution. All the properties and rules for this distribution are the same from positively skewed or right skewed distribution.

positive skew frequency distribution

In positively skewed frequency distribution on average the frequency of the information are smaller as compared to the distribution so the positive skew frequency distribution is nothing but the positively skewed or right skewed distribution where the curve is right tailed curve.

positive vs negative skewed distribution|positively skewed distribution vs negatively skewed

positive skewed distributionnegative skewed distribution
In the positively skewed distribution the information is distributed as mean is the largest and mode is smallest In the negatively skewed distribution the information is distributed as mean is the smallest and mode is largest
the curve is right tailed the curve is left tailed
mean>median>modemean<median<mode

FAQs

How do you know if a distribution is positively or negatively skewed

The skewness is positive if mean>median>mode and negative if mean<median<mode,

From the distribution curve also we can judge if the curve is right tailed it is positive and if the curve is left tailed it is negative

How do you determine positive skewness

By calculating the measure of coefficient of skewness if positive then skewness is positive or by plotting the curve of distribution if right tailed then positive or by checking mean>median>mode

What does a positive skew represent

The positive skewness represent that the score of the distribution lies nearer to large values and the curve is right tailed and the mean is the largest measure

How do you interpret a right skewed histogram

if the histogram is right skewed then the distribution is positively skewed distribution where mean>median>mode

In distributions that are skewed to the right what is the relationship of the mean median and mode

The relationship is mean>median>mode

Conclusion:

The skewness is important concept of statistics which gives the asymmetry or lack of symmetry present in the distribution of probability depending on the positive or negative value it is classified as positively skewed distribution or negatively skewed distribution, in the above article the brief concept with examples discussed   , if you require further reading go through

https://en.wikipedia.org/wiki/skewness

For more post on mathematics, please follow our Mathematics page

Hermite Polynomial: 9 Complete Quick Facts

image 145

  The Hermite polynomial is widely occurred in applications as an orthogonal function. Hermite polynomial is the series solution of Hermite differential equation.

Hermite’s Equation

    The differential equation of second order with specific coefficients as

d2y/dx2 – 2x dy/dx + 2xy = 0

is known as Hermite’s equation, by solving this differential equation we will get the polynomial which is Hermite Polynomial.

Let us find the solution of the equation

d2y/dx2 – 2x dy/dx + 2ny = 0

with the help of series solution of differential equation

101 1

now substituting all these values in the Hermite’s equation we have

image 136

This equation satisfies for the value of k=0 and as we assumed the value of k will not be negative, now for the lowest degree term xm-2 take k=0 in the first equation as the second gives negative value, so the coefficient xm-2 is

a0m (m-1)=0 ⇒ m=0,m=1

as a0 ≠ 0

now in the same way equating the coefficient of xm-1 from the second summation

104

and equating the coefficients of xm+k to zero,

ak+2(m+k+2)(m+k+1)-2ak(m+k-n) = 0

we can write it as

ak+2 = 2(m+k-n)/(m+k+2)(m+k+1) ak

if m=0

ak+2 = 2(k-n)/(k+2)(k+1) ak

if m=1

ak+2 = 2(k+1-n)/(k+3)(k+2) ak

for these two cases now we discuss the cases for k

When $m=0, ak+2= 2(k-n)/ (k+2)(k+1)} ak$

If, $k=0 a2 =-2 n/2 a0=-n a0$

$k=1, a3=2(1-n)/6 a1 =-2(n-1)/3 ! a1$

If $k=2, a4 =2(2-n)/12 a2 =2 (2-n)/12 (-n a0) =22 n(n-2)/4 ! a0$

108

so far m=0 we have two conditions when a1=0, then a3=a5=a7=….=a2r+1=0 and when a1 is not zero then

image 140

by following this put the values of a0,a1,a2,a3,a4 and a5 we have

image 141

and for m=1 a1=0 by putting k=0,1,2,3,….. we get

ak+2 = 2(k+1-n)/(k+3)(k+2)ak

image 142

so the solution will be

image 143

so the complete solution is

image 144

where A and B are the arbitrary constants

Hermite Polynomial

   The Hermite’s equation solution is of the form y(x)=Ay1(x)+By2(x) where y1(x) and y2(x) are the series terms as discussed above,

image 145
image 146

one of these series end if n is non negative integer if n is even y1 terminates otherwise y2 if n is odd, and we can easily verify that for n=0,1,2,3,4…….. these polynomials are

1,x,1-2x2, x-2/3 x3, 1-4x2+4/3x4, x-4/3x3+ 4/15x5

so we can say here that the solution of Hermite’s equation are constant multiple of these polynomials and the terms containing highest power of x is of the form 2nxn denoted by Hn(x) is known as Hermite polynomial

Generating function of Hermite polynomial

Hermite polynomial usually defined with the help of relation using generating function

image 150
image 149

[n/2] is the greatest integer less than or equal to n/2 so it follows the value of Hn(x) as

image 151
image 152

this shows that Hn(x) is a polynomial of degree n in x and

Hn(x) = 2nxn + πn-2 (x)

where πn-2 (x) is the polynomial of degree n-2 in x, and it will be even function of x for even value of n and odd function of x for odd value of n, so

Hn(-x) = (-1)n Hn(x)

some of the starting Hermite polynomials are

H0(x) = 1

H1(x) = 2x

H2(x) = 4x2 – 2

H3(x) = 8x3-12

H4(x) = 16x4 – 48x2+12

H5(x) = 32x2 – 160x3+120x

Generating function of Hermite polynomial by Rodrigue Formula

Hermite Polynomial can also be defined with the help of Rodrigue formula using generating function

image 153

since the relation of generating function

image 154

  Using the Maclaurin’s theorem, we have

image 155

or

by putting z=x-t and

for t=0,so z=x gives

this we can show in another way as

differentiating

with respect to t gives

taking limit t tends to zero

now differentiating with respect to x

taking limit t tends to zero

from these two expressions we can write

in the same way we can write

 differentiating n times put t=0, we get

from these values we can write

from these we can get the values

Example on Hermite Polynomial           

  1. Find the ordinary polynomial of

Solution: using the Hermite polynomial definition and the relations we have

2. Find the Hermite polynomial of the ordinary polynomial

Solution: The given equation we can convert to Hermite as

and from this equation equating the same powers coefficient

hence the Hermite polynomial will be

Orthogonality of Hermite Polynomial | Orthogonal property of Hermite Polynomial

The important characteristic for Hermite polynomial is its orthogonality which states that

To prove this orthogonality let us recall that

which is the generating function for the Hermite polynomial and we know

so multiplying these two equations we will get

multiplying and integrating within infinite limits

and since

so

using this value in above expression we have

which gives

now equate the coefficients on both the sides

which shows the orthogonal property of Hermite polynomial.

  The result of orthogonal property of Hermite polynomial can be shown in another way by considering the recurrence relation

Example on orthogonality of Hermite Polynomial

1.Evaluate the integral

Solution: By using the property of orthogonality of hermite polynomial

since the values here are m=3 and n=2 so

2. Evaluate the integral

Solution: Using the orthogonality property of Hermite polynomial we can write

Recurrence relations of Hermite polynomial

The value of Hermite polynomial can be easily find out by the recurrence relations

Hermite polynomial
Hermite polynomial recurrence relations

These relations can easily obtained with the help of definition and properties.

Proofs:1. We know the Hermite equation

y”-2xy’+2ny = 0

and the relation

image 174

by taking differentiation with respect to x partially we can write it as

image 175

from these two equations

image 176
image 177

now replace n by n-1

image 178
image 179

by equating the coefficient of tn

image 180
image 181

so the required result is

image 182

2. In the similar way differentiating partially with respect to t the equation

image 183

we get

image 184
image 185

n=0 will be vanished so by putting this value of e

image 186
image 187

now equating the coefficients of tn

image 188

thus

image 189

3. To prove this result we will eliminate Hn-1 from

image 190

and

image 191

so we get

image 192

thus we can write the result

image 193

4. To prove this result we differentiate

image 194

we get the relation

image 195

substituting the value

image 196

and replacing n by n+1

image 197

which gives

image 173

Examples on Recurrence relations of Hermite polynomial

1.Show that

H2n(0) = (-1)n. 22n (1/2)n

Solution:

To show the result we have

image 172

H2n(x) =

taking x=0 here we get

image 171

2. Show that

H’2n+1(0) = (-1)n 22n+1 (3/2)2

Solution:

Since from the recurrence relation

H’n(x) = 2nHn-1(X)

here replace n by 2n+1 so

H’2n-1(x) = 2(2n+1) H2n(x)

taking x=0

image 170

3. Find the value of

H2n+1(0)

Solution

Since we know

image 169

use x=0 here

H2n-1(0) = 0

4. Find the value of H’2n(0).

Solution :

we have the recurrence relation

H’n(x) = 2nHn-1(x)

here replace n by 2n

H’2n(x) = =2(2n)H2n-1(x)

put x=0

H’2n(0) = (4n)H2n-1(0) = 4n*0=0

5. Show the following result

image 168

Solution :

Using the recurrence relation

H’n(x) = 2nHn-1 (x)

so

image 167

and

d3/dx3 {Hn(x)} = 23n(n-1)(n-2)Hn-3(x)

differentiating this m times

image 166

which gives

image 165

6. Show that

Hn(-x) = (-1)n Hn(x)

Solution :

we can write

image 163
image 164

from the coefficient of tn we have

image 162

and for -x

image 161

7. Evaluate the integral and show

Solution : For solving this integral use integration parts as

image 160

Now differentiation under the Integral sign differentiate with

respect to x

image 159

using

H’n(x) = 2nHn-1 (x)

and

H’m(x) = 2mHm-1 (x)

we have

image 157

and since

???? n,m-1 = ????n+1, m

so the value of integral will be

image 156

Conclusion:

The specific polynomial which frequently occurs in application is Hermite polynomial, so the basic definition, generating function , recurrence relations and examples related to Hermite Polynomial were discussed in brief here   , if you require further reading go through

https://en.wikipedia.org/wiki/Hermite_polynomials

For more post on mathematics, please follow our Mathematics page

13 Facts On Chebyshev’s Inequality & Central Limit Theorem

In the probability theory the Chebyshev’s Inequality & central limit theorem deal with the situations where we want to find the probability distribution of sum of large numbers of random variables in approximately normal condition, Before looking the limit theorems we see some of the inequalities, which provides the bounds for the probabilities if the mean and variance is known.

Markov’s inequality

The Markov’s inequality for the random variable X which takes only positive value for a>0 is

gif

to prove this for a>0 consider

Since

gif

now taking expectation of this inequality we get

gif

the reason is

gif

which gives the Markov’s inequality for a>0 as

gif

Chebyshev’s inequality

 For the finite mean and variance of random variable X the Chebyshev’s inequality for k>0 is

gif

where sigma and mu represents the variance and mean of random variable, to prove this we use the Markov’s inequality as the non negative random variable

gif

for the value of a as constant square, hence

gif

this equation is equivalent to

gif

as clearly

gif

Examples of Markov’s and Chebyshev’s inequalities :

  1. If the production of specific item is taken as random variable for the week with mean 50 , find the probability of production exceeding 75 in a week and what would be the probability if the production of a week is between 40 and 60 provided the variance for that week is 25?

Solution: Consider the random variable X for the production of the item for a week then to find the probability of production exceeding 75 we will use Markov’s inequality as

gif

Now the probability for the production in between 40 to 60 with variance 25 we will use Chebyshev’s inequality as

gif

so

gif

this shows the probability for the week if the production is between 40 to 60 is 3/4.

2. Show that the chebyshev’s inequality which provides upper bound to the probability is not particularly nearer to the actual value of the probability.

Solution:

Consider the random variable X is uniformly distributed with mean 5 and variance 25/3 over the interval (0,1) then by the chebyshev’s inequality we can write

gif.latex?P%28%7CX

but the actual probability will be

gif.latex?P%28%7CX 5%7C%26gt%3B4%5C%7D%3D0

which is far from the actual probability likewise if we take the random variable X as normally distributed with mean and variance then Chebyshev’s inequality will be

gif

but the actual probability is

gif.latex?P%28%7CX %5Cmu%7C%26gt%3B2%20%5Csigma%5C%7D%3DP%5Cleft%5C%7B%5Cleft%7C%5Cfrac%7BX %5Cmu%7D%7B%5Csigma%7D%5Cright%7C%26gt%3B2%5Cright%5C%7D%3D2%5B1

Weak Law of Large Numbers

The weak law for the sequence of random variables will be followed by the result that Chebyshev’s inequality can be used as the tool for proofs for example to prove

gif

if the variance is zero that is the only random variables having variances equal to 0 are those which are constant with probability 1 so by Chebyshev’s inequality for n greater than or equal to 1

gif

as

gif

by the continuity of the probability

gif

which proves the result.

to prove this we assume that variance is also finite for each random variable in the sequence so the expectation and variance

gif

now from the Chebyshev’s inequality the upper bound of the probability as

gif

which for n tending to infinity will be

gif

Central Limit theorem

The central limit theorem is one of the important result in probability theory as it gives the distribution to the sum of large numbers which is approximately normal distribution in addition to the method for finding the approximate probabilities for sums of independent random variables central limit theorem also shows the empirical frequencies of so many natural populations exhibit bell-shaped means normal curves, Before giving the detail explanation of this theorem we use the result

“If the sequence of random variables Z1,Z2,…. have the distribution function and moment generating function as FZn and Mzn then

gif

Central Limit theorem: For the sequence of identically distributed and independent random variables X1,X2,……. each of which having the mean μ and variance σ2 then the distribution of the sum

gif

tends to the standard normal as n tends to infinity for a to be real values

Proof: To prove the result consider the mean as zero and variance as one i.e μ=0 & σ2=1 and the moment generating function for Xi exists and finite valued so the moment generating function for the random variable Xi/√n will be

gif

hene the moment generating function for the sum ΣXi/√n will be

gif

Now let us take L(t)=logM(t)

so

gif

to show the proof we first show

by showing its equivalent form

%202

since

hence this shows the result for the mean zero and variance 1, and this same result follows for the general case also by taking

%20%5Csigma

and for each a we have

gif

Example of the Central Limit theorem

To calculate the distance in light year of a star from the lab of an astronomer, he is using some measuring techniques but because of change in atmosphere each time the distance measured is not exact but with some error so to find the exact distance he plans to observe continuously in a sequence and the average of these distances as the estimated distance, If he consider the values of measurement identically distributed and independent random variable with mean d and variance 4 light year, find the number of measurement to do to obtain the 0.5 error in the estimated and actual value?

Solution: Let us consider the measurements as the independent random variables in sequence X1,X2,…….Xn so by the Central Limit theorem we can write

gif

which is the approximation into standard normal distribution so the probability will be

CodeCogsEqn 76

so to get the accuracy of the measurement at 95 percent the astronomer should measure n* distances where

gif.latex?2%20%5CPhi%5Cleft%28%5Cfrac%7B%5Csqrt%7Bn%5E%7B*%7D%7D%7D%7B4%7D%5Cright%29

so from the normal distribution table we can write it as

which says the measurement should be done for 62 number of times, this also can be observed with the help of Chebyshev’s inequality by taking

gif

so the inequality results in

gif.latex?P%5Cleft%5C%7B%5Cleft%7C%5Csum %7Bi%3D1%7D%5E%7Bn%7D%20%5Cfrac%7BX %7Bi%7D%7D%7Bn%7D d%5Cright%7C%26gt%3B0.5%5Cright%5C%7D%20%5Cleq%20%5Cfrac%7B4%7D%7Bn%280

hence for n=16/0.05=320 which gives certainity that there will be only o.5 percent error in the measurement of the distance of the star from the lab of observations.

2. The number of admitted students in engineering course is Poisson distributed with mean 100, it was decided that if the admitted students are 120 or more the teaching will be in two sections otherwise in one section only, what will be the probability that there will be two sections for the course?

Solution: By following the Poisson distribution the exact solution will be

gif

which is obviously not give the particular numerical value, If we consider the random variable X as the students admitted then by the central limit theorem

which can be

gif.latex?%5Cbegin%7Barray%7D%7Bl%7D%20%3DP%5Cleft%5C%7B%5Cfrac%7BX 100%7D%7B%5Csqrt%7B100%7D%7D%20%5Cgeq%20%5Cfrac%7B119.5 100%7D%7B%5Csqrt%7B100%7D%7D%5Cright%5C%7D%20%5C%5C%20%5Capprox%201

which is the numerical value.

3. Calculate the probability that the sum on ten die when rolled is between 30 and 40 including 30 and 40?

Solution: Here considering the die as Xi for ten values of i. the mean and variance will be

gif

thus following the central limit theorem we can write

gif.latex?%5Cbegin%7Baligned%7D%20P%5B29.5%20%5Cleq%20X%20%5Cleq%2040.5%5C%7D%20%26amp%3B%3DP%5Cleft%5C%7B%5Cfrac%7B29.5 35%7D%7B%5Csqrt%7B%5Cfrac%7B350%7D%7B12%7D%7D%7D%20%5Cleq%20%5Cfrac%7BX 35%7D%7B%5Csqrt%7B%5Cfrac%7B350%7D%7B12%7D%7D%7D%20%5Cleq%20%5Cfrac%7B40.5 35%7D%7B%5Csqrt%7B%5Cfrac%7B350%7D%7B12%7D%7D%7D%5Cright%5C%7D%20%5C%5C%20%26amp%3B%20%5Capprox%202%20%5CPhi%281.0184%29

which is the required probability.

4. For the uniformly distributed independent random variables Xi on the interval (0,1) what will be the approximation of the probability

gif

Solution: From the Unifrom distribution we know that the mean and variance will be

gif

Now using the central limit theorem we can

gif.latex?%5Cbegin%7Baligned%7D%20P%5Cleft%5C%7B%5Csum %7B1%7D%5E%7B10%7D%20X %7Bi%7D%26gt%3B6%5Cright%5C%7D%20%26amp%3B%3DP%5Cleft%5C%7B%5Cfrac%7B%5Csum %7B1%7D%5E%7B10%7D%20X %7Bi%7D 5%7D%7B%5Csqrt%7B10%5Cleft%28%5Cfrac%7B1%7D%7B12%7D%5Cright%29%7D%7D%26gt%3B%5Cfrac%7B6 5%7D%7B%5Csqrt%7B10%5Cleft%28%5Cfrac%7B1%7D%7B12%7D%5Cright%29%7D%7D%5Cright%5C%7D%20%5C%5C%20%26amp%3B%20%5Capprox%201

thus the summation of the random variable will be 14 percent.

5. Find the probability for the evaluator of the exam to give grades will be 25 exams in starting 450 min if there are 50 exams whose grading time is independent with mean 20 min and standard deviation 4 min.

Solution: Consider the time require to grade the exam by the random variable Xi so the random variable X will be

gif

since this task for 25 exam is withing 450 min so

gif
gif
gif

here using the central limit theorem

gif.latex?%5Cbegin%7Baligned%7D%20P%5BX%20%5Cleq%20450%5C%7D%20%26amp%3B%3DP%5Cleft%28%5Cfrac%7BX 500%7D%7B%5Csqrt%7B400%7D%7D%20%5Cleq%20%5Cfrac%7B450 500%7D%7B%5Csqrt%7B400%7D%7D%5Cright%29%20%5C%5C%20%26amp%3B%20%5Capprox%20P%28Z%20%5Cleq 2.5%5C%7D%20%5C%5C%20%26amp%3B%3DP%28Z%20%5Cgeq%202.5%5C%7D%20%5C%5C%20%26amp%3B%3D1 %5CPhi%282.5%29%3D0

which is the required probability.

Central Limit theorem for independent random variables

For the sequence which is not identically distributed but having independent random variables X1,X2,……. each of which having the mean μ and variance σ2 provided it satisfies

  1. each Xi is uniformly bounded
  2. sum of the variances is infinite, then
gif

Strong Law of Large Numbers

Strong Law of Large numbers is very crucial concept of the probability theory which says that the average of sequence of commonly distributed random variable with probability one will converge to the mean of that same distribution

Statement: For the sequence of identically distributed and independent random variables X1,X2,……. each of which having the finite mean with probability one then

gif

Proof: To prove this consider the mean of each of random variable is zero, and the series

gif

now for this consider power of this as

after taking the expansion of the right hand side terms we have the terms of the form

gif

since these are independents so the mean of these will be

gif

with the help of combination of the pair the expansion of the series now will be

gif

since

gif

so

gif

we get

gif

this suggest the inequality

gif

hence

gif

By the convergence of the series since the probability of each random variable is one so

gif

since

gif

if the mean of each random variable is not equal to zero then with deviation and probability one we can write it as

gif

or

gif

which is required result.

One Sided Chebyshev Inequality

The one sided Chebysheve inequality for the random variable X with mean zero and finite variance if a>0 is

Chebyshev's Inequality
chebyshev inequality

to prove this consider for b>0 let the random variable X as

gif

which gives

gif

so using the Markov’s inequality

Chebyshev's Inequality
one sided chebyshev

which gives the required inequality. for the mean and variance we can write it as

gif

This further can be written as

gif

Example:

Find the upper bound of the probability that the production of the company which is distributed randomly will at least 120, if the production of this certain company is having mean 100 and variance 400.

Solution:

Using the one sided chebyshev inequaility

gif

so this gives the probability of the production within a week atleast 120 is 1/2, now the bound for this probability will be obtained by using Markov’s inequality

which shows the upper bound for the probability.

Example:

Hundred pairs are taken from two hundred persons having hundred men and hundred women find the upper bound of the probability that at most thirty pair will consist a men and a women.

Solution:

Let the random variable Xi as

gif

so the pair can be expressed as

gif

Since every man can equally likely to be pair with remaining people in which hundred are women so the mean

gif

in the same way if i and j are not equal then

gif

as

%20197

hence we have

gif
gif
gif
gif
gif

using the chebyshev inequality

gif.latex?P%5C%7BX%20%5Cleq%2030%5C%7D%20%5Cleq%20P%5C%7B%7CX

which tells that the possibility of pairing 30 men with women is less than six, thus we can improve the bound by using one sided chebyshev inequality

gif.latex?%5Cbegin%7Baligned%7D%20P%5BX%20%5Cleq%2030%5C%7D%20%26amp%3B%3DP%5BX%20%5Cleq%2050.25

Chernoff Bound

If the moment generating function already known then

gif

as

gif

in the same way we can write for t<0 as

gif

Thus the Chernoff bound can be define as

gif

this inequality stands for all the values of t either positive or negative.

Chernoff bounds for the standard normal random variable

The Chernoff bounds for the standard normal random variable whose moment generating function

%202%7D

is

so minimizing this inequality and right hand side power terms gives for a>0

%202%7D

and for a<0 it is

%202%7D

Chernoff bounds for the Poisson random variable

The Chernoff bounds for the Poisson random variable whose moment generating function

gif

is

gif

so minimizing this inequality and right hand side power terms gives for a>0

%20%5Clambda 1%29%7D%5Cleft%28%5Cfrac%7B%5Clambda%7D%7Bi%7D%5Cright%29

and it would be

gif

Example on Chernoff Bounds

In a game if a player is equally likely to either win or lose the game independent of any past score, find the chernoff bound for the probability

Solution: Let Xi denote the winning of the player then the probability will be

gif

for the sequence of n plays let

gif

so the moment generating function will be

gif

here using the expansions of exponential terms

CodeCogsEqn 77

so we have

%202%7D

now applying the property of moment generating function

%202%7D%20%5Cend%7Baligned%7D

This gives the inequality

%202%20n%7D%20%5Cquad%20a%26gt%3B0

hence

Conclusion:

The inequalities and limit theorem for the large numbers were discussed and the justifiable examples for the bounds of the probabilities were also taken to get the glimpse of the idea, Also the the help of normal, poisson random variable and moment generating function is taken to demonstrate the concept easily, if you require further reading go through below books or for more Article on Probability, please follow our Mathematics pages.

A first course in probability by Sheldon Ross

Schaum’s Outlines of Probability and Statistics

An introduction to probability and statistics by ROHATGI and SALEH

Problems On Probability & Its Axioms

lagrida latex editor 33

Probability is a fundamental concept in mathematics that allows us to quantify uncertainty and make predictions about the likelihood of events occurring. It plays a crucial role in various fields, including statistics, economics, physics, and computer science. In this section, we will explore the definition of probability and its importance in mathematics, as well as the axioms that form the foundation of probability theory.

Definition of Probability and Its Importance in Math

Probability can be defined as a measure of the likelihood of an event occurring. It is represented as a number between 0 and 1, where 0 indicates impossibility and 1 indicates certainty. The concept of probability is essential in mathematics because it helps us analyze and understand uncertain situations.

In real life, we encounter probabilistic situations every day. For example, when flipping a fair coin, we know that the probability of it landing on heads is 0.5. Similarly, when rolling a fair six-sided die, the probability of rolling a specific number, say 3, is 1/6. By understanding and applying probability, we can make informed decisions and assess risks in various scenarios.

Probability theory provides a systematic framework for studying and analyzing uncertain events. It allows us to mathematically model and analyze random phenomena, such as coin flips, dice rolls, and card games. By using probability theory, we can calculate the likelihood of different outcomes, estimate the expected value of random variables, and make predictions based on available data.

Axioms of Probability Theory

To ensure a consistent and coherent approach to probability, mathematicians have established a set of axioms that form the foundation of probability theory. These axioms provide a rigorous framework for defining and manipulating probabilities. Let’s take a closer look at the three axioms of probability:

  1. Non-negativity: The probability of any event is always a non-negative number. In other words, the probability of an event cannot be negative.

  2. Additivity: For any collection of mutually exclusive events (events that cannot occur simultaneously), the probability of the union of these events is equal to the sum of their individual probabilities. This axiom allows us to calculate the probability of complex events by considering the probabilities of their constituent parts.

  3. Normalization: The probability of the entire sample space (the set of all possible outcomes) is equal to 1. This axiom ensures that the total probability of all possible outcomes is always 1, providing a consistent framework for probability calculations.

By adhering to these axioms, we can ensure that our calculations and reasoning about probabilities are logically sound and consistent. These axioms, along with other probability concepts, such as conditional probability, independence, and Bayes’ theorem, form the building blocks of probability theory.

In the upcoming sections, we will delve deeper into probability theory, exploring various probability concepts, examples, exercises, and calculations. By understanding the axioms and principles of probability, we can develop a solid foundation for tackling more complex probability problems and applying probability in real-world scenarios.

Problems on Probability and Its Axioms

Example 1: Restaurant Menu Combinations

Imagine you’re at a restaurant with a diverse menu, offering a variety of appetizers, entrees, and desserts. Let’s say there are 5 appetizers, 10 entrees, and 3 desserts to choose from. How many different combinations of a meal can you create?

To solve this problem, we can use the fundamental principle of counting. The principle states that if there are m ways to do one thing and n ways to do another, then there are m * n ways to do both.

In this case, we can multiply the number of choices for each course: 5 appetizers * 10 entrees * 3 desserts = 150 different combinations of a meal.

Example 2: Probability of Item Purchases

Suppose you’re running an online store and you want to analyze the probability of customers purchasing certain items together. Let’s say you have 100 customers, and you track their purchase history. Out of these customers, 30 have bought item A, 40 have bought item B, and 20 have bought both items A and B. What is the probability that a randomly selected customer has bought either item A or item B?

To solve this problem, we can use the principle of inclusion-exclusion. This principle allows us to calculate the probability of the union of two events by subtracting the probability of their intersection.

First, we calculate the probability of buying item A or item B separately. The probability of buying item A is 30/100 = 0.3, and the probability of buying item B is 40/100 = 0.4.

Next, we calculate the probability of buying both item A and item B. This is given by the intersection of the two events, which is 20/100 = 0.2.

To find the probability of buying either item A or item B, we add the probabilities of buying each item and subtract the probability of buying both items: 0.3 + 0.4 – 0.2 = 0.5.

Therefore, the probability that a randomly selected customer has bought either item A or item B is 0.5.

Example 3: Probability of Card Occurrences

Let’s consider a standard deck of 52 playing cards. What is the probability of drawing a heart or a diamond from the deck?

To solve this problem, we need to determine the number of favorable outcomes (drawing a heart or a diamond) and the total number of possible outcomes (drawing any card from the deck).

There are 13 hearts and 13 diamonds in a deck, so the number of favorable outcomes is 13 + 13 = 26.

The total number of possible outcomes is 52 (since there are 52 cards in a deck).

Therefore, the probability of drawing a heart or a diamond is 26/52 = 0.5.

Example 4: Probability of Temperature Occurrences

Suppose you are interested in predicting the weather for the next day. You have observed that over the past year, the probability of a hot day is 0.3, the probability of a cold day is 0.2, and the probability of a rainy day is 0.4. What is the probability that tomorrow will be either hot or cold, but not rainy?

To solve this problem, we can use the probability addition rule. The rule states that the probability of the union of two mutually exclusive events is the sum of their individual probabilities.

In this case, the events “hot day” and “cold day” are mutually exclusive, meaning they cannot occur at the same time. Therefore, we can simply add their probabilities: 0.3 + 0.2 = 0.5.

Therefore, the probability that tomorrow will be either hot or cold, but not rainy, is 0.5.

Example 5: Probability of Card Denominations and Suits

Consider a standard deck of 52 playing cards. What is the probability of drawing a card that is either a king or a spade?

To solve this problem, we need to determine the number of favorable outcomes (drawing a king or a spade) and the total number of possible outcomes (drawing any card from the deck).

There are 4 kings and 13 spades in a deck, so the number of favorable outcomes is 4 + 13 = 17.

The total number of possible outcomes is 52 (since there are 52 cards in a deck).

Therefore, the probability of drawing a card that is either a king or a spade is 17/52 ≈ 0.327.

Example 6: Probability of Pen Colors

lagrida latex editor 33

Suppose you have a bag containing 5 red pens, 3 blue pens, and 2 green pens. What is the probability of randomly selecting a red or blue pen from the bag?

To solve this problem, we need to determine the number of favorable outcomes (selecting a red or blue pen) and the total number of possible outcomes (selecting any pen from the bag).

There are 5 red pens and 3 blue pens in the bag, so the number of favorable outcomes is 5 + 3 = 8.

The total number of possible outcomes is 5 + 3 + 2 = 10 (since there are 5 red pens, 3 blue pens, and 2 green pens in the bag).

Therefore, the probability of randomly selecting a red or blue pen from the bag is 8/10 = 0.8.

Example 7: Probability of Committee Formation

Suppose there are 10 people, and you need to form a committee of 3 people. What is the probability that you select 2 men and 1 woman for the committee?

To solve this problem, we need to determine the number of favorable outcomes (selecting 2 men and 1 woman) and the total number of possible outcomes (selecting any 3 people from the group of 10).

First, we calculate the number of ways to select 2 men from a group of 5 men: C(5, 2) = 10.

Next, we calculate the number of ways to select 1 woman from a group of 5 women: C(5, 1) = 5.

To find the total number of favorable outcomes, we multiply the number of ways to select 2 men by the number of ways to select 1 woman: 10 * 5 = 50.

The total number of possible outcomes is the number of ways to select any 3 people from a group of 10: C(10, 3) = 120.

Therefore, the probability of selecting 2 men and 1 woman for the committee is 50/120 ≈ 0.417.

Example 8: Probability of Suit Occurrences in a Card Hand

Consider a standard deck of 52 playing cards. What is the probability of drawing a hand of 5 cards that contains at least one card of each suit (hearts, diamonds, clubs, and spades)?

To solve this problem, we need to determine the number of favorable outcomes (drawing a hand with at least one card of each suit) and the total number of possible outcomes (drawing any hand of 5 cards from the deck).

First, we calculate the number of ways to select one card from each suit: 13 * 13 * 13 * 13 = 285,316.

Next, we calculate the total number of possible outcomes, which is the number of ways to draw any 5 cards from a deck of 52: C(52, 5) = 2,598,960.

Therefore, the probability of drawing a hand of 5 cards that contains at least one card of each suit is 285,316/2,598,960 ≈ 0.11.

Example 9: Probability of choosing the same letter from two words

When it comes to probability, we often encounter interesting problems that challenge our understanding of the subject. Let’s consider an example that involves choosing the same letter from two words.

Suppose we have two words, “apple” and “banana.” We want to determine the probability of randomly selecting the same letter from both words. To solve this problem, we need to break it down into smaller steps.

First, let’s list all the letters in each word:

Word 1: “apple”
Word 2: “banana”

Now, we can calculate the probability of choosing the same letter by considering each letter individually. Let’s go through the process step by step:

  1. Selecting a letter from the first word:
  2. The word “apple” has five letters, namely ‘a’, ‘p’, ‘p’, ‘l’, and ‘e’.
  3. The probability of selecting any particular letter is 1 out of 5, as there are five letters in total.

  4. Selecting a letter from the second word:

  5. The word “banana” has six letters, namely ‘b’, ‘a’, ‘n’, ‘a’, ‘n’, and ‘a’.
  6. Similarly, the probability of selecting any particular letter is 1 out of 6.

  7. Calculating the probability of choosing the same letter:

  8. Since each letter has an equal chance of being selected from both words, we multiply the probabilities together.
  9. The probability of selecting the same letter is (1/5) * (1/6) = 1/30.

Therefore, the probability of choosing the same letter from the words “apple” and “banana” is 1/30.

What are the important properties of conditional expectation and how do they relate to problems on probability and its axioms?

The concept of conditional expectation is a fundamental concept in probability theory, and it has important properties that can help us solve problems related to probability and its axioms. To understand these properties and their relationship to probability problems, it is essential to delve into the Properties of conditional expectation explained. These properties provide insights into how conditional expectations behave and can be used to calculate expectations and probabilities in various scenarios. By understanding these properties, we can bridge the gap between the concept of probability and its axioms and the idea of conditional expectation, enabling us to tackle complex probability problems with confidence.

Frequently Asked Questions

1. What is the importance of probability in math?

Probability is important in math because it allows us to quantify uncertainty and make predictions based on available information. It provides a framework for analyzing and understanding random events and their likelihood of occurrence.

2. How would you define probability and its axioms?

Probability is a measure of the likelihood of an event occurring. It is defined using three axioms:

  1. The probability of any event is a non-negative number.
  2. The probability of the entire sample space is 1.
  3. The probability of the union of mutually exclusive events is equal to the sum of their individual probabilities.

3. What are the three axioms of probability?

The three axioms of probability are:

  1. Non-negativity: The probability of any event is a non-negative number.
  2. Normalization: The probability of the entire sample space is 1.
  3. Additivity: The probability of the union of mutually exclusive events is equal to the sum of their individual probabilities.

4. What are the axioms of expected utility theory?

The axioms of expected utility theory are a set of assumptions that describe how individuals make decisions under uncertainty. They include the axioms of completeness, transitivity, continuity, and independence.

5. What are the axioms of probability theory?

The axioms of probability theory are the fundamental principles that govern the behavior of probabilities. They include the axioms of non-negativity, normalization, and additivity.

6. Can you provide some solved problems on axioms of probability?

Certainly! Here is an example:

Problem: A fair six-sided die is rolled. What is the probability of rolling an even number?

Solution: Since the die is fair, it has six equally likely outcomes: {1, 2, 3, 4, 5, 6}. Out of these, three are even numbers: {2, 4, 6}. Therefore, the probability of rolling an even number is 3/6 = 1/2.

7. Where can I find probability problems and answers?

You can find probability problems and answers in various resources such as textbooks, online math websites, and educational platforms. Additionally, there are specific websites that provide probability problems and solutions, such as Math-Aids Answers.

8. Are there any probability examples available?

Yes, there are many probability examples available. Some common examples include flipping a coin, rolling dice, drawing cards from a deck, and selecting balls from an urn. These examples help illustrate how probability concepts can be applied in different scenarios.

9. What are some probability formulas and rules?

There are several probability formulas and rules that are commonly used, including:

  • Addition Rule: P(A or B) = P(A) + P(B) – P(A and B)
  • Multiplication Rule: P(A and B) = P(A) * P(B|A)
  • Complement Rule: P(A’) = 1 – P(A)
  • Conditional Probability: P(A|B) = P(A and B) / P(B)
  • Bayes’ Theorem: P(A|B) = P(B|A) * P(A) / P(B)

10. Can you suggest some probability exercises for practice?

Certainly! Here are a few probability exercises you can try:

  1. A bag contains 5 red balls and 3 blue balls. What is the probability of drawing a red ball?
  2. Two dice are rolled. What is the probability of getting a sum of 7?
  3. A deck of cards is shuffled and one card is drawn. What is the probability of drawing a heart?
  4. A jar contains 10 red marbles and 5 green marbles. If two marbles are drawn without replacement, what is the probability of getting two red marbles?
  5. A spinner is divided into 8 equal sections numbered 1 to 8. What is the probability of landing on an even number?

These exercises will help you practice applying probability concepts and calculations.

Conditional Expectation: 7 Facts You Should Know

image 5 1

For the random variable dependent on one another requires the calculation of conditional probabilities which we already discussed, now we will discuss some more parameters for such random variables or experiments like conditional expectation and conditional variance for different types of random variables.

Conditional Expectation

   The definition of conditional probability mass function of discrete random variable X given Y is

image

here pY(y)>0 , so the conditional expectation for the discrete random variable X given Y when pY (y)>0 is

image 1

in the above expectation probability is the conditional probability.

  In similar way if X and Y are continuous then the conditional probability density function of the random variable X given Y is

image 2

where f(x,y) is joint probability density function and for all yfY(y)>0 , so the conditional expectation for the random variable X given y will be

MT2

for all yfY(y)>0.

   As we know that all the properties of probability are applicable to conditional probability same is the case for the conditional expectation, all the properties of mathematical expectation are satisfied by conditional expectation, for example conditional expectation of function of random variable will be

image 3

and the sum of random variables in conditional expectation will be

image 4

Conditional Expectation for the sum of binomial random variables

    To find conditional expectation of the sum of binomial random variables X and Y with parameters n and p which are independent, we know that X+Y will be also binomial random variable with the parameters 2n and p, so for random variable X given X+Y=m the conditional expectation will be obtained by calculating the probability

image 5

since we know that

image 6

thus the conditional expectation of X given X+Y=m is

image 7

Example:

Find the conditional expectation

image 8

if the joint probability density function of continuous random variables X and Y is given as

image 9

solution:

To calculate the conditional expectation we require conditional probability density function, so

image 10

since for the continuous random variable the conditional expectation is

image 11

hence for the given density function the conditional expectation would be

image 12

Expectation by conditioning||Expectation by conditional expectation

                We can calculate the mathematical expectation with the help of conditional expectation of X given Y as

image 13

for the discrete random variables this will be

image 14

which can be obtained as

image 15

and for the continuous random we can similarly show

image 16

Example:

                A person is trapped in his building underground as the entrance is blocked due to some heavy load fortunately there are three pipelines from which he can come out the first pipe take him safely out after 3 hours, the second after 5 hours and the third pipeline  after 7 hours, If any of these pipeline chosen equally likely by him, then what would be the expected time he will come outside safely.

Solution:

Let X be the random variable that denote the time in hours until the person came out safely and Y denote the pipe he chooses initially, so

image 17

since

image 18

If the person chooses the second pipe , he spends 5 hous in that but  he come outside with expected time

image 19

so the expectation  will be

image 20

Expectation of sum of random number of random variables using conditional expectation

                Let N be the random number of random variable and sum of random variables is     then the expectation  

image 21

since

image 22

as

MT11

thus

MT12

Correlation of bivariate distribution

If the probability density function of the bivariate random variable X and Y is

image 23

where

image 24

then the correlation between random variable X and Y for the bivariate distribution with density function is

since correlation is defined as

image 25

since the expectation using conditional expectation is

image 26

for the normal distribution the conditional  distribution X given Y is having mean

image 27

now the expectation of  XY given Y is

image 28

this gives

image 29

hence

image 30

Variance of geometric distribution

    In the geometric distribution let us perform successively independent trials which results in success with probability p , If N represents the time of first success in these succession then the variance of N as by definition will be

image 31

Let the random variable Y=1 if the first trial results in success and Y=0 if first trial results in failure, now to find the mathematical expectation here we apply the conditional expectation as

image 32

since

image 33

if success is in first trial then N=1 and N2=1 if failure occur in first trial , then to get the first success  the total number of trials will have the same distribution as 1 i.e the first trial that results in failure with  plus the necessary number of additional trials,  that is

image 34

Thus the expectation will be

image 35

since the expectation of geometric distribution is so

image 36

hence

image 37

and

E

image 38

so the variance of geometric distribution will be

image 39

Expectation of Minimum of sequence of uniform random variables

   The sequence of uniform random variables U1, U2 … .. over the interval (0, 1) and N is defined as

image 40

then for the expectation of N, for any x ∈ [0, 1] the value of N

image 41

we will set the expectation of N as

image 42

to find the expectation we use the definition of conditional expectation on continuous random variable

lagrida latex editor 6

now conditioning for the first term of the sequence  we have

image 43

here we get

image 44

the remaining number of uniform random variable is same at the point where the first uniform value is y,in starting and then were going to add uniform random variables until their sum surpassed x − y.

so using this value of expectation the value of integral will be

image 45

if we differentiate this equation

image 46

and

image 47

now integrating this gives

image 48

hence

image 49

the value of k=1 if x=0 , so

m

image 50

and m(1) =e, the expected number of uniform  random variables over the interval (0, 1) that need to be added until their sum surpasses 1, is equal to e

Probability using conditional Expectation || probabilities using conditioning

   We can find the probability also by using conditional expectation like expectation we found with conditional expectation, to get this consider an event and a random variable X as

image 51

from the definition of this random variable and expectation clearly

image 52

now by conditional expectation in any sense we have

image 53

Example:

compute the probability mass function of random variable X , if U is the uniform random variable on the interval (0,1), and consider the conditional distribution of X given U=p as binomial  with parameters n and p.

Solution:

For the value of U the probability by conditioning is

image 54

we have the result

lagrida latex editor 15

so we will get

image 55

Example:

what is the probability of X < Y, If X and Y are the continuous random variables with probability density functions fX and fY respectively.

Solution:

By using conditional expectation and conditional probability

image 56

as

image 57

Example:

Calculate the distribution of sum of continuous independent random variables X and Y.

Solution:

To find the distribution of X+Y we have to find the probability of the sum by using the conditioning as follows

image 58

Conclusion:

The conditional Expectation for the discrete and continuous random variable with different examples considering some of the types of these random variables discussed using the independent random variable and the joint distribution in different conditions, Also the expectation and probability how to find using conditional expectation is explained with examples, if you require further reading go through below books or for more Article on Probability, please follow our Mathematics pages.

https://en.wikipedia.org/wiki/probability_distribution

A first course in probability by Sheldon Ross

Schaum’s Outlines of Probability and Statistics

An introduction to probability and statistics by ROHATGI and SALEH

Moment Generating Functions: 13 Important Facts

Moment generating function    

Moment generating function is very important function which generates the moments of random variable which involve mean, standard deviation and variance etc., so with the help of moment generating function only, we can find basic moments as well as higher moments, In this article we will see moment generating functions for the different discrete and continuous random variables. since the Moment generating function(MGF) is defined with the help of mathematical expectation denoted by M(t) as

gif

and using the definition of expectation for the discrete and continuous random variable this function will be

gif.latex?M%28t%29%3D%5Cleft%5C%7B%5Cbegin%7Barray%7D%7Bll%7D%20%5Csum %7Bx%7D%20e%5E%7Bt%20x%7D%20p%28x%29%20%26%20%5Ctext%20%7B%20if%20%7D%20X%20%5Ctext%20%7B%20is%20discrete%20with%20mass%20function%20%7D%20p%28x%29%20%5C%5C%20%5Cint %7B

which by substituting the value of t as zero generates respective moments. These moments we have to collect by differentiating this moment generating function for example for first moment or mean we can obtain by differentiating once as

gif

This gives the hint that differentiation is interchangeable under the expectation and we can write it as

gif

and

gif

if t=0 the above moments will be

gif

and

gif

In general we can say that

gif

hence

gif

Moment generating function of Binomial distribution||Binomial distribution moment generating function||MGF of Binomial distribution||Mean and Variance of Binomial distribution using moment generating function

The Moment generating function for the random variable X which is Binomially distribution will follow the probability function of binomial distribution with the parameters n and p as

gif

which is the result by binomial theorem, now differentiating and putting the value of t=0

gif

which is the mean or first moment of binomial distribution similarly the second moment will be

gif

so the variance of the binomial distribution will be

gif

which is the standard mean and variance of Binomial distribution, similarly the higher moments also we can find using this moment generating function.

Moment generating function of Poisson distribution||Poisson distribution moment generating function||MGF of Poisson distribution||Mean and Variance of Poisson distribution using moment generating function

 If we have the random variable X which is Poisson distributed with parameter Lambda then the moment generating function for this distribution will be

gif

now differentiating this will give

gif

this gives

gif

which gives the mean and variance for the Poisson distribution same which is true

Moment generating function of Exponential distribution||Exponential distribution moment generating function||MGF of Exponential distribution||Mean and Variance of Exponential distribution using moment generating function

                The Moment generating function for the exponential random variable X by following the definition is

gif

here the value of t is less than the parameter lambda, now differentiating this will give

gif

which provides the moments

gif

clearly

gif

Which are the mean and variance of exponential distribution.

Moment generating function of Normal distribution||Normal distribution moment generating function||MGF of Normal distribution||Mean and Variance of Normal distribution using moment generating function

  The Moment generating function for the continuous distributions also same as the discrete one so the moment generating function for the normal distribution with standard probability density function will be

this integration we can solve by adjustment as

%202%7D%20%5Cend%7Barray%7D

since the value of integration is 1. Thus the moment generating function for the standard normal variate will be

%202%7D

from this we can find for any general normal random variable the moment generating function by using the relation

gif

thus

%202%7D%20%5C%5C%20%26amp%3B%3De%5E%7B%5Cleft%5C%7B%5Cfrac%7B%5Csigma%5E%7B2%7D%20t%5E%7B2%7D%7D%7B2%7D+%5Cmu%20t%5Cright%5C%7D%7D%20%5Cend%7Baligned%7D

so differetiation gives us

gif

thus

gif

so the variance will be

gif

Moment generating function of Sum of random variables

The Moment generating function of sum of random variables gives important property that it equals the product of moment generating function of respective independent random variables that is for independent random variables X and Y then the moment generating function for the sum of random variable X+Y is

Moment generating function
MGF OF SUM

here moment generating functions of each X and Y are independent by the property of mathematical expectation. In the succession we will find the sum of moment generating functions of different distributions.

Sum of Binomial random variables

If the random variables X and Y are distributed by Binomial distribution with the parameters (n,p) and (m,p) respectively then moment generating function of their sum X+Y will be

gif

where the parameters for the sum is (n+m,p).

Sum of Poisson random variables

The distribution for the sum of independent random variables X and Y with respective means which are distributed by Poisson distribution we can find as

gif

Where

gif

is the mean of Poisson random variable X+Y.

Sum of Normal random variables

     Consider the independent normal random variables X and Y with the parameters

gif

then for the sum of random variables X+Y with parameters

gif

so the moment generating function will be

gif

which is moment generating function with additive mean and variance.

Sum of random number of random variables

To find the moment generating function of the sum of random number of random variables let us assume the random variable

gif

where the random variables X1,X2, … are sequence of random variables of any type, which are independent and identically distributed then the moment generating function will be

gif
gif

Which gives the moment generating function of Y on differentiation as

gif

hence

gif

in the similar way the differentiation two times will give

gif

which give

gif

thus the variance will be

gif

Example of Chi-square random variable

Calculate the moment generating function of the Chi-squared random variable with n-degree of freedom.

Solution: consider the Chi-squared random variable with the n-degree of freedom for

gif

the sequence of standard normal variables then the moment generating function will be

gif

so it gives

%202%7D%20%5Cend%7Baligned%7D

the normal density with mean 0 and variance σ2 integrates to 1

%202%7D

which is the required moment generating function of n degree of freedom.

Example of Uniform random variable

Find the moment generating function of random variable X which is binomially distributed with parameters n and p given the conditional random variable Y=p on the interval (0,1)

Solution: To find the moment generating function of random variable X given Y

gif

using the binomial distribution, sin Y is the Uniform random variable on the interval (0,1)

gif.latex?%5Cbegin%7Barray%7D%7Bl%7D%20E%5Cleft%5Be%5E%7Bt%20X%7D%5Cright%5D%3D%5Cint %7B0%7D%5E%7B1%7D%5Cleft%28p%20e%5E%7Bt%7D+1 p%5Cright%29%5E%7Bn%7D%20d%20p%20%5C%5C%3D%5Cfrac%7B1%7D%7Be%5E%7Bt%7D 1%7D%20%5Cint %7B1%7D%5E%7Be%5E%7Bt%7D%7D%20y%5E%7Bn%7D%20d%20y%5C%5C%20%3D%5Cfrac%7B1%7D%7Bn+1%7D%20%5Cfrac%7Be%5E%7Bt%28n+1%29%7D 1%7D%7Be%5E%7Bt%7D

Joint moment generating function

The Joint moment generating function for the n number of random variables X1,X2,…,Xn

gif

where t1,t2,……tn are the real numbers, from the joint moment generating function we can find the individual moment generating function as

gif

Theorem: The random variables X1,X2,…,Xn are independent if and only if the joint mement generating function

gif

Proof: Let us assume that the given random variables X1,X2,…,Xn are independent then

gif

Now assume that the joint moment generating function satisfies the equation

gif
  • to prove the random variables X1,X2,…,Xn are independent we have the result that the joint moment generating function uniquely gives the joint distribution(this is another important result which requires proof) so we must have joint distribution which shows the random variables are independent, hence the necessary and sufficient condition proved.

Example of Joint Moment generating function

1.Calculate the joint moment generating function of the random variable X+Y and X-Y

Solution : Since the sum of random variables X+Y and subtraction of random variables X-Y are independent as for the independent random variables X and Y so the joint moment generating function for these will be

%202%7D%20%5C%5C%20%26amp%3B%3De%5E%7B2%20%5Cmu%20t+%5Csigma%5E%7B2%7D%20t%5E%7B2%7D%7D%20e%5E%7B%5Csigma%5E%7B2%7D%20s%5E%7B2%7D%7D%20%5Cend%7Baligned%7D

as this moment generating function determine the joint distribution so from this we can have X+Y and X-Y are independent random variables.

2. Consider for the experiment the number of events counted and uncounted distributed by poisson distribution with probability p and the mean λ, show that the number of counted and uncounted events are independent with respective means λp and λ(1-p).

Solution: We will consider X as the number of events and Xc the number of counted events so the number of uncounted events is X-Xc,the joint moment genrating function will generate moment

gif

and by the moment generating function of binomial distribution

gif

and taking expectation off these will give

gif

Conclusion:

By using the standard definition of moment generating function the moments for the different distributions like binomial, poisson, normal etc were discussed and the sum of these random variables either the discrete or continuous the moment generating function for those and joint moment generating function were obtained with suitable examples , if you require further reading go through below books.

For more articles on Mathematics, please see our Mathematics page.

A first course in probability by Sheldon Ross

Schaum’s Outlines of Probability and Statistics

An introduction to probability and statistics by ROHATGI and SALEH

Covariance, Variance Of Sums: 7 Important Facts

COVARIANCE, VARIANCE OF SUMS, AND CORRELATIONS OF RANDOM VARIABLES

  The statistical parameters of the random variables of different nature using the definition of expectation of random variable is easy to obtain and understand, in the following we will find some parameters with the help of mathematical expectation of random variable.

Moments of the number of events that occur

    So far we know that expectation of different powers of random variable is the moments of random variables and how to find the expectation of random variable from the events if number of event occurred already, now we are interested in the expectation if pair of number of events already occurred, now if X represents the number of event occurred then for the events A1, A2, ….,An define the indicator variable Ii as

gif

the expectation of X in discrete sense will be

gif

because the random variable X is

gif

now to find expectation if number of pair of event occurred already we have to use combination as

gif

this gives expectation as

gif
gif

from this we get the expectation of x square and the value of variance also by

gif

By using this discussion we focus different kinds of random variable to find such moments.

Moments of binomial random variables

   If p is the probability of success from n independent trials then lets denote Ai for the trial i as success so

gif
gif
gif
gif

and hence the variance of binomial random variable will be

gif

because

gif

if we generalize for k events

gif
gif

this expectation we can obtain successively for the value of k greater than 3  let us find for 3

gif
gif

gif
gif

using this iteration we can get

gif

Moments of hypergeometric random variables

  The moments of this random variable we will understand with the help of an example suppose n pens are randomly selected from a box containing N pens of which m are blue, Let Ai denote the events that i-th pen is blue, Now X is the number of blue pen selected is equal to the number of events A1,A2,…..,An that occur because the ith pen selected is equally likely to any of the N pens of which m are blue

gif

and so

A %7Bi%7D%29%20%3D%5Cfrac%7Bm%7D%7BN%7D%20%5Cfrac%7Bm 1%7D%7BN 1%7D
gif
gif

this gives

gif

so the variance of hypergeometric random variable will be

gif
gifgif

in similar way for the higher moments

gif
gif

hence

gif

Moments of the negative hypergeometric random variables

  consider the example of a package containing n+m vaccines of which n are special and m are ordinary, these vaccines removed one at a time, with each new removal equally likely to be any of the vaccine that remain in the package. Now let random variable Y denote the number of vaccines that need to be withdrawn until a total of r special vaccines have been removed, which is negative hypergeometric distribution, this is somehow similar with negative binomial to binomial as to hypergeometric distribution. to find the probability mass function if the kth draw gives the special vaccine after k-1 draw gives r-1 special and k-r ordinary vaccine

gif

now the random variable Y

Y=r+X

for the events Ai

gif
gif

as

gif

hence to find the variance of Y we must know the variance of X so

gif
gif
gif
gif

hence

gif

COVARIANCE             

The relationship between two random variable can be represented by the statistical parameter covariance, before the definition of covariance of two random variable X and Y recall that the expectation of two functions g and h of random variables X and Y respectively gives

gif
gif
gif
gif
gif

using this relation of expectation we can define covariance as

   “ The covariance between random variable X and random variable Y denoted by cov(X,Y)  is defined as

gif

using definition of expectation and expanding we get

gif
gifgif

it is clear that if the random variables X and Y are independent then

gif
gif

but the converse is not true for example if

gif

and defining the random variable Y as

gif

so

gif

here clearly X and Y are not independent but covariance is zero.

Properties of covariance

  Covariance between random variables X and Y has some properties as follows

gif
gif
gif
gif

using the definition off the covariance the first three properties are immediate and the fourth property follows by considering

em%3E%7Bj%3D1%7D%5E%7Bm%7D%20Y %7Bj%7D%20%5Cright%20%5D%20%3D%5Csum %7Bj%3D1%7D%5E%7Bm%7D%20v %7Bj%7D

now by definition

covariance

Variance of the sums

The important result from these properties is

gif

as

gif
gif
gif
gif

If Xi ‘s are pairwise independent then

Example: Variance of a binomial random variable

  If X is the random variable

gif

where Xi are the independent Bernoulli random variables such that

gif

 then find the variance of a binomial random variable X with parameters n and p.

Solution:

since

gif
gif

so for single variable we have

gif
gif
gif

so the variance is

gif

Example

  For the independent random variables Xi with the respective means and variance and a new random variable with deviation as

gif

then compute

gif

solution:

By using the above property and definition we have

gif
gif
gif

now for the random variable S

COVARIANCE

take the expectation

gif

Example:

Find the covariance of indicator functions for the events A and B.

Solution:

for the events A and B the indicator functions are

gif
gif

so the expectation of these are

gif
gif
gif
gif

thus the covariance is

gif
B%29%20 %20P%28A%29%5D

Example:

     Show that

gif

where Xi are independent random variables with variance.

Solution:

The covariance using the properties and definition will be

gif
gif
gif
gif

Example:

  Calculate the mean and variance of random variable S which is the sum of n sampled values if set of N people each of whom has an opinion about a certain subject that is measured by a real number v that represents the person’s “strength of feeling” about the subject. Let  represent the strength of feeling of person  which is unknown, to collect information a sample of n from N is taken randomly, these n people are questioned and their feeling is obtained to calculate vi

Solution

let us define the indicator function as

gif

thus we can express S as

gif

and its expectation as

gif

this gives the variance as

gif
gif

since

gif
gif

we have

gif
gif
gif
gif
gif

we know the identity

gif

so

gif
gif
gif
gif

so the mean and variance for the said random variable will be

gif
gif

Conclusion:

The correlation between two random variables is defined as covariance and using the covariance the sum of the variance is obtained for different random variables, the covariance and different moments with the help of definition of expectation is obtained  , if you require further reading go through

https://en.wikipedia.org/wiki/Expectation

A first course in probability by Sheldon Ross

Schaum’s Outlines of Probability and Statistics

An introduction to probability and statistics by ROHATGI and SALEH.

For more post on mathematics, please follow our Mathematics page

Conditional Variance & Predictions: 7 Important Facts

lagrida latex editor 21 300x70 1

In this article the conditional Variance and predictions using conditional expectation for the different kind of random variable with some examples we will discuss.

Conditional Variance

The conditional variance of random variable X given Y is defined in similar way as conditional Expectation of random variable X given Y as

(X|Y)=E[(X-E[X|Y])2|Y]

here variance is the conditional expectation of difference between random variable and square of conditional expectation of X given Y when the value of Y is given.

The relation between the conditional variance and conditional expectation is

(X|Y) = E[X2|Y] – (E[X|Y])2

E[(X|Y)] = E[E[X2|Y]] – E[(E[X|Y])2]

= E[X2] – E[(E[X\Y])2]

since E[E[X|Y]] = E[X], we have

(E[X|Y]) = E[(E[X|Y])2] – (E[X])2

this is somehow similar from the relation of unconditional variance and expectation which was

Var(X) = E[X2] – (E[X])2

and we can find the variance with the help of conditional variance as

Var(X) = E[var(X|Y] + var(E[X|Y])

Example of conditional variance

Find the mean and variance of the number of travelers who enters into the bus if the people arrived at bus depot is Poisson distributed with mean λt and the initial bus arrived at bus depot is uniformly distributed over the interval (0,T) independent of people arrived or not.

Solution:

To find the mean and variance let for any time t , Y is the random variable for the time bus arrive and N(t) is the number of arrivals

E[N(Y)|Y = t] = E[N(t)|Y = t]

by the independence of Y and N(t)

=λt

since N(t) is Poisson with mean \lambda t
Hence

E[N(Y)|Y]=λY

so taking expectations gives

E[N(Y)] = λE[Y] = λT/2

To obtain Var(N(Y)), we use the conditional variance formula

lagrida latex editor 21

thus

(N(Y)|Y) = λY

E[N(Y)|Y] = λY

Hence, from the conditional variance formula,

Var(N(Y)) = E[λY]+(λY)

=λT/2 + λ2T2/12

where we have used the fact that Var(Y)=T2 / 12.

Variance of a sum of a random number of random variables

consider the sequence of independent and identically distributed random variables X1,X2,X3,………. and another random variable N independent of this sequence, we will find variance of sum of this sequence as

CodeCogsEqn 92

using

lagrida latex editor 48

which is obvious with the definition of variance and conditional variance for the individual random variable to the sum of sequence of random variables hence

CodeCogsEqn 93

Prediction

In prediction the value of one random variable can be predicted on the basis of observation of another random variable, for prediction of random variable Y if observed random variable is X we use g(X) as the function which tells the predicted value, obviously we try to choose g(X) closed to Y for this the best g is g(X)=E(Y|X) for this we must have to minimize the value of g by using the inequality

lagrida latex editor 49

This inequality we can get as

lagrida latex editor 22

However, given X, E[Y|X]-g(X), being a function of X, can be treated as a constant. Thus,

lagrida latex editor 23

which gives the required inequality

lagrida latex editor 50

Examples on Prediction

1. It is observed that the height of a person is six feet, what would be the prediction of his sons height after grown up if the height of son which is x inches now is normally distributed with mean x+1 and variance 4.

Solution: let X be the random variable denoting the height of the person and Y be the random variable for the height of son, then the random variable Y is

Y=X+e+1

here e represent the normal random variable independent of random variable X with mean zero and variance four.

so the prediction for the sons height is

lagrida latex editor 24

so the height of the son will be 73 inches after growth.

2. Consider an example of sending signals from location A and location B, if from location A a signal value s is sent which at location B received by normal distribution with mean s and variance 1 while if the signal S sent at A is normally distributed with mean \mu and variance \sigma^2, how we can predict that the signal value R sent from location A will be received is r at location B?

Solution: The signal values S and R denote here the random variables distributed normally, first we find the conditional density function S given R as

lagrida latex editor 25

this K is independent of S, now

lagrida latex editor 26

here also C1 and C2 are independent on S, so the value of conditional density function is

WhatsApp Image 2022 09 10 at 11.02.40 PM

C is also independent on s, Thus the signal sent from location A as R and received at location B as r is normal with mean and variance

lagrida latex editor 27

and the mean square error for this situation is

lagrida latex editor 28

Linear Predictor

Every time we can not find the joint probability density function even the mean, variance and the correlation between two random variables is known, in such a situation linear predictor of one random variable with respect to another random variable is very helpful which can predict the minimum, so the for the linear predictor of random variable Y with respect to random variable X we take a and b to minimize

lagrida latex editor 29

Now differentiate partially with respect to a and b we will get

lagrida latex editor 26 1

solving these two equations for a nd b we will get

lagrida latex editor 31

thus minimizing this expectation gives the linear predictor as

lagrida latex editor 32

where the means are the respective means of random variables X and Y, the error for the linear predictor will be obtained with the expectation of

conditional variance
conditional variance: Error in prediction

This error will be nearer to zero if correlation is perfectly positive or perfectly negative that is coefficient of correlation is either +1 or -1.

Conclusion

The conditional variance for the discrete and continuous random variable with different examples were discussed, one of the important application of conditional expectation in prediction is also explained with suitable examples and with best linear predictor, if you require further reading go through below links.

For more post on Mathematics, please refer to our Mathematics Page

A first course in probability by Sheldon Ross

Schaum’s Outlines of Probability and Statistics

An introduction to probability and statistics by ROHATGI and SALEH