You are on page 1of 17

05 Regression and Correlation (1)

05 Regression and Correlation


We are mainly concern with the use of associations among variables. These associations may be
useful in many ways, and one of the most important and most common is prediction. This method is used
for predicting the value of one quantity by using values of related quantities. Such method may also lead
to methods for controlling the value of one variable by adjusting the values of related variables.
Regression analysis offers us a sensible and sound approach for examining associations among variables
and for obtaining good rules for prediction.
Regression Analysis
we make decisions based on prediction of future events
we develop relationship between what is already known and what is to be estimated
There always exists a tendency towards the average. For example, the height of children born to
tall parents tends towards the shorter height. So we make use of regression that is a process of predicting
one variable (the height of the children) from another (the height of parents). So we develop an estimating
equation. That is a mathematical formula that relates the known variable to the unknown variable.

Scattered diagram can give us two type of information. Visually we can look for pattern that
indicate that the variables are related. Then, if the variables are related, we can see what kind of line, or
estimating equation, describes this equation.
Later on we will study correlation analysis to determine the degree to which the variables are
related. It tells us how well the estimating equation actually describes the relationship.
We find a CAUSAL relationship between variables.
i.e. how does the independent variable causes the dependent variable to change.
Deterministic and Probabilistic Relations or Models
A formula that relates quantities in the real world is called a model. Recall that in physics we
have studies that if a body is moving under uniform motion with an initial velocity u and uniform
acceleration a, the velocity after time t is given by:
v = u + at
This is a model for uniform motion. This model has the property that when a value of t is
substituted in the above equation, the value of v is determined without any error. Such models are called
deterministic models. An important example of the deterministic model is the relationship between
Celsius and Fahrenheit scales in the form of F = 32+9/5C. Other examples of such models are Boyles
law, Newtons law of gravitation, ohms law etc.
Consider an other example to investigate the relationship between the yield of potatoes y and the
level of fertilizer application x. An investigator divides the field into eight plots of equal size with equal
fertility and applied varying amounts of fertilizer to each. The yield of potatoes (in kg) and the fertilizer
application (in kg) was recorded for each plot. This data is given below:
Fertilizer Applied: =x 1 1.5 2 2.5 3 3.5 4 4.5
Yield of Potatoes: =y 25 31 27 28 36 35 32 34

Muhammad Naeem Sandhu, Assistant Professor Department of Mathematics


05 Regression and Correlation (2)

Suppose the investigator believes that the relation between y and x is exactly given by:
y = 22 + 2.5 x
If this is true we must obtain the exact value of yield y for a given value of x. Thus when x = 1,
the yield must be:
Y = 22 + 2.5 (1) = 24.5
But it is 25. There is an error of 24.5 25.0 = - 0.5. Hence no deterministic model can be
constructed to represent this experiment. This type of error is known as probabilistic model. The
deterministic relation in such cases is then modified to include both a deterministic component and a
random error component given as
Yi = a = bXi + i , where is are the unknown random errors.
Regression Model
There are many statistical investigations in which the main objective is to determine whether a
relationship exists between two or more variables. If such a relationship can be expressed by a
mathematical formula, we will then be able to use it for the purpose of making predictions. The reliability
of any prediction will, of course, depend on the strength of the relationship between the variables included
in the formula.
A mathematical equation that allows us to predict values of one dependent variable from known
values of one or more independent variables is called a regression equation. Today the term regression is
applied to all types of prediction problems and does not necessarily imply a regression towards the
population mean.
Linear Regression
We consider here the problem of estimating or predicting the value of a dependent variable Y on
the basis of a known measurement of an independent and frequently controlled variable X. The variable
intended to be estimated or predicted is termed as dependent variable or response variable and the
variable on the basis of which the dependent variable is to be estimated is called the independent variable,
the regressor or the predictor.
e.g. If we want to estimate the heights of children on the basis of their ages, the heights would
be the dependent variable and the ages would be the independent variable. In estimating the yields of a
crop, on the basis of the amount of the fertilizer used, the yield will be the dependent variable and the
amount of fertilizer would be the independent variable.
Scatter Diagram
Let us consider the distribution of statistics grades corresponding to intelligence test scores of 50,
55, 65 and 70. The statistics grades for a sample of 12 freshmen having these intelligence test scores are
presented in the following table

Test score (Xi) 65 50 55 65 55 70 65 70 55 70 50 55

Statistics grade (Yi) 85 74 76 90 85 87 94 98 81 91 76 74

The data table has been plotted in figure to give a scattered


diagram. In the scattered diagram, the points follow closely a straight
line indicate that the two variables are to some extend linearly related.
Once a reasonable linear relationship is obtained, we usually try to
express this mathematically by a straight-line equation Y = a + bX,
called the linear regression line, where the constants a and b represent
the y-intercept and slope respectively. Such a regression line has been
drawn in the following figure. This linear regression line can be used
to predict the value Y corresponding to any given value X.

Muhammad Naeem Sandhu, Assistant Professor Department of Mathematics


05 Regression and Correlation (3)

Using the regression line in figure, we can predict a statistics grade of 88 for a student whose
intelligence test score is 60. However, we would be extremely fortunate if a student with an intelligence
test score of 60 made a statistics grade of exactly 88. In fact, the original data of table show that three
students with this intelligence test score received grades of 85, 90 and 94. we must therefore, interpret the
predicted statistics grade of 88 as an average or expected value for all students taking the course who have
an intelligence test score of 60.
Many possible regression lines could be fitted to the sample data, but we choose that particular
line which best fits that data. The best regression line is obtained by estimating the regression parameters
by the most commonly used method of least squares.
Estimation of a Straight Line using the Method of Least Squares
The basic linear relationship between the dependent variable Yi and the value Xi is
Yi = a + b Xi + i
where a and b are called the unknown population parameters (b is also called the coefficient of
regression), Yi are the observed values and i are the error components.
We write the equation for the estimating line as

Yi = a + b Xi
The line will have a good fit if it minimizes the error between the estimated points on the line and
the actual observed points that were used to draw it.
One way to measure the error of the estimating line is to sum all the individual differences or
errors, between the estimated points and the observed points. In the figure we have two estimated lines
that have been fitted to the same set of three data points. Two very different lines have been drawn to
describe the relationship between the two variables.


We have calculated the individual differences between the corresponding Y and Y, and then we
have found the sum of these differences. The total error in both the cases is zero. This means that both the
lines describe the data equally well. Thus we must conclude that the process of summing individual
differences for calculating the error is not a reliable way to judge the goodness of fit of an estimating line.
Actually the sum of the individual errors is cancelling effect of the positive and negative values.
However if we consider the sum of the absolute values of the errors seems to be the better criteria
to find a good fit but it does not stress the magnitude of the errors. But if we consider the sum of the
squares of the errors, it magnifies the larger errors and it cancels the effect of the positive and negative
values. So finally we looking for an estimating line that minimizes the sum of the squares of the errors.
That is called the method of the least squares.
Method of Least Squares
The method of least squares determines the values of the unknown parameters that minimize the
sum of squares of the errors where errors are defined as the difference between observed values and the
corresponding predicted or estimated values.

Muhammad Naeem Sandhu, Assistant Professor Department of Mathematics


05 Regression and Correlation (4)

It is denoted by
n n n
S(a,b) = ei2 = (Yi Yi)2 = (Yi a b Xi)2
i =1 i =1 i =1
minimizing S(a,b), we put first partial derivatives w. r. t. a and b
equal to zero. Therefore
S(ab) n
= 2 (Yi a b Xi)(1) = 0
a i =1
S(ab) n
= 2 (Yi a b Xi)(Xi) = 0
b i =1
by simplifying, we have
Yi = na + bXi
Xi Yi = aXi + bXi 2
by solving, we have
(Y)(X2) (X)(XY) nXY (X)(Y)
a= b=
n(X2) (X)2 n(X2) (X)2
If the variable X is taken as dependent variable , then the least square line is given by
X = c + dY
and the normal equations are
X = nc + dY
XY = cY + dY2
By solving simultaneously, w have the values of c and d
(X)(Y2) (Y)(XY) nXY (X)(Y)
c= 2 2 ,d=
n(Y ) (Y) n(Y2) (Y)2
Example (1)
Fit a least square line of regression to the following data taking X as independent variable.

Vehicle no. 0 1 2 3 4
Life in Years X 5 3 3 1
Repair expenses during
last year in hundreds of $ Y 7 7 6 4

Solution
(i) The equation of the least square line is
Y = a + bX
and the normal equations are
y = na + bX
XY = aX + bX2
From given data we have n = 4

X Y XY X2
5 7 35 25
3 7 21 9
3 6 18 9
1 4 4 1
X = 12 Y = 24 XY= 78 X 2 =64

Substituting the values of n, X, Y, XY, X2 in the normal equations, we have


y = na + bX
XY = aX + bX2

Muhammad Naeem Sandhu, Assistant Professor Department of Mathematics


05 Regression and Correlation (5)

24 = 4a + 12b
78 = 12a + 64b
By solving simultaneously, we get
a = 3.75, b = 0.75
We can also find the values of a and b using the formulas
(Y)(X2) (X)(XY) nXY (X)(Y)
a= ,b=
n(X2) (X)2 n(X2) (X)2
Hence the required straight line is

Y = 3.75 + 0.75 X
Example (2)
Following is given the data of 10 randomly selected areas in each area number of oil stoves and
the annual consumption of oil in barrels is given. Fit a regression equation of annual oil consumption on
number of stoves.
No. of stoves: =x 1 1.5 2 2.5 3 3.5 4 4.5
Annual Consumption of oil: =y 25 31 27 28 36 35 32 34
Solution
Necessary calculations are given below
x y xy x2
27 142 3834 729
32 170 5440 1024
38 200 7600 1444
42 194 8148 1764
48 224 10752 2304
54 256 13824 2916
60 261 15660 3600
67 270 18090 4489
73 304 22192 5329
79 349 27571 6241
Total 520 2370 133111 29840
(Y)(X2) (X)(XY) nXY (X)(Y)
a= ,b=
n(X2) (X)2 n(X2) (X)2
we have a = 52.6792 and b = 3.5254
^
so the linear regression equation is: y = 52.68 + 3.53 x
Positive value of b shows that for an increase of one stove, the oil consumption is increased by 3.53
barrels on the average.
Example (3)
Consider the data on experiment on potatoes. We want to fit a line of regression to the data using
the method of least squares.
The necessary calculations are given in the table below:
x y Xy x2 y2 y^ e
1 25 25.0 1.00 625 26.83 -1.83
1.5 31 46.5 2.25 961 28.02 2.98
2 27 54.0 4.00 729 29.21 -2.21
2.5 28 70.0 6.25 784 30.40 -2.40
3 36 108.0 9.00 1296 31.59 4.41
3.5 35 122.5 12.25 1225 32.79 2.21
4 32 128.0 16.00 1024 33.98 -1.98
4.5 34 153.0 20.25 1156 35.17 -1.17
Total 22 248 707.0 71.00 7800

Muhammad Naeem Sandhu, Assistant Professor Department of Mathematics


05 Regression and Correlation (6)

(Y)(X2) (X)(XY) nXY (X)(Y)


a= 2 2 ,b=
n(X ) (X) n(X2) (X)2
we have a = 24.452 and b = 2.38095
The fitted line which is the simple regression line: ^y = 24.452+ 2.38095 x
Conclusions
(i) The above regression line provides the estimated values (or the mean values) of y for
given values of x.
(ii) b = 2.381 indicates an increase of 2.381 kg in yield per unit increase fertilizer. Such
interpretation is valid only when x lies between 1 and 4.5.
(iii) an extension of the model beyond these values may lead to unreasonable results.
(iv) If we assume the model to extend to x = 0, a = 24.452 is the estimated expected yield
when no fertilizer is used.
Accuracy of the Regression Curve
a) Using scattered diagram, we can guess, how accurate the regression curve is.
b) If sum of the residuals is zero then the curve is supposed to be good one.
c) If the points in the plot of the residuals are close to the x-axis and scattered in a random
way, the model appears to provide a good fit.
d) If the points in the plot of the residuals are distributed in a systematic manner we should
try some other model.

Example (4)
(1) Given below the data relating to the thermal energy generated in Pakistan 1981-94. The energy
generation is in billion kwh.
Year 1981 1982 1983 1984 1985 1986 1987
Energy Generated 4.2 5.2 5.1 5.2 6.5 7.3 8.4
Year 1988 1989 1990 1991 1992 1993 1994
Energy Generated 10.8 11.9 14.5 16.1 19.4 19.7 23.0
Fit a straight line to the data. Find the residuals. Plot the residuals and comment on your result.
(2) Following is the annual installation of computers in labs in UET. Fit a linear regression equation
of the computers on years and give the annual rate of installation of them.
Year: 2001-2003 2003-2005 2005-2007 2007-2009 2009-2011
No of Computers installed: 139 144 150 154 158
Note
For each situation where the independent variable is a time factor, the values assigned to
2001-2003, may be taken as 1,2,3,

Muhammad Naeem Sandhu, Assistant Professor Department of Mathematics


05 Regression and Correlation (7)

Example (5)
Fit a parabola y = ax2 + bx + c in least square sense to the data
X= 10 12 15 23 20
Y= 14 17 23 25 21
Solution
The normal equations to the curve are
Y = a X2 + b X + 5c
XY = a X3 + b X2 + c X
X2Y = a X4 + b X3 + c X2
Table:
X Y X2 X3 X4 XY X2Y
10 14 100 1000 10000 140 1400
12 17 144 1728 20736 204 2448
15 23 225 3375 50625 345 5175
23 25 529 12167 279841 575 13225
20 21 400 8000 160000 420 8400
X = Y = x2 = x3 = x4 XY = X2Y=
80 100 1398 26270 =521202 1684 30648

Substituting the obtained values from the table in normal equations, we have
100 = 1398 a + 80 b + 5c
1684 = 26270 a + 1398 b + 80 c
30648 = 521202 a + 26270 b + 1398 c
on solving, a = 0.07, b = 3.03, c = 8.89
Hence the required equation is
Y = 0.07 X2 + 3.03 X 8.89
Example (6)
Use method of least squares to determine the constants a and b such that Y = a ebX fits the
following data.
X= 0.0 0.5 1.0 1.5 2.0 2.5
Y= 0.10 0.45 2.15 9.15 40.35 180.75
Solution
The curve to be fitted is Y = a ebX or y = A+Bx, where y = log 10Y, A = log10 a and B = b log10 e
the normal equations are
y = 6A + BX
Xy = A x + B x2
Table construction:
X Y Y = log10 y X2 Xy
0 0.10 -1 0 0
0.5 0.45 - 0.3468 0.25 -0.1734
1.0 2.15 0.3324 1.0 0.3324
1.5 9.15 0.9614 2.25 1.4421
2.0 40.35 1.6058 4.0 3.2116
2.5 180.75 2.2571 6.25 5.6428
2
X=7.5 y = 3.8099 X = 13.75 Xy = 10.4555

Muhammad Naeem Sandhu, Assistant Professor Department of Mathematics


05 Regression and Correlation (8)

Substituting the values of the summations in the normal equations, we get


3.8099 = 6A + 7.5 B and 10.4555 = 7.5 A + 13.75 B
On solving, A = 0.9916, B = 1.3013
a = anti log10 A = 0.1019, b = B/log10e = 2.9963
and the curve is Y = 0.1019 e(2.9963)X
Example (7)
Obtain a relation of the form Y = a bX for the following data by the method of least squares.
X= 2 3 4 5 6
Y= 8.3 15.3 33.1 65.2 127.4
Solution
The curve to be fitted is Y = a bX or y = A + BX, where A = log10 a, B= log10b and y = log10Y.
the normal equations are: y = 5A + BX and Xy = a X + B X2

X Y y = log10Y X2 Xy
2 8.3 0.9191 4 1.8382
3 15.4 1.1872 9 3.5616
4 33.1 1.5198 16 6.0792
5 65.2 1.8142 25 9.0710
6 127.4 2.1052 36 12.6312
X = 20 y = 7.5455 X2 = 90 Xy = 33.1812

Substituting the values in normal equations, we get


7.5455 = 5A + 20B and 33.1812 = 20A + 90 B
On solving, A = 0.31 and B = 0.3
a = anti-log 10 A = 2.04 and b = anti-log10 B = 1.995
Hence, the required curve is Y = 2.04 (1.995)X
Exercise 1.

Fit a least squares line for 20 pairs of observations having X = 2, Y = 8, X2 = 180 and XY=404
Example (8)
For 5 pairs of observations, it is given that A.M of X is 2 and A.M of Y is 15. It is also known
that X = 30, X3 = 100, X4 =354, XY = 242, X2Y = 850. Fit a second degree parabola taking X as
2

an independent variable.
Example (9)
Given the following sets of values:
X 6.5 5.3 8.6 1.2 4.2 2.9 1.1 3.9
Y 3.2 2.7 4.5 1.0 2.0 1.7 0.6 1.9

(b) Compute the least squares regression equation for Y values on X values.
(c) Compute the least squares regression equation for X values on Y values.
Example (10)
For each of the following data, determine the estimated regression equation Y = a + bX:

(a) Y = 20, X = 10, XY = 1000, X2 = 2000, n = 10.
(b) X = 528, Y = 11720, XY = 193640, X2 = 11440, n = 32

Muhammad Naeem Sandhu, Assistant Professor Department of Mathematics


05 Regression and Correlation (9)

Example (11)
For the following set of data: (AIOU)

(a) plot the scatter diagram.


(b) Develop the estimating equation that best describes the data
(c) Predict Y for X = 10, 15, 20
X 13 16 14 11 17 9 13 17 18 12
Y 6.2 8.6 7.2 4.5 9.0 3.5 6.5 9.3 9.5 5.7
A Motivation
For the data in Example (3) relating to potatoes yields, remove the first pair (x = 1, y = 25) and fit
a line of regression to the remaining seven pairs. Is the line same as already determined for the eight
pairs? Do the same by removing the second pair (x=1, y = 31) instead of the first. Are the three lines
different?
You will observe that a change in data leads to a different line. We say that a least squares line
has zero breakdown point. There are methods in which a change of as many as 50% data points does not
cause any change in equation of the fitted line.
This factor will be discussed in the following section.
Standard Deviation of Regression or Standard Error of Estimate
Now we study how to measure the reliability of the estimating equation we have developed. In a
scattered diagram, we realize that a line as an estimator is more accurate when the data points lie close to
the line than when the points are farther away from the line. To measure the reliability of the estimating
equation, statisticians have developed the standard error of estimate which is similar to the standard
deviation.
The standard error of estimates measures the variability of or scatter of the observed values
around the regression line. The standard deviation of regression or the standard error of estimate of Y on
X denoted by sY.X and defined by
^
(YY)2 ^
sY.X = where Y = a + bX, the estimated regression line.
n2
Alternative formula,
Y i2 aYi bX iYi
sY.X = where n is the number of pairs.
n2
Dividing by n-2
Because the values a and b were obtained from a sample of data points, we lose 2 degree of
freedom when we use points to estimate the regression line.
In statistics, the number of degree of freedom is the number of values in the final calculation of a
statistic that are free to vary.
Example (12)

X Y XY X2
Y = 3.75 + 0.75 X
5 7 35 25
3 7 21 9
3 6 18 9
1 4 4 1
X = 12 Y = 24 XY= 78 X2 =64

Muhammad Naeem Sandhu, Assistant Professor Department of Mathematics


05 Regression and Correlation (10)

Example (13)
Cost accountant often estimates overhead based on the level of production. At the Standard
Knitting Co. they have collected information on overhead expenses and units produced at different plants,
and want to estimate a regression equation to predict future overhead. (AIOU)
Overhead 191 170 272 155 280 173 234 116 153 178
Units 40 42 53 35 56 39 48 30 37 40

(a) Develop the regression equation for the cost accountants.


(b) Predict overhead when 50 units are produced.
(c) Calculate the standard error of estimate.
Coefficient of Determination
The variability among the values of the dependent variable Y, called the total variation, is given

by (Y Y)2. This composed of two parts:
^
(i) one is explained by (associated with) the regression line. i.e (Y Y)2
^
(ii) other which is not explained by (not associated with) the regression line. i.e. (Y Y)2.
Symbolically,
^ ^
(Y Y)2 = (Y Y)2 + (Y Y)2
Total Variation = Unexplained Variation + Explained Variation
See the diagram;

The coefficient of determination which measures the proportion of variability in the values of the
dependent variable (Y) associated with its linear relation with the independent variable (X) is defined by:
^ 2 ^ 2
Explained Variation
(Y . Y) (Y Y .)
r2 = Total Variation = =1

(Y Y)2 (Y Y)2
Alternate formula for Coefficient of Determination:

2 a Y + bXY nY2
r =
2
Y2 nY

Muhammad Naeem Sandhu, Assistant Professor Department of Mathematics


05 Regression and Correlation (11)

Example (14) years R&D Annual


expenses (X) Profit (Y)
Calculate 1st 5 31
Coefficient of 2nd 11 40
Determination 3rd 4 30
using both the 4th 5 34
formulas 5th 3 25
6th 2 20
X = 30 Y=180

Muhammad Naeem Sandhu, Assistant Professor Department of Mathematics


05 Regression and Correlation (12)

Covariance of Random Variables


The covariance between two random variables is a measure of the nature of the association
between the two. If large values of X often result in large values of Y or small values of X result in small
values of Y, positive X X will often result in positive Y Y and negative XX will often result in
negative Y Y. Thus, the product (X X )(Y Y ) will tend to be positive. On the other hand, if large
X values often result in small Y values, the product (X X)(Y Y) will tend to be negative. The sign of
the covariance indicates whether the relationship between two dependent random variables is positive or
negative. When X and Y are statistically independent, it can be shown that the covariance is zero. The
converse, however, is not generally true. Two variables may have zero covariance and still not be
statistically independent.
Note that the covariance only describes the linear relationship between two random variables.
Therefore, if a covariance between X and Y is zero, X and Y may have a nonlinear relationship, which
means that they are not necessarily independent.
Let X and Y be random variables with joint probability distribution f(x, y). The correlation of X
and Y is
_ _ _ _
XY = E[(X X )( Y Y )] = (x x )( y y ) f(x,y) dx dy , if X and Y are discrete
n=1 n=1
_ _ _ _
and XY = E[(X X )( Y Y )] = (X X )( Y Y ) f(x,y) , if X and Y are continuous
n=1 n=1
Book Probability and Statistics by Walpole 8th Ed Page 119 Chapter 4
Correlation
Two variables are said to be correlated if they tend to simultaneously vary in some direction; if
both the variables tend to increase (or decrease) together, the correlation is said to be direct or positive.
e.g. the length of an iron bar will increase as temperature increases. If one variable tend to increase as the
other variable decreases, the correlation is said to be negative or inverse. e.g. the volume of gas will
decrease as the pressure increases.
Correlation in fact is the strength of the relationship that is the interdependence between the two
variables that is there is no distinction between dependent and independent variable. In regression, by
contrast, we are interested in determining the dependence of one variable upon the other variable.
The numerical measure of strength in the linear relationship between any two variables is called
the correlation coefficient, usually denoted by r, is defined by
_ _
(XX) (Y Y)
r=
_ _
(XX)2 (YY) 2
called Pearson Product Moment Correlation Coefficient.
alternatively
XY(X)( Y)/n
r=
[X (X)2/n][ Y2(Y)2/n]
2

It assumes values that range from +1 for perfect positive linear relationship, to 1, for perfect
negative linear relationship and r = 0 indicates no linear relationship between X and Y.
It is important to note that r = 0 does not mean that there is no relationship at all. e.g. if all the
observed values lie exactly on a circle, there is a perfect non-linear relationship between the variables.

Muhammad Naeem Sandhu, Assistant Professor Department of Mathematics


05 Regression and Correlation (13)

Analysis of Correlation and Regression


(1) The correlation answers the STRENGTH of linear association between paired variables, say X
and Y. On the other hand, the regression tells us the FORM of linear association that best predicts
Y from the values of X.
(2) Correlation is calculated whenever:
o both X and Y is measured in each subject and quantify how much they are linearly
associated.
o in particular the Pearson's product moment correlation coefficient is used when the
assumption of both X and Y are sampled from normally-distributed populations are
satisfied
o or the Spearman's moment order correlation coefficient is used if the assumption of
normality is not satisfied.
o correlation is not used when the variables are manipulated, for example, in experiments.
(3) Linear regression is used whenever:
at least one of the independent variables (Xi's) is to predict the dependent variable Y.
Note: Some of the Xi's are dummy variables, i.e. Xi = 0 or 1, which are used to code
some nominal variables.
if one manipulates the X variable, e.g. in an experiment.
(4) Linear regression are not symmetric in terms of X and Y. That is interchanging X and Y will give
a different regression model (i.e. X in terms of Y) against the original Y in terms of X. On the
other hand, if you interchange variables X and Y in the calculation of correlation coefficient you
will get the same value of this correlation coefficient.
(5) The "best" linear regression model is obtained by selecting the variables (X's) with at least strong
correlation to Y, i.e. >= 0.80 or <= -0.80
(6) The same underlying distribution is assumed for all variables in linear regression. Thus, linear
regression will underestimate the correlation of the independent and dependent when they (X's
and Y) come from different underlying distributions.
Exercise 2.
Determine covariance and correlation for the following joint probability distribution
X: -1 -0.5 0.5 1
Y: -2 -1 1 2
f(x, y) 1/8 1/4 1/2 1/8
Exercise 3.
Determine the value for c and the covariance and correlation for the joint probability mass
function fXY (x, y) = c(x + y) for x = 1, 2, 3 and y = 1, 2, 3.
Exercise 4.
The joint probability distribution is
X: -1 0 0 1
Y: 0 -1 1 0
f(x, y) 1/4 1/4 1/4
Show that the correlation between X and Y is zero, but X and Y are not independent
Exercise 5.
Determine the covariance and correlation for the joint probability density function fXY = e x y
over the range 0 < x and 0 < y.

Rank Correlation

Muhammad Naeem Sandhu, Assistant Professor Department of Mathematics


05 Regression and Correlation (14)

Sometimes, the actual measurements of individuals or objects are either not available or accurate
assessment is not possible. They are then arranged in order according to some characteristic of interest..
Such an ordered arrangement is called a ranking and the order given to an individual or object is called its
rank. The correlation between two such sets of ranking is called Rank Correlation.
Let we have n pairs of two data sets ranked with respect to some characteristic. Say, (x1, y1),
(x2,y2), (x3, y3), , (xn, yn). Since both xi and yi are the first n natural numbers, therefore we have
n(n+1)
xi = 1 + 2 + + n =
2
n(n+1)(2n+1)
x 2 = y2 = 12 + 22 + + n2 =
6
2 2 2
- = (yi - y)2
(xi - x)2 - = yi2 - (yi) = n(n+1)(2n+1) - n(n+1) = n(n -1)
n 6 4 12
Let di = xi - yi
Then
di2 = (xi - yi)2 = x i2 + yi2 - 2x i yi
n(n+1)(2n+1) n(n+1)(2n+1)
= + - 2xi yi
6 6
n(n+1)(2n+1) 1
xi yi = - di2
6 2
The product moment coefficient of correlation is:
XY(X)( Y)/n
r=
[X (X)2/n][ Y2(Y)2/n]
2

by substitution we have
6di2
r=1-
n(n2 - 1)
This is also ranging from 1 to + 1
Note
If two objects or observations are tied (having same value), lets say for fourth and fifth, then they
are both given the mean rank of 4 and 5. i.e. 4.5.
This situation is given in the following example.
Example (15)
The following table shows the number of hours studied (X) by a random sample of ten students
and their grades in examination (Y):
X: 8 5 11 13 10 5 18 15 2 8
Y: 56 44 79 72 70 54 94 85 33 65
Calculate Spearmans rank correlation coefficient.
Solution
We rank the X values by giving rank 1 to the highest value 18, rank 2 to 15, rank 3 to 13, rank 4
to 11, rank 5 to 10, rank 6.5 (mean of rank 6 and 7) to both 8, rank 8.5 (mean of rank 8 and 9) to both 5
and rank 10 to 2. Similarly we rank the values of Y by giving 1 to the highest value 94, rank 2 to 85, rank
3 to 79, , and rank 10 to 33 which is the smallest.
Table given below:

Muhammad Naeem Sandhu, Assistant Professor Department of Mathematics


05 Regression and Correlation (15)

X Y Rank of X Rank of Y di d2
8 56 6.5 7 - 0.5 0.25
5 44 8.5 9 - 0.5 0.25
11 79 4 3 1.0 1
13 72 3 4 - 1.0 1
10 70 5 5 0.0 0
5 54 8.5 8 0.5 0.25
18 94 1 1 0.0 0
15 85 2 2 0.0 0
2 33 10 10 0.0 0
8 65 6.5 6 0.5 0.25
d2 = 3
The value of n is 10.
Hence
6di2
r=1-
n(n2 - 1)
6(3)
=1-
10(102 - 1)
= 0.98
Compare this value with the correlation coefficient for the original values.
Example (16)
Ten competitors in a beauty contest are ranked by three judges in the following order
1st Judge 1 6 5 10 3 2 4 9 7 8
2nd Judge 3 5 8 4 7 10 2 1 6 9
3rd Judge 6 4 9 8 1 2 3 10 5 7
Use the rank correlation coefficient to discuss which pair of judges have the nearest approach to
common tastes in beauty.
Example (17)
The body mass index (BMI) For an individual is found as follows. Using your weight in pounds
and your height in inches, multiply your weight by 705, divide the result by your height, and divide again
by your height. The desirable body mass index varies between 19 and 25. Table gives the body mass
index and the age for 20 individuals. Find the coefficient rank correlation for data shown in the table.

BMI AGE MBI AGE


22.5 27 19.0 44
24.6 32 17.5 18
28.7 45 32.5 29
30.1 49 22.4 29
18.5 19 28.8 40
20.0 22 21.3 39
24.5 31 25.0 21
25.0 27 19.0 20
27.5 25 29.7 52
30.0 44 16.7 19

Muhammad Naeem Sandhu, Assistant Professor Department of Mathematics


05 Regression and Correlation (16)

Example (18)
Table gives the percent of calories from fat and the micrograms of lead per decilitter of blood for
a sample of preschoolers. Find the coefficient of rank correlation for the data shown in the table.

PERCENT FAT LEAD


CALORIES
40 13
35 12
33 11
29 8
35 13
36 15
30 11
36 14
33 12
28 9
39 15
26 7

Ans. Coefficient of Rank Correlation = 0.919

Muhammad Naeem Sandhu, Assistant Professor Department of Mathematics


05 Regression and Correlation (17)

Assignment (2)
Given the following data:

X1: 41 31 26 43 21 33 41 31 46 31 36 32 38 27 35 40
X2: 62 52 50 56 51 52 63 50 47 40 56 54 60 57 57 58
X3: 41 33 33 38 35 36 43 33 37 33 33 31 30 37 35 31

(a) Calculate all the six simple regression coefficients.


(b) Calculate all the simple correlation coefficients.
(c) Calculate the correlation coefficients between X1 and a combined effect of X2 and
X3.
(d) Calculate the correlation coefficients between X1 and X2 when the effect of X3.is
held constant.

X1 X2 X3 X12 X22 X32 X1X2 X1X3 X2X3




(a) Calculate six simple regression coefficients.

nX1 X2 (X 1)(X2) nX1 X2 (X1)(X2) nX1 X3 (X1)(X3)


b12 = b 21 = b13 =
nX22 (X2)2 nX12 (X1)2 nX32 (X3)2
nX1 X3 (X 1)(X3) nX2 X3 (X2)(X3) nX2 X3 (X2)(X3)
b31 = b 23 = b32 =
nX12 (X1)2 nX32 (X3)2 nX22 (X2)2

(b) Calculate all the simple correlation coefficients.


nX1 X2 X1 X2
r12 =
nX1 (X1)2 nX22 (X2)2
2

nX1 X3 X1 X3
r13 =
nX1 (X1)2 nX32 (X3)2
2

nX2 X3 X2 X3
r23 =
nX2 (X2)2 nX32 (X3)2
2

(c) correlation coefficients between X1 and a combined effect of X2 and X3.

r122 + r132 - 2 r12 r13 r23


R1.23 =
1 - r232

(d) The correlation coefficients between X1 and X2 when the effect of X3.is held constant

r12 - r13 r23


r12.3 =
1 - r132 1 - r232

Muhammad Naeem Sandhu, Assistant Professor Department of Mathematics

You might also like