Transcript

CHAPTER-14

INTRODUCTION TO CORRELATION & REGRESSION ANALYSIS

ByDR. PRASANT SARANGI

Key concepts:

Introduction to Correlation Analysis Rank Correlation Linear Regression Analysis Multiple Regression Analysis

CORRELATION ANALYSIS

• Positive Correlation• Negative Correlation• Linear Correlation and • Non-linear Correlation

Positive Correlation

• Two variables are said to be positively correlated when the movement of the one variable leads the movement of the other variable in the same direction.

• There exists direct relationship between the two variables.

Negative Correlation

• Correlation between two variables is said to be negative when the movement of one variable leads to the movement in the other variable in the opposite direction.

• Here there exists inverse relationship between the two variables.

Linear Correlation

• The correlation between two variables is said to be linear where the points when drawn is a graph represents a straight line.

• Non-linear Correlation A relationship between two variables is said to be non-linear if a unit change

in one variable causes the other variable to change in fluctuations. If X is changed then corresponding values of Y will not change in the same

proportion.

Methods of Measuring Correlation

• The Graphical MethodThe correlation can be graphically shown by using scatter diagrams. Scatter diagram reveals two important useful information. Firstly, through this diagram, one can observe the patterns between two

variables which indicate whether there exists some association between the variables or not.

Secondly, if an association between the variables is found, then it can be easily identified regarding the nature of relationship between the two (whether two variables are linearly related or non-linearly related).

• Karl Pearson’s Coefficient of Correlation Karl Pearson’s coefficient of correlation (developed in 1986) measures

linear relationship between two variables under study. Since, the relationship is expressed is linear, hence, two variables change in a fixed proportion. This measure provides the answer of the degree of relationship in real number, independent of the units in which the variables have been expressed, and also indicates the direction of the correlation.

• Direct method

22ii

iiXY

yx

yxr

Assumed Mean Method

2222 )()(

))((

YYXX

YXYXXY

ddnddn

ddddnr

• Grouped Data

2222 )()(

))((

YYXX

YXYXXY

fdfdnfdfdn

fdfddfdnr

Assumptions of Coefficient of Correlation

1. The Value of the Coefficient of Correlation Lies between -1 (minus one) to +1 (plus one).

2. The Value of the Coefficient of Correlation is Independent of the Change of Origin and Change of Scale of Measurement

2222 )()(

)()(

iiii

iiiiXY

kknhhn

khkhnr

Rank Correlation Coefficient

There are three different situations of applying the Spearman’s rank correlation coefficient.

• When ranks of both the variables are given• When ranks of both the variables are not given and • When ranks between two or more observations in a series are

equal

• When Ranks of Both the Variables are Given

)(

61

61

2

2

3

2

nnn

dor

nn

dRXY

When Ranks of both the Variables are not Given

•In such cases, each observation in the series is to be ranked first.

•The selection of highest value depends on the researcher.

• In other words, either the highest value or the lowest value will be ranked 1 (one) depends upon the decision of the researcher.

• When Ranks between Two or More Observations in a Series are Equal• The ranks to be assigned to each observation are an average of the ranks

which these observations would have got, if they differed from each other.

)1(

......)(12

1)(

12

1)(

12

16

12

3332

321

31

2

nn

mmmmmmd

RXY

Simple Linear Regression Model

What do we use regression models for:

1. Estimate a relationship among economic variables, such as y = f(x).

2. Test hypotheses

3. Forecast or predict the value of one variable, y, based on the value of another variable, x.

Dependent and Independent Variables

Dependent variable - the variable we are trying to explain

Independent (or explanatory) variables - variables that we think cause movements in the dependent variable

Simple Regression Model

Y = dependent variableX = independent variable

Model is: Y = α + Xα is the intercept or constant is the slope coefficient

Linearity

Models that are linear in the variables and in the coefficients:Y = α + X

Models that are nonlinear in the variables but linear in the coefficients: Y = α + X2

Models that are nonlinear in the variables and in the coefficients:Y = α + X

Some models that are nonlinear can be made linear in the coefficients:

Y = e α X

take logs:ln Y = α + ln X

r

E(Y|X)

E(Y|X)

AverageExpenditure

X (income)

E(Y|X)= α + X

=E(Y|X)

X

An Example showing income and average expenditure

Error Term

Y is a random variable composed of two parts:

I. Systematic component: E(Y) = α + X This is the mean of Y.

II. Random component: u = Y - E(Y | X) = Y - α - X

u is called the stochastic or random error.

Together E(Y) and u form the model: Y = α + X + u

Sources of error term

• Dependent variable measured with error• Model left out relevant variables• Wrong functional form• Inherent randomness of behaviour

True Relationship

u4

Y

X

E(Y)= α + X

Y4

Y1

Y3

Y2

X1 X2 X3 X4

u1

u2

u3

The Estimated Model

We use the data on Y and X to come up with guesses for α and . These estimated parameters or coefficients are

α and cap

^ ^

Our estimated, or “fitted”, model gives the predicted value for Y for any given X:

Yi = α + Xi

The residual is the difference between the actual or observed value of Y and the predicted value:

ui = Yi - Yi = Yi - α - Xi

^ ^ ^

^ ^ ^ ^

top related