Regression analysis

667 views 20 slides Aug 08, 2017
Slide 1
Slide 1 of 20
Slide 1
1
Slide 2
2
Slide 3
3
Slide 4
4
Slide 5
5
Slide 6
6
Slide 7
7
Slide 8
8
Slide 9
9
Slide 10
10
Slide 11
11
Slide 12
12
Slide 13
13
Slide 14
14
Slide 15
15
Slide 16
16
Slide 17
17
Slide 18
18
Slide 19
19
Slide 20
20

About This Presentation

Good presentation for university or college students ....


Slide Content

WELCOME TO OUR PRESENTATION
PRESENTED BY:
Md . Sohag
Em@il : [email protected]
Daffodil international University

INTRODUCTION. . .
•FATHER OF REGRESSION ANALYSIS
CARL F. GAUSS (1777-1855).
•CONTRIBUTIONS TO PHYSICS, MATHEMATICS &
ASTRONOMY.
•THE TERM “REGRESSION” WAS FIRST USED IN
1877 BY FRANCIS GALTON.

REGRESSION ANALYSIS. . .
• IT IS THE STUDY OF THE
RELATIONSHIP BETWEEN
VARIABLES.
• IT IS ONE OF THE MOST
COMMONLY USED TOOLS FOR
BUSINESS ANALYSIS.
• IT IS EASY TO USE AND APPLIES
TO MANY SITUATIONS.

REGRESSION TYPES. . .
•SIMPLE REGRESSION : SINGLE
EXPLANATORY VARIABLE
•MULTIPLE REGRESSION : INCLUDES ANY
NUMBER OF EXPLANATORY VARIABLES.

• DEPENDANT VARIABLE : THE SINGLE VARIABLE BEING EXPLAINED/
• PREDICTED BY THE REGRESSION MODEL
• INDEPENDENT VARIABLE : THE EXPLANATORY VARIABLE(S) USED TO
• PREDICT THE DEPENDANT VARIABLE.
• COEFFICIENTS (Β): VALUES, COMPUTED BY THE REGRESSION TOOL,
• REFLECTING EXPLANATORY TO DEPENDENT VARIABLE RELATIONSHIPS.
• RESIDUALS (Ε): THE PORTION OF THE DEPENDENT VARIABLE THAT ISN ’T
• EXPLAINED BY THE MODEL; THE MODEL UNDER AND OVER PREDICTIONS.

REGRESSION ANALYSIS. . .
•LINEAR REGRESSION: STRAIGHT-LINE
RELATIONSHIP
– Form: y=mx+b
•NON-LINEAR: IMPLIES CURVED RELATIONSHIPS

logarithmic relationships

REGRESSION ANALYSIS. . .
•CROSS SECTIONAL: DATA GATHERED FROM THE
SAME TIME PERIOD
•TIME SERIES: INVOLVES DATA OBSERVED OVER
EQUALLY SPACED POINTS IN TIME.

SIMPLE LINEAR REGRESSION MODEL. . .

TYPES OF REGRESSION MODELS. . .

xbbyˆ
10i
+=
The sample regression line provides an estimate of
the population regression line
ESTIMATED REGRESSION MODEL. . .
Estimate of
the regression
intercept
Estimate of the
regression slope
Estimated
(or predicted)
y value
Independent
variable
The individual random error terms e
i
have a mean of zero

REGRESSION ANALYSIS: MODEL BUILDINGREGRESSION ANALYSIS: MODEL BUILDING
General Linear Model
Determining When to Add or Delete Variables
Analysis of a Larger Problem
Multiple Regression Approach
to Analysis of Variance

GENERAL LINEAR MODELGENERAL LINEAR MODEL
Models in which the parameters (β
0
, β
1
, . . . , β
p
) all have exponents of one are called
linear models.
First-Order Model with One Predictor Variable
y x=+ +bbe
0 11
y x=+ +bbe
0 11

VARIABLE SELECTION PROCEDURESVARIABLE SELECTION PROCEDURES
Stepwise Regression
Forward Selection
Backward Elimination
Iterative; one independent
variable at a time is
added or
deleted
Based on
the F statistic

VARIABLE SELECTION PROCEDURESVARIABLE SELECTION PROCEDURES
F Test
To test whether the addition of x
2
to a
model involving x
1 (or the deletion of x
2
from a model involving x
1and x
2) is
statistically significant


F
0
=MS
R
/MS
Res
(MS
R
=SS
R
/K)
The p-value corresponding to the F statistic
is the criterion used to determine if a variable
should be added or deleted
(SSE(reduced)-SSE(full))/number of extra terms
MSE(full)
F=

FORWARD SELECTIONFORWARD SELECTION
This procedure is similar to stepwise-regression,
but does not permit a variable to be deleted.
This forward-selection procedure starts with no
independent variables.
It adds variables one at a time as long as a
significant reduction in the error sum of squares
(SSE) can be achieved.

BACKWARD ELIMINATIONBACKWARD ELIMINATION
This procedure begins with a model that includes all the
independent variables the modeler wants considered.
It then attempts to delete one variable at a time by
determining whether the least significant variable currently
in the model can be removed because its p-value is less than
the user-specified or default value.
Once a variable has been removed from the model it cannot
re enter at a subsequent step.

Example1-:From the following data obtain the two regression equations
using the method of Least Squares.
X 3 2 7 4 8
Y 6 1 8 5 9
Solution-:
X Y XY X
2
Y
2
3 6 18 9 36
2 1 2 4 1
7 8 56 49 64
4 5 20 16 25
8 9 72 64 81
å=24X å=29Y å =168XY 142
2
=åX 207
2
=åY

Example2-: from the previous data obtain the regression equations by
Taking deviations from the actual means of X and Y series.
X 3 2 7 4 8
Y 6 1 8 5 9
X Y x
2
y
2
xy
3 6 -1.8 0.2 3.24 0.04 -0.36
2 1 -2.8 -4.8 7.84 23.04 13.44
7 8 2.2 2.2 4.84 4.84 4.84
4 5 -0.8 -0.8 0.64 0.64 0.64
8 9 3.2 3.2 10.24 10.24 10.24
XXx -= YYy -=
å=24Xå=29Y 8.26
2
=åx 8.28=åxy8.38
2
=åyå=0x 0å=y
Solution-:

Example-: From the data given in previous example calculate regression
equations by assuming 7 as the mean of X series and 6 as the mean of Y series.
X Y
Dev. From
assu.
Mean 7
(d
x
)=X-7
Dev. From
assu. Mean
6 (d
y
)=Y-6
d
x
d
y
3 6 -4 16 0 0 0
2 1 -5 25 -5 25 +25
7 8 0 0 2 4 0
4 5 -3 9 -1 1 +3
8 9 1 1 3 9 +3
Solution-:
2
x
d
2
yd
å=24X å=29Y å -=11
x
d å -=1
y
då =51
2
x
d å=39
2
y
d å =31
yx
dd
Tags