REGRESSION, REGRESSION ANALYSIS,BASIS OF REGRESSION, TYPES OF REGRESSION, SIMPLE REGRESSION, COMPUTATION OF SIMPLE REGRESSION, MULTIPLE REGRESSION, COMPUTATION OF MULTIPLE REGRESSION, MODELS OF REGRESSION, ASSUMPTIONS OF SIMPLE AND MULTIPLE REGRESSION, STANDARDIZED REGRESSION COMPUTATION, GRAPHICAL ...
REGRESSION, REGRESSION ANALYSIS,BASIS OF REGRESSION, TYPES OF REGRESSION, SIMPLE REGRESSION, COMPUTATION OF SIMPLE REGRESSION, MULTIPLE REGRESSION, COMPUTATION OF MULTIPLE REGRESSION, MODELS OF REGRESSION, ASSUMPTIONS OF SIMPLE AND MULTIPLE REGRESSION, STANDARDIZED REGRESSION COMPUTATION, GRAPHICAL REPRESENTATION, CRITERION AND PREDICTOR VARIABLES, STATISTICS, PARTIAL CORRELATION
Size: 603.24 KB
Language: en
Added: Oct 28, 2019
Slides: 28 pages
Slide Content
Application No. : 38464073e50d11e99aa6f572b452740e Name: SAYANTAN SARKAR Affiliation: Banaras Hindu University Email: [email protected]
Acknowledgement Course Name: Academic Writing Every author owes a great deal to others, and I am no exception. First I would like to acknowledge Dr. Ajay Semalty of Garhwal University, for teaching us academic writing in such a nice manner in the SWAYAM programme. Next I would like to acknowledge the assistance of books, Google scholar and my friend Bapi Biswas . Finally, I am deeply appreciative and dedicate to my supportive family.
REGRESSION This Photo by Unknown author is licensed under CC BY-SA .
REGRESSION The term regression was first used by Francis Galton with reference to inheritance of stature. Galton found that children of tall parents tend to be less tall and children of short parents less short , than their parents. In other words ,the height of the offspring tends to "move back" towards the mean height of general population. This tendency towards the " mean height". This principle is known as REGRESSION.
What is Regression Regression is a prediction statistics. It predicts the most likely value of a variable (criterion variable) on the basis of given values of another variable or variables( predictor). The variable whose values are predicted is known as dependent or criterion variable and whose values from the basis of prediction is called independent variable or predictor. Regression can be worked out only if the dependent variable and the independent variable possess significant correlation with each other. It translate the relation between two or more variables in to an expression showing one of them as function of others(summary line).
What is Regression Regression ,like correlation holds good only in particular population to which the sample belongs, and only for that limited range of scores of the variable from which it has been derived; it cannot be extend beyond this limit. The regression line is called the "best-fitting" line because "on average" it passes through the center of criterion variable.
Types of Regression SIMPLE REGRESSION -The criterion or dependent variable is a function of single independent variable or predictor. The scores of the criterion are predicted from the given scores of the single predictor. e.g., the regression of examination marks of a candidate in mathematics on his/her numerical aptitude test score. MULTIPLE REGRESSION -The criterion is function of two or more predictors. Its scores are predicted from the scores of more than one predictor. e.g., The regression of math’s mark of an examinee on his her numerical aptitude and abstract reasoning test score.
Models of Regression MODEL 1 – It is the regression of dependent variable or criterion(y) on an independent or predictor(x) which is fixed treatment variable. It is a value of y for the specified value of x when the predictor is varied by the investigators at precise and predetermined manner and rate. The value of y suffers from error due to random variation but the value of x are free from random errors due to because they vary under plan and control of investigators and are not random. It can explore causation ( cause and effect relation). e.g., the regression of blood sugar level on pre determined doses of injecting insulin.
Models of Regression MODEL 2 – It is the regression of criterion Y on the predictor X which is "classified variable" which is beyond the control of the investigator. Predict mostly likely value of Y on the basis of already existing value of X in the individual. Here X is measured but not applied by the investigator. Its value of both X and Y suffers from random errors. It cannot explore the cause and effect relationship between the variable. e.g.; examination makes in a language on the verbal abilty test of an score of an examinee.
Models of regression MODEL 3 – this is always a multiple regression predicting value of dependent variable, from given value of two or more predictors, where the predictor variable can be either fixed treatment variable or classification variable. e.g.: regression of surface area on height and weight of the individual.
Assumptions of simple linear regression The variables involved in regression are continuous measurement variable. Both variables have unimodal and fairly symmetrical distribution in population. The scores of criterion variable is in linear function of scores of the predictor variable. The Y scores of criterion, measured in a large number of the predictors are distributed normally independent of each other. The predictor variable is either "fixed" experimental treatment or a classification variable.
Computation simple linear equation Simple linear regression expresses a single dependent variable Y as the linear function of a single Independent variable X. First, we establish the relationship in a sample .Then we use the regression line to determine the Y' (likely Y score with respect to X score) for each X.. We can measure the X scores of individual who are not in our sample, and the corresponding Y' is the best prediction of Y scores. e.g.: when SAT score are used to predict a student's future college grades, SAT scores are predictor variable and college grades is criterion variable.
Computation simple linear equation Regression Equation : The equation describes two characteristics of regression line. The slope Y intercept SLOPE:- The slope is the number that indicate how slanted the regression line is and the direction in which it slants. Y intercept:- It is the vaue of Y at the point where regression line crosses the Y- axis. So, the intercept is the value of Y score where x equals 0.
Computation simple linear equation LINEAR REGRESSION EQUATION Y' = bX + a Here, b = regression coefficient(slope of the line), Regression coefficient : regression coefficient is the average rate of increase or decrease in score of criterion for unit rise in score of pred ictor. a = the y intercept(where the line crosses Y axis,where X=0) X = value of predictor variable. Y = value of criterion variable. STEP 1 - Compute Pearson Product Moment r . STEP 2 – Computation of slope Where,
Computation simple linear equation If the b the slope is positive which will be fitting the score we get from computing the correlation coefficient. STEP 3 – a = Y"-(b)(X"). Here, Y" = sum of values of Y score divided by the number of cases N. X" =sum of value of X divided by number of cases N. STEP 4 - After getting the slope of (b) and the Y intercept (a) we can now compute the Y' score for each value of X score by the formula. By pacing the score of X (the score we have already) in the equation of linear regression we can get the Y' score for each value of X . To graph regression line , we need to plot the data points for the previous X-Y' pairs and draw the line.
From the graph the regression line clearly shows that when we have the DAT (X-predictor) score 70 then the predicted maths (Y-criterion) score is 43 (Y'=predicted score).Again, when we have the DAT score 110 then the predicted maths score is 61.8.The Y intercept is 10.11 and the slope is +0.47.
Multiple Regression Multiple regression, a method of multivariate statistics, predict likely value of a variable (criterion or independent variable) from values of two or more variable ( predictor or independent variable). Can be only computed if the variable possess a significant co-relationship with each other. It shows criterion variable as a function of the predictor variables. The predictor variables can be "fixed" treatment variable or classification variable.
Multiple regression Multiple regression even limited to two predictors can be relatively complex. Usually, different predictor variables are related to each other which means they are often measuring and predicting the same thing. Because variables may overlap with each other adding another predictor variable does not always add to accuracy of prediction.
Multiple regression and partial correlation A direct procedure for controlling correlation third variable is by partial correlation. Its allows researcher to measure the relationship between two variables while eliminating or holding constant the effect of the third variable. Here three variables are X,Y and Z it is possible to compute three individual Pearson correlation , r XY meaning correlation between X and Y. , r XZ meaning correlation between X and Z. , r YZ meaning correlation between Y and Z.
Multiple regression and partial correlation Where , r XY.Z = the pearson correlation product moment r when X and Y when the Z is hold constant. , r XY = the correlation between X and Y , r square = coefficient of determination
Assumptions of multiple linear regression All the variables involved should be continuous measurement variables. Their scores have a unimodal and fairly symmetrical distribution in population. The paired scores of each pair of variables in an individual are independent of all other such paired scores in the sample. There is a linear association between the scores of each pair of variable.
Computation multiple linear regression Computation of linear regression with three variables- Regression predicts the most likely value of Y' of the criterion Y from the given value of two predictors X1 and X2.The general regression equation for the straight line showing Y as the linear function of X1 and X2 is as Y' = a + b1X1 + b2X2 Here , Y' = predicted value of Y , a = Y intercept , b1 = slope of regression line of Y on X1 when X2 is constant Partial regression coefficient. , b2 = slope of regression line of Y on X2 when X1 is constant Partial regression coefficient. , X1 = predictor variable 1 , X2 = predictor variable 2
Computation of multiple linear regression Here, b1 and b2 are computed using the SDs (SY,S1 and S2) of respective variable and their beta coefficient ( β1 and β2). So here , And So , a = Y" - b 1 X" 1 – b 2 X" 2 . Here, a= the y intercept. , Y" = the mean score of the variable Y , X"1 = the mean score of the variable X1 , X"2 = the mean score of the variable X2. Putting the values in the linear regression equation we can get the predicted Y' value with respect to both the independent or predictors variables. Click to add text
Computation of multiple regression Here β is known as standard regression coefficient which are in normalized units. So one standard score unit increase in IV is associated with how many standard score units in DV. So β are those proportion of total variance of criterion variable that are associated with variance of respective predictor variables.
Computation of multiple regression Computation of multiple regression more than three variable- so if we consider that number of variable involved is "n" then the regression equation will be Y' = a + b 1 X 1 +b 2 X 2 +….....+ b n X n Where, a = the Y intercept, , b1x1 = the regression slope of X1 on Y while all other predictor variables are hold constant , bnxn = the regression slope of Xn on Y while all other predictor variables are hold constant
Further Reading Garrett, H.E. Statistics In Psychology And Education (6 th ed.). Gravetter , F.J., & Wallnau , L.B. Statistics For The Behavioral Science (10 th ed.).
REFERENCES Das, D., & Das, A.(2017). Statistics In Biology And Psychology (6 th ed.). pg. 193-204. Kolkata: Academic Publisher. Tabachnick , B.J., & Fidell , L.S. (2012). Using Multivariate Statistics (6 th ed.). Pg. 117-118. London: Pearson. Image source: Das, D., & Das, A.(2017). Statistics In Biology And Psychology (6th ed.). pg. 201. Kolkata: Academic Publisher.