STATISTICAL TEST Statistical tests are intended to decide whether a hypothesis about distribution of one or more populations or samples should be rejected or accepted.
PARAMETRIC TESTS Parametric test is a statistical test that makes assumptions about the parameters of the population distribution(s) from which one’s data is drawn.
APPLICATIONS Used for Quantitative data. Used for continuous variables. Used when data are measured on approximate interval or ratio scales of measurement. Data should follow normal distribution.
PARAMETRIC TESTS t-test
Contd.. 2. ANOVA 3. Pearson’s r correlation 4. Z test
STUDENT’S T-TEST Developed by Prof.W.S.Gossett A t-test compares the difference between two means of different groups to determine whether the difference is statistically significant.
One Sample t-test Assumptions: Population is normally distributed Sample is drawn from the population and it should be random We should know the population mean Conditions : Population standard deviation is not known Size of the sample is small (<30 )
Contd.. In one sample t-test , we know the population mean. We draw a random sample from the population and then compare the sample mean with the population mean and make a statistical decision as to whether or not the sample mean is different from the population.
Let x 1 , x 2 , ……., x n be a random sample of size “n” has drawn from a normal population with mean (µ) and variance . Null hypothesis (H ): Population mean (μ) is equal to a specified value µ 0. Under H , the test statistic is
Two sample t-test Used when the two independent random samples come from the normal populations having unknown or same variance. We test the null hypothesis that the two population means are same i.e., µ 1 = µ 2
Contd… Assumptions : Populations are distributed normally Samples are drawn independently and at random Conditions : 1 . Standard deviations in the populations are same and not known 2. Size of the sample is small
If two independent samples x i ( i = 1,2,….,n 1 ) and y j ( j = 1,2, …..,n 2 ) of sizes n 1 and n 2 have been drawn from two normal populations with means µ 1 and µ 2 respectively . Null hypothesis H : µ 1 = µ 2 Under H , the test statistic is
Paired t-test U sed when measurements are taken from the same subject before and after some manipulation or treatment. Ex: To determine the significance of a difference in blood pressure before and after administration of an experimental pressure substance.
Assumptions: 1 . Populations are distributed normally 2. Samples are drawn independently and at random Conditions : 1. Samples are related with each other 2. Sizes of the samples are small and equal 3. Standard deviations in the populations are equal and not known
Null Hypothesis: H0 : µ d = 0 Under H0, the test statistic Where, d = difference between x1 and x2 d ̅ = Average of d s = Standard deviation n = Sample size
Z-Test Z-test is a statistical test where normal distribution is applied and is basically used for dealing with problems relating to large samples when the frequency is greater than or equal to 30. It is used when population standard deviation is known.
Contd… Assumptions: Population is normally distributed The sample is drawn at random Conditions: Population standard deviation σ is known Size of the sample is large (say n > 30)
Let x 1 , x 2 , ……… x, n be a random sample size of n from a normal population with mean µ and variance σ 2 . Let x ̅ be the sample mean of sample of size “n” Null Hypothesis: Population mean (µ) is equal to a specified value µ ο H : µ = µ ο
Under H ο , the test statistic is If the calculated value of Z < table value of Z at 5% level of significance, H is accepted and hence we conclude that there is no significant difference between the population mean and the one specified in H as µ ο .
Pearson’s ‘r’ Correlation Correlation is a technique for investigating the relationship between two quantitative, continuous variables. Pearson’s Correlation Coefficient (r) is a measure of the strength of the association between the two variables
Types of correlation Type of correlation Correlation coefficient Perfect positive correlation r = +1 Partial positive correlation < r < +1 No correlation r = Partial negative correlation > r > -1 Perfect negative correlation r = -1
ANOVA (Analysis of Variance) Analysis of Variance (ANOVA) is a collection of statistical models used to analyse the differences between group means or variances. Compares multiple groups at one time Developed by R.A.Fischer
One way ANOVA Compares two or more unmatched groups when data are categorized in one factor Ex: Comparing a control group with three different doses of aspirin Comparing the productivity of three or more employees based on working hours in a company
Two way ANOVA Used to determine the effect of two nominal predictor variables on a continuous outcome variable. It analyses the effect of the independent variables on the expected outcome along with their relationship to the outcome itself. Ex: Comparing the employee productivity based on the working hours and working conditions.
Assumptions of ANOVA: The samples are independent and selected randomly. Parent population from which samples are taken is of normal distribution. Various treatment and environmental effects are additive in nature. The experimental errors are distributed normally with mean zero and variance σ 2.
ANOVA compares variance by means of F-ratio F = It again depends on experimental designs Null hypothesis: H ο = All population means are same If the computed F c is greater than F critical value, we are likely to reject the null hypothesis. If the computed Fc is lesser than the F critical value , then the null hypothesis is accepted.
ANOVA Table Sources of Variation Sum of squares (SS) Degrees of freedom (d.f) Mean squares (MS) F - Ratio Between samples or groups (Treatments) Treatment sum of squares ( TrSS) (k-1) Within samples or groups ( Errors ) Error sum of squares (ESS) (n-k) Total Total sum of squares (TSS) (n-1) Sources of Variation Sum of squares (SS) Degrees of freedom (d.f) F - Ratio Between samples or groups (Treatments) Treatment sum of squares ( TrSS) (k-1) Within samples or groups ( Errors ) Error sum of squares (ESS) (n-k) Total Total sum of squares (TSS) (n-1)
S.No Type of group Parametric test 1. Comparison of two paired groups Paired t-test 2. Comparison of two unpaired groups Unpaired two sample t-test 3. Comparison of population and sample drawn from the same population One sample t-test 4. Comparison of three or more matched groups but varied in two factors Two way ANOVA 5. Comparison of three or more matched groups but varied in one factor One way ANOVA 6. Correlation between two variables Pearson Correlation