Top Banner
Chapter 6: Analysing the Data Part III: Common Statistical Tests Nonparametric tests Occasionally, the assumptions of the t-tests are seriously violated. In particular, if the type of data you have is ordinal in nature and not at least interval. On such occasions an alternative approach is to use nonparametric tests. We are not going to place much emphasis on them in this unit as they are only occasionally used. But you should be aware of them and have some familiarity with them. Nonparametric tests are also referred to as distribution-free tests. These tests have the obvious advantage of not requiring the assumption of normality or the assumption of homogeneity of variance. They compare medians rather than means and, as a result, if the data have one or two outliers, their influence is negated. Parametric tests are preferred because, in general, for the same number of observations, they are more likely to lead to the rejection of a false hull hypothesis. That is, they have more power. This greater power stems from the fact that if the data have been collected at an interval or ratio level, information is lost in the conversion to ranked data (i.e., merely ordering the data from the lowest to the highest value). The following table gives the non-parametric analogue for the paired sample t-test and the independent samples t-test. There is no obvious comparison for the one sample t-test. Chi-square is a one-sample test and there are alternatives to chi-square but we will not consider them further. Chi-square is already a non-
31
Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: Non Parametric Test

Chapter 6: Analysing the DataPart III: Common Statistical Tests

Nonparametric testsOccasionally, the assumptions of the t-tests are seriously violated. In particular, if the type of data you have is ordinal in nature and not at least interval. On such occasions an alternative approach is to use nonparametric tests. We are not going to place much emphasis on them in this unit as they are only occasionally used. But you should be aware of them and have some familiarity with them.

Nonparametric tests are also referred to as distribution-free tests. These tests have the obvious advantage of not requiring the assumption of normality or the assumption of homogeneity of variance. They compare medians rather than means and, as a result, if the data have one or two outliers, their influence is negated.

Parametric tests are preferred because, in general, for the same number of observations, they are more likely to lead to the rejection of a false hull hypothesis. That is, they have more power. This greater power stems from the fact that if the data have been collected at an interval or ratio level, information is lost in the conversion to ranked data (i.e., merely ordering the data from the lowest to the highest value).

The following table gives the non-parametric analogue for the paired sample t-test and the independent samples t-test. There is no obvious comparison for the one sample t-test. Chi-square is a one-sample test and there are alternatives to chi-square but we will not consider them further. Chi-square is already a non-parametric test. Pearson's correlation also has non-parametric alternative (Spearman's correlation) but we will not deal with it further either.

There are a wide range of alternatives for the two group t-tests, the ones listed are the most commonly use ones and are the defaults in SPSS. Generally, running nonparametric procedures is very similar to running parametric procedures, because the same design principle is being assessed in each case. So, the process of identifying variables, selecting options, and running the procedure are very similar. The final p-value is what determines

Page 2: Non Parametric Test

significance or not in the same way as the parametric tests. SPSS gives the option of two or three analogues for each type of parametric test, but you need to know only the ones cited in the table. Same practice with these tests is given in Assignment II.

Parametric testNon-parametric analogue

One-sample t-test Nothing quite comparable

Paired sample t-test Wilcoxon T Test

Independent samples t-test Mann-Whitney U Test

Pearson's correlation Spearman's correlation

ReadingsHowell describes several measures which assess the degree of relationship between the two variables in chi-square. This material is worth reading but for this unit we will not be discussing these at all. Howell also describes a correction for continuity that is sometimes used and the use of likelihood ratios. Again we will not be dealing with these issues. However, you should read Howell's discussion of assumptions.

Ray does not appear to discuss chi-square or contingency tables.

Research Methods and StatisticsPESS202Lecture and Commentary NotesThese notes have a long history. Most of them were originally written for UNE by Andrew F. Hayes, Ph.D. in 1997. They were updated in Jan. 1998 by Travis L. Gee, Ph.D. and again by Ian R. Price, Ph.D. in late 1998. Many of them can also be traced back to

Page 3: Non Parametric Test

Ray Cooksey and Adam Patrech from 1993 to 1997. So all are acknowledged as making a contribution.

These commentary notes are designed to form the backbone of the material to be learnt in this unit. You should read and understand these notes first and foremost. The information you need to complete your assignments and study for your exam is essentially to be found in here. You should then try to get a broader, and also more detailed, understanding of the topics and concepts by consulting your textbooks. In those texts you will also find more worked examples. During internal practical sessions external residential schools you will get practice at carrying out the analyses that are discussed in these notes and are required for the assignments.

Associated with these materials is a disk that contains a number of programs to assist your understanding. The WebStat web site is also intended to do this. So, the plan is for you to develop your understanding of research methods and statistics using a many-pronged attack. Some ways will suit some people more than others but if you faithfully try them all, you should find something that causes the penny to drop or the light to come on. If so, you should then start to appreciate the exciting and rewarding nature of research.

o

Nonparametric Tests

Nonparametric statistical tests

Nonparametric statistical tests are used instead of the parametric tests we have considered thus far (e.g. t-test; F-test), when:

The data are nominal or ordinal (rather than interval or ratio).

Page 4: Non Parametric Test

The data are not normally distributed, or have heterogeneous variance (despite being interval or ratio).

The following are some common nonparametric tests:

Chi square: 2

1. used to analyze nominal data

2. compares observed frequencies to frequencies that would be expected under the null hypothesis

Mann-Whitney U

1. compares two independent groups on a DV measure with rank-ordered (ordinal) data

2. nonparametric equivalent to a t-test

Wilcoxon matched-pairs test

1. used to compare two correlated groups on a DV measured with rank-ordered (ordinal) data

2. nonparametric equivalent to a t-test for correlated samples

Kruskal-Wallis test

1. used to compare two or more independent groups on a DV.with rank-ordered (ordinal) data

2. nonparametric alternative to one-way ANOVA

Zzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzz

Kruskal-Wallis non-parametric ANOVA

Page 5: Non Parametric Test

Data types that can be analysed with Kruskal-Wallis

the data points must be independent from each other

the distributions do not have to be normal and the variances do not have to be equal

you should ideally have more than five data points per sample

all individuals must be selected at random from the population

all individuals must have equal chance of being selected

sample sizes should be as equal as possible but some differences are allowed

Limitations of the test

if you do not find a significant difference in your data, you cannot say that the samples are the same

if significant differences are found when comparing more than two samples there are non-parametric multiple comparison tests available but they are only found in UNISTAT and otherwise have to be performed manually or calculated long-hand in Excel.

Introduction to Kruskal-Wallis

Kruskal-Wallis compares between the medians of two or more samples to determine if the samples have come from different populations. For instance it is a well known aspect of natural history that the littorinid species (snails) that are found on sheltered and exposed shores have different shell morphologies.

Page 6: Non Parametric Test

This could be tested by measuring the shell thickness of each individual in samples taken from a sheltered, an exposed and an intermediate shore. If the distributions prove not to be normal and/or the variances are different then the Kruskal-Wallis should be used to compare the groups. If a significant difference is found then there is a difference between the highest and lowest median. A non-parametric multiple comparison test must then be used to ascertain whether the intermediate shore also is significantly different. These are found in UNISTAT but must be set up on a spreadsheet in Excel or done by hand from the examples given in Zar (1984).

In the above example only one factor is considered (level of shore exposure) and so is termed a one-way Kruskal-Wallis. There are examples in Zar (1984) of a two-way Kruskal-Wallis test but again must be set up in Excel or done by hand.

Hypotheses

Data arrangement

Once you have established that your data suits Kruskal-Wallis, your data must be arranged thus for use in one of the statistical packages (SPSS, UNISTAT):

Results and interpretation

(Degrees of Freedom = number of samples/treatments - 1)

On completion of the 1-way Kruskal-Wallis the results will look something like this:

Page 7: Non Parametric Test

Although it looks a bit daunting do not be worried. There is only one value that concerns the selection of one of the hypotheses. The Right-Tail Probability (0.0052) is the probability of the differences between the data sets occurring by chance. Since it is lower than 0.05 the HO must be rejected and the HA accepted.

Two-way Kruskal-Wallis results would appear in a similar format to the two-way ANOVA.

Zzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzz

Non-parametric statisticsIn statistics, the term non-parametric statistics covers a range of topics:

distribution free methods which do not rely on assumptions that the data are drawn from a given probability distribution. As such it is the opposite of parametric statistics. It includes non-parametric statistical models, inference and statistical tests.

non-parametric statistic can refer to a statistic (a function on a sample) whose interpretation does not depend on the population fitting any parametrized distributions. Statistics cased on the ranks of observations are one example of such statistics and these play a central role in many non-parametric approaches.

non-parametric regression refers to modelling where the structure of the relationship between variables is treated non-parametrically, but where nevertheless there may be parametric assumptions about the distribution of model residuals.

Page 8: Non Parametric Test

Contents

1 Applications and purpose 2 Non-parametric models 3 Methods 4 General references 5 See also

[edit] Applications and purpose

Non-parametric methods are widely used for studying populations that take on a ranked order (such as movie reviews receiving one to four stars). The use of non-parametric methods may be necessary when data has a ranking but no clear numerical interpretation, such as when assessing preferences; in terms of levels of measurement, for data on an ordinal scale.

As non-parametric methods make fewer assumptions, their applicability is much wider than the corresponding parametric methods. In particular, they may be applied in situations where less is known about the application in question. Also, due to the reliance on fewer assumptions, non-parametric methods are more robust.

Another justification for the use of non-parametric methods is simplicity. In certain cases, even when the use of parametric methods is justified, non-parametric methods may be easier to use. Due both to this simplicity and to their greater robustness, non-parametric methods are seen by some statisticians as leaving less room for improper use and misunderstanding.

The wider applicability and increased robustness of non-parametric tests comes at a cost: in cases where a parametric test would be appropriate, non-parametric tests have less power. In other words, a larger sample size can be required to draw conclusions with the same degree of confidence.

Page 9: Non Parametric Test

[edit] Non-parametric models

Non-parametric models differ from parametric models in that the model structure is not specified a priori but is instead determined from data. The term nonparametric is not meant to imply that such models completely lack parameters but that the number and nature of the parameters are flexible and not fixed in advance.

A histogram is a simple nonparametric estimate of a probability distribution Kernel density estimation provides better estimates of the density than histograms. Nonparametric regression and semiparametric regression methods have been developed based

on kernels, splines, and wavelets. Data Envelopment Analysis provides efficiency coeficients similar to those obtained by

Multivariate Analysis without any distributional assumption.

[edit] Methods

Non-parametric (or distribution-free) inferential statistical methods are mathematical procedures for statistical hypothesis testing which, unlike parametric statistics, make no assumptions about the probability distributions of the variables being assessed. The most frequently used tests include

Anderson-Darling test Cochran's Q Cohen's kappa Efron-Petrosian test Friedman two-way analysis of variance by ranks Kendall's tau Kendall's W Kolmogorov-Smirnov test Kruskal-Wallis one-way analysis of variance by ranks Kuiper's test Mann-Whitney U or Wilcoxon rank sum test Maximum parsimony for the development of species relationships using computational

phylogenetics median test Pitman's permutation test Rank products Siegel-Tukey test Spearman's rank correlation coefficient Student-Newman-Keuls (SNK) test Van Elteren stratified Wilcoxon Rank Sum Test Wald-Wolfowitz runs test

Page 10: Non Parametric Test

Wilcoxon signed-rank test .

[edit] General references

Corder, G.W. & Foreman, D.I, "Nonparametric Statistics for Non-Statisticians: A Step-by-Step Approach", Wiley (2009) (ISBN: 9780470454619)

Wasserman, Larry, "All of Nonparametric Statistics", Springer (2007) (ISBN: 0387251456) Gibbons, Jean Dickinson and Chakraborti, Subhabrata, "Nonparametric Statistical Inference", 4th

Ed. CRC (2003) (ISBN: 0824740521)

[edit] See also

Parametric statistics Resampling (statistics) Robust statistics Particle filter for the general theory of sequential Monte Carlo methods

Zzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzz

Nonparametric methods

Nonparametric Tests

Wilcoxon Mann-Whitney Test

Wilcoxon Signed Ranks Test

Sign Test

Runs Test

Kolmogorov-Smirnov Test

Kruskal-Wallis Test

Main Contents page | Index of all entries

Page 11: Non Parametric Test

Nonparametric Tests

Nonparametric tests are often used in place of their parametric counterparts when certain assumptions about the underlying population are questionable. For example, when comparing two independent samples, the Wilcoxon Mann-Whitney test does not assume that the difference between the samples is normally distributed whereas its parametric counterpart, the two sample t-test does. Nonparametric tests may be, and often are, more powerful in detecting population differences when certain assumptions are not satisfied.

All tests involving ranked data, i.e. data that can be put in order, are nonparametric.

Wilcoxon Mann-Whitney Test

The Wilcoxon Mann-Whitney Test is one of the most powerful of the nonparametric tests for comparing two populations. It is used to test the null hypothesis that two populations have identical distribution functions against the alternative hypothesis that the two distribution functions differ only with respect to location (median), if at all.

The Wilcoxon Mann-Whitney test does not require the assumption that the differences between the two samples are normally distributed.

In many applications, the Wilcoxon Mann-Whitney Test is used in place of the two sample t-test when the normality assumption is questionable.

This test can also be applied when the observations in a sample of data are ranks, that is, ordinal data rather than direct measurements.

Page 12: Non Parametric Test

Wilcoxon Signed Ranks Test

The Wilcoxon Signed Ranks test is designed to test a hypothesis about the location (median) of a population distribution. It often involves the use of matched pairs, for example, before and after data, in which case it tests for a median difference of zero.

The Wilcoxon Signed Ranks test does not require the assumption that the population is normally distributed.

In many applications, this test is used in place of the one sample t-test when the normality assumption is questionable. It is a more powerful alternative to the sign test, but does assume that the population probability distribution is symmetric.

This test can also be applied when the observations in a sample of data are ranks, that is, ordinal data rather than direct measurements.

Sign Test

The sign test is designed to test a hypothesis about the location of a population distribution. It is most often used to test the hypothesis about a population median, and often involves the use of matched pairs, for example, before and after data, in which case it tests for a median difference of zero.

The Sign test does not require the assumption that the population is normally distributed.

In many applications, this test is used in place of the one sample t-test when the normality assumption is questionable. It is a less powerful alternative to the Wilcoxon signed ranks test, but does not assume that the population probability distribution is symmetric.

Page 13: Non Parametric Test

This test can also be applied when the observations in a sample of data are ranks, that is, ordinal data rather than direct measurements.

Runs Test

In studies where measurements are made according to some well defined ordering, either in time or space, a frequent question is whether or not the average value of the measurement is different at different points in the sequence. The runs test provides a means of testing this.

Example Suppose that, as part of a screening programme for heart disease, men aged 45-65 years have their blood cholesterol level measured on entry to the study. After many months it is noticed that cholesterol levels in this population appear somewhat higher in the Winter than in the Summer. This could be tested formally using a Runs test on the recorded data, first arranging the measurements in the date order in which they were collected.

Kolmogorov-Smirnov Test

For a single sample of data, the Kolmogorov-Smirnov test is used to test whether or not the sample of data is consistent with a specified distribution function. When there are two samples of data, it is used to test whether or not these two samples may reasonably be assumed to come from the same distribution.

The Kolmogorov-Smirnov test does not require the assumption that the population is normally distributed.

Compare Chi-Squared Goodness of Fit Test.

Page 14: Non Parametric Test

Kruskal-Wallis Test

The Kruskal-Wallis test is a nonparametric test used to compare three or more samples. It is used to test the null hypothesis that all populations have identical distribution functions against the alternative hypothesis that at least two of the samples differ only with respect to location (median), if at all.

It is the analogue to the F-test used in analysis of variance. While analysis of variance tests depend on the assumption that all populations under comparison are normally distributed, the Kruskal-Wallis test places no such restriction on the comparison.

It is a logical extension of the Wilcoxon-Mann-Whitney Test.

Page 15: Non Parametric Test

Non-Parametric Tests

In this section...

The Mann-Whitney U The Kruskal-Wallis H Test Chi-Square

Like all of the statistical tests discussed up to this point, non-parametric tests are used to investigate the relationship between two or more variables. Recall from our discussion at the start of this module that one of the key factors in determining which statistical test to run is the nature of the data to be analyzed. All of the statistical techniques you have learned up to now have made assumptions regarding the data (in particular regarding the population parameters estimated by the data). Correlation, ANOVA, independent and paired-samples t-tests, and regression all assume that the population parameters captured by the data are (1) normally distributed (values on all variables correspond roughly to the bell shaped normal curve); (2) quantitative in nature (the values can be manipulated arithmetically in a meaningful manner); (3) and, at the very least, interval (differences between values are captured by equal intervals). Indeed, these are conditions that must be met in order to run parametric tests.

But if you reflect for a moment on the nature of data in general, you will realize that not all data sets meet these assumptions. Consider, for example, the following: what if in our fictitious compensation study salary levels for our sample "bunch" around the extremes (high salary and low salary), with very few people earning amounts in the "average" range. Data such as these are not normally distributed--they "violate the normality assumption." Or say one of our questions is "are you a college graduate?" and we offer only two response options, "yes" or "no." This dichotomous variable is not quantitative in nature (how do you determine the mean of "yes"?). Lastly, there are many variables that are not captured on an interval or ratio scale. Some data simply divide the values into two mutually exclusive groups--USF graduates and non-USF graduates, for example. Such data are called "nominal" or "categorical": they "name" a group or create a mutually exclusive "category". Or, you might discover that your research question involves a variable that is ordinal in nature, that is, where the relationship between cases or subjects is expressed in terms of rank (first, second, third, etc.). A research question from our

Page 16: Non Parametric Test

compensation study that includes an ordinal variable might be: Is there a relationship between current salary and the national ranking of the college or university from which the subject received the highest degree?

Because not all data are the same and because research questions should not be limited by available data analytic techniques, alternate methods of statistical analysis are necessary. Non-parametric tests are one such useful alternative. Non-parametric tests, unlike their parametric equivalents, do not make any assumptions regarding the population parameters (hence the name). You should consider using non-parametrics in the situations listed below (if the truth be told, in some cases you have no choice but to use them). Remember, you always have the choice to run a parametric or a non-parametric test to answer your research questions. Parametric tests pack more statistical muscle, though, so examine your data closely and make this choice wisely. A consideration of the statistical power behind your analysis will help (or could hurt) your results.

Use non-parametric tests when:

1. One or more variables in your data set, including the dependent variable, is measured on a nominal or ordinal scale.2. One or more variables in your data set, including the dependent variable, violates the normality assumption.3. The sample size is small (< 20 cases or subjects).

There are many non-parametric tests. In fact, non-parametric equivalents exist for most "standard" parametric techniques. In this section we will describe only three common non-parametric tests: the Mann-Whitney U (the non-parametric equivalent of an independent samples t-test), the Kruskal Wallis H test (the non-parametric parallel to a one-way ANOVA), and Chi-square.

As you progress through the next few sections you should not forget that research questions answered by non-parametric tests are not much different than research questions answered by parametric tests. The purpose of all of these techniques is to investigate (compare) the relationship between two or more groups on a dependent measure. Parametric tests answer research questions by calculating and comparing the means of groups under study on the dependent variable. Non-parametric tests accomplish the same task by comparing the mean rank (or median) between groups. (In a clever trick, non-parametrics combine data on the dependent measure from all groups under consideration, then rank all subjects or cases based on the total sample, and then separate the groups out again. You need do nothing but interpret the

Page 17: Non Parametric Test

results!)

The Mann-Whitney U (Wilcoxon-Mann-Whitney)

As an illustrative example in the discussion of the the independent samples t-test at the beginning of this module, we considered an investigation of the relationship between gender and current salary level. The research question asked was this: is there a difference in salary level between men and women? In that case we were interested in comparing the mean differences between two groups on a dependent measure. As you will see, the Mann-Whitney U is interested in the same question but works with the data in a slightly different way.

In the earlier discussion we chose to run an independent samples t-test to compare the mean salary level for men with that of women. Why did we decide against the non-parametric equivalent, the Mann-Whitney U? Review (and then apply) the three conditions outlined above in which a non-parametric test is indicated. Is at least one variable nominal or ordinal? Certainly gender is an obvious dichotomous nominal variable (male, female). This alone suggests that a non-parametric test is at least plausible in this case. Condition 2 asks if at least one variable violates the normality assumption. Does the distribution of current salary levels approximate the normal curve? Possibly, but since we don't know for certain we would need to run Descriptives to check. Finally, condition 3 deals with the sample size. Are we dealing with an unsually small sample size in this instance? Let's assume we've done due diligence and that we have a sufficient sample.

Reflect now on the parametric versus non-parametric dilemma in this case. What was the determining factor in the choice of the independent samples t-test over the Mann-Whitney U to compare the differences between the two groups?

The answer is found in the assumption made by the researcher on the nature of the distribution of the dependent variable (current salary level). The very act of choosing the independent samples t-test over the Mann-Whitney U indicates that it was assumed current salary level is normally distributed. If we abandon that assumption or prove that it is false (here's where running Descriptives is critical), we are left with a categorial variable (gender) and a

Page 18: Non Parametric Test

quantitative variable in violation of the normality assumption. These are two important conditions that call for the use of a non-parametric test.

Analyze --> Nonparametric Tests --> 2 Independent Samples

Choose from the complete list of variables displayed at the left of the dialog box the dependent variable and move it to the "test variable" box on the right (from our compensation study, this would be "current salary"). Next, choose from the list the independent variable and move it to the "grouping variable" box ("gender" from the compensation study). (Note: The term "grouping variable" should suggest to you that this variable should be categorical.)

Just like the independent samples t-test you will now have to define the groups to be used in the analysis. Click on the "define groups" button. Enter under "Group 1" the value you wish to assign to the first group, and then tab and repeat the procedure for the second group. In our running example, this might be "1" for females and "2" for males. The values you assign here have no inherent meaning, rather they serve only to "mark" or indicate group membership. Once you have assigned the values, click continue. Notice in the test type box at the lower left that Mann-Whitney U is already marked.

For thoroughness sake, click now on the Options button and mark Descriptives.

The Output Editor shows three tables. The first table contains the traditional descriptives on your data set: number of cases and means and standard deviations on both variables. Look closely at the mean of the dependent variable, since that might give you a sense of the differences between the two groups. (On a side note, consider the mean and standard deviation for your dichotomous variable. What does this mean?) Before leaving this table, notice there is no information in this table to support a violation of normality

Page 19: Non Parametric Test

assumption. Running more detailed Descriptives as suggested above is necessary for that.

The second table is labelled "ranks." This should not surprise you when you recall that non-parametric tests investigate differences between groups by comparing ranks (and sometimes medians). Just as you examined the means on the dependent variable in the independent samples t-test output, look now at the relative ranks of the dependent variable in table 2. In our hypothetical example from the compensation study, the mean rank of current salary level for males is 149.7 and for females it's 309.2. Do you suspect there is a difference in salary level for men and women?

The final table in this output confirms our hunch. The first line of table 3 presents a calculated value for the Mann Whitney U test. In order to determine if this value is significant or not, check the value on line 4. If it is less than or equal to your predetermined significance level (.05 or .01), you have found a statistically significant difference between the groups.

The Kruskal-Wallis H Test

The Mann Whitney U is used to compare differences between two independent groups. When investigating differences between more than two groups, the Kruskal-Wallis H test is the more appropriate choice. The relationship between these two tests is the same as between their parametric equivalents, the independent samples t-test and the one-way ANOVA.

To illustrate this procedure, we propose to investigate the relationship between ethnicity and salary level from our made up compensation study. Our four ethnic categories are: African American, Hispanic, Asian Pacific Islander, and Caucasian. Here ethnicity is a categorical (nominal) independent variable with four levels; hence, to answer our reserach question we will be considering four groups. In the case of salary level, we have two options. The first is to run Descriptives on current salary level to determine if the data violates the normality assumption. If that's the case, we simply proceed to analysis. The second option is a bit more complicated. We can recode current salary level into a dichotomous variable ("high" and "low") and then run the analysis. In doing so we clearly meet the criteria for running a non-parametric test. (To refresh your memory on how to recode variables, click here. In this example, to determine where to draw the line between high and low salary,

Page 20: Non Parametric Test

running Frequencies to determine the median of the distribution is helpful.)

Analyze --> Nonparametric Tests --> K Independent Samples

Choose from the complete list of variables displayed at the left of the dialog box the dependent variable and move it to the "test variable" list on the right (from our compensation study, this would be the recoded dichotomous variable"high salary" and "low salary."). Next, choose from the list the independent variable and move it to the "grouping variable" box ("ethnicity" from the compensation study). You must now define the range of values for the groups to be used in the analysis. (A range of values is necesary for this test because you are examining more than two groups.) Click on the "define range" button, and enter the range of values for your variable (in the salary study, the range of values if 1 to 4, one value for each ethnic group.) Remember, these values have no inherent meaning, instead they only "mark" or indicate group membership. Once you have assigned the values, click Continue. Notice in the test type box at the lower left that Kruskal-Wallis H is already marked.

For thoroughness sake click now on the Options button and mark Descriptives.

The Output Editor shows three tables. The first table contains the traditional descriptives on your data set: number of cases, means, and standard deviations on both variables. Since in our fictitious example from the compensation study both variables are categorical, this table is not particularly helpful. Still, always review the contents of this table.

The second table is the rank table. Notice a rank on the dependent variable has been calculated for all levels of your independent variable. In our ongoing example, we have ranks for four groups--Hispanic, Asian, Caucasian and African American. To get a sense of differences between groups, look at the relative ranks of the dependent variable. In our hypothetical example from the compensation study, the mean rank of in salary level for Caucasians is 239.01, for Asians, 245.95; for African Americans, 284.76; and for Hispanics, 279.14. Given this information, do you suspect there is a difference in salary category (high versus low) based on ethnicity?

Table 3 contains the answer to that question. If the Chi-square value reported

Page 21: Non Parametric Test

on line 1 is not significant, there is no statistically significant difference between the groups on the dependent measure under study. If the Chi-square value reported is significant there is a difference between the groups.

Like the results from an ANOVA, these tell you only that a difference exists somewhere among the groups; it does not tell you exactly where thse difference(s) is (are). So, like ANOVA, post hoc tests are required to determine which groups differ on the dependent variable. Unfortunately, there is not a quick and easy procedure for running non-parametric post hocs in SPSS. Rather, the researcher must run a Mann Whitney U on all pairs of groups in order to determine if there are differences between them. In our example, it's likely that there is a difference in rank salary level between Asians and African Americans and Asians and Hispanics. But is there a difference between Caucasians and Asians? Or between Hispanis and African Americans? It's good practice to run a post hoc test for each pair of groups under study; you will not want to report in your thesis, disseration, or article that "visual inspection of mean ranks indicates there is no difference between the groups." In our example, we would run a Mann Whitney U comparing group 1 with groups 2, 3, and 4; group 2 with 3 and 4; and group 3 with 4. To refresh your memory on the Mann Whitney U, click here.

Chi-Square

Probably the most commonly used non-parametric test is chi-square. Chi-square is quite versatile and can be applied in a variety of situations. Two of the more frequent uses of chi-square are described in this section.

The most common use of the chi-square test is to examine the relationship between two nominal variables. If this sounds like correlation to you, it should. This application of Chi-square is the non-parametric equivalent of a Pearson correlation coefficient. From our compensation study, for example, we could use Chi-square to examine the relationship between gender and salary group ("high" and "low").

Analyze --> Descriptives --> Crosstabs

Identify from the complete list of variables displayed on the left of the dialog box the two variables you wish to study. Move one of the variables to the

Page 22: Non Parametric Test

"row" box and the other to the "column" box. (It doesn't really matter which one goes where.) Click on the Statistics botton at the bottom of the dialog box. Mark Chi-square and then click Continue. Click OK.

The first table in the output, Case Processing Summary, indicates how many cases were used in the analysis. It also shows how many cases were missing. Remember, it is important to always review this information. Consider this: what if your results are based on data from only half the total sample? Would you have as much confidence in results?

The next table is the crosstabulation of the two variables under study (in our case, gender and salary group). This table shows the distribution of one variable across the levels of the other. In our example, the crosstabs function tells us exactly how many men and how many women fall into "high salary" and "low salary" groups, respectively.

The final table, the Chi-Square Tests table, tells you whether the two variables under study are related. Direct your attention to the very first line of that table: the Pearson Chi-Square. (The fortuitous naming of this coefficient is a handy reminder of the purpose of this test!) If the value reported in column four on the first line is equal to or less than your level of significance (.05 or .01), you can conclude that the two variables are related.

A second, more advanced use of Chi-square is for hypothesis testing. You can use Chi-square to test a hunch or suspision you have about the relationship between two variables. In this application of the test, Chi-square examines the frequency of occurence within a group. Specifically, Chi-square evaluates whether the observed (or actual) frequency of occurence within a category or group is different from an expected frequency of occurence within that group. That is, is the real world different from what you hypothesize it to be?

Let's go back to our compensation study to illustrate. Let's assume we are interested in examining the relationship between gender (male, female) and salary level (high, low). Given what we know from research about the discrepancy in earning power between men and women, it would be foolish to expect an equal number (percentage) of men and women to be in the "high" or "low" group. A 50-50 expected frequency of occurence therefore seems unlikely. We may hypothesize, however, based upon reading in the management/compensation literature, that women will comprise roughly 25%

Page 23: Non Parametric Test

of the "high salary" group and 75% of the "low salary" group. Chi-square can be used to test this hypothesis: do women truly make up 25% of the high salary group and 75% of the low salary group? A non-significant Chi-square result would support the hypothesis, suggesting that our hypothesized (or expected) frequencies match the observed frequencies, and women do in fact comprise only one quarter of the high salary group and three quarters of the low salary group. On the other hand, a significant result would suggest that our expected frequencies were different from the observed frequencies, meaning that women do not make up 25% of the high salary or 75% of the low salary groups. The exact proportion of women in either group is unclear, but the results of the Chi-square test tell us it is not 25-75. (Your hypothesis must be grounded in the literature of your field. It would not be appropriate to "guess" your way to the result you want by just randomly plugging in numbers.)

Analyze --> Non-Parametrics --> Chi-Square

Choose from the complete list of variables displayed on the left of the dialog box the two variables under study and move them to the Test Variable List on the right (from our compensation study example, this would be gender and the recoded variable "salary_hilo"). In the Expected Value box directly below the Test Variable List, enter the values corresponding to the frequencies you suspect or hypothesize in each group. From the fictitious compensation study, the expected values are "25" and "75". Click on the Values option, and enter the first expected frequency (25) and click Add; repeat this procedure for the second expected value (75). Click on the Options button in the lower right hand corner. Choose Descriptives and then click on Continue. Back on the main dialog box, click OK.

If you haven't been following along with your own data up to this point, take a moment now to catch up.

The first table in the Output, Descriptive Statistics, contains the standard descriptive statistics you have come to expect by now. Make it a habit to review this information before continuing.

The next three tables are directly related to the Chi-Square test. The first two are labeled individually with the name of the variables you have chosen to study ("gender" and "salary_hilo" in our ongoing example). The first column of

Page 24: Non Parametric Test

each table shows the number of mutually exclusive categories within each variable. In the salary study we are looking at gender and salary level (defined as "high salary" and "low salary"), two clearly dichotomous variables. We should not be surprised then to see that the first column in both tables shows two values.

The second column of each table shows us the number of observed occurences within each category of the variable. Remember, this is the count (frequency) of actual occurences within the data set of the category of the variable. The third column contains a calculated value corresponding to the hypothesized frequency we expected to see (either 25% or 75% from the compensation study). The fourth column reports the difference between observed and expected frequencies.

The fourth and final table, Test Statistics, answers the question: is there a difference between what is and what we expected? If the significance level reported on the third line of this table is less than or equal to your predetermined level of significance (.05 or .01), you can conclude that the frequency of actual occurence and expected occurence are different. In the compensation study, for example, if we had a significant result we could report that there is a difference between what we expected (25% of females in the high salary group and 75% in the low salary group) and what in fact was observed.