Top Banner
Reliability and Validity
38

Reliability and Validity. Reliability When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing. Degree.

Mar 31, 2015

Download

Documents

Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: Reliability and Validity. Reliability  When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing.  Degree.

Reliability and Validity

Page 2: Reliability and Validity. Reliability  When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing.  Degree.

Reliability

When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing.

Degree to which scores are free of “measurement error”

Consistency of measurement

Page 3: Reliability and Validity. Reliability  When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing.  Degree.

VALIDITY

• The extent to which measures indicate what they are intended to measure.

• The match between the conceptual definition and the operational definition.

Page 4: Reliability and Validity. Reliability  When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing.  Degree.

RELATIONSHIP BETWEEN

RELIABILITY AND VALIDITY

• Necessary but not sufficient

• Reliability is a prerequisite for measurement validity

• One needs reliability, but it’s not enough

Page 5: Reliability and Validity. Reliability  When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing.  Degree.

Example

• Measuring height with reliable bathroom scale

• Measuring “aggression” with observer agreement by observing a kid hitting a Bobo doll

Page 6: Reliability and Validity. Reliability  When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing.  Degree.

Types of Reliability Measurement

•1. Stability Reliability

•2. Equivalence Reliability

Page 7: Reliability and Validity. Reliability  When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing.  Degree.

Stability Reliability• Test-retest • SAME TEST – DIFFERENT TIMES• Testing phenomenon at two different

times;• The degree to which the two

measurements of “Sam Ting,” using same measure, are related to one another

• Only works if phenomenon is unchanging

Page 8: Reliability and Validity. Reliability  When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing.  Degree.

Example of Stability

• Administering same questionnaire at 2 different times

• Re-examining client before deciding on intervention strategy.

• Running “trial” twice (e.. g. errors in tennis serving)

Page 9: Reliability and Validity. Reliability  When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing.  Degree.

Notes on Stability Reliability

• When ratings are by an observer rather than the subjects themselves, this is called Intraobserver Reliability or Intrarater Reliability.

• Answers about the past are less reliable when they are very specific, because the questions may exceed the subjects’ capacity to remember accurately.

Page 10: Reliability and Validity. Reliability  When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing.  Degree.

Equivalence Reliability

1. Inter-item (split ½)

2. Parallel forms [Different types of measures]

3. Interobserver Agreement -Is every observer scoring the same ?

Page 11: Reliability and Validity. Reliability  When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing.  Degree.

1. Inter-item Reliability

• (Internal consistency): The association of answers to a set of questions designed to measure the same concept.

Page 12: Reliability and Validity. Reliability  When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing.  Degree.

Note on Inter-item Validity• The stronger the association among

individual items and the more items included, the higher the reliability of an index

• Cronbach’s alpha is a statistic commonly used to measure inter-item reliability

• Cronbach’s alpha is based on the average of all the possible correlations of all the split 1/2s of a set of questions on a questionnaire

Page 13: Reliability and Validity. Reliability  When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing.  Degree.

2. Parallel forms of Reliability

• Split ½ (inter-item)

• Different types of measures

• Interobserver Reliability– Is everyone measuring the same thing ?

• Different measures – same time

Page 14: Reliability and Validity. Reliability  When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing.  Degree.

3.Interobserver Reliability

• Correspondence between measures made by different observers.

Page 15: Reliability and Validity. Reliability  When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing.  Degree.

Note for Stat Students Only

• The text inadvertently describes a 3rd type of reliability that we’re not concerned with in this class: ‘goodness of fit’ about a slope line. It’s sometimes referred to as random measurement error.

• Save this for Grad School =)

Page 16: Reliability and Validity. Reliability  When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing.  Degree.

Note on Reliability

• For Statistics people, the following quote refers to ‘goodness of fit’ around a slope line due to measurement error.

Secondary Definition of Reliability from a previous slide

“…or that the measured scores changes in direct correspondence to actual changes in the phenomenon”

Page 17: Reliability and Validity. Reliability  When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing.  Degree.

And Now Onto Validity…..

Page 18: Reliability and Validity. Reliability  When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing.  Degree.

Types of Validity• 1. Content Validity

– Face Validity– Sampling Validity (content validity)

• 2. Empirical Validity– Concurrent Validity– Predictive Validity

• 3. Construct Validity

Page 19: Reliability and Validity. Reliability  When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing.  Degree.

Face Validity

• confidence gained from careful inspection of a concept to see if it’s appropriate “on its face;”

• In our [collective] intersubjective, informed judgment, have we measured what we want to measure?

• (N.B. use of good judgment)

Page 20: Reliability and Validity. Reliability  When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing.  Degree.

Example of Face Validity

• Rosenberg’s self esteem scale questions:

Page 21: Reliability and Validity. Reliability  When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing.  Degree.

Content validity

• Also called “sampling validity”

• establishes that the measure covers the full range of the concept’s meaning, i.e., covers all dimensions of a concept

• N.B depends on “good “ judgment

Page 22: Reliability and Validity. Reliability  When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing.  Degree.

Example of content validity

• Earlier SES scale in class

• Authoritarian personality questions from Walizer & Wienir

Page 23: Reliability and Validity. Reliability  When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing.  Degree.

*Note *

• Actually I think face and content validity are probably Sam Ting

Page 24: Reliability and Validity. Reliability  When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing.  Degree.

EMPIRICAL Validity

• Establishes that the results from one measure match those obtained with a more direct or already validated measure of the same phenomenon (the “criterion”)

• Includes – Concurrent– Predictive

Page 25: Reliability and Validity. Reliability  When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing.  Degree.

Concurrent Validity

• Validity exists when a measure yields scores that are closely related to scores on a criterion measured at the same time

• Does the new instrument correlate highly with an old measure of the same concept that we assume (judge) to be valid? (use of “good” judgment)

Page 26: Reliability and Validity. Reliability  When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing.  Degree.

Example of concurrent validity

• Aronson’s doodle measure of achievement motivation.

• Act vs. SAT

Page 27: Reliability and Validity. Reliability  When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing.  Degree.

Predictive Validity

• Exits when a measure is validated by predicting scores on a criterion measured in the future

• Are future events which we judge to be a result of the concept we’re measuring anticipated [predicted] by the scores we’re attempting to validate

• Use of “good” judgment

Page 28: Reliability and Validity. Reliability  When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing.  Degree.

Examples of Predictive Validity

• Bronson screening test for “at risk” parenting followed up by interviewing and observing family members and school staff later

• Sat / ACT scores and later college “performance” (grades)

• Grades are “judged” to be measured validly

Page 29: Reliability and Validity. Reliability  When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing.  Degree.

What’s a Construct? [NSB]*• Multidimensional concept

– SES– Industrialization

• Fuzzy concept / hard to define– Ego strength– Love

• Concept build out of other concepts– Force=mass * acceleration* Ya better know these!!!!!

Page 30: Reliability and Validity. Reliability  When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing.  Degree.

Consider This:

• If a construct is hard to conceptualize doesn’t it make sense that it’ll be more difficult to operationalize and validate?

Page 31: Reliability and Validity. Reliability  When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing.  Degree.

Construct validity• : established by showing that a measure is

(1) related to a variety of other measures as specified in a theory, used when no clear criterion exists for validation purposes (2) that the operationalization has a set of interrelated items and (3) that the operationalization has not included separate concepts

Page 32: Reliability and Validity. Reliability  When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing.  Degree.

Construct validity• Check the intercorrelation of items used to

measure construct judged to be valid• Use theory to predict a relationship and use

a judged to be valid measure of the other variable then check for relationship

• Demonstrate that your measure isn’t related to judged to be valid measures of unrelated concepts

Page 33: Reliability and Validity. Reliability  When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing.  Degree.

Convergent Validity

• Convergent validity: achieved when one measure of a concept is associated with different types of measures in the same concept (this relies on the same type of logic as measurement triangulation)

• Measures intercorrelated

Page 34: Reliability and Validity. Reliability  When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing.  Degree.

Example of questions that Interrelate

Questions for Companionate…intimacy

• We get along well

• We communicate

• We like the same stuff

• Our chemistry is good

• We support each other

Page 35: Reliability and Validity. Reliability  When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing.  Degree.

Discriminant Validity• Discriminant validity: scores on the

measure to be validated are compared to scores on measures of different but related concepts and discriminant validity is achieved if the measure to be validated is NOT strongly associated with the measures of different concepts

• Measure not related to unrelated concepts

Page 36: Reliability and Validity. Reliability  When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing.  Degree.

Questions for Passion

• I think my partner is HOT

• My partner turns me on

• When I’m with my partner I just feel the electricity

Page 37: Reliability and Validity. Reliability  When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing.  Degree.

Using theory

• Measure of constructs predicts what theory says it should

Page 38: Reliability and Validity. Reliability  When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing.  Degree.

• Companionate rel longevity

satisfaction