Professional Documents
Culture Documents
CH 3 - 9-25-2013
CH 3 - 9-25-2013
measurement Examples: Intelligence, psychological disorders, personality, relationship processes, attachment, aggression, etc. Accurately and appropriately measuring variance in behaviour or mental processes
Are we really measuring the construct we want to measure? Is the measure consistent over time? Is the measure practical to use in research?
Types of Measures
Observational Measures Involve the direct observation of behavior Can be used to measure anything an organism does that can be observed Researchers can either directly observe or use audio and video recordings Examples:
Types of Measures
Physiological and Neuroscientific Measures Facilitate the study of associations between biological processes and behavior Involves the use of specialized equipment to measure heart rate, brain activity, hormonal changes, and other bodily responses Examples: fMRI, EEG, cortisol, oxytocin
Types of Measures
Self-report measures Responses to questionnaires and interviews Can use self-reports to measure:
measurement
Converging operations: Measuring a construct in
E.g., combining observational, physiological, and self-report measures Combining measures within a study allows us to better separate signal from measurement noise
Nominal Scale numbers assigned are simply labels for characteristics or behaviors
For example: Males = 1, Females = 2 Democrat = 1, Republican = 2, Independent =3
Interval Scale equal differences between the numbers reflect equal differences between participants, but there is no true zero point Examples: Scores on an IQ test Ratings on a 5-point agree-disagree scale
Ratio scales provide the greatest amount of information and should be used whenever possible.
Nominal
Ordinal
Interval
Ratio
Number of correct answers Time to complete task Gain in height since last year
Scales of Measurement
Scale of Measurement influences the choice of
statistical analysis Typically interval and ratio vs. ordinal and nominal
Measurement Error
Observed Score = True Score + Measurement Error
True Score the score that a participant would have
obtained if the measure were perfect and we were able to measure without error Measurement Error variability in scores due to factors that distort the true score
Transient States a temporary, unstable state of the participant E.g., mood, health, fatigue, anxiety
2.
Stable Attributes enduring traits of the participant such as illiteracy, paranoia, hostility
4.
ambiguous wording
5.
Reliability
Reliability the consistency or dependability of a
measuring technique
The reliability of a measure is an inverse function
of measurement error.
If a measurement has high reliability, participants
Reliability
Testing reliability requires analyzing the
variance that is associated with participants true scores: Reliability = True-score variance / Total variance
Assessing Reliability
Researchers estimate reliability by assessing the
extent to which two or more measurements of the same behavior, object, or event yield similar scores.
Researchers usually use a correlation coefficient to
Correlation Coefficients
Correlation coefficient expresses the strength
Can range from -1.00 to +1.00 Correlation of .00 indicates no relationship between the variables The sign indicates whether the relationship between the variables is positive or negative.
Test-Retest Reliability
Test-Retest Reliability consistency of
Administer measure on two separate occasions Examine the correlation between the scores obtained on the two occasions Correlation > .70 indicates acceptable reliability Useful only if the attribute being measured should not change over time
Interitem Reliability
Interitem Reliability assesses the degree of
on a scale are measuring the same thing. If not, summing scores across the items creates measurement error and lowers reliability.
a particular item and the sum of all the other items on the scale (ideally > .30) Split-half reliability divide the items on a scale into two sets and examine the correlation between the sets Cronbachs alpha coefficient () equivalent to the average of all possible split half reliabilities
Most frequently used Adequate interitem reliability if exceeds .70 Influenced by number of items
Interrater Reliability
Interrater Reliability -- the consistency among
3. Train observers.
4. Minimize errors in coding data.
(basically, try to minimize variance due to measurement error)
Validity
Validity the degree to which a measurement
procedure actually measures what it is intended to measure rather than measuring something else (or nothing other than error)
Face Validity
Face Validity the extent to which a measure
Just because something has face validity doesnt mean that it is valid. Many measures without face validity are valid. Some measures are designed to lack face validity so as to disguise their purpose.
Construct Validity
Hypothetical constructs entities that cannot be
Construct Validity
Convergent Validity a measure correlates
Embarrassability should be positively correlated with shyness but negatively correlated with selfconfidence.
Criterion-Related Validity
Criterion-related validity the extent to which a
measure allows us to distinguish among participants on the basis of a particular behavioral criterion
Researchers examine whether behavioral outcomes
divorce
Criterion-Related Validity I
Concurrent validity scores on a measure
are related as expected to a criterion that is assessed at the time the measure is administered
Example: an Embarrassability Scale (administered today) predicts blushing in the current situation
Criterion-Related Validity II
Predictive Validity scores on a measure
Example: an Embarrassability Scale (administered today) predicts whether students sign-up for public speaking classes next semester
The question is not whether various groups score differently on the test. Rather, test bias is present when the validity of a measure is lower for some groups than for others.
Any Questions?