Sunteți pe pagina 1din 25

Defining and Measuring

Variables
Chapter 3

Dusana Rybarova
Psyc 290B
May 17 2006
Outline:

1. An overview of measurement
2. Constructs and operational definitions
3. Validity and reliability of measurement
4. Scales of measurement
5. Modalities of Measurement
6. Other Aspects of Measurement
1. An overview of measurement
two aspects of measurement are particularly
important in planning a research study or
reading a research report:
often there is not a one-to-one relationship between
the variable measured and the measurement obtained
(knowledge, performance and exam grade)
there are usually several different options for
measuring any particular variable (types of exams
and questions on exams)
Direct measurement (height, weight) vs indirect
measurement (motivation, knowledge, memory,
marital satisfaction)
2. Constructs and operational
definitions
Theories summarize our observations, explain
mechanisms underlying a particular behavior and
make predictions about the behavior.
many research variables, particularly variables of
interest to behavioral scientists, are hypothetical
attributes or mechanisms explaining and predicting
some behavior in a theory are called constructs

external
stimulus construct behavior
factor

reward motivation performance


constructs can not be directly observed or measured
however, researchers can measure external, observable
events as an indirect method of measuring the construct
itself
operational definition
is a procedure for measuring and defining a construct, indirect
method of measuring something that can not be measured directly
an operational definition specifies a measurement procedure for
measuring an external, observable behavior and uses the resulting
measurements as a definition and a measurement of the
hypothetical construct
e.g. IQ test is an operational definition for the construct intelligence
- provide and example of a theoretical construct and its
operational definition
You dont always have to come up with your own
operational definition of the construct, you can use some
conventional measurement procedure from previous
studies
3. Validity and reliability of
measurement
How do you decide which method of
measurement (operational definition of a
construct) is the best?
there are two general criteria for
evaluating the quality of any measurement
procedure
validity
reliability
Validity of measurement

Validity of measurement
concerns the truth of the measurement
it is the degree to which the measurement
process measures the variable it claims to
measure
Is the IQ score truly measuring intelligence?
What about size of the brain and bumps on
the scull?
Different kinds of validity
face validity
the simplest and least scientific definition of validity
it is demonstrated when a measure superficially appears to
measure what it claims to measure
Based on subjective judgment and difficult to quantify
e.g. intelligence and reasoning questions on the IQ test
Problem - participants can use the face validity to change
their answers
concurrent validity (criterion validity)
is demonstrated when scores obtained from a new measure
are directly related to scores obtained from a more
established measure of the same variable
e.g. new IQ test correlates with an older IQ test
Different kinds of validity (cont.)

Different kinds of validity


predictive validity
when scores obtained from a measure accurately predict
behavior according to a theory
e.g. high scores on need for achievement test predict
competitive behavior in children (ring toss game)
construct validity
is demonstrated when scores obtained from a measure are
directly related to the variable itself
Reflects how close the measure relates to the construct
(height and weight example)
in one sense, construct validity is achieved by repeatedly
demonstrating every other type of validity
Different kinds of validity (cont.)
Different kinds of validity
convergent validity
is demonstrated by a strong relationship between the scores
obtained from two different methods of measuring the same
construct
e.g. an experimenter observing aggressive behavior in children
correlated with teachers ratings of their behavior
divergent validity
is demonstrated by using two different methods to measure two
different constructs
convergent validity must be shown for each of the two
constructs and little or no relationship exists between the scores
obtained from the two different constructs when they are
measured by the same method
e.g. aggressive behavior and general activity level in children
Convergent validity, divergent
validity and construct validity
By demonstrating strong convergent validity for two different
constructs and then showing divergent validity between the
two constructs, you obtain strong construct validity of the
two constructs
Aggressive Teachers ratings
High convergent Experimenters

behavior
validity observation
Related scores
High Diver High Diver
gent Vali gent Vali
dity dity
Active Unrelated Unrelated
behavior scores scores

Teachers ratings
High convergent
Experimenters
validity
Related scores observation
Reliability of measurement
Reliability of measurement
a measurement procedure is said to be reliable if
repeated measurements of the same individual under
the same conditions produce identical (or nearly
identical) values
reliability is the stability or the consistency of
measurement

measured score = true score + error

IQ score = true IQ score + mood, fatigue etc.


Reliability and error of
measurement
Inconsistency (lack of reliability) of measurement comes
from error
The higher the error the more unreliable the
measurement
Sources of error
observer error
the individual who makes the measurements can introduce simple
human error into the measurement process
environmental changes
small changes in the environment from one measurement to another
(e.g. time of the day, distraction in the room, lighting)
participant changes
participants change between measurements (mood, hunger,
motivation)
Types and measures of reliability
successive measurements
Obtaining scores from two successive measurements and calculating
a correlation between them
the same group, the same measurement at two different times
test-retest reliability
simultaneous measurements
obtained by direct observation of behaviors (two or more separate
observers at the same time), consistency across raters
inter-rater reliability
internal consistency
degree of consistency of scores from separate items on a test or
questionnaire consisting of multiple items
you want all the items or groups of items tapping the same
processes
researchers commonly split the set of items in half, compute a
separate score of each half, and then evaluate the degree of
agreement between the two scores
split-half reliability
The relationship between reliability
and validity
they are partially related and partially
independent
reliability is a prerequisite for validity
(measurement procedure can not be valid
unless it is reliable e.g. IQ, huge variance of
repeated measurements is impossible if we
are truly measuring intelligence)
it is not necessary for a measurement to be
valid for it to be reliable (e.g. height as a
measure of intelligence)
4. Scales of measurement
Scales define the type categories we use in measurement
and the selection of a scale has direct impact on our
ability to describe relationships between variables
the nominal scale
simply represents qualitative difference in the variable measured
can only tell us that a difference exists without the possibility
telling the direction or magnitude of the difference
e.g. majors in college, race, gender, occupation
the ordinal scale
the categories that make up an ordinal scale form an ordered
sequence
can tell us the direction of the difference but not the magnitude
e.g. coffee cup sizes, socioeconomic class, T-shirt sizes, food
preferences
Scales of measurement (cont.)
the interval scale
categories on an interval scale are organized
sequentially, and all categories are the same size
we can determine the direction and the magnitude of
a difference
May have an arbitrary zero (convenient point of
reference)
e.g. temperature in Farenheit, time in seconds
the ratio scale
consists of equal, ordered categories anchored by a
zero point that is not arbitrary but meaningful
(representing absence of a variable
allows us to determine the direction, the magnitude,
and the ratio of the difference
e.g. reaction time, number of errors on a test
5. Modalities of measurement
One can measure a construct by selecting
a measure from three main categories
There are three basic modalities of
measurement:
self-report
physiological measurement
behavioral measurement
behavioral observation
content analysis and archival research
Self-report measures
you ask a participant to describe his behavior,
to express his opinion or characterize his
experience in an interview or by using a
questionnaire with ratings
Positive aspects
Only the individual has direct access to information
about his state of mind
More direct measure
Negative aspects
Participants may distort the responses to create a
better self-image or to please the experimenter
The response can also be influenced by wording of
the questions and other aspects of the situation
Physiological measures
Physiological manifestations of the underlying
construct
e.g. EEG, EKG, galvanic skin response,
perspiration, PET, fMRI
advantages
provides accurate, reliable, and well-defined
measurements that are not dependent on subjective
interpretation
disadvantages
equipment is usually expensive or unavailable
Presence of monitoring devices may create unnatural
situation
question: Are these procedures a valid measure of the
construct (e.g. increase in heart rate to fear, arousal)
Behavioral measures
behaviors that can be observed and measured (e.g.
reaction time, reading speed, focus of attention,
disruptive behavior, number of words recalled on a
memory test)
How to select the right behavioral measure?
Depends on the purpose of the study
In clinical setting the same disorder can reveal itself through
different symptoms
In studying memory we want to have the same measure for all
subjects to be able to compare them
Beware of situational changes in behavior (e.g.
disruptive behavior in school vs when observed) and
different behavioral indicators of a construct
6. Other aspects of measurement
multiple measures
sometimes you can use two (or more) different
procedures to measure the same variable (e.g. heart
rate and questionnaire as a measure of fear)
problems (the two variables may not behave in the
same way)
e.g. a specific therapy for treating fear may have large effect
on behavior but no effect on heart rate
the lack of agreement between two measures is called
desynchrony
One measure can be more sensitive than other
Different measures may indicate different dimensions of the
variable and change at different times during the treatment
Sensitivity and range effects
are the measures sensitive enough to respond to the
type and magnitude of the changes that are
expected? (e.g. seconds vs. milliseconds, difficult or
easy exams)
range effects
a ceiling effect (the clustering of scores at the high end of a
measurement scale, allowing little or no possibility of
increases in value, e.g. test that is too easy)
a floor effect (the clustering of scores at the low end of a
measurement scale, allowing little or no possibility of
decreases in value, e.g. test that is too difficult)
Range effects are usually a consequence of using a measure
that is inappropriate for a particular group (e.g. 4-grade test
for college students)
Participant reactivity and
experimenter bias
participant reactivity is the way how participant reacts to
the experimental situation (e.g. overly cooperative, overly
defensive, or hostile)
To avoid these problems one can try to disguise the true purpose of
the experiment or observe individuals without their awareness
(beware ethical issues)
experimenter bias is the way experimenter influences
results (e.g. by being warm and friendly with one group of
participants vs. cold and stern with other group)
to avoid participant reactivity and experimenter bias we
use:
standardized procedures (e.g. instructions recorded on a tape)
a research study is single blind if the researcher does not know the
predicted outcome
a research study is double blind if both the researcher and the
participants are unaware of the predicted outcome
Participant reactivity and
experimenter bias

to avoid participant reactivity and


experimenter bias we use blind experiments
a research study is single blind if the researcher
does not know the predicted outcome
a research study is double blind if both the
researcher and the participants are unaware of the
predicted outcome

S-ar putea să vă placă și