0% found this document useful (0 votes)
7 views74 pages

Types of Research

Just some documents on the types of research
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
7 views74 pages

Types of Research

Just some documents on the types of research
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 74

Quantitative Methods

Defining Surveys and Experiments


A survey design provides a quantitative description
of trends, attitudes, and opinions of a population,
or tests for associations among variables of a
population, by studying a sample of that
population. Survey designs help researchers
answer three types of questions: (a) descriptive
questions (e.g., What percentage of practicing
nurses support the provision of hospital abortion
services?);
(b) questions about the relationships
between variables (e.g., Is there a
positive association between
endorsement of hospital abortion
services and support for implementing
hospice care among nurses?); or in
cases where a survey design is
repeated over time in a longitudinal
study;
(c) questions about predictive
relationships between
variables over time (e.g., Does
Time 1 endorsement of support
for hospital abortion services
predict greater Time 2 burnout
in nurses?).
An experimental design systematically
manipulates one or more variables in order to
evaluate how this manipulation impacts an
outcome (or outcomes) of interest. Importantly, an
experiment isolates the effects of this
manipulation by holding all other variables
constant. When one group receives a treatment
and the other group does not (which is a
manipulated variable of interest), the experimenter
can isolate whether the treatment and not other
factors influence the outcome.
For example, a sample of nurses could be
randomly assigned to a 3-week expressive
writing program (where they write about their
deepest thoughts and feelings) or a matched
3-week control writing program (writing about
the facts of their daily morning routine) to
evaluate whether this expressive writing
manipulation reduces job burnout in the
months following the program (i.e., the writing
condition is the manipulated variable of
interest, and job burnout is the outcome of
interest).
Whether a quantitative study employs a
survey or experimental design, both
approaches share a common goal of
helping the researcher make inferences
about relationships among variables, and
how the sample results may generalize to a
broader population of interest (e.g., all
nurses in the community).
The Survey Design
Identify the purpose of survey research. The
primary purpose is to answer a question (or
questions) about variables of interest to you. A
sample purpose statement could read: “The
primary purpose of this study is to empirically
evaluate whether the number of overtime hours
worked predicts subsequent burnout symptoms in
a sample of emergency room nurses.”
Indicate why a survey method is the preferred type
of approach for this study. In this rationale, it can
be beneficial to acknowledge the advantages of
survey designs, such as the economy of the design,
rapid turnaround in data collection, and
constraints that preclude you from pursuing other
designs (e.g., “An experimental design was not
adopted to look at the relationship between
overtime hours worked and burnout symptoms
because it would be prohibitively difficult, and
potentially unethical, to randomly assign nurses to
work different amounts of overtime hours.”).
Indicate whether the survey will be
cross-sectional—with the data
collected at one point in time—or
whether it will be longitudinal—
with data collected over time.
Specify the form of data collection. Fowler (2014)
identified the following types: mail, telephone, the
Internet, personal interviews, or group administration
(see also Fink, 2016; Krueger & Casey, 2014). Using an
Internet survey and administering it online has been
discussed extensively in the literature (Nesbary, 2000;
Sue & Ritter, 2012). Regardless of the form of data
collection, provide a rationale for the procedure, using
arguments based on its strengths and weaknesses,
costs, data availability, and convenience.
The Population and Sample
The population. Identify the population in the
study. Also state the size of this population, if
size can be determined, and the means of
identifying individuals in the population.
Questions of access arise here, and the
researcher might refer to availability of
sampling frames—mail or published lists—of
potential respondents in the population.
Sampling design
Identify whether the sampling design for this population is single
stage or multistage (called clustering). Cluster sampling is ideal
when it is impossible or impractical to compile a list of the
elements composing the population (Babbie, 2015). A single-stage
sampling procedure is one in which the researcher has access to
names in the population and can sample the people (or other
elements) directly. In a multistage or clustering procedure, the
researcher first identifies clusters (groups or organizations), obtains
names of individuals within those clusters, and then samples within
them.
Type of sampling
Identify and discuss the selection process for
participants in your sample. Ideally you aim to
draw a random sample, in which each
individual in the population has an equal
probability of being selected (a systematic or
probabilistic sample). But in many cases it may
be quite difficult (or impossible) to get a
random sample of participants.
Alternatively, a systematic sample can have
precision equivalent random sampling (Fowler,
2014). In this approach, you choose a random start
on a list and select every X numbered person on the
list. The X number is based on a fraction determined
by the number of people on the list and the number
that are to be selected on the list (e.g., 1 out of every
80th person). Finally, less desirable, but often used,
is a nonprobability sample (or convenience sample),
in which respondents are chosen based on their
convenience and availability.
Stratification
Identify whether the study will involve stratification
of the population before selecting the sample. This
requires that characteristics of the population
members be known so that the population can be
stratified first before selecting the sample (Fowler,
2014). Stratification means that specific
characteristics of individuals (e.g., gender—
females and males) are represented in the sample
and the sample reflects the true proportion in the
population of individuals with certain
characteristics.
When randomly selecting people from a
population, these characteristics may or
may not be present in the sample in the
same proportions as in the population;
stratification ensures their representation.
Also identify the characteristics used in
stratifying the population (e.g., gender,
income levels, education).
Within each stratum, identify
whether the sample contains
individuals with the characteristic
in the same proportion as the
characteristic appears in the entire
population.
Sample size determination

Indicate the number of people in the


sample and the procedures used to
compute this number. Sample size
determination is at its core a tradeoff: A
larger sample will provide more accuracy
in the inferences made, but recruiting more
participants is time consuming and costly.
In survey research, investigators
sometimes choose a sample size
based on selecting a fraction of the
population (say, 10%) or selecting a
sample size that is typical based on
past studies. These approaches are
not optimal; instead sample size
determination should be based on
your analysis plans (Fowler, 2014).
Instrumentation
Name the survey instruments used to collect data. Discuss
whether you used an instrument designed for this research, a
modified instrument, or an instrument developed by someone
else. For example, if you aim to measure perceptions of stress
over the last month, you could use the 10-item Perceived
Stress Scale (PSS) (Cohen, Kamarck, & Mermelstein, 1983) as
your stress perceptions instrument in your survey design. Many
survey instruments, including the PSS, can be acquired and
used for free as long as you cite the original source of the
instrument.
But in some cases, researchers have made the
use of their instruments proprietary, requiring a
fee for use. Instruments are increasingly being
delivered through a multitude of online survey
products now available (e.g., Qualtrics, Survey
Monkey). Although these products can be
costly, they also can be quite helpful for
accelerating and improving the survey research
process.
For example, researchers can create
their own surveys quickly using custom
templates and post them on websites
or e-mail them to participants to
complete. These software programs
facilitate data collection into organized
spreadsheets for data analysis,
reducing data entry errors and
accelerating hypothesis testing.
Validity of scores using the instrument. To
use an existing instrument, describe the
established validity of scores obtained
from past use of the instrument. This
means reporting efforts by authors to
establish validity in quantitative research—
whether you can draw meaningful and
useful inferences from scores on the
instruments.
The three traditional forms of validity to
look for are (a) content validity (Do the
items measure the content they were
intended to measure?), (b) predictive or
concurrent validity (Do scores predict a
criterion measure? Do results correlate
with other results?), and (c) construct
validity (Do items measure hypothetical
constructs or concepts?).
In more recent studies, construct validity has
become the overriding objective in validity, and it
has focused on whether the scores serve a useful
purpose and have positive consequences when
they are used in practice (Humbley & Zumbo,
1996). Establishing the validity of the scores in a
survey helps researchers to identify whether an
instrument might be a good one to use in survey
research. This form of validity is different from
identifying the threats to validity in experimental
research, as discussed later.
Reliability of scores on the instrument
Also mention whether scores resulting from past
use of the instrument demonstrate acceptable
reliability. Reliability in this context refers to the
consistency or repeatability of an instrument. The
most important form of reliability for multi-item
instruments is the instrument’s internal
consistency—which is the degree to which sets of
items on an instrument behave in the same way.
This is important because your instrument
scale items should be assessing the same
underlying construct, so these items should
have suitable intercorrelations. A scale’s
internal consistency is quantified by a
Cronbach’s alpha (α) value that ranges
between 0 and 1, with optimal values ranging
between .7 and .9.
For example, the 10-item PSS has
excellent internal consistency across
many published reports, with the
original source publication reporting
internal consistency values of α =.84–
.86 in three studies (Cohen, Kamarck,
and Mermelstein, 1983).
It can also be helpful to evaluate a second
form of instrument reliability, its test-retest
reliability. This form of reliability concerns
whether the scale is reasonably stable over
time with repeated administrations. When you
modify an instrument or combine instruments
in a study, the original validity and reliability
may not hold for the new instrument, and it
becomes important to establish validity and
reliability during data analysis.
Sample items

Include sample items from the instrument


so that readers can see the actual items
used. In an appendix to the proposal,
attach sample items or the entire
instrument (or instruments) used.
Content of instrument
Indicate the major content sections in the instrument,
such as the cover letter (Dillman, 2007, provides a useful
list of items to include in cover letters), the items (e.g.,
demographics, attitudinal items, behavioral items,
factual items), and the closing instructions. Also mention
the type of scales used to measure the items on the
instrument, such as continuous scales (e.g., strongly
agree to strongly disagree) and categorical scales (e.g.,
yes/no, rank from highest to lowest importance).
Pilot testing
Discuss plans for pilot testing or field-testing the survey and
provide a rationale for these plans. This testing is important to
establish the content validity of scores on an instrument; to
provide an initial evaluation of the internal consistency of the
items; and to improve questions, format, and instructions. Pilot
testing all study materials also provides an opportunity to
assess how long the study will take (and to identify potential
concerns with participant fatigue). Indicate the number of
people who will test the instrument and the plans to
incorporate their comments into final instrument revisions.
Administering the survey
For a mailed survey, identify steps for
administering the survey and for following up to
ensure a high response rate. Salant and Dillman
(1994) suggested a four-phase administration
process (see Dillman, 2007, for a similar three-
phase process). The first mail-out is a short
advance-notice letter to all members of the
sample, and the second mail-out is the actual mail
survey, distributed about 1 week after the advance-
notice letter.
The third mail-out consists of a postcard follow-up
sent to all members of the sample 4 to 8 days after
the initial questionnaire. The fourth mail-out, sent
to all nonrespondents, consists of a personalized
cover letter with a handwritten signature, the
questionnaire, and a preaddressed return envelope
with postage. Researchers send this fourth mail-
out 3 weeks after the second mail-out. Thus, in
total, the researcher concludes the administration
period 4 weeks after its start, providing the returns
meet project objectives.
Components of an Experimental Study
Method Plan
An experimental method plan follows a
standard form: (a) participants and design, (b)
procedure, and (c) measures. These three
sequential sections generally are sufficient
(often in studies with a few measures, the
procedure and measures sections are
combined into a single procedure section).
Participants
Readers need to know about the selection,
assignment, and number of participants who will
take part in the experiment. Consider the following
suggestions when writing the method section plan
for an experiment:
Describe the procedures for recruiting participants
to be in the study, and any selection processes
used.
Often investigators aim to recruit a
study sample that shares certain
characteristics by formally stating
specific inclusion and exclusion study
criteria when designing their study
(e.g., inclusion criterion: participants
must be English language speaking;
exclusion criterion: participants must
not be children under the age of 18).
Recruitment approaches are wide-
ranging, and can include random digit
dialing of households in a community,
posting study recruitment flyers or e-
mails to targeted communities, or
newspaper advertisements. Describe
the recruitment approaches that will be
used and the study compensation
provided for participating.
One of the principal features distinguishing
an experiment from a survey study design
is the use of random assignment. Random
assignment is a technique for placing
participants into study conditions of a
manipulated variable of interest. When
individuals are randomly assigned to
groups, the procedure is called a true
experiment.
If random assignment is used, discuss how
and when the study will randomly assign
individuals to treatment groups, which in
experimental studies are referred to as
levels of an independent variable. This
means that of the pool of participants,
Individual 1 goes to Group 1, Individual 2 to
Group 2, and so forth so that there is no
systematic bias in assigning the
individuals.
This procedure eliminates the
possibility of systematic differences
among characteristics of the
participants that could affect the
outcomes so that any differences in
outcomes can be attributed to the
study’s manipulated variable (or
variables) of interest (Keppel &
Wickens, 2003).
Often experimental studies may be interested in
both randomly assigning participants to levels of a
manipulated variable of interest (e.g., a new
treatment approach for teaching fractions to
children versus the traditional approach) while also
measuring a second predictor variable of interest
that cannot utilize random assignment (e.g.,
measuring whether the treatment benefits are
larger among female compared to male children; it
is impossible to randomly assign children to be
male or female).
Designs in which a researcher has only
partial (or no) control over randomly
assigning participants to levels of a
manipulated variable of interest are
called quasi-experiments.
At the end of the participants section, it is
helpful to provide a formal experimental
design statement that specifies the
independent variables and their
corresponding levels. For example, a formal
design statement might read, “The experiment
consisted of a one-way two-groups design
comparing burnout symptoms between full-
time and part-time nurses.”
Experimental Procedures
Identify the type of experimental design to be used
in the proposed study. The types available in
experiments are pre-experimental designs, quasi-
experiments, and true experiments. With pre-
experimental designs, the researcher studies a
single group and implements an intervention
during the experiment. This design does not have a
control group to compare with the experimental
group.
In quasi-experiments, the investigator uses
control and experimental groups, but the
design may have partial or total lack of
random assignment to groups. In a true
experiment, the investigator randomly
assigns the participants to treatment
groups. A single-subject design or N of 1
design involves observing the behavior of a
single individual (or a small number of
individuals) over time.
Identify what is being compared in the experiment.
In many experiments, those of a type called
between-subject designs, the investigator
compares two or more groups (Keppel & Wickens,
2003; Rosenthal & Rosnow, 1991). For example, a
factorial design experiment, a variation on the
between group design, involves using two or more
treatment variables to examine the independent
and simultaneous effects of these treatment
variables on an outcome (Vogt & Johnson, 2015).
This widely used experimental design explores
the effects of each treatment separately and
also the effects of variables used in
combination, thereby providing a rich and
revealing multidimensional view. In other
experiments, the researcher studies only one
group in what is called a within-group design.
For example, in a repeated measures design,
participants are assigned to different
treatments at different times during the
experiment.
Another example of a within-group design
would be a study of the behavior of a single
individual over time in which the
experimenter provides and withholds a
treatment at different times in the
experiment to determine its impact.
Finally, studies that include both a
between-subjects and a within-subjects
variable are called mixed designs.
Provide a diagram or a figure to illustrate the specific
research design to be used. A standard notation system
needs to be used in this figure. As a research tip, we
recommend using the classic notation system provided by
Campbell and Stanley (1963, p. 6):
• X represents an exposure of a group to an experimental
variable or event, the effects of which are to be measured.
• O represents an observation or measurement recorded
on an instrument.
• Xs and Os in a given row are applied to the same specific
persons. Xs and Os in the same column, or placed
vertically relative to each other, are simultaneous.
• The left-to-right dimension indicates the temporal
order of procedures in the experiment (sometimes
indicated with an arrow).
• The symbol R indicates random assignment.
• Separation of parallel rows by a horizontal line
indicates that comparison groups are not equal
(or equated) by random assignment. No
horizontal line between the groups displays
random assignment of individuals to treatment
groups.
One-Shot Case Study
This design involves an exposure of a group to a
treatment followed by a measure.
Group A X_____________________O
One-Group Pretest-Posttest Design
This design includes a pretest measure
followed by a treatment and a posttest for
a single group.
Group A O1————X————O2
Static Group Comparison or Posttest-Only With
Nonequivalent Groups
Experimenters use this design after
implementing a treatment. After the treatment,
the researcher selects a comparison group and
provides a posttest to both the experimental
group(s) and the comparison group(s).
Group A X______________________O
Group B _______________________O
Alternative Treatment Posttest-Only With
Nonequivalent Groups Design
This design uses the same procedure as
the Static Group Comparison, with the
exception that the nonequivalent
comparison group received a different
treatment.
Group A X1_____________________O
Group B X2_____________________O
Nonequivalent (Pretest and Posttest) Control-
Group Design
In this design, a popular approach to quasi-
experiments, the experimental Group A and the
control Group B are selected without random
assignment. Both groups take a pretest and
posttest. Only the experimental group receives the
treatment.
Group A O————X————O
___________________________
Group B O—————————O
Single-Group Interrupted Time-Series Design
In this design, the researcher records
measures for a single group both before and
after a treatment.
Group A O—O—O—O—X—O—O—O—O
Control-Group Interrupted Time-Series Design
This design is a modification of the Single-Group
Interrupted Time-Series design in which two groups
of participants, not randomly assigned, are observed
over time. A treatment is administered to only one of
the groups (i.e., Group A).
Group A O—O—O—O—X—O—O—O—O
__________________________________
Group B O—O—O—O—O—O—O—O—O
Pretest–Posttest Control-Group Design
A traditional, classical design, this procedure
involves random assignment of participants to
two groups. Both groups are administered both
a pretest and a posttest, but the treatment is
provided only to experimental Group A.
Group A R——–––O———X———O
Group B R———O———————O
Posttest-Only Control-Group Design
This design controls for any confounding
effects of a pretest and is a popular
experimental design. The participants are
randomly assigned to groups, a treatment is
given only to the experimental group, and both
groups are measured on the posttest.
Group A R——————X—————O
Group B R————————————O
Solomon Four-Group Design
A special case of a 2 × 2 factorial design, this
procedure involves the random assignment of
participants to four groups. Pretests and
treatments are varied for the four groups. All
groups receive a posttest.
Group A R————O———X———O
Group B R————O———————O
Group C R———————X————O
Group D R——————————–—O
A-B-A Single-Subject Design
This design involves multiple observations of a
single individual. The target behavior of a single
individual is established over time and is
referred to as a baseline behavior. The baseline
behavior is assessed, the treatment provided,
and then the treatment is withdrawn.
Baseline A Treatment B Baseline A
O–O–O–O–O–X–X–X–X–X–O–O–O–O–O–O
Threats to Validity
There are several threats to validity that will raise
questions about an experimenter’s ability to conclude
that the manipulated variable(s) of interest affect an
outcome and not some other factor. Experimental
researchers need to identify potential threats to the
internal validity of their experiments and design them so
that these threats will not likely arise or are minimized.
There are two types of threats to validity: (a) internal
threats and (b) external threats.
Internal validity threats are experimental procedures,
treatments, or experiences of the participants that threaten
the researcher’s ability to draw correct inferences from the
data about the population in an experiment. There are
those involving participants (i.e., history, maturation,
regression, selection, and mortality), those related to the
use of an experimental treatment that the researcher
manipulates (i.e., diffusion, compensatory and resentful
demoralization, and compensatory rivalry), and those
involving procedures used in the experiment (i.e., testing
and instruments).
External validity threats arise when experimenters draw
incorrect inferences from the sample data to other
persons, other settings, and past or future situations.
These threats arise because of the characteristics of
individuals selected for the sample, the uniqueness of
the setting, and the timing of the experiment. For
example, threats to external validity arise when the
researcher generalizes beyond the groups in the
experiment to other racial or social groups not under
study, to settings not examined, or to past or future
situations.
References
Creswell, J. W. & Creswell, J. D. (2022). Research design:
Qualitative, quantitative, and mixed methods approaches (6th ed.).
Thousand Oaks, CA: Sage.
Fraenkel, J. R., Wallen, N. E., & Hyun, H. E. (2023). How to design
and evaluate research in education (11th ed.). New York: McGraw-
Hill.

You might also like