Research and Statistics
Research and Statistics
1. Good research is empirical: primarily concerned with one or more aspects of a real world
situation and deals with concrete data that serves as a foundation for the external validity of
research findings.
2. Good research is replicable: research findings must be validated by replicating the study,
resulting in a solid foundation for decision making
Dimensions of research
Purpose of research
Principal objective or purpose of research in any field of inquiry is to add to what is known about the
phenomenon under the investigation through the application of scientific methods.
The 3 major purposes are exploring, explaining and describing any problem/event under study.
No control over the variables is intended. The research methods employed include
Descriptive observation, interviews, case study, introspection. The goal is to merely describe the
research phenomena under study and may involve categorization, frequencies and descriptive
statistics.
Finding the direction and strength of the association between the variables under
Correlational study. The manipulation of the variables is not possible as we study the covariance a hi
research nd not the causation. The goal is not to predict but to find the degree of association
between the dependent and the independent variable, it helps us to understand
characteristics and understand behavior
We manipulate the independent variable to study the effect on the dependent variable.
Experimental It helps us to analyse the relationship between the variables and explain changes and
research test a theory
Paradigms in Psychology
1. Essentially a worldview, a whole framework of beliefs, values and methods within which
research takes place.
2. Research philosophy combined with research methodology comprises research paradigm
3. Define what falls within and outside the limits of legitimate research and how members of
research communities view both the phenomena that they study and research methodology
that should be employed to study these phenomena
4. Another use of the word is in the form or worldview
5. The term is used to describe the set of experiences, beliefs and values that affect the way an
individual perceives reality and responds to that perception.
6. Kuhnian phrase: paradigm shift, to denote a change in how a given society goes about
organizing and understanding reality.
7. Word refers to pattern in greek
8. Concerned with epistemology, ontology, methodology and axiology
ontological Nature of our beliefs about reality. Concerns with the notion of what actually exists, the
nature of reality and what can be known about it. Comes from Greek ontos meaning
being. Relates to the nature of reality and its characteristics.
epistemological Concerned with the search for the foundations of human knowledge which can offer
some assuredness of the truth of our knowledge claims. Comes from the Greek root
episteme, meaning knowledge and can be understood as the study of knowledge itself.
This word root was originally applied to what greeks considered rigorous knowledge or
knowledge gained from studying the world
methodological Strategy, plan of action, process or design that informs one’s choice of research
methods. Guides the researcher in deciding what type of data is required for the study
and which data collection tools will be most apt for the purpose of his/her study.
Methods refers to the specific means of collecting and analysing data such as
questionnaires and open ended interviews. What methods to use for a research project
will depend on the design of that project and the researcher’s theoretical mindset.
axiological Refers to ethical issues that need to be considered when planning a research proposal.
Consider the philosophical approach to making decisions of value or the right
decisions. Involves defining concepts of right and wrong behavior relating to the
research. Consider what value we shall attribute to different aspects of our research,
participants, and the data.
Deductive Inductive
Deductive approach involves the search of Replaces the previously held major premise by
knowledge in a scientific way. hypothesis or assumptions.
The scientific enquiry that we seek today is Hypothesis is then tested by collection of data and
possible due to the two key contributions in logical analysis of data
establishing and popularizing the “the
scientific method’ of inquiry.
Based on syllogistic reasoning establishing a Start with a set of observations looking for patterns in
relationship between a major premise, minor observations and then they move from those specific
premise and a major conclusion experiences to a more general set of propositions about
those experiences.
Moves from general assumptions to the Moves from specific observations to generalizations
specific application
Starts with a theory, developing hypotheses Utilized to discover patterns and construct
from that theory and then collecting and generalisations and theories based on the specific
analysing data to test those hypotheses observations
Developed based on prior research and Study would usually begin with data and then try to find
theory pattern in it and then attempt to make a general theory
Operational definitions
1. Detailed explanation of the technical or ambiguous terms and measurements used in the
research process
2. Helps to better address subjective and ambiguous constructs in psychology when they are
used in scientific research
3. Avoiding confusion regarding the intended meaning, addressing the distortion of meaning due
to subjective experiences
4. The insistence that all abstract scientific terms must be operationally defined is called
operationalism or operationism
Research variables
1. A variable is a property that takes on different values as expectations and circumstances
change.
2. Types of variables
3. Independent variable are manipulated by the experimenter
4. Types of independent varible
5. Task variables: refer to characteristics which are associated with a behavioral task given to the
participant. Includes physical characteristics or complexity of the apparatus used in
experiment.
6. Environment variable: characteristics of env that tend to produce changes in DV
7. Subject variables: characteristics of subject that tend to produce DV. 2 major types are natural
subject (age, sex) and induced or instructional subject variable (induced by instructions given
by the researcher)
8. Outcome variable: DV that the researcher measured to observe any resulting change made by
the IV. also called response variable
9. Unmeasured variable: all the extraneous variables that are the focus of the research but they
affect the relationship under study.
10. 3 types: subject relevant (constitute the characteristics of the participant that affects the
researcher’s study in an undesired manner). For eg, age intelligence
11. Situational relevant: environmental and task variables whose undesired effect is controlled by
the research
12. Sequence relevant: occur due to different ordinal positions that conditions of the experiment
occupy in a sequence.
13. Connecting variable: intervening/mediator variables that are essential in completing the
relationship between the independent and dependent variable when a direct relationship
cannot be observed in certain cases.
14. Dependent variable is observed and recorded by the experimenter
15. It depends on the behaviour of the participant which is supposed to depend on the independent
variable
16. There are 2 levels of independent variable: experimental and control
17. Control/ extraneous variable
18. Potential independent variable that is held constant during an experiment
19. Qualitative vs quantitative variable
20. Quantitative are numbers
21. Qualitative are attitudes (good or bad). They can be compared but not measured
(Nominal/Ordinal).
22. Active vs Attribute Variables
23. Active Variables are those that can be manipulated, changed and controlled experimentally.
24. Attribute variables are those that cannot be manipulated, or controlled rather reflect the
characteristics of the study population. They are thus, the pre-existing qualities of the
population.
25. Continuous vs Discrete variables
26. The Continuous variables exist between a range, say from 30 to 40 the value can be 30.1 to
39.9. A continuous variable is one that may take on an infinite number of intermediate values
along a specified interval.
27. The discrete Variables are the absolute values. A discrete variable, restricted to certain values,
usually consists of whole numbers, such as the family size, number of defective items in a
box.
28. Can be divided into 3 types
29. Constant: when a variable can have only one value
30. Dichotomous: can have 2 values
31. Polytomous: can have many values.
32. Other types of variables
33. Confounding variables also known as a third variable affects the dependent variable despite
the fact that it is not the independent variable being studied. This can cause problems in a
study
34. Extraneous variable is any variable present in the experiment that may cause the relationship
between the independent and dependent variable to be weaker than expected or observed. Any
env clue that may push the participant to behave or act in a certain way is referred to as a
demand characteristic. Experimenter effect is any hint provided by the experimenter to
persuade or sway the results in some way. Situational variables are any variables
corresponding to the noise level, the temperature or anything else present in the situation.
35. To control extraneous variables, it is important to ensure elimination, noise can be eliminated
by using sound proof situations or settings. Constancy, by holding the extraneous values
constant in all situations. Balancing, participants are made equal in all aspects in both
controlled and experimental groups. Counterbalancing, used to control variables occurring as
a result of practice or fatigue together called as order effect. Randomization, each member of
the population, having an equal chance to be selected. This technique is applied where the
extraneous variables are known, but these effects can’t be controlled by known techniques.
36. Control variable is something that the researcher manipulates to keep it constant across
something that the researcher manipulates to keep it constant across conditions allowing the
results to be more homogenous and/or valid by preventing it from becoming confounded.
37. A moderator variable modifies the strength of the relationship between the 2 variables by
changing how much the independent variable influences the dependent variable.
38. Moderator variable could be anything related to a person’s categorical variables or quantitative
variables
39. Depending on the causation, variables are classified as:
40. Change variables: are the independent variables that are manipulated, measured and selected
by the research to produce some observable effect on dependent variables. Also called
stimulus variables. Alternatively be called explanatory, predictor, right hand side or X variable
NOTE:
1. When the env, experimenter, participant hinders the research objective, it becomes necessary
to control them:
2. Using deception: creating an artificial situation/story to disguise the procedures or objectives
of the study
3. Placebo effect: real reaction to a fake treatment
4. Controlling expectation of subject and experimenter: participants should not know which
treatment they are being given, or whether they are being given a placebo, single blind. In
double blind, both participant and experimenter are unaware of the specific conditions being
presented
Hypothesis
1. Speculation or theory based on insufficient evidence that lends itself to further testing and
experimentation.
2. A hypothesis can usually be prove true or false
3. The hypothesis is written in 2 forms
4. Null and alternative hypothesis
5. Null hypothesis is defined as the prediction that there is no interaction between variables. It
says that there is no statistical significance between the 2 variables in the hypothesis. It the
hypothesis that the researcher is trying to disprove or reject
6. Represents the status quo or the prevailing knowledge about a situation
7. Ho: p=1.
8. Null hypothesis must always contain equality.
9. In rare cases may also contain < and > signs
10. If a researcher is unable to disprove or reject the null hypothesis, then it means that the sample
data results are due to chance factors and are not significant in terms of supporting the idea
being investigated.
11. The null hypothesis contains the not equal sign. This indicates we are testing whether or not
there is some effect without specifying the direction of that effect.
12. Alternative hypothesis is inverse or opposite of the null hypothesis
13. States that the results are not due to chance and that they are significant in terms of supporting
the theory being investigated.
14. Strict inequality which is usually the suspicion or claim being tested
15. Also known as research hypothesis and sometimes directional hypothesis
16. Alternative hypothesis contains the less than or greater than signs which indicates whether or
not there is a positive or negative effect
17. Hypothesis testing is the procedure of comparing a null hypothesis and an alternative
hypothesis against each other to determine validity
18. The entire goal of our research now becomes to verify and reject the null hypothesis and in
turn accept the alternative hypothesis
Characteristics of hypothesis
testable Must be able to be tested using scientific methods so that data can be gathered
falsifiable Hypothesis must be able to be proven false. Data supporting the falseness of the
hypothesis must be able to be collected
logical Must be based on reasoning. If a theory already exists that is related to the
hypothesis, deductive reasoning can be used to generate the hypothesis so that it does
not disagree with the theory. If theory does not exist, inductive reasoning can be used
to design a hypothesis that is in agreement with general observations.
positive A hypothesis must be worded in such a way that it proposes the existence of a
relationship between the subjects of the study. Very rarely scientists set out to show
that there is no relationship.
Aim of a hypothesis
● Difference
● Relationship (statistically significant implies difference in result did not occur by chance)
● Interaction
Collection of data
● Universe or population is the set of individuals from which a statistical sample is drawn for a
study. Any selection of individuals grouped by a common characteristic can be said to be a
population.
● Population or the universe can be finite or infinite
● The population is said to be finite if it is made up of a fixed number of elements so it is
possible to list it in its entirety.
● N used to indicate how many elements or items there are in the case of a finite population
● Sample is the subset of the population that is representative of the entire population
● Sources of data can be:
● primary source in which one itself collects the data. Gathered first hand. Quite expensive.
Data collection is under direct control and supervision of the investigator.
● More reliable and accurate
● secondary data is the one that makes available data that were already collected by some other
agency
● Collected from other published and unpublished sources like census, govt publications,
internal record, reports, books, websites. Easily available, saves time and cost. Usefulness of
the data may be limited in a number of ways like relevance and accuracy.
Techniques of data collection
● 2 types of techniques used for collecting data
● Census technique: one where data is collected from each and every member of the population.
Data is obtained from each member of the population. More representative, accurate and
reliable
● Sample technique: only a part of the population is studied and conclusions are drawn on that
basis for the entire population.
Sampling methods
● Process of choosing or selecting the subjects of the research study that consists of persons,
events, objects or behaviors.
● Group of subjects in the study
● Sampling is the process by which the researcher chooses his or her sample
● Involves 3 elements
● Selecting the sample
● Collecting the info
● Making an inference about the population
● Values obtained from the study of samples such as average and dispersion are known as
statistics.
● Characteristics of a good sample are:
● Representativeness: the sample selected for the research must be best representative of the
population under study.
● Accurate or unbiased: a sample is free from any influence that causes any differences or
doubts on their true representativeness
● Adequacy: size must be adequate in order for it to truly represent the entire population
● Independence: all members of the sample must be chosen independently of one another and
each member should have an equal chance of being selected in the sample
● Homogeneity: there is complete similarity in the nature of the universe and that of the sample.
If 2 samples from the same population are taken they should yield similar results.
● Size of sample means the number of sampling units selected from the population for
investigation
● If sample size is large, it can be burdensome financially and might require more time.
● If the sample is too small, it might not be adequate enough to rightly represent the population.
● If the population consists of homogenous members, then small sample may serve the purpose
● Larger samples reduce error or the difference between the sample and the population.
● Randomly chosen samples usually require fewer participants
● Real world issues that influence the actual size of a sample, including time, cost and practical
considerations.
● There are 2 types of sampling: probability and non probability
Probability non-probability
Every unit in the population has a chance Often used when time and money are limited or
(non-zero probability) of being selected in the more information is needed about a particular
sample and this chance can be accurately population
determined
4 basic types: simple random, stratified random, Frequently utilised in qualitative or exploratory
systematic random and cluster random. research to gather data and form a hypothesis
Equal probability of being selected. Avoids bias Types of non probability: convenience, quota,
in the overall choice. Simple random involves snowball.
randomly selecting respondents from a
sampling frame but with large sampling frames,
a table of random or computerized random
number generators is used.
Stratified sampling: stratification is the process Convenience sampling is also called accidental
of classifying sampling units of the population or opportunity sampling, in which sample is
into homogeneous units. The sampling frame is drawn from that part of the population that is
divided into homogeneous and non-overlapping close to hand, readily available or convenient.
subgroups called strata.
Cluster sampling: if the population dispersed Snowball is when you start identifying a few
over a wide geographic region. Large respondents that match the criteria for inclusion
populations divide into smaller groups known as in your study and then ask them to recommend
clusters and then select randomly among the others they know who also meet your selection
clusters to form a sample. Target population is criteria. This hardly leads to representative
too large or spread out and studying each would samples, but sometimes the only way to reach
be costly, time consuming and improbable hard to reach population or when no sampling
frame is available
Methods of research
Method 1: observation
● Act of meticulously viewing another’s interaction with his or her surroundings
● Supreme technique for nonverbal behavior
● Collecting facts that are direct knowledge of the investigator
● Perception with the purpose aka, regulated perception
● Procedure of observation: by mechanical or electronic device (audio or visual recording),
checklist and schedules (objectifies the observation, a score will be provided which will
facilitate comparative analysis), time sampling (certain behavior occurs during a sample of
short time intervals), event sampling (records all instances of a particular event or behavior
during specific time period ignoring all other behavior), specimen sampling (Researcher
record the description of the subject’s entire scheme of behavior for a specific period)
● Types of observation
● With respect to role of investigator, participant and non participant
● With respect to the method of observation, direct, indirect
● With reference to control on the system to be observed, controlled or uncontrolled
● You can conduct observational research ranging from a complete observer, when the
researcher is not seen or noticed by the participants as a detached observer. Reduces
Hawthorne effect because participants are more likely to act naturally when they don’t realise
they are being watched.
● Observer as participant, the researcher is known and recognized by the participants and the
participants are aware of the observer’s research goals.
● There is some interaction with the participants but it is brief. The researcher’s goal is to
maintain as much neutrality as possible.
● Participant as observer: the researcher is fully engaged with the participants in this situation.
Participants aware that he is a researcher
● Complete participant : full embedded researcher almost if he or she were a spy. Fully interact
with the researcher but unaware that they are being observed and studied.
● When performing indigenous fieldwork it is sometimes referred to as going native
Method 2 : Interviews
● Involve social interaction
● Need training in how to interview
● Interview schedule, set of prepared ques designed to be asked exactly as worded.
● Have a standardized format which means the same ques are asked to each interviewee in the
same order
● Interviews will be recorded by the researchers and the data is written up as a transcript which
can be analyzed at a later date
● Language and special care must be undertaken
● Types of interviews
● Structured interview: also known as formal and ques are asked in a set/standardized order and
the interviewer will not deviate from the interview schedule or probe beyond the answers
received. Based on structures, close ended ques. Easy to replicate as a fixed set of closed
questions are used. Quick. Not flexible. Lack detail
● Unstructured: discovery interviews and are more like guided conservation. Informal. Interview
schedules might not be used and even if used they will contain open ended ques that can be
asked in any order. Added or missed questions as the interview progresses. More flexible.
Generate qualitative data with open ended ques. Increased validity. Probes deeper
understanding.time consuming. Expensive and extensive
● Group interviews, dozens are interviewed altogether also known as focus groups. Group
interacts with each other and does not drift off topic. Highly skilled. Generate qualitative data.
Have increased validity because some participants may feel more comfortable being with
others as they are used to talking in groups in real life. Details are confidential and respect
their privacy. Less reliable and may sometimes lack validity as they conform to peer pressure
and give false answers.
Method 3: Questionnaires
● Only one opportunity to collect data from each informant
● Question needs to be clear, comprehensive and effective
● Closed ques also called mcq, straight forward, quick to naswer and lie within intellectual
range of majority of population
● Checklists, questions, graded response questions, open ended questions
● Drawback of fixed alternative ques is putting answers in people’s mouth
● Advisable to conduct a pilot study for testing the questionnaires due to their limitations in true
representation.
Descriptive
● Descriptive statistics refers to various statistical calculation that are used to describe a data set
as it appears
● It performs 2 operations: organising data and summarising data
● 4 major techniques involve classification, tabulation, graphical presentation and diagrammatic
presentation
● 2 techniques for summarising: measures of tendency and dispersion
● We make use of 4 types of descriptive stats
● Measures of frequency (concerned with how many items are there in data sets,; frequency,
counts and relative frequency)
● Central tendency (mean, median, mode , dispersion (range) and measures of position
(percentile rank and quartile rank)
Measures of central tendency
● Refer to middle or average of a data set
Mean
● Average of a data set
● Calculatef by taking sum of the data dividing by sum of the size of data set.
● Rigdily defined, easy to calculate and simple, representative, can be computed even if detailed
distribution is not knwn and is least affected by fluctuation of sampling
● Cannot be determined by visual observation, cannot be computed for open ended CI, cannot
be computed for qualitative and too much affected by extreme observations.
Median
● Middle value
● Not the middle point but the point that divides distribution in half.
● Calculated differently when even
● Free from effect of extreme values and preferred for extreme values or outliers
● Real value and is better representative value of the series
● Can be estimated through graphic presentation of data
● Calculated even in the case of open ended classes (incomplete series)
● Not based on all the items in the series
Mode
● Value of the variable which occurs most frequently
● Simple and popular
● Less affected by marginal values and ignores them.
● Located graphically using histogram
● Best representative value
● Does not require knowlege of all items and frequencies
● Difficult to identify when all items are identical
● Involves cumbersome procedure of grouping the data
● Is not representative of all the items in a series.
Mean and median equal: same in any set where the terms are consecutive or equally paced. Mean
skews in the direction that the set spreads out. If the terms greater than the median are more spread
out than those less than the median, the mean is greater than the median. Inverse also true.
Outlier
● Value deviates from all other values in a data set
Measures of dispersion
● Also called measures of variability
● Tell us about how scores are arranged in relation to the centre
● Dispersion refers to how the data is spread out, how widely or narrowly is it scattered on a
plot or how much variability is present in the data points when compared to the mean or
average value of all data points
● The dispersion tells us about the vairability or distance between the data points and the avg
value of data set.
● If there is too much difference between the distances between each data point and the avg
value then the data can be considered is volatile or unstable
Range (R )
● difference between the largest and the smallest value in the data set
● Also known as distance between highest and lowest value in the data set
● Good measure if the data at ordinal level
● Not a good measure if the distribution is highly skewed
Interquartile range
● Best way to determine the consistency of scores within the data set is to identify the quartiles
in the data
● A quartile is a one quarter of the data
● It is possible to see the distribution of scores within a data set more clearly
● Difference between the upper and lower quartile is known as interquartile range
Standard deviation
● Standard means typical or average and deviation means the difference between the score and
their mean
● Stable index of variability since signs are considered because square of deviations is taken
● Also known as root mean square deviation
● If data points are further from the mean, there is higher deviation within the data det, the more
spread out the data, the higher the SD
● The wider the curve’s width, the larger the data set’s standard deviation from the mean
● All scores identical in the sample, SD is 0
● Addition or subtraction from each individual score does not change SD
● Multiplication or division of each score by constant number produced identical change in SD
● Least affected by fluctuations of sampling
● When we pick small sample from large population, there is an exclusion of many extreme
values which will naturally lower SD
Bessel’s correction
● When sample variance is used to estimate population variance then it has a denominator of
n-1, known as bessel’s correction. Sample variance will be less precise than population and -1
corrects this biasness
● Can be used with highly small sample although correction will have no impact
● Excessively large sample. Corrections will have no meaning
● When you have large sample size and want to approximate the population mean
● When you only need to find sample mean and not generalize with population, you can omit
the correction.
T and z scores
1. When we do not known the SD of the distribution, the student’s t distribution or t-distribution
is a hypothetical distribution used to calculate population parameters when the sample size is
small and when population variance is unknown
2. Much like normal distribution but has fatter traits
3. Used when sample size is 30 or less than 30
4. Population SD is unknown
5. Population distribution must be unimodal and skewed
6. The variable in t-distribution ranges from -infinity to +infinity
7. Less peaked than normal distribution at the centre and higher peaked in the tails.
8. Degree of freedom is n-1
9. William Sealy Gosset in 1908 developed t-test and t-distribution
10. He found major techniques for large samples were not useful when used in case of small
samples
11. Published his findings under the pen name “student”, student’s t-test
12. Degrees of freedom depend on the sample, the bigger the sample, the closer we get to the
actual numbers. As the sample size and so does the degrees of freedom increases, the t
distribution approaches (appears to be) the bell shape of the standard normal distribution
Inferential statistics
1. Used to draw conclusions and make predictions about an entire population based on the data
from a representative sample
2. Allows to make inferences beyond the data set collected from the sample
3. Deduces whether a give data sample is similar to the population or not
4. Makes use of random samples for testing and allows us to have confidence that the sample
represents the population
5. Goal is to make generalization about a population
6. Hypothesis testing, confidence intervals, correlation and regression analysis. Statistical tests
like t-tests, ANOVA and ANCOVA provide additional info about data collected for inferential
analysis
7. The 1st basic type is called t-test. Used to compare the average scores between 2 different
groups in a study to see if the groups are different from each other.
8. the 2nd basic type is called an analysis of variance. Use the nickname ANOVA for this test.
Compares the average scores between 3 or more different groups in a study to see if the
groups are different from each other.
9. Bonferroni correction states that if one is testing ‘n’ independent hypotheses, one should use a
significant level of 0.05/n.
10. Thus if there are 2 independent hypotheses result would be declared significant only if P<
0.025
Parametrics and non parametrics
● Inferential statistics fall into 2 possible categorizations: parametric and non-parametric.
● Parametric tests rely on the assumption that the data you are testing resembles a particular
distribution (often a normal or bell shaped distribution).
● Non parametric refer to distribution free tests because there are not strict assumptions to check
in regards to the distribution of the data
● As a general rule of thumb, when the dependent variable’s level of measurement is nominal or
ordinal, then a non-parametric test should be selected
● When the dependent variable is measured on a continuous scale, then the parametric test
should be selected
● Common parametric tests: Kolmogorov Smirnov test (KS test), shapiro Wilk test, anderson
darling test
Hypothesis test.
Produce high quality actionable data. Less precise but much easier to facilitate
Student’s t-test
● Was developed by Prof W.S Gossett in 1908
● Published statistical papers under the pen name of student
● This test is used when the samples are small and population variances are unknown
● T-test compares the difference between the 2 means of different (independent) groups to
determine whether the difference is statistically significant
● Also used to compare population mean and sample mean
● Used when samples are small and Population variance are not known
● Test makes various assumptions: samples are randomly selected, data utilised is quantitative,
variable follow normal distribution, sample variance are mostly same in both the groups under
study and samples are small and mostly lower than 30
● Used for different purposes giving out tests called
● one sample: if there is a group being compared against any standard value. df= n-1
● 2 sample: if the groups are coming from 2 different populations, also known as independent
t-test. Df =( n-1) (n-1)
● paired t-test: used to determine whether the mean difference between two dependent (or
paired) groups is statistically significant. df= n-1
● t= Xbar- assumed mean/ standard error of mean
ANOVA
● When we need to compare more than 2 groups
● Given by Sir Ronald Fisher
● Explain the variation in measurements
● Involves a test of significance of the difference in mean values of the variable between 2
groups
● If there are more than 2 groups, ANOVA is used
● Assumptions: Sample population can be easily proximate to normal distribution, all
populations have same SD, individuals in population are selected randomly, independent
samples
● ANOVA compares variance by means of a simple ratio called F-ratio. Measured as variance
between groups/variance within groups
● Resulting F stats is then compared with the critical value of F obtained from F tables as done
with t
● If the calculated value exceeds the critical value for the apt level of alpha, the null hypothesis
will be rejected.
● F= ratio of variances
● F tests can also be used independently of the ANOVA technique to test hypothesis of
variances
2 types of ANOVA
Various experimental groups differ in terms of If the various groups differ in terms of 2 or more
only factor at a time factors at a time
A study used to assess effectiveness of 4 diff A study used to assess 4 diff antibiotics on
antibiotics on younger adults adults in three diff age groups
Z-test
● Used for testing significance difference between 2 means
● Compares sample mean with population mean
● Compares 2 sample means
● Compares sample proportion with population proportion
● Compare 2 sample proportions
● Sample must be random and quantitative
● Should be larger than 30
● Follow normal distribution
● Sample variances should almost be same in both groups of the study
● If SD is known, a z-test can be applied even if sample is smaller than 30
● One tailed and two tailed z -tests
● A result larger than difference between sample mean will give +z and smaller than difference
between mean will give -z
Z proportionality test
● Used for testing significant differences between 2 proportions
MANOVA
● Simply an ANOVA with several dependent variables
● Should represent continous measures
● Moderately correlated
Factorial DOE with one factor and one blocking Friedman test
variable
Kruskal Wallis
● Use this test instead of a one way ANOVA to find out if 2 or more medians are different
● Ranks of data points used instead of data points themselves
● Your variables should have one independent variable with 2 or more levels, the test is more
commonly used when you have 3 or more levels
● Ordinal scale, ratio scale or interval scale dependent variables
● Independent observations.
● All groups should have same shape distribution
Freidman test
● Used for differences between groups with ordinal dependent variables
● Can also be used for continuous data if the one way ANOVA with repeated measures in apt
● Finding differences in treatment across multiple attempts
● Extension of sign test
● Used when there are multiple treatments
● Null hypothesis: All the treatments have identical effects or that the samples differ in some
way
● Alternative hypothesis: treatments do have diff effects
Power of a test
● Ability to detect an effect if there is one present
● Reduces the chance of type 2 error
● More power, more we are likely to detect an effect if there
● Probability of correctly rejecting null hypothesis if it is false
● Power = 1- beta = 1-P (type 2 error)
● Test with highest power considered best
● Opposite of level of significance (probability of rejecting null when true)
● Commonly done using G power software
Effect size
● Objective and standardized measure of the size
● How much impact our effect has on our test population
● Larger the size, more is the effect
● Effect sizes of different hypothesis can be used to determine which gives greatest effect size or
least effective size
● The larger the size the stronger the relationship between 2 variables since it is quantitative
measure of the magnitude of the experimenter effect
● Helps us determine if difference is real or due to chance factor.
● To understand the strength of the difference between 2 groups, a researcher needs to calculate
the effect size
● Effect size calculates the power of a relationship amongst the variables given on the numeric
scale.
● 3 ways to measure the effect size:
● Odd ratio
● The standardized mean difference
● Correlation coefficient
● Calculated by dividing the difference between the mean of 2 variables with the standard
deviation
● The larger the effect size, the more imp the effect
● The more imp the effect, the more easily it can be seen by just looking
● Because the effect size is an objective and standardized way of measuring effect, we can use it
to compare different hypothesis tests to each other.
Degrees of freedom
● 1st appeared in the works of Carl Fredrich Gauss in early 1821
● Defined and popularized by William Sealy Gosset in 1908 and Ronald Fisher in 1922
● Defines the number of values in a dataset having the freedom to vary.
● Estimates parameters in statistical analysis or finds the missing or unknown value when
making the final calculation
● Equals sample size minus the number of parameters or relationships
● df=n-P
● If mean of a set of scores is fixed then df is one less than the number of scores df=n-1
● Df for 2 sample t test= (n1+ n2 )-2
● For paired t test is n-1
● For f test complicated, calculates variance within groups and between groups
● For between groups = p (total number of groups ) -1
● Within groups= subtracting total number of people in all groups by number of groups, N-P
● Chi square: rows and column are used to calculate df = (number of C-1) (number of rows-1)
● ANOVA: N-k, N is the data sample size and k is the number of cell means, groups or
conditions.
Regression analysis
● Predicts one variable from another
Key Terminologies
1. Factor loading:
● Factor loading is basically the correlation coefficient for the variable and factor. Factor
loadings are merely correlation coefficients.
● Hence they range from - 1.0 through .0 to +1.0. Factor loading shows the variance explained
by the variable on that particular factor.
● In the SEM approach, as a rule of thumb, 0.7 or higher factor loading represents that the factor
extracts sufficient variance from that variable.
2. Eigenvalues:
● When factor analysis is going to generate the factors, each and every factor has ab associated
eigenvalue which will give the total variance explained by each factor.
● Usually, the factors having eigenvalues greater than 1 are useful. Eigenvalues show variance
explained by that particular factor out of the total variance.
● From the commonality column, we can know how much variance is explained by the first
factor out of the total variance.
● If eigenvalues are greater than zero, then it's a good sign.
● Since variance cannot be negative, negative eigenvalues imply the model is ill-conditioned.
● Eigenvalues close to zero imply there is item multicollinearity, since all the variance can be
taken up by the first component.
● The sum of eigenvalues for all the components is the total variance.
Types of Factoring
1. Principal Component Analysis (PCA):
● It is a statistical procedure that uses an orthogonal transformation that converts a set of
correlated variables to a set of uncorrelated variables.
● PCA is the most widely used tool in exploratory data analysis and in machine learning for
predictive models.
● It is a dimensionality-reduction method that is often used to reduce the dimensionality of large
data sets, by transforming a large set of variables into a smaller one that still contains most of
the information in the large set.
● Principal components are new variables that are constructed as linear combinations or mixture
of the initial variables
● These combinations are uncorrelated and most of the info within the initial variables is
squeezed or compressed into the 1st components
Cohort studies
● used to investigate the causes of disease and to establish links between risk factors and health
outcomes.
● The word cohort means a group of people.
● These types of studies look at groups of people.
● They can be forward-looking (prospective) or backward-looking (retrospective).
● Prospective studies are planned in advance and carried out over a future period of time.
Retrospective cohort studies look at data that already exist and try to identify risk factors for
conditions. Interpretations are limited because the researchers cannot go back and gather
missing data.
● These long-term studies are sometimes called longitudinal studies.
● In a prospective cohort study, researchers raise a question and form a hypothesis about what
might cause a disease.
● Then they observe a group of people, known as the cohort, over a period of time. This may
take several years. They collect data that may be relevant to the disease. In this way, they aim
to detect any changes in health linked to the possible risk factors they have identified.
● Cohort studies are also good at finding relationships between health and environmental factors
such as chemicals in the air, water, and food. These are issues that the World Health
Organization (WHO) helps researchers to investigate with large-scale cohort studies.
● Cohort studies are graded as the most robust form of medical research after experiments such
as randomized controlled trials, but they are not always the best form of observational work.
● They are less suited to finding clues about rare diseases.
● Typically unsuitable for identifying the causes of a sudden outbreak of disease
● They are expensive to run and usually take many years, often to produce results
● They can only offer clues about the causes of disease, rather than definitive proof of links
between risk factors and health. This is true of any observational medical research.
● Participants may leave the cohort, perhaps move away, lose touch, or die from a cause that is
not being studied. This can bias the results
Time Series
● Technique for analyzing time series data, or variables that continually change with time. This
technique is popular in econometrics, mathematical finance, and signal processing.
● Special techniques are used to correct for autocorrelation, or correlation within values of the
same variable across time.
● appropriate for longitudinal research designs that involve single subjects or research units that
are measured repeatedly at regular intervals over time.
● TSA can provide an understanding of the underlying naturalistic process and the pattern of
change over time, or it can evaluate the effects of either a planned or unplanned intervention.
● According to Daniel T. Kaplan and Leon Glass (1995), there are two critical features of a time
series that differentiate it from cross-sectional data-collection procedures:
● Repeated measurements of a given behavior are taken across time at equally spaced intervals.
Taking multiple measurements is following characteristics: essential for understanding how
any given behavior unfolds over time, and doing so at equal intervals affords a clear
investigation of how the dynamics of that behavior manifest at distinct time scales.
● The temporal ordering of measurements is preserved. Doing so is the only way to fully
examine the dynamics governing a particular process. If we expect that a given stimulus will
influence the development of a behavior in a particular way, utilizing summary statistics will
completely ignore the temporal ordering of the data and likely occlude one's view of
important behavioral dynamics.
● If the future is expected to be similar to the past, time series analysis can be useful in
projecting future events.
● Historical data analysis is likely to yield three distinct curves: trend, cyclical, and seasonal.
● A trend is defined as a consistent shift in data in one direction or the other.
● A cycle will begin with an increase, followed by a decrease, and then resume with an increase.
A cycle has a tendency to repeat itself on a regular basis. Seasonal variations are also cycles,
but they are limited to a single season of the year.
ANOVA
● Locating differences between multiple levels of a single inpdendent group mean.
● Calculate F-ratio defined as the score to determine level of difference between the means.
● Has many forms
● One way has one independent variable with multiple levels and one dependent variable
● It has 2 flavors
● Between subjects, subjects are placed in mutually exclusive groups and will be compared to
each other
● Repeated measures defined as the study that uses the same group of participants for each level
of the variable.
● With repeated measure you may get a carryover effect
● Carryover effect is defined as previous levels or conditions that may cause subsequent
Two-way ANOVA has two independent variables, and three-way ANOVA has three
independent variables. Although the number of independent variables is unlimited, factorial
ANOVA is limited to one dependent variable.
● Covariance designs: Sometimes, measures of dependent variables may be influenced by
extraneous variables called covariates.
● Covariates are those variables that are not of central interest to an experimental study, but
should nevertheless be controlled in an experimental design in order to eliminate their
potential effect on the dependent variable and therefore allow for a more accurate detection of
the effects of the independent variables of interest.
MANOVA
● Multivariate analysis of variance (MANOVA) is simply an ANOVA with several dependent
variables.
● are appropriate when multiple dependent variables are included in the analysis.
● The dependent variables should represent continuous measures (i.e., interval or ratio data).
● Dependent variables should be moderately correlated.
● If there is no correlation at all, MANOVA offers no improvement over an analysis of variance
(ANOVA)
● if the variables are highly correlated, the same variable may be measured more than once.
● In many MANOVA situations, multiple independent variables, called factors, with multiple
levels are included.
● The independent variables should be categorical (qualitative).
● Unlike ANOVA procedures that analyze differences across two or more groups on one
dependent variable, MANOVA procedures analyze differences across two or more groups on
two or more dependent variables.
● Advantages:
● Normal distribution, linearity, homogeneity of variances and covariances
● Limitations: Outliers - MANOVA is extremely sensitive to outliers. Multicollinearity and
Singularity - When there is high correlation between dependent variables, one dependent
variable becomes a near-linear combination of the other dependent variables.
Longitudinal
● Individual or one group over a continued period of time to observe the effect of time
● Good causal relationship
● Growth increments and patterns
● Typically takes a much longer period of time to get results
● Expensive and difficult to track
● Fewer participants required
● Controlled group differences and cohort effects
● Confounded by variables that they might experience
● Highly prone to carryover effects
Cross sectional
● Between subject quasi experimental
● Researcher observes the subject at different ages or at diff points in temporal sequence
● Immediate snapshot comparison of subjects
● Conducted quicker and cost efficiently
● No demand to observe participants for a continued long term duration
● Easier and cost efficient way to collect way since all of the data is collected at once
● Cohort effect
● Mutiple variables can be studied
● Easier to control
● Often requires more participants
● More time consuming
Cross sequential
● Mix of cross sectional and longitudinal designs
● Involves multiple groups multiple times over a set time period
● Much more complicated, expensive and time consuming
● Rarely used
Ex post facto
● After the fact
● Conducting the study after the event had occurred
● Heart attack
● No direct control on the manipulation of IV since they have already occurred and random
assignment of participants is not possible
Meta analysis
● A review is searching for databases to find existing literature on a topic
● Meta analysis goes one step further and uses data from existing studies to derive conclusions
about the body of research
● Each study is statistically weighted depending on the number of subjects and number of
variables
● Studies on the same treatment can produce contradictory results
Latin square
● A Latin square design is a type of experimental design used in research, particularly in
situations where controlling for multiple sources of variability is important.
● It's a method for arranging experimental units in a manner that ensures each treatment
condition appears exactly once in each row and each column of a square grid.
Spss
● Statistical package for social science
● Now change to statistical product and service solutions
● Original launched in 1968
● By SPSS inc.later acquired by IBM international business machine corporation in oc 2009
● 2 types of variance available in SPSS
● Variable view: in the form of variables
● Data view : rows and column
Important points
● Normal distribution table is the z score table
● Chi square is calculated for goodness of fit that is to see whether data from a sample matches
the population from which the data was taken
● Expected frequency is the probability count that appears in contingency table calculations
including the chi square test
● Observed frequency are the counts made from the experimental data
● Goodness of fit test include chi square, Kolmogorov smirnoff and shapiro wilk
● 1st uses medium to large sample sizes. Discrete distributions are not allowed. 2nd is used with
a small population and its purpose is to test the normality of a random sample.