0% found this document useful (0 votes)
28 views21 pages

1305 69038KPReliability

Reliability of test and its factors

Uploaded by

Braq Javed
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF or read online on Scribd
0% found this document useful (0 votes)
28 views21 pages

1305 69038KPReliability

Reliability of test and its factors

Uploaded by

Braq Javed
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF or read online on Scribd
You are on page 1/ 21
Reliability and Affecting Factors By Dr. Kshama Pandey Associate Professor Reliability |) | the specific day and time of the test (as compared with other possible days and times of testing), Reliability is the extent to | which test scores are not the specific questions or problems that were affected by chance |on the edition of the test that the test taker factors—by the luck of the {took (as compared with those on other draw. It is the extent to __ |editions), and which the test taker’s score — | does not depend on: | the specific raters who rated the test taker’s ! responses (if the scoring process involved [i lany judgment). = Reliability refers to the consistency of a measure and Psychologists consider three types of consistency: time (test-retest reliability), items (internal consistency), ‘ent researchers (inter-rater reliability). Types of consistency in Measuring reliability When researchers measure a construct that they assume to be consistent across time, then the scores they obtain should also be consistent across time. Test-retest reliability is the extent to which this is actually the case. For example, intelligence is generally thought to be consistent across time. Test-Retest Reliability ‘| A person who is highly intelligent today will be highly intelligent next week. This means that any good measure of intelligence should produce roughly the same scores for this individual next week as it does today. scores over time cannot be a very good measure of a construct that is supposed to be consistent. Clearly, a measure that produces highly inconsistent ‘ Interrater Reliability * Many behavioural measures involve significant judgment on the part of an observer or a rater. * Inter-rater reliability is the extent to which different observers are consistent in their judgments. * For example, if you were interested in measuring university students’ social skills, you could make video recordings of them as they interacted with another student whom they are meeting for the first time. * Then you could have two or more observers watch the videos and rate each student’s level of social skills. ff ° interrater Reliability To the extent that each participant does in fact have some level of social skills that can be detected by an attentive observer, different observers’ ratings should be highly correlated with each other. Inter-rater reliability would also have been measured in Bandura’s Bobo doll study. Tn this case, the observers’ ratings of how many acts of aggression a particular child committed while playing with the Bobo doll should have been highly positively correlated. Interrater reliability is often assessed using Cronbach’s a when the judgments are quantitative or an analogous statistic called Cohen’s « when they are categorical. Factors Influencing the Reliability of Test Scores B. ™ Factors * Group variability A) Intrinsic Factors: + Length of the Test * Homogeneity of Items + Difficulty Value of Items + Discriminative Value + Test instructions + Item selection = Reliability of the scorer * Guessing and chance errors * Environmental conditions + Momentary fluctuations A) Intrinsic Factors: The principal intrinsic factors (i.e. those factors which lie within the test itself) which affect the reliability are: Length of the Test | + Reliability has a definite relation with the length of the test . * The more the number of items the test contains, the greater will be its reliability and vice-versa. + However, it is difficult to ensure the maximum length of the test to ensure an appropriate value of reliability. = > Re The number of times a test should be lengthened to get a desirable level of Teliability is given by the formula: = m(l-ri) ri(l ta) Where 4, = the desired reliability r, = the obtained reliability and n = number of times a test is to be lengthened. n Homogeneity of Items * Homogeneity of items has two aspects: Item reliability and the homogeneity of traits measured from one item to another. + If the items measure different functions and the inter-correlations of items are ‘zero’ or near to it, then the reliability is ‘zero’ or very low and vice-versa. Difficulty Value of Items The difficulty level and clarity of expression of a test item also affect the reliability of test scores. If the test items are too easy or too difficult for the group members, it will tend to produce scores of low reliability. Because both the tests | have a restricted spread of scores. = ~~! Discriminative Value When items can discriminate well between superior and inferior, the item total-correlation is high, the reliability is also likely to be high and vice-versa. — =_— ™~! Test instructions Clear and concise instructions increase reliability. Complicated and ambiguous directions give rise to difficulties in understanding the questions and the nature of the response expected from the teste ultimately leading to low reliability. Ttem selection If there are too many interdependent items in a test, the reliability is found to be low. — Reliability of the scorer The reliability of the scorer also influences reliability of the test. If he is moody, fluctuating type, the scores will vary from one situation to another. Mistake in him give rises to mistake in the score and thus leads to reliability. — Extrinsic Factors The important extrinsic factors (i.e. the factors which remain outside the test itself) influencing the reliability are: When the group of pupils being tested is homogeneous in ability, the reliability of the test scores is likely to be lowered and vice-versa. Guessing and chance errors Guessing in test gives rise to increased error variance and as such reduces reliability. For example, in two-alternative response options there is a 50% chance of answering the items correctly in terms of guessing. Environmental conditions As far as practicable, testing environment should be uniform. Arrangement should be such that light, sound, and other comforts should be equal to all testees, otherwise it will affect the reliability of the test scores. Momentary fluctuations Momentary fluctuations may raise or lower the reliability of the test scores. ’ Broken pencil, momentary distraction by sudden sound of a train running outside, anxiety regarding non-completion of home-work, mistake in giving the answer and knowing no way to change it are the factors which may affect the reliability of test score

You might also like