Categorías: Todo - performance - correlation - reliability - validity

por Kristine Komorowski hace 12 años

1066

Validity and Reliability

The text discusses various aspects of learning and assessment, focusing on reliability and validity methods used to evaluate the effectiveness of tests. Reliability is explored through different methods such as internal consistency, split-half method, Kuder-Richardson methods, test-retest or stability, and alternative forms or equivalence.

Validity and Reliability

Components of Learning and Assessment

Reliability

Internal Consistency... refers to similar test items correlated with each other that result in answers that correlate with each other. This method is best used when the exam is measuring one consistent concept.
Kuder-Richardson Methods... measure the correspondence between multiple tests of different forms. The strength of reliability is greater when an entire test measures a single concept.
Split-half Method... analyzes the correlation between two equivalent halves of a test. This method works best when using an odd-even reliability approach to split the test in order to evenly distribute easy and difficult questions.
Alternative Forms or Equivalence... is obtained by giving two equivalent tests to the same group of students and then comparing the results. If there is a correlation between the results then the the tests are considered reliable. The main problem with this type of reliability test is that is difficult to design two well-written equivalent tests.
Test-retest or Stability... is performed by giving the test to the students twice within a certain time period and comparing the results. This is valid when no instruction of the tested material has been performed and the time period between tests is not too long that the results become invalid. Generally, the longer the period between tests the lower the reliability coefficient will be.

Validity

Contruct Validity... is used when measuring something that has not been measured before, has invalid results or when no criterion exists with which to compare it to. This type of validity evidence is based on theory that provides rationale as to the expectation of performance on a test. For example, if one was teaching a lesson on long division, the expectation would be that after intensive instruction in long division, the scores would improve. However, more intense instruction in phonics would not yield the same outcome.
Criterion-related Validity... is established by correlating scores from a test with an external criterion such as a prior exam.
Predictive Validity... refers to how well the test predicts the future performance of the test takers. Aptitude tests such as the SAT or GRE are examples of this type of validity evidence.
Concurrent Criterion-related Validity... is determined by administering both a new test and a prior assessment to a group and then finding the correlation between the two sets of scores. If the scores indicate a strong correlation in a criterion that people have confidence in, then the new test can be considered as a good alternative.
Content Validity... is the most basic evidence and refers to the review of the test to ensure that the test items correspond to the test objectives. It gives a basic overview of whether the test of valid but may not address extraneous factors such as reading level of the exam or aptitude.