Validity Analysis Reliability analysis

2 Empirical Validity To determine how test scores are related to some independent, ouside criterion such as marks given at the end of a course or instructors’ or supervisors’ rating. If the evidence shows that there is a high correlation between test scores and a trustworthy external criterion, we are justified in putting our confidence in the empirical validity of the test. Empirical validity depends in the large part on the reliability of both test and criterion measure. 3 Face Validity It means simply the way the test looks- to the examinees, test administrators, educators and the like. This is not validity in the technical sense and it can never be permitted to take the place of empirical or content validity. Yet, if the content of the test appear irrelevant, silly or inappropriate, knowledgeable administrators will hesitate to adopt the test and examinees will lack the proper motivation.

3.10.2 Reliability analysis

To measure reliability of question object used alpha formula as follows: Formula: 11 r = ⎟ ⎟ ⎠ ⎞ ⎜ ⎜ ⎝ ⎛ − ⎟ ⎠ ⎞ ⎜ ⎝ ⎛ − ∑ 2 t 2 i σ σ 1 1 n n In which N N x x 2 2 2 − = ∑ σ Explanation: 11 r = the reliability ∑ 2 i σ = the sums of score variants every items 2 t σ = total of variant

3.10.3 Level of Difficulties Analysis

Considered from difficulty aspect, the good question is neither too easy nor difficult. The very easy question not reinforce the students to the make their effort higher in finishing the question. The very difficult question will cause the students give up and don’t have spirit to try again because the question is out of their ability. The level of difficulties could be measured by the formula: Explanation: P : level difficulties of the question ∑ x : The sum of students answer correct m S : maximum score N : The sum of students m S N x P . ∑ = Table 3.1 Criteria Interval P Criteria 0.00 P 0.30 Difficult 0.30 P 0.70 Medium 0.70 P 1.00 Easy

3.10.4 Discriminating Power

The discriminating power will measure how well the test items arranged to identify the differences in the student’s competence. The formula is : Where, DP : The Discriminating Index RU : The number of students in upper group who answered the item correctly RL : The number of students in lower group who answered the item correctly ½ n : The number of students in one group n 12 RL - RU DP = 33