Reliability Item Analysis Index of Difficulty Discriminating Power

26 Each test item in this study especially its content was designed based on the material had taught in order to be valid. To get the validity of each item, the writer analyzed statistically. The writer used the γ pbis formula as follows: In which: γ pbis : Coefficient Biseral correlation M p : Average of students who answered item no.1 correctly. M t : Average of total score S t : Standard deviation of total score p : number of students who answered item no.1 correctly. q : number of students who answered item no.1 incorrectly. Criteria: γ pbis r table, so the item is valid

3.7.4 Reliability

Reliability was a general quality of stability of scores regardless of what the test measured. In this study, the writer decided to use K-R.20 in measuring reliability of the test. This formula uses the number of items and the test, the mean of the scores, and the square of the deviation. Consulting the K-R. 20, the computation of the reliability of the test used the following formula: q p S M M γ t t p pbis − = 27 In which: r 11 : Reliability of total test. p : number of students who answered item no.1 correctly. q : number of students who answered item no.1 incorrectly. ∑pq : The sum of the result p x q n : number of students s : Standard deviation of the test.

3.7.5 Item Analysis

After determining and scoring the try-out test, an item analysis was made to evaluate the effectiveness of the items. It was meant to check whether each item met the requirement of a good test item. This item analysis concentrated two vital features, level of difficulty and discriminating power. Heaton 1975: 172 said that all item should be examination from the point of view of their difficulty level of discrimination.

3.7.6 Index of Difficulty

The index of difficulty level is generally expressed as the percentage of the students who answered the item correctly. It can be calculated by using the formula according to Arikunto 2006: 178 as follows: ⎟⎟⎠ ⎞ ⎜⎜⎝ ⎛ ∑ − ⎟ ⎠ ⎞ ⎜ ⎝ ⎛ = 2 2 11 S pq S 1 - k k r 28 P= JS B Where: P : index of difficulty B : number of the students who answered item correctly. JS : number of the students in a class. Criteria ID interval Criteria 0.00 ID ≤ 0.30 0.30 ID ≤ 0.70 0.70 ID 1.00 difficult medium easy

3.7.7 Discriminating Power

The discriminating power is a measure of the effectiveness of an item discriminating between high and low scores of the whole test. The higher the values of discriminating power are, the more effective the item will be. Heaton 1975:173 states the discrimination index of an item indicated the extent to which the item discriminated between the tested, separating the more able tested from the less able. The index of discriminating told us whether those students who performed well on the whole test tended to do well or badly on each item in the test. Discriminating power can be obtained by using this following formula: 29 Where: DP : Discriminating Power JB A : number of student in upper group who answered the item correctly JB B : number of student in lower group who answered the item incorrectly JS A : number of students in upper group Criteria DP interval Criteria 0.00 DP ≤ 0.20 0.20 DP ≤ 0.40 0.40 DP ≤ 0.70 0.70 DP ≤ 1.00 poor satisfactory good excellent

3.8 The Pre Test