Response to Rasch & quality control: Controlling data, forgetting quality?by Edward Schaefer and Takaaki Kumazawa |
Abstract |
[ p. 47 ]
"Rasch measurement . . . basically test[s] the notion of unidimensionality, the idea that a test should be testing one thing at a time," |
[ p. 48 ]
"To make valid decisions, teachers should evaluate both students' product and process, and set multiple criteria." |
[ p. 49 ]
". . . in spite of the absence of a clearly defined construct for writing ability, FACETS was able to reliably calibrate writer ability and rater severity, but that the raters did in fact interpret the scale in different ways." |
[ p. 50 ]
Here we are entirely in agreement, and his points are well taken, as shown in the discussion on validity at the beginning of this article. Schaefer in fact used a similar scoring rubric in a previous FACETS study of native English speaker ratings of Japanese EFL essays (Schaefer, 2004). But in the present study, that was simply the preexisting reality of the situation, regardless of whether Rasch analysis was used or not. One reason Schaefer undertook the study was to persuade the faculty of the need to elucidate an explicit construct. It was therefore surprising that the data turned out to be as sound as it was, with the only case of significant bias being the researcher himself. One possible explanation for this is that most of the faculty members have been working together for many years, and there is an implicit construct – their views on what constitutes good writing were thus not as divergent as Lassche worries they might be. This possibility does not, of course, mitigate the need for a clearly stated and theoretically justified construct for good thesis writing.[ p. 51 ]