导 师: 何莲珍
学科专业: E0211
授予学位: 硕士
作 者: ;
机构地区: 浙江大学
摘 要: as a comparatively direct measure of learners' communicative language ability,performance test (typically writing and speaking test) claims construct validity and a strong power for predictive utility of test scores and is gaining increasing interest both in practice and research in language testing. despite the seemingly direct relationship between the underlying constructs and scores, a routine but more complicated procedure should still be followed in test validation. the present thesis aims to give a general overview of justifications for and problems involved in test validation in performance testing context and illustrate how many-facet rasch model (mfrm)could be adopted as an effective tool in addressing problems in test validation.as an illustration, the present study investigates and models possible sources of score variability within the framework of mfrm, based upon raw scores of cet-set (college english test-spoken english test) administered on 19th and 20th, nov. 2005in hangzhou. the results demonstrate that cet-set could successfully separate candidates into statistically distinct levels of proficiency and facets in the model such as rater, task, rating domains and rating scales generally function well within certain limits. however, tendency to give central ratings along the scale, less variability across three rating domains and certain patterns of bias in raters' rating pattern are detected. fair scores for each candidate are also provided, minimizing the variability due to facets other than candidates' ability.mfrm manifests itself as effective in detecting whether each test method facet functions as intended in such performance test setting and providing useful feedback for test improvement.