A comparison of multiple-choice test response modes and scoring methods

dc.contributor.authorThayer, Ninaen
dc.contributor.committeechairFrary, Robert B.en
dc.contributor.committeememberCross, Lawrence H.en
dc.contributor.committeememberHinkle, Dennis E.en
dc.contributor.committeememberKehoe, Jerard F.en
dc.contributor.committeememberMartin, Edwin P.en
dc.contributor.committeememberMcLaughlin, John A.en
dc.contributor.departmentEducational Research and Evaluationen
dc.date.accessioned2017-03-10T21:56:17Zen
dc.date.available2017-03-10T21:56:17Zen
dc.date.issued1982en
dc.description.abstractThis study reports the comparison of seven different response/scoring methods used with multiple-choice tests in an academic setting with students enrolled in a required undergraduate art appreciation course at a medium-sized state university. The number of correct responses, based on a single choice per item, constituted the score on a preliminary test in each section of 35 to 40 students. Then, a distinct response/scoring method was randomly assigned to each section for two additional tests. One section continued to use number-right scoring and six sections used methods requiring or permitting multiple marks per item. All test scores counted toward course grades and examinees were so informed. Responses were used to determine estimates of internal consistency reliability and of validity with the previous quarter's grade-point average and subscores of the Scholastic Aptitude Test. An Evaluation Questionnaire, administered in each section, obtained for each response/scoring method self-report information about examinee study habits, testing preferences and experience, and responding behaviors. Observations of examinee behavior, subjective in nature, helped explain or confirm various empirical findings from this study or from the literature. Based on the findings from this study, it was concluded that: 1. observed differences in estimates of reliability or validity were not sufficient to justify the effort expended in administering and hand-scoring multiple-mark tests. 2. examinees experienced substantial difficulty becoming familiar with response/scoring methods which permit multiple marks. These methods require more than casual explanation and practice before examinees become adept in their use. 3. Item mark totals from methods scoring all the levels of information can provide more feedback to instructors regarding the performance of examinees on specific test items than could be acquired from the NR method. This feedback can help the instructor identify areas of content needing revision or test items needing rewording. When the examinees receive feedback, they have the opportunity to observe, score and learn from their own item performance, Pressey's initial consideration in 1950.en
dc.description.degreePh. D.en
dc.format.extentviii, 144, [2] leavesen
dc.format.mimetypeapplication/pdfen
dc.identifier.urihttp://hdl.handle.net/10919/76605en
dc.language.isoen_USen
dc.publisherVirginia Polytechnic Institute and State Universityen
dc.relation.isformatofOCLC# 8734332en
dc.rightsIn Copyrighten
dc.rights.urihttp://rightsstatements.org/vocab/InC/1.0/en
dc.subject.lccLD5655.V856 1982.T529en
dc.subject.lcshEducational tests and measurementsen
dc.titleA comparison of multiple-choice test response modes and scoring methodsen
dc.typeDissertationen
dc.type.dcmitypeTexten
thesis.degree.disciplineEducational Research and Evaluationen
thesis.degree.grantorVirginia Polytechnic Institute and State Universityen
thesis.degree.leveldoctoralen
thesis.degree.namePh. D.en

Files

Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
LD5655.V856_1982.T529.pdf
Size:
6.25 MB
Format:
Adobe Portable Document Format