Show simple item record

The Relationship of Expert-System Scored Constrained Free-Response Items to Multiple-Choice and Open-Ended Items

dc.contributor.authorBennett, Randyen_US
dc.contributor.authorRock, Donalden_US
dc.contributor.authorBraun, Henryen_US
dc.contributor.authorFrye, Douglasen_US
dc.contributor.authorSpohrer, Jamesen_US
dc.contributor.authorSoloway, Ellioten_US
dc.date.accessioned2010-04-13T20:16:19Z
dc.date.available2010-04-13T20:16:19Z
dc.date.issued1990en_US
dc.identifier.citationBennett, Randy; Rock, Donald; Braun, Henry; Frye, Douglas; Spohrer, James; Soloway, Elliot (1990). "The Relationship of Expert-System Scored Constrained Free-Response Items to Multiple-Choice and Open-Ended Items." Applied Psychological Measurement 14(2): 151-162. <http://hdl.handle.net/2027.42/68260>en_US
dc.identifier.issn0146-6216en_US
dc.identifier.urihttps://hdl.handle.net/2027.42/68260
dc.description.abstractThis study examined the relationship of an expert- system scored constrained free-response item (requir ing the student to debug a faulty computer program) to two other item types: (1) multiple-choice and (2) free- response (requiring production of a program). Confir matory factor analysis was used to test the fit of a three-factor model to these data and to compare the fit of the model to three alternatives. These models were fit using two random-half samples, one given a faulty program containing one bug and the other a program with three bugs. A single-factor model best fit the data for the sample taking the one-bug constrained free re sponse and a two-factor model fit the data somewhat better for the second sample. In addition, the factor intercorrelations showed this item type to be highly re lated to both the free-response and multiple-choice measures.en_US
dc.format.extent3108 bytes
dc.format.extent869221 bytes
dc.format.mimetypetext/plain
dc.format.mimetypeapplication/pdf
dc.publisherSage Publicationsen_US
dc.subject.otherIndex Terms: Artificial Intelligenceen_US
dc.subject.otherCon Structed-response Itemsen_US
dc.subject.otherExpert-system Scoringen_US
dc.subject.otherFree- Response Itemsen_US
dc.subject.otherOpen-ended Items.en_US
dc.titleThe Relationship of Expert-System Scored Constrained Free-Response Items to Multiple-Choice and Open-Ended Itemsen_US
dc.typeArticleen_US
dc.subject.hlbsecondlevelEducationen_US
dc.subject.hlbsecondlevelPsychologyen_US
dc.subject.hlbtoplevelSocial Sciencesen_US
dc.description.peerreviewedPeer Revieweden_US
dc.contributor.affiliationumUniversity of Michiganen_US
dc.contributor.affiliationotherEducational Testing Serviceen_US
dc.contributor.affiliationotherEducational Testing Serviceen_US
dc.contributor.affiliationotherEducational Testing Serviceen_US
dc.contributor.affiliationotherYale Universityen_US
dc.contributor.affiliationotherYale Universityen_US
dc.description.bitstreamurlhttp://deepblue.lib.umich.edu/bitstream/2027.42/68260/2/10.1177_014662169001400204.pdf
dc.identifier.doi10.1177/014662169001400204en_US
dc.identifier.sourceApplied Psychological Measurementen_US
dc.identifier.citedreferenceAckerman, T.A., & Smith, P.L. (1988). A comparison of the information provided by essay, multiple-choice, and free-response writing tests. Applied Psychological Measurement, 12, 117-128.en_US
dc.identifier.citedreferenceBennett, R.E. (in press). Toward intelligent assessment: An integration of constructed-response testing, artificial intelligence, and model-based measurement. In N. Frederiksen, R. J. Mislevy, & I. Bejar (Eds.), Test theory for a new generation of tests. Hillsdale NJ : Erlbaum.en_US
dc.identifier.citedreferenceBirenbaum, M., & Tatsuoka, K.K. (1987). Open-ended versus multiple-choice response formats—It does make a difference for diagnostic purposes. Applied Psychological Measurement, 11, 385-395.en_US
dc.identifier.citedreferenceBleistein, C., Maneckshana, B., & Mc Lean, D. (1988). Test analysis: College Board Advanced Placement Examination Computer Science 3JBP (SR-88-63. Princeton NJ: Educational Testing Service.en_US
dc.identifier.citedreferenceBraun, H.I. (1988). Understanding scoring reliability: Experiments in calibrating essay readers. Journal of Educational Statistics, 13, 1-18.en_US
dc.identifier.citedreferenceBraun, H.I., Bennett, R.E., Frye, D., & Soloway, E. (in press). Scoring constructed-responses using expert systems. Journal of Educational Measurement.en_US
dc.identifier.citedreferenceCollege Board. (1988). Advanced Placement course description : Computer science. New York: Author.en_US
dc.identifier.citedreferenceJohnson, W.L., & Soloway, E. (1985). PROUST: An automatic debugger for Pascal programs. Byte, 10(4), 179-190.en_US
dc.identifier.citedreferenceJöreskog, K., & Sörbom, D. (1986). PRELIS: A program for multivariate data screening and data summarization. Mooresville IN: Scientific Software, Inc.en_US
dc.identifier.citedreferenceJöreskog, K., & Sörbom, D. (1988). LISREL 7: A guide to the program and applications. Chicago: SPSS, Inc.en_US
dc.identifier.citedreferenceLoehlin, J.C. (1987). Latent variable models. Hillsdale NJ: Erlbaum.en_US
dc.identifier.citedreferenceLord, F.M. (1980). Applications of item response theory to practical testing problems. Hillsdale NJ: Erlbaum.en_US
dc.identifier.citedreferenceMarsh, H.W., & Hocevar, D. (1985). Application of confirmatory factor analysis to the study of self-concept : First and higher order factor models and their invariance across groups. Psychological Bulletin, 97, 562-582.en_US
dc.identifier.citedreferenceMazzeo, J., & Bleistein, C. (1986). Test analysis: College Board Advanced Placement Examination Computer Science 3IBP (SR-86-105. Princeton NJ: Educational Testing Service.en_US
dc.identifier.citedreferenceMazzeo, J., & Flesher, R. (1985). Test analysis: College Board Advanced Placement Examination Computer Science 3HBP (SR-85-180. Princeton NJ: Educational Testing Service.en_US
dc.identifier.citedreferenceSobel, M.E., & Bohrnstedt, G.W. (1985). Use of null models in evaluating the fit of covariance structure models. In N. B. Tuma (Ed.), Sociological methodology (pp. 152-178). San Francisco CA: Jossey-Bass.en_US
dc.identifier.citedreferenceSpohrer, J.C. (1989). MARCEL: A generate-test-and-debug (GTD) impasse/repair model of student programmers (CSD/RR No. 687. New Haven CT: Yale University, Department of Computer Science.en_US
dc.identifier.citedreferenceSternberg, R.J. (1980). Factor theories of intelligence are all right almost. Educational Researcher, 9, 6-18.en_US
dc.identifier.citedreferenceTraub, R.E., & Fisher, C.W. (1977). On the equivalence of constructed-response and multiple-choice tests. Applied Psychological Measurement, 1, 355-369.en_US
dc.identifier.citedreferenceTucker, L.R., & Lewis, C. (1973). A reliability coefficient for maximum likelihood factor analysis. Psychometrika, 38, 1-10.en_US
dc.identifier.citedreferenceWard, W.C. (1982). A comparison of free-response and multiple-choice forms of verbal aptitude tests. Applied Psychological Measurement, 6, 1-11.en_US
dc.identifier.citedreferenceWard, W.C., Frederiksen, N., & Carlson, S.B. (1980). Construct validity of free-response and machine-scorable forms of a test. Journal of Educational Measurement, 17, 11-29.en_US
dc.owningcollnameInterdisciplinary and Peer-Reviewed


Files in this item

Show simple item record

Remediation of Harmful Language

The University of Michigan Library aims to describe library materials in a way that respects the people and communities who create, use, and are represented in our collections. Report harmful or offensive language in catalog records, finding aids, or elsewhere in our collections anonymously through our metadata feedback form. More information at Remediation of Harmful Language.

Accessibility

If you are unable to use this file in its current format, please select the Contact Us link and we can modify it to make it more accessible to you.