72
11111 2.2 L 11111 ' 111111.25 11111 1.4 11111' .6 MICROCOPY RESOLUTION TEST CHART NATIONAL BUREAU OF STANDARDS STANDARD REFERENCE MATERIAL 1010a (ANSI and ISO TEST CHART No. 2)

11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

Embed Size (px)

Citation preview

Page 1: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

11111 2.2

L

11111 '

111111.25 11111 1.4 11111' .6

MICROCOPY RESOLUTION TEST CHARTNATIONAL BUREAU OF STANDARDS

STANDARD REFERENCE MATERIAL 1010a(ANSI and ISO TEST CHART No. 2)

Page 2: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

DOCUMENT RESUME

M 240 313 CE 038 157

'ITLE Assess Student Performance: Knowledge. SecondEdition. Module D-2 of Category D--InstructionalEvaluation. Professional Teacher Education ModuleSeries.

NSTITUTION Ohio State Univ., Columbus. National Center forResearch in Vocational Education.

PONS AGENCY Department of Education, Washington, DC.;SPORT NO ISBN-0-89606-143-4WM DATE 84ROTE 71p.; For related documents, see ED 228 460, ED 236

356, and CE 038 158.NAMABLE FROM American Association for Vocational Instructional

Materials, 120 Driftmier Engineering Center,University of Georgia, Athens, GA 30602.

WM TYPE Guides Classroom Use Materials (For Learner)(051)

MRS PRICEESCRIPTORS

MF01/PC03 Plus Postage.*Academic Achievement; Behavioral Objectives;*Cognitive Measurement; *Competency Based TeacherEducation; Evaluation Criteria; *Evaluation Methods;Learning Modules; Measurement Techniques;Postsecondary Education; *Student Evaluation;Teaching Skills; *Test Construction; Test Items;Vocational Education Teachers

ABSTRACTThis module, one in a series of 127 performance-based

.eacher education learning packages focusing upon specific)rofessional competencies of vocational teachers, deals withassessing student performance. Addressed in the individual learning!xperiences are the following topics: understanding the:onsiderations involved in assessing student cognitive performance;:onstructing multiple-choice, matching, completion, true-and-false,:ssay, oral, case study, and problem-solving items to measure the' chievement of selected student performance objectives; and assessingftudent cognitive performance in an actual teaching situation. Each.earning experience contains an objective, instructional text, one orlore learning activities, and a feedback activity. (MN)

.**********************************************************************

Reproductions supplied by EDRS are the best that can be madefrom the original document.

.**********************************************************************

Page 3: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

NIODOSanCe,V2rO

Ot Pee\ studseSSeSS ec Nedg-- knov.wcm

Second..,z4A\\*.lon

ModOe 0-2 oi Category0--WistrunaN

evaNuaton

PlIOVESS040-1EACtlect

EDUCKOON

1100.0Woes

*TheoNastotte.

Centettot fkeseatclA

In Vocatonaleducaton

Key rogoonStott

Jam 5. 1-lar611.on,Prograrn

°Vector

4110

VWbect E.Norton, Pssoaate

Prograrn°Vector

Glen E. Fardlg,SPedallst.

l...o ls G. Han'Igion,Programs

e6stanl.

KarentA.

au nn, ProgramPIstanl.

SecondEcnnon.

Copyt\gilt L0196.41)111eNanona\

Centetiot i;eeaoll\n\locanona\

Educanon,

ilie OlnoState

n'ofect1960Kenny

i;oad,Co\unOus,

Olio 43210.

Copytight 1s dened

en0 la tertn:Cheteallet

Be podkonsof Os woda

covetedby Os coplOght

vike be1n the

peb0c

domain.

110wodc Was

deyeoped

%MOT econttact

vdth theDepedinent

of Edecetkon.tkovieyet.

e boontlenbtdkoes

not necesretM

fetedthe pogeon

of Poof thet

AVIACY,end no

Ol11C1A%endocsecnent

of theseMelee

s s oe 0 e nletted.

1934

ISSN049606-13-4

Pub\19110end clIsulbuted

by the runeticanAssocletlon

totVocatlonellnstructlonel

Meterlels

001141),120

OciltmketEnrimeeting

entec,Onlvec6ty

ol Geor4a,Abens,

Geor4a30602,

k4041

542-2.5e6.

U.S. DEPARTMENT OF EDUCATIONNATIONAL INSTITUTE OF EDUCATIONUCATIONAL RESOURCES INFORMATION

CENTER IERICIThis document has been reproduc1 asreceived from the person or organitationoriginating itMinor change: have been made to improvereproluction quality.

Points of view or opinions stated in this doctiment do not necessarily represent official MEposition or policy

1

2

-PERMISSION TO REPRODUCE THISMATERIAL HAS BEEN GRANTED BY

r

TO THE EDUCATIONAL RESOURCESINFORMATION CENTER (ERIC).-

Page 4: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

DetermineWhat Is toBe Taught

I

h/Establish

StudentPerformance

Criteria

i

--IP-

AssessStudent

Performance(Knowledge)

AssessStudent

Performance EllII"(Skills)

AssessStudent

Performance(Attitudes)

DetermineStudentGrades

....-40.-Evaluate YourInstructionalEffectiveness

INSTRUCTIONAL EVALUATION PROCESS

2

I

I

41

Page 5: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

INTRODUCTION

How will you know whether your students haveacquired the technical knowledge implied or statedin the student performance objectives for the pro-gram? How can you be sure that they have learnedthe required theory, facts, data, or other informa-tion? The answer is that you can assess their knowl-edge level using some type of cognitive test items:true-fals3, multiple-choice, matching, completion,essay, or oral items.

Of course, measurement of student knowledge isonly one part of the picture. In fact, there are threekinds of performanceknowledge, skills, and atti-tudesthat you will want to measure, and these arenot truly discrete. In performing a skill, for example,a student is not only demonstrating psychomotor skill,but also certain attitudes (e.g., concern for safetyand cooperation) and the possession of certainknowledge (e.g., the proper procedure to follow inmaking change for a customer).

However, there are many times when the mea-surement of knowledge alone is important. You maywant to determine whether students know the stepsin a procedure before allowing them to carry out thatprocedure for the first time. You may need to deter-

mine if students have understood a particular read-ing assignment or class lecture. You may want totest their ability to solve practice problems. And suon.

By assessing student knowledge regularly, you cankeep them informed concerning their progress inlearning the technical knowledge required for theoccupation. You can determine their readiness to goon to subsequent learning activities. Assessment canalso provide you with vwuable information about theeffectiveness of your o.vn teachingwhether yourstudents are learning, or whether you need to changeyour teaching strategies. And assessment also pro-vides a basis for assigning grades in traditional vo-cational-technical programs.

This module is designed to help you gain the skillsyou need (1) to construct reliable, valid, and usabletest items that will assess your students' achieve-ment of the cognitive student performance objec-tives that make up your program; (2) to put thoseitems together into an effective testing device; and(3) to create a testing environment that will allowstudents to demonstrate their knowledge.

3

Page 6: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

ABOUT THIS MODULEObjectivesTerminal Object''sess student ccperformance w A beusing the Teacher 1=er o68 (Learning Exper,'enct.

Enabling Objectives:1. After completing the rah .d reading, demonstrate

knowledge of the conside. ns involved in assessingstudent cognitive performance (Learning Experience1).

2. After completing the retured reading, construct fivemultiple-choice items to neasure the achievement ofselected cognitive stuc nt performance objectives(Learning Experience I!

3. After completing the quired reading, construct amatching item to measure the achievement of se-lected cognitive student performance objectives(Learning Experience III).

4. After completing the required reading, construct fivecompletion items to measure the achievement of se-lected cognitive student performance objectives(Learning Experience IV).

5. After completing the required reading, construct ten true-false items to measure the achievement of selectedcognitive student performance objectives (LearningExperience V).

6. After completing the required reading, construct threeessay items and two oral items to measure theachievement of selected cognitive student perfor-mance objectives (Learning Experience VI).

7. After completing the required reading, construct onecase study or problem-solving item to measure theachievement of a selected cognitive student perfor-mance objective (Learning Experience VII).

ResourcesA list of the outside resources that supplement those con-tained within the module follows. Check with your re-source person (1) to determine the availability and the lo-cation of these resources, (2) to locate additional referencesin your occupational specialty, and (3) to get assistancein setting up activities with peers or observations of skilledteachers, if necessary. Your resource person may also becontacted if you have any difficulty with directions or inassessing your progress at any time.

Learning Experience IOptional

Reference: Green, John A. Teacher-Made Tests.Second Edition. New York, NY: Harper & Row Pub-lishers, 1975.Reference: Gronlund, Norman E. Measurement andEvaluation in Teaching. Fourth Edition. New York, NY:Macmillan Publishing Co., 1981.Reference: Karmel, Louis J., and Karmel, Marylin 0.Measurement and Evaluation in the Schools. Sec-ond Edition. New York, NY: Macmillan Publishing Co.,1978,

actual teaching situation, as-lwledge) performance. Your

by your resource person,kssessment Form, pp. 67

4

Learning Experience IIOptional

Peers to critique your multiple-choice items.

Learning Experience IIIOptional

Peers to critique your matching item.

Learning Experience IVOptional

Peers to critique your completion items.

Learning Experience VOptional

Peers to critique your true-false items.

Learning Experience VIOptional

Peers to critique your essay and oral items.

Learning Experience VIIOptional

Peers to critique your case study or problem-solvingitem.

Learning Experience VIIIRequired

An actual teaching situation in which you can assessstudent cognitive (knowledge) performance.A resource person to assess your competency in as-sessing student cognitive (knowledge) performance.

General InformationFor information about the general organization of eachperformance-based teacher education (PBTE) module,general procedures for its use, and terminology that iscommon to all the modules, see About Using the NationalCenter's PBTE Modules on the inside back cover. For morein-depth information on how to use the modules in teacher/trainer education programs, you may wish to refer to threerelated documents:

The Student Guide to Using Performance-Based TeacherEducation Materials is designed to help orient preservice andinservice teachers and occupational trainers to PBTE in gen-eral and to the PBTE materials.

The Resource Person. Guide to Using Performance-BasedTeacher Education Materials can help prospective resourcepersons to guide and assist preservice and inservice teachersand occupational trainers in the development of professionalteaching competencies through use of the PBTE modules. Italso includes lists of all the module competencies, as well asa listing of the supplementary resources and the aadreaseswhere they can be obtained.

The Guide to the Implementation of Performance-BasedTeacher Education is designed to help those who will admin-ister the PBTE program. It contains answers to implementa-tion questions, possible solutions to problems, and alternativecourses of action.

S

S

Page 7: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

Learning Experience IOVERVIEW

Activity

IOptionalkActivity44128.

After completing the required reading, demonstrate knowledge of the consid-erations involved in assessing student cognitive performance.

You will be reading the information sheet, Student Cognitive PerformanceAssessment, pp. 6-16.

You may wish to read one or more of the following supplementary references:Green, Teacher-Made Tests, pp. 19-41; Gronlund, Measurement and Eval-uation in Teaching, pp. 65-120; and/or Karmel and Karmel, Measurementand Evaluation in the Schools, pp. 397 -409.

You will be demonstrating knowledge of the considerations involved in as-sessing student cognitive performance by completing the Self-Check, pp.17-18.

You will be evaluating your competency by comparing your completed Self-Check with the Model Answers, pp. 19-20.

5

Page 8: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

Activity Assessing student cognitive performance, under one name or another, is asold as education. Experts agree that it is one of the vocational teacher's mostimportant responsibilities. For information on why, when, and how to assessyour vocational students' cognitive performance, read the following informa-tion sheet.

STUDENT COGNITIVE PERFORMANCE ASSESSMENTVocational-technical instruction should be based

on identified student performance objectives, whichstate the intended outcomes of the educational pro-cess in terms of the specific knowledge, skills, andattitudes students need for entry into the world ofwork.' Learning in all three domainsknowledge(cognitive), skills (psychomotor), and attitudes (af-fective)is important, but in this module we areconcerned with the cognitive domain.

Benjamin Bloom, divides the cognitive domain intosix levels. Cognitive objectives at the lowest level(i.e., knowledge) require students to recognize orrecall correct facts, data, or information. Such objec-tives usually call for students to list, define, identifyitems, or in some other way recognize or recall aparticular piece or body of information.

Student performance objectives at the higher lev-els in the cognitive domain require students to domore than simply recognize or recall factual infor-mation correctly. They also require students to use

1. To gain skill in developing student performance objectives. you maywish to refer to Module B-2, Develop Student Performance Objectives.To gain skill in determining the performance criteria that underlie yourmeasurement of student achievement. you may wish to refer to ModuleD-1. Establish Student Performance Criteria.2. Benjamin S. Bloom, ed., Taxonomy of Educational Objectives.Handbook I: Cognitive Domain (New York, NY: David McKay Company,1956). Reprinted by permission of David McKay Company, Inc.

6

that information in some way. Objectives at the sec-ond level of the cognitive domain (i.e., comprehen-sion), for example, may call for students to summa-rize, interpret, translate, or paraphrase facts, data,or information. Objectives at still higher levels mayrequire application, analysis, synthesis,.or evalua-tion of factual information.

Student cognitive performance assessment is theprocess of determining whether students haveachieved these cognitive performance objectives. Itis the process of testing students' knowledge andtheir ability to use that knowledge in a manner ap-propriate to your occupational area.

Purposes of AssessmentInformation about students' cognitive perfor-

mancewhat knowledge they have acquired andhow they can use that knowledgecan help in pro-viding high-quality instruction for the students en-rolled in your program. You and other people can putthis information to several important uses.

Student readiness. You can use this informationto determine whether your students are ready for aparticular learning activity. Often, students must knowone thing before they proceed to another activity. Let'ssay, for example, that a home economics teacher isdemonstrating to students how to separate egg yolksfrom whites. The teacher has finished showing themthe procedure and is now ready to have them trytheir skill on some real eggs.

Before the teacher turns students loose on thoseeggs, however, he or she might want to make surethat they're ready for the task. Eggs do cost moneyand they can make a nasty mess if mishandled. Theteacher would be likely to quickly test students'knowledge of the proceduretheir readiness for thehands-on task. He/she would want to ensure thatstudents know (1) the steps in the procedure (e.g.,first you break the egg); and (2) the critical factorsin any steps (e.g., you must break the egg carefully,so that the two halves of the shell remain intact).

For that matter, you will undoubtedly test stu-dents' knowledge of safety rules and practices be-fore students actually begin to use the tools andequipment in the laboratory. And, you will often teststudents' knowledge of theory before they proceedto practical exercises using that theory. Perhaps

Page 9: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

carpentry students are ready to begin constructionprojects including the purchase of lumber. The car-pentry teacher would probably want to test students'knowledge of formulas for converting running feet toboard feet before they make their purchases.

Instructional improvement. You can also use theinformation you get through assessing student cog-nitive performance to improve your instruction. Let'stake the egg-separating example again. The teachermight test students' knowledge of the procedure anddiscover that they don't know what he/she thoughtthey did. This could indicate that the instructionalactivities need improvement. Perhaps the demon-stration was unclear or confusing. Perhaps somestudents were just sitting too far away to see whatthe teacher was doing. Finding out that the studentshaven't learned what you were teaching can alertyou to possible weaknesses in your instruction.

Information about student progress. Otherpeople can make good use of this information aboutstudents' learning, too. Students, for instance, needto know how they are progressing in your program.They will want to know if they are successfully ac-quiring the technical knowledge they will need onthe job.

Providing students with feedback concerning theircognitive performance can help them identify theirown weaknesses so thai they can work more suc-cessfully to acquire needed technical knowledge.Positive feedback on the progress they are makingcan help to motivate them to progress further.

Administrators will also be interested in the infor-mation you gain by assessing student cognitive per-formance. One of your responsibilities, as a voca-tional-technical instructor, will no doubt be to reportstudents' progress or achievement. Almost all vo-cational teachers must submit student grades or otherprogress reports at specified intervals. Part of stu-dents' overall progress, of course, is their progressin acquiring needed technical knowledge.

The information about student progress that youreport to administrators may also be shared with otherpeople in the school or community. Counselors mightuse this information in making placement decisionswith students. Parents of secondary students areusually concerned with how their children are doingin the vocational-technical program.

Finally, employers frequently want to know aboutyour students' competence. They commonly want toknow students' course grades. They might even wantto talk to you about the technical content of your pro-gram or about the knowledge, attitudes, and skills ofa particular student. Of course, you can't tell poten-tial employers what your students know, and whatthey can do with what they know, until you have foundout yourself by assessing their cognitive perfor-mance.

7

Assessment iecnniquesYou might assess your students' cognitive perfor-

mance at almost any time. It would not be exagger-ating to say that you will assess cognitive perfor-mance, in one way or another, for one purpose oranother, every day in your classes. Ensuring thatstudents have mastered the technical content of yourprogram should be an ongoing part of your instruc-tion.

There are a number of different ways to assessstudent cognitive performance. In some situations,you might use very informal means to do the job.You might, for example, give a short oral quiz, re-quiring students to provide brief written answers thatyou can quickly check, during class, before they moveinto the laboratory. A final course examination, onthe other hand, would almost certainly involve a full-blown, formal written test.

Regardless of how formal or informal the testingsituation is, however, the test items that you will useto measure student knowledge will be of two types:objective and subjective.

Some kinds of cognitive test items are called ob-jective by testing experts because scoring them isalmost entirely an objective process. The followingtypes of items are considered to be objective:

Multiple-choiceMatchingCompletionTrue-false

Scoring these items is primarily objective becausethe scorer (i.e., you) does not need to use judgmentto determine whether a student's answer is right orwrong. Students' answers can be compared to asimple scoring or answer key. If the key says thatthe answer to an item should be "true," a student'sanswer either does or does not agree. Neither opin-ion nor judgment are needed for scoring.

Other kinds of test items are called subjectivebecause they do require the use of judgment andinterpretation in scoring answers. The following typesof items are considered to be subjective:

EssayOral

As you are no doubt aware, scoring essay items isnot as simple a matter as scoring objective items. Isa given answer the right answer? !s it partially right?If an essay question is worth 20 points, how manypoints is a partially right answer worth? Scorers mustuse their own judgment and opinion to resolve thesequestions in scoring subjective items.

There is another basic difference in test items.Some items require students to supply the answerfrom their own memory. The subjective types of items

Page 10: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

are both supply tests, as you certainly know if you'veever taken an essay test. Of the objective types oftest items, only one is of the supply typecomple-tion items. Completion s require students to fillin the blank or give a short answer (word, numeral,symbol, or phrase) to a question, again from theirown memory.

The other types of objective test items (multiple-choice, matching, and true-false), however, are se-lection types. In all of these, students are given twoor more possible answers from which to select thecorrect answer or answers. Selection types test stu-dents' ability to recognize correct information, whilesupply types test students' ability to recall correctinformation.

Finally, a word about terminology. You may havenoticed in the previous explanation that we talkedabout test items rather than test questions and aboutmultiple-choice items rather than multiple-choicetests. The reasons that the word it' m is the pre-ferred word in both cases are as follows:

Items on a test may or may not be stated ininterrogativequestionform. True-false items,for example, are not questions but declarativestatements.A given test may include many different typesof items. While you might devise a test that in-cluded, for example, only essay items (i.e., anessay test), you will also devise tests that in-clude one section of essay items, one of true-false items, one of multiple-choice items, or someother combination.

Consequently, the word item is a more precise de-scriptor in both situations.

8

Thi: Good Test: In TheoryWhat makes tests good or bad? What qualities

should your tests have? Objective vs. subjectiveitems, supply vs. selection itemshow do you choosewhich to use? There is a wealth of literature in whichtesting experts agree that a good test should havethe following characteristics.

The test must be valid. Validity is the extent towhich a test measures what it is supposed to mea-sure. To the extent that a test measures what it issupposed to, it is valid. This notion seems so simpleand straightforward that you might wonder why it iseven mentioned. How could you possibly go wrong?

You might go wrong if your test measures morethan you intended. When you assess student cog-nitive performance, you want your tests to measurestudents' knowledge of specific facts, data, or in-formation. Your test becomes less valid if it alsomeasures other knowledge as well.

Assume, for example, that it is very important thatstudents have read an assigned chapter beforemoving on to the next learning activity. You thus want'to administer a test to ensure that they have readand understood the assigned material. One of yourtest items is the following multiple-choice item:

What country is currently pioneering work inthe area of friction welding?

1. India2. Greenland3. Nicaragua4. Soviet Union

Some students would be able to answer this itemcorrectly using genera: knowledge, whether they hadread the assigned material or not. Some students.could be likely to know that, of the four countrieslisted, the Soviet Union has by far the most techno-logically advanced industrial economy. The SovietUnion would, therefore, be the most likely of the fourcountries to be engaged in pioneering work in atechnological area. A guess based on this reason-ing would be correct.

This item, then, measures students' generalknowledge as well as their knowledge of the readingassignment. Could the item be repaired to measurewhat it is supposed to? If all the possible answerswere equally plausible (i.e., all countries listed hadtechnologically advanced industrial economies),general knowledge would be useless. Students wouldneed to know the material covered in the assignedreading in order to answer the item. The item wouldthen measure what it is supposed to, thereby in-creasing its validity.

9

Page 11: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

Another way in which validity can be affected is ifitems on a test are answered by other items. Forexample, consider the following two compietion items:

Harmful live and grow on inanimateobjects.When harmful germs get inside a human body,

can begin.

The first correct answergermsis provided in thesecond item. Students who are discerning enoughto recognize this can do well on the test without hav-ing truly learned the material. Thus, test validity isreduced.

The validity of a test can also be reduced if stu-dents must use skills other than those you intend totest in "rder to answer the items. For example, anywritten test reouires students to use reading skills.Essay tests require the use of writing skills. Oral testsrequire the use of oral communication skills.

The need for students to use these other skillscan be a serious problem, causing reduced validity,if some students cannot read or understand the itemsto begin with. If this happens, your test is measuringtwo things whether you intend it to or notstudents'technical knowledge and their communication skills.

Students will need some level of communicationskills to succeed in the world of work. We are notsuggesting, by any means, that such skills are notimportant. But if you wish to measure technicalknowledge, then you must ensure that your test al-lows students to show how much technical knowl-edge they actually possess. If a student's commu-nication skills are weak, that problem should beidentified and remediated, but it should not be al-lowed to cloud the measurement of technical skills.

The problem can be avoided if your items arecarefully developed with students' communication skilllevels in mind. In other words, written items shouldbe at students' reading level. All items should beclearly and simply stated so that students' commu-nication skills are sufficient to allow them to dem-onstrate their technical knowledge. In this way, theeffects of communication skills on validity will not besignificant.

The length of a test can also affect its validity. Let'ssay, for example, that you are preparing a final ex-amination for your students covering the entire se-mester's technical content. You intend to test stu-dents' acquisition of a considerable body of technicalknowledge and their ability to use that knowledge.Your final examination, then, must be long enoughto cover all the knowledge and use of knowledgethat you intend to test.

9

If your final examination has only five multiple-choice items, you are probably only testing stu-dents' knowledge of five small bits of information andnot of the whole of the technical content coveredduring the semester. The five-item final examinationthus would have very low validity. If the test is tomeasure knowledge of the many bits that comprisethe whole technical content, it has to be long enoughto include a representative sampling of all those bits.

On the other hand, a five-item test could have veryhigh validity in a different situation. Perhaps you wantto determine if students are ready to go on to hands-on activities after information activities. A five-itemtest might be quite sufficient if there are only five bitsof information to check on. Since, in this case, youare sampling a small body of technical content, ashort test should be sufficient.

The test should be reliable. Reliability is theconsistency with which a test measures achieve-ment. Experts in the theory of testing tell us that avalid test is always reliablethat is, it consistentlymeasures what it is supposed to measure. Sometheorists, in fact, treat reliability as a part of validity.In any case, it is important to note that a test can bereliable (i.e., can get a consistent measurement) evenif it is not valid (i.e., does not measure what it isintended to measure).

This point can be illustrated with an extreme ex-ample. Let's say that you decide to measure stu-dents' knowledge of some information. For someunexplained reasonperhaps you've had a veryhectic dayyou write your test in German, eventhough none of your students understand a word ofGerman. Now, obviously, this test would not be validfor any student who didn't speak German. Thus, itwouldn't be valid for your students. It would notmeasure what it is supposed to; instead, it wouldmeasure students' knowledge of the German lan-guage.

In spite of this, however, the test could be reliable.Let's say that you actually administer this German-language test to your students. There is a goodchance that all of them would score zero. Further-more, they would get this same score consistently.You could give them the same test again the nextday and they would still score zero (unless they allspent the previous evening closeted with a Germantutor). Hence, although this German-language testwould have no validity for English speakers, it wouldyield reliable resultsconsistent scores.

One factor that has a great effect on test reliabilityis the subjectivity of the scoring. If a test has highreliability, you should be able to administer and re-administer it to a group of students and get roughlythe same set of scores (making allowance for such

10

Page 12: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

things as increased learning in the intervening time).Or, two different scorers should be able to sit downwith a set of student tests, rate them independently,and arrive at very similar scores.

Subjective scoring, however, can lower the relia-bility of tests because it can allow inconsistent mea-surement. Two different teachers could each inde-pendently score student essay items and end up withcompletely different scores for the same answers.For that matter, one scorer could score a test oneway one time yet score it completely differently an-other time.

Psychological studies have confirmed time andtime again that inconsistent scoring can be a seri-ous problem with subjective test items. Scorers mustconstantly make decisionson the basis of theirpersonal judgment, opinion, and preferenceaboutthe worth of students' answers.

Thus, it is important to objectify subjective testscoring insofar as possible. For example, beforeadministering an essay test, you need to determinewhether points will be deducted for incomplete sen-tences or for spelling errors. (Are these skills part ofwhat you wish to measure? How many points willyou deduct for each error or as a maximum?) Youalso need to prepare a scoring key, which lists thekey points you expect to be made in each answer,and you need to determine how much value to as-sign to each key point covered. A thorough scoringkey can minimize reliability problems with subjectivetests.

Objective tests, on the other hand, do minimize oreven eliminate subjectivity, in scoring (hence, theirname). If a true-false item is well written, for ex-ample, there can be only one answer to it. Further,no interpretation or judgment is required on the partof the scorer to determine if a student's answer iscorrect. The item is either true or false, and the stu-dent either did or didn't mark the correct response.

Different scorers, consequently, should have nodifficulty in arriving at the same score for an individ-ual student's test. Or, you couid score the same setof papers twice and get exactly the same set of re-sults. For that matter, with a scoring key, anyoneshould be able to score an objective test consis-tently, whether he/she knew one bit of the contenttested or not.

As important as reliability is, however, it must bebalanced with other conce, ns. Objective test items,in general, are good at measuring lower-level cog-nition (e.g., recall of facts). Higher-level cognition ismore effectively measured through subjective testitems. Thus, you must not, in the interests of relia-bility (and ease of scoring), use only objective mea-sures. Rather, as mentioned before, you must en-sure that your procedures for scoring subjective testitems are as objective as possible.

The test should be usable. No matter how validor reliable a test might be, it will be of little use toyou if it is unreasonably difficult or time-consumingto prepare, administer, or score. A final examinationwith 1,000 objective items could be quite valid andreliable, for instance. Yet such a test would take solong to prepare, administer, and score, that it simplywould not be practical to use.

Likewise, essay or oral items can in theory be usedto test student recall of facts, data, or informationthe lowest level of the cognitive domain. However,essay and oral items can be quite time-consuming.It takes students a longer time to write essay an-swers and you a longer time to score them. You can'tgive an oral item to the whole class at once (wouldthey all answer tonet1 in unison?), so consider-able time would be needed for individual administra-tion and scoring.

Consequently, essay and oral items would not bethe most practical, usable means of assessing stu-dent cognitive performance at the recall level. Muchmore practical and usable in this case would be ob-jective items.

Other factors may affect the usability of tests aswell. Ultimately, you will need to use your own judg-ment to determine any other qualities that would makea test practical and usable in your own situation.

The Good Test: In PracticeHow can you develop tests that are valid, reliable,

and usable in your own program? The followingguidelines can help you ensure that your tests mea-sure what they are supposed to measure, anti mea-sure it consistently, while remaining practical andusable.

Base your test en student performance objec-tives. Just as your student performance objectivesidentify the knowledge, skills, and attitudes to be

10

Page 13: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

AD

IP'

P

taught, they also establish a basis for student eval-uation. Logically, a given test should cover each ofthe student performance objectives that studentsshould have achieved at a particular point. As youidentify cognitive objectives to be covered in yourtest, you can also identify the specific criteria, de-scribed in the objectives, against which students'achievement should be measured.

You can then develop test items that also requirestudents to demonstrate the required knowledge atthe specified level (e.g., recall, 'synthesis). This willbe of great importance in ensuring that your test hashigh validitythat it actually measures what it issupposed to.

The test should differentiate. If your test mea-sures what it is supposed to measure, it should dif-ferentiate between students who know the materialbeing tested and those who don't. If it does not dif-ferentiate, its validity will be low.

To tell if a test differentiates in this way, you willneed to look at other indicators of student knowl-edge or achievement. These could be general in na-turethe grades students usually get in your pro-gram, for example. If students who usually get goodgrades do well on a test, while students who usuallyget poor grades do badly on the test, it is probablyreasonable to conclude that the test differentiates.Other general indicators of knowledge or achieve-ment that you can use for comparison might bescores on standardized tests, general intelligencetests, or other tests of your own that you know tohave high validity and reliability.

I

You should keep in mind, however, that many ofthese general indicators should be taken with a grainof salt. Siandardized achievement tests and intelli-gence tests ai..? often criticized by testing expertsbecause of their cultural and language bias. Mostappropriate for your purposes would be other spec-ific indicators of knowledge of the particular contentbeing tested.

For example, you might be able to compare stu-dents' test scores with their scores on daily quizzescovering the same content. Your own observation ofstudent performance in the classroom or laboratorycan often tell you which students know the materialand which don't. For example, a student who doeswell on a test, and then also does well on hands-onpractice activities that require using that specificknowledge, can reasonably be assumed to possessthat knowledge.

Minimize the effects of communication skills.We have already discussed that every type of testite-n requires students to use some communicationskills to understand and answer the items. A testbecomes less valid if it requires students to usecommunication skills they don't have. When thishappens, students cannot demonstrate their knowl-edge because they cannot communicate it to you.The test is not measuring what it is supposed to.

Thus, you should choose your jest items carefully.For example, the essay item probably requires thehighest level of communication skills. If a cognitiveperformance objective requires that students actu-ally organize and present information themselves,

11

12

Page 14: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

2 v . 04r- ,

1 .1 Fv:;,c17.7 r .x Site.. I .1"

Z.. 'I , r c

. rx : .:-.f

tok,e irt, ei e..,t tee -e: t* .alr- .1' 04.11

.17'' :.,141C. (,;.1.41-;,, 0711. 1+t.

V I.: I :toy( :tk :41."":41:, 7.1 ',10t*,0

v.1.

: r er Irk .

is .

1;1r 01:4,00. I titlea t-f !....!,e1r--

" t:t7 4.1'11c, 1.:

-tc-;14, .1.a41 ,! 7, A ',v.. !r.-!teor,-..4,er !'irs

, t-Zt i C 71' "

40, c 4F121:11. trt!* . iri I rr 'et! Ft, I

c ?' 110,., 1:r* 1101 ,t,r. 16,4.

( !

. c .;..Ii1.47 ".::/rt. It"

' i Ltl tiirtit-ite :4'1.r- ei

C .14, t*. ,r

. r t-",1,41'.1 f cs.',11c -'01t, ./"St. 1' .1'1 1117 J #..J ;!t.-"1'7.

21'. 1.--10:-..; ! st ,;.iP',1114. K.

tc S

I. . ,!c,, lir 1

'.17: ,F

z v 5C: t1.1.14

,c 141:".C1, A 3'0 irj- 111 ki,5 .3

Id+ t 7 1:...,0 cc- 4.11.1.i :F1

'1.. 1111.-.,'71.17":.: 0' t071:t'..1C,(1 I

41, -I: ,11.1 V :11V .1 p

rfcel/fetra t ketto two aped bi.riptti dietrctleoltixI 4r. qt-

r 1 y 1,-

n .? 1

I 10, C

it the directions we unclear or incomplete,be reduced because the measurement wont

tx eetesistent Mary Lou will answer an item this waybvicaiuse she thinks she is supposed to, while Johnnyivi answer it a completely different way, again be-:.ause he thinks nes supposed to Should this occur..4.edaeres scores will not truly indicate whether theyki-,ew !he material er they didn t

09e c;:xxi way to ensure that your test directions./41- cear sirep;e is to test them out Givetne ;)'regions and a sample item !O a student who is

cy.iing to be taking the test Ask the student tothe 0,rections and answer "ti_ item accordingly

Ye. may want to lest you. direct ens witn severala e, gent students Just to be on tne safe side. Once

have developed and tested your directions for aea, le u44 kind of test 'tern. /0..: can reuse those di-,ielioes lot other ii ens of the same kind

ldtton it is eeen a good idea to include a;,,r,plit .4(re the 'A-meet arr,wer appropriately1,1,;1 .c41,5 or)(i w,tri you' !st directions An example

etaGents with ark..:.her s,.;hortunity to be sureeecitestand you dire-7,.71 and a nr,wP! Rprq)

p1 7,114, is especidiy necessary if students areiet to the Kind of test item. the way you want

.1 7' 11(I.$ rl:4fred on their ciperr. and so or

e);,;c o.rections Cxp!ain adminis-teievre dela pertaining to the whop'; test. 'rbu shouldle -Auricles how mice time is all,iwed for the test:flew rnuoi time they should spend on various parts;,oee iii , point 441ac . )I the test the parts, and the

items in the parts

t,...oents neer! to whether they arettc!ir answe,r, in the test paper or on a:Ari%Wcir !..ficret and exactly how to mark theirYou should tel them whether they are to

4,,4 the appropriate word in the bank in each item,tee nurre- sr of the correct response, place an

(...te T a; ,-)lurnn if the statement ir; true.;40,31e.-1:1 give separate directions for

ei,e0iee aritwee, witn examples if appropriate, forr.',Irterriet kind et ilern that you use on the te3t.

-..,HOuld also be told what to do during thetest 1 thce. houll nave additional questions Theire,aeetioos !.houle be answered, out in such a wayell! it dere. ont disturb It rest of the class For t.IX-;1,,:4);( yOu ;.;(3,,J1(1 ask students to fai5cy their hands

ter, nave Qi.nr'.110fi that you can go tothe, sitene and answer Such questions

.01.f aeit quiet. basis

Page 15: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

Do not use too many different kinds of items.Authorities generally agree that you should use nomore than three different kinds of items (e.g., true-false, essay, matching) on a single test. In addition,all items of the same' kind should be grouped to-gether.

Students need to develop a particular mindset inorder to answer a particular kind of item. Achievingthat mindset may be difficult and time-consuming forsome students. You shouldn't, therefore, requirestudents to change their mindset more often than isnecessary--unless you also want to measure theirability to change modes of reasoning and answeringrapidly on command.

If you find that you have prepared a test with toomany kinds of items, you have two possible solu-tions, both of which are simple and easy. First, youmight rewrite some of the items to convert them fromone type to another. You could inc J de these reviseditems with others of the same type, ending up withfewer kinds of items. Second, you could split yourtest into two tests and give students the two testson two different days.

And note, please, that when you use different typesof items on a single test, you should number all itemsconsecutively (1, 2, 3, and so on) from beginning toend. Do not start numbering items in each new sec-tion with the number 1. When you are discussing theitems with students, either during or after the test, itis much simpler if there is only one item per number.Otherwise, there can be confusion"Which num-ber 1 do you mean, the essay item or the true-falseitem?"

The test should be just the right length. Deter-mining the right length for a test involves a compro-mise between validity, on the one hand, and reliabil-ity and usability, on the other. We discussed previouslyhow the validity of a test depends on how compre-hensively the content is sampled. The larger the bodyof knowledge to be tested, the lengthier the testshould be if it is to be valid.

In some situations, then, you might naturally betempted to prepare very long tests. A final exami-nation, for example, might test for knowledge of largequantities of information. ',/ou could logically con-clude that the best approach is to use a large quan-tity of items. The more comprehensive the sample,the more valid the testright?

Page 16: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

That is true. However, if the test is too long, relia-bility can suffer. The longer the test, the more likelyit is that students will become bored or fatigued. Fa-tigue and boredom are two student variables thataffect reliability, because they affect the consistencyof the students' performance in test-taking. If stu-dents are tired or bored, they wil! be less likely toremember what they know. They may confuse theirmemories of one bit of knowledge with another.

You will need to use your own judgment and yourknowledge of your students, their needs, and theirabilities to determine the right leinth for a given test.You should know your students well elough to knowapproximately at what point fatigue and boredom willbegin to adversely affect their test-taking perfor-mance.

Remember, too. that the best compromise onlength often depends on the situation. The test mustbe long enough to be valid for your purposes, yetnot so long that reliability is reduced.

Watch for the guessing factor. There is alwaysthe possibility that a student could take a blind guessat any test item and get the right answer out of thinair. If a multiple-choice item has four choices, stu-dents have a 25 percent chance of getting the cor-rect answer without even reading the item.

Students have a 50-50 chance on true-false items,which only offer two choices. Even in a matchingitem listing five inventions and the names of five in-ventors, the chances are only one in fiveandsometimes the last item is free, by process of elimi-nation.

One kind of guessing that you can and shouldprevent, however, is reaso ed guessing. When youmake a reasoned guessan educated guessyouuse whatever clues you can find in the item, to-gether with your general knowledge, to figure outwhat the answer should be.

Multiple-choice items, for instance, often allowstudents to guess the right answer by process ofelimination "Well, I know for sure that it's not A,and I've never heard of B before, and C just doesn'tsound right, so it must be D." To minimize the suc-cess of educated guesses, you will need to ensurethat each of your items is carefully devised and freeof clues to the correct response.

14

Now, in som situations, you may want to ensurethat students cannot guess the correct answer be-cause it is eEssential that they know the informationbeing tested. Let's say tt.at you're testing students'Knuwiedge of safety rules and practices at the be-ginning of the year or term. You want to be abso-lutely sure that students know their safety rules andpractices before they try using any tools, equipment,or machinery. In this situation, if they don't know, theymight injure themselves.

The best approach when you need to be sure thatthey know is to avoid selection items. This meansthat you will have to decide whether to use essayitems, oral items, or completion items. Completionitems would be the most likely candidate. Your testcan be valid, reliable, and usableand studentscannot guess.

Make good copies for everyone. If you are ad-ministering a written test, each student should havehis or her own copy. So, be sure that you have enoughwell-produced copies of the testneat, clean, andfree from grammatical or typographical errors.

A sloppy production job (e.g., misspelled words,incomplete sentences, poor copies) invites studentsto approach the test haphazardly. It also creates onemore opportunity for students to misunderstand yourtest directions or items.

Create a favorable testing environment. Thephysical environment in your classroom or labora-tory should allow students to concentrate on takingthe test. The room should be reasonably quiet andfree from distractions. The temperature should becomfortable, and lighting should be adequate. Ven-tilation should be sufficient to keep the air in the roomfresh.

The testing environment includes psychologicalfactors as well. Some students become very anx-ious and apprehensive when taking a test, regard-less of whether they know the material being testedor not. Or, students may be apprehensive becauseyours is a new kind of test for themperhaps you'regiving them their first essay items. You will often beable to minimize such apprehension by making thetesting environment nonthreatening.

One way to do this is to make sure that studentsunderstand the purpose of the test. Is the test to be

15

I

Page 17: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

graded, and are grades to be recorded? Or is thetest an ungraded review of material? You should alsotell students in advance exactly what material is tobe covered in the test. Otherwise, some studentswill undoubtedly waste time trying to decide what tostudy--becoming, in the process, more and moreapprehensive about the testing situation.

Another way to make the environment nonthreat-ening is to give students a practice test. This maybe especially important if you are using a type of testitem (e.g., essay) that your students are not used to.Practice tests give students the opportunity to gainexperience with the type of test item and see thatthey can be s=cbssful at it.

Many testing experts also recommend puttingeasier questions at the beginning of your test. Theirlogic is that, since all students should be able to an-swer these easier questions, all students should beable to experience some initial success in taking thetest, thereby decreasing their anxiety.

You will probably never be able to eliminate all un-favorable psychological factors in test-taking. Youcannot control what might happen to your studentsoutside your own classroom or laboratory. Any stu-dent may have just had an argument with friends orfamily and arrive to take the test in a distraught state.The possibility of these random psychological fac-tors makes it all the more important that you do whatyou can to create a favorable testing environment.

Plan your test carefully. You will recall that thesingle most important characteristic of a good test isvalidity. A major factor affecting the validity of a testis how comprehensively it samples students' knowl-edge. Careful planning can help you to develop teststhat are comprehensive samples of your students'knowledge.

The steps in planning a test are few and simple.First, you should identify the specific cognitive stu-dent performance objectives for which you want tomeasure student achievement. Second would be tolist those objectives on a sheet of paper or planningform. This list of cognitive student performance ob-jectives should describe the content to be coveredby your test.

Third, you should review your daily lesson plansor learning packages, if necessary, to identify addi-tional, more specific content to be included in the

test. Sometimes, student performance objectives arenot stated in enough detail to identify the specificcontent to be tested. Should you find that to be case,the greater amount of detail provided in your lessonplans or learning packages will be helpful.

Fourth, you will need to determine how many itemsto develop for each objective you have listed. Differ-ent objectives may well need to be covered by dif-ferent numbers of items. How many items to includefor a single objective depends on how important thatobjective is compared to the other objectives cov-ered in the test.

Some indicators of this relative importance are (1)the amount of time spent on instruction for each ob-jectiva, (2) the amount of emphasis placed on in-struction for each objective, and (3) the amount ofmaterial covered by each objective.

There may be other indicators of relative impor-tance to consider in your own situation. In any event,the more important or extensive an objective is, ac-cording to these indicators, the more items you shouldinclude to cover it. Furthermore, the relative impor-tance of objectives should not come as a surprise toyour stun

In determining the number of items to 'nclude foreach oojective, you are determining the relativeweight to be given to different content in your test. Ifyou choose, you can carry your planning forwardanother logical step. You can determine the pointvalue or other weight to be assigned to each item orcluster of items.

As you do so, it is very important to rememberthat point value should reflect the same relative im-portance of the objectives. If Objective A is coveredby twice as many items as Objective B because it istwice as important, then the total point value of theitems for A should be twice the point value of theitems for B.

Sample 1 illustrates one planning form that youmight find convenient to use in organizing your testplanning efforts. You can list your cognitive studentperformance objectives or other descriptors of con-tent in the left-hand column. The number of items tobe developed for each objective can be listed in thesecond column; the type of items (e.g., multiple-choice), in the third column; and the total point valuefor these items, in the right-hand column.

Page 18: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

SAMPLE 1

TEST PLANNING FORM

Cognitive Student PerformanceObjectives

No. ofItems

Type ofItems Point Value

161?

I

Page 19: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

OptionalActivity

4%11 /10

You may wish to read one or more of the following supplementary references:

Green, Teacher-Made Tests, has many useful chapters. Chapter 2, pp.19-41, discusses the levels of Bloom's taxonomy and the general prin-ciples of test construction.

Gronlund, Measurement and Evaluation in Teaching, is a scholarly anddetailed look at all aspects of evaluation. Chapter 3, pp. 65-92, examinesvalidity in depth and discusses methods for determining validity. Chapter4, pp. 93-120, treats reliability and formulas for determining coefficients ofreliability.

Karmel and Karmel, Measurement and Evaluation in the Schools, is areadable and informative compendium on evaluation. The chapter onteacher-made tests, pp. 379-409, compares the advantages and disad-vantages of different test types.

The following items check your comprehension of the material in the infor-mation sheet, Student Cognitive Performance Assessment, pp. 6-16. Eachof the four items requires a short, essay-type response. Please respond fully,but briefly, and make sure you respond to all parts of each item.

SELF-CHECK1. In chatting with a colleague one day, the question of tests comes up. The colleague reports,with obvious

pride, that no student ever scores 100 percent on a test in his program. How would you respond to thiscomment?

2. Another of your colleagues reports, in jest, that she plans to test her students' knowledge of a particularbody of content soon. However, she has become tired of using that same old time-worn measurementtool, the written test. Instead, her tool will be a yardstickshe will take each student's height as ameasurement of his/her knowledge.

A silly suggestion, granted, but how valid would her technique be? How reliable would it be?

17

Page 20: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

3. Today on your way out of the building, you overheard two students talking about a unit test they justtooka test consisting of one essay item. One student WES pleased because she had concentratedher studies on the topic covered by the item.

The other student was very angry and was complaining to the whole world about it. He had studied allthe material covered in the unit and, consequently, didn't have as much to say about that one topic.

Which student should get your sympathyand why?

4. One day, a colleague stops to ask you a question. She is very concerned about the possible effects ofcommunication skills on students' performance in taking written tests. Consequently, she wonders if itwouldn't be better to test knowledge solely through observation of students' performance in hands-onactivities in the classroom or laboratory?

"Wouldn't it be logical," she says, lo assume that, if students performed hands-on activities correctly,they must have known whatever factual information was required for the activity? Why not have stu-dents demonstrate their knowledge only in this way?" How would you respond to this suggestion?

19

18

Page 21: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

Compare your written responses to the self-check items with the model an-swers given below. Your responses need not exactly duplicate the modelresponses; however, you should have covered the same major points.

MODEL ANSWERS1. You could probably assume that there is some-

thing wrong with the teacher's tests. Even thoughyou don't know anything about what type of testitems the teacher uses or how he administers orscores his tests, the following line of reasoningshould apply.

Instruction in technical content should be appro-priate for students' needs, abilities, and interests.Instruction should be of high quality. Tests shouldcover the same content as instruction. Test itemsshould be clearly and simply stated. If these con-ditions have been met, then some students shouldscore 100 percent at least some of the time.

Not all students will get all answers correct, ofcourse. Variations in individual needs, abilities,and interests will result in variations in test scores,as long as the test differentiates between thosewho know the material and those who don't.However, if no student ever gets 100 percent, youare forced to suspect that some of the three pre-vious conditions are not being met.

Possibly, none of the conditions are being met. Ifinstruction is inappropriate or of poor quality, stu-dents' test performance will be just as poor. Ifstudents are being tested on content for whichthey received no instruction, they are not likely todo well. Finally, if test items are purposely ambig-uous and misleading, then students will in allprobability be misled by them and perform poorlyon tests.

You might give this teacher some recommenda-tions for improving his tests. You could tell theteacher that he should base his tests carefully onhis student performance objectives, just as heshould base his instruction on them. The teachershould also avoid trick items that must be readand interpreted in only one obscure way. Atten-tion to these two guidelines could improve thevalidity of his tests.

Finally, the teacher should probably evaluate hisinstructional effectiveness. Although the problemmight lie with his tests, it might just as well liewith his teaching.

2. How tall is knowledge? This teacher's test wouldhave no validity. Whatever measurement tool theteacher uses must measure what it is supposedto in order to be valid. The tool she has chosenin jest could not measure knowledge. All theyardstick can measure is feet and inches, butknowledge doesn't come that way. How couldJohnny be 5'6" smart?

In spite of nonexistent validity, however, her tech-nique would probably be highly reliable. The toolmight not measure what it is supposed to, but itwould probably measure what it does measurevery consistently. She could give that sameyardstick to another person and have him/herperform the same measurement on the samestudents. Both sets of results would most likelybe very similar, if not exactly the same.

Or, if she performed the same measurement onher students a second time, the results shouldagain be very similar or the same. (This assumesthat students haven't grown an inch or two smarterin the intervening time.) The problem of no valid-ity would always remain, however. No matter howmany times the teacher measured a student withthe yardstick, she would still be measuring heightand not knowledge.

3. In this instance, your sympathy should probablygo to the student with the complaint. The youngman most likely has a very good point. First ofall, the instructor's directions in announcing theupcoming test should have made very clear thescope of material to be covered and the types ofquestions to be included.

Assuming that the instructor simply indicated thata unit test was to be administered, then the youngman would have been justified in studying, andexpecting to be tested on, the full range of ma-terial covered in the unit.

Any test must contain a comprehensive sampleof the content being covered if it is to be valid. Ifthe student's test was supposed to measure hisknowledge of an entire unit, a sampling of all thekey content of that unit should, in some reason-able way, be included in the test.

19<0n.

Page 22: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

frrankiy, it is ainicuit to imagine now ail the con-tent for an entire unit could be covered in a singletest item. Admittedly, an essay item does haveroom for a good bit of content. But can one es-say item require students to demor strate all thelearning outcomes described in all the cognitivestudent performance objectives included in thatunit?

Logic alone should be sufficient to answer thatquestion: No. The teacher question has prob-ably prepared a test with low validity because itwas not a comprehensive sample of the contentto be tested. Perhaps one essay item could mea-sure achievement of one cognitive student per-formance objectiveone day's lesson, for in-stance. But a whole unit's worth of contentdeserves a whole unit's worth of test.

4. Your answer to this colleague should be that theobservation idea sounds good on the surface butcould be extremely difficult to carry out in prac-tice. The teacher's logic is pretty sound but prob-ably just hasn't been carried far enough.

You could construct a strong logical argument forusing the observation approach. If it is certain thatstudents must have certain knowledge to per-form certain activities, then it is logical to con-clude that successful performance indicatesknowledge.

Let's say that a business and office occupationsstudent has successfully typed a given sequenceof letters twenty times without error. This wouldsurely seem to indicate that the student hadknowledge of the location of the keys on thekeyboard.

But what if the student doesn't perform correctly?Should this be taken as an indication of noknowledge? Can specific bits of knowledge belinked definitively with specific aspects of perfor-

mance'? A great aeal or time ana work wow] neeato be spent establishing the relationship betweenbits of knowledge and bits of performance.

Furthermore, although ability to perform relies toa great extent on acquisition of the relevantknowledge base, not all vocational-technicalcontent can be measured through performance.When the auto mechanics instructor provides in-formation on the evolution of today's engine, heor she probably doesn't expect students to per-form on those early engines described. But he/she might want to test their knowledge, througha cognitive test, of how today's engine evolved.

In addition, instructors frequently want to testknowledge before permitting students to per-form in order to further ensure that they performcompetently and safely. Granted, a student canknow the safety procedures and choose to ig-nore them in practice, so you would want to mea-sure safety performance also. But first, you wouldwant to make sure that knowledge of the pro-cedures had been acquired.

Another objection to the colleague's approach isthat it would have low usability compared to otherforms of knowledge testing. Observation wouldrequire large amounts of instructor time, sinceeach student's performance would need to beobserved individually.

In the final analysis, other types of tests are knownto be effective in testing students' knowledge withfar greater efficiency and practicality. Communi-cation skills can indeed cause problems in test-ingthe teacher was rightfully concerned aboutthis issue. Nonetheless, those problems can besuccessfully addressed by constructing test itemsand directions using a language level appro-priate to the students taking the test.

Level of Performance: Your written responses to the self-check items should have covered the samemajor points as the model answers. If you missed some points or have questions about any additionalpoints you made, review the material in the information sheet, Student Cognitive Performance Assessment,pp. 6-16, or check with your resource person if necessary.

21

20

Page 23: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

Learning Experience IIOVERVIEW

EnablingObjective

After completing the required reading, construct five multiple-choice items tomeasure the achievement of selected cognitive student performance objec-tives.

You will be reading the information sheet, Multiple-Choice Items, pp. 22-27.

Activity

2

You will be selecting one or more student performance objectives that ex-press or require cognitive learning.

Activity

3

You will be constructing five multiple-choice items to measure achievementof the student performance objectives you select.

You may wish to ask peers to give suggestions for the improvement of themultiple-choice items you develop.

You will be evaluating your competency in constructing multiple-choice items,using the Checklist for Multiple-Choice Items, p. 29.

2122

Page 24: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

Multiple-choice items are among the most popular and versatile objectivetest types. For information on howand how notto construct multiple-choice items, read the following information sheet.

MULTIPLE-CHOICE ITEMSMultiple-choice items are selection items in which

students are asked to choose the correct responseto a given item from a short list of given responses.Each item consists of a stem, which supplies thecentral question or problem to which students are torespond. Each item also contains responses, fromamong which students must choose the correct re.sponse. Incorrect responses are called distracters,and the correct response is simply called the an-swer.

Multiple-choice items can appear in a variety offorms. First, the stem can be a complete directquestion, as in the following:

Which of the following systems does the stom-ach belong to?

1. skeletal2. respiratory3. nervous4. digestive

Or, the stem can be in the form of an incompletestatement, as follows:

A neutered male horse is called a1. gelding2. pinto3. palomino4. roan

Likewise, students can be asked to respond in anumber of ways. Most commonly, students are askedto choose the one correct answer to the item. In thiscase, only one of the responses can be correct. Allof the distracters must be completely incorrect. Inthe two examples used previously, there is only onecorrect answer.

Students can also be asked to identify the bestresponse from among those listed. In this case, thedistracters may be partially correct, but they are notas good as the answer. Students might be asked, aswell, to identify all correct responses from amongthose listed.

Uses of Multiple-Choice ItemsOne reason that multiple-choice items are popu-

lar is that they are versatile. They can be used tomeasure student achievement in all six levels of the

cognitive domain. It is true that the higher the levelof knowledge to be measured, the more difficult it isto construct an item properly.

You will find, however, that with practice and care,you can write multiple-choice items that require stu-dents to go far beyond mere recognition of specificfacts. Sample 2 contains multiple-choice items gearedto each of the six levels of the cognitive domain.

Constructing Multiple-Choice ItemsIf you have ever tried to write multiple-choice items,

you probably know how difficult that can be. Like-wise, you probably know how frustrating it can be totry to figure out the meaning of an unclear multiple-choice item if you are the test-taker. You will find ithelpful to use the following basic guidelines in con-structing multiple-choice items.

The stem should be significant. Some testingexperts say that the best stem for a multiple-choiceitem is one that would make a good completion itemif you didn't supply responses to choose from. Thefirst item in sample 2 has a stem that fits this rule.As you can see, the item could be converted to acompletion item simply by dropping the choices andadding a blank at the end.

Not all experts agree that it is necessary to bequite this strict in writing stems, but all do agree thatthe stem should state the question or problem suffi-ciently so that only the right choice is justified as theanswer. Only if the stem states the problem or ques-tion clearly can students be sure of what the item isasking and how to answer. The following is a poorexample of a stem:

The ignition in a car1. supplies current to the spark plugs2. is part of the electrical system3. consists of distributor, points, and coil4. is likely to be electronic today

In fact, all the responses in this example are cor-rect. Students would have an extremely hard timeanswering the item if they had to choose the onecorrect response. They would have to rely on theirknowledge of the teacher's own way of thinking andteaching in order to guess which of these correctresponses was the "right" one. This, of course, would

22 23

Page 25: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

SAMPLE 2

MULTIPLE-CHOICE ITEMSFOR SIX COGNITIVE LEVELSLevel ItemKnowledge Frying is a form of cooking by contact with

1. dry heat2. hot oil3. flame4. steam

Comprehension A corner joint in which all crosscut surfaces are concealed is a1. butt2. dovetail3. miter4. rabbet

Application If you have determined that you will need 10 six-foot lengths of 1" x 6" stock,how many board feet will you need to buy?1. 152. 303. 604. 120

Analysis The mare is to the stallion as the ewe is to the1. ram2. lamb3. wether4. mutton

Synthesis If you were preparing a chocolate pudding using high heat, no stirring, and un-beaten eggs, the result would be1. lumpy texture2. smooth texture3. curdling4. soft consistency

Evaluation Which of the following breakfast menus is nutritionally well balanced?1. orange juice, frosted cereal, skim milk, apricot Danish2. fried eggs, hash browns, donuts, coffee3. tomato juice, coffee with cream, pancakes and syrup4. orange juice, soft-cooked egg, whole wheat toast, skim milk

Page 26: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

not test students' knowledge of ignitions, but rathertheir knowledge of the teacher's personality. In otherwords, the item would not be valid.

This stem could be improved by asking somequestions about it. For example, what knowledgeabout the ignition are you trying to evaluate with theitemits function, its component parts, or its tech-nology? The stem could be rewritten as follows toevaluate students' knowledge of the function of ig-nitions:

The function of the ignition in a car is to

If the stem contains opinions, say so. You mayhave occasion to test students' knowledge of opin-ion. You may justifiably want students in an auto-motive mechanics program to know that someexperts think that diesels will replace internal com-bustion engines within 25 years. Consider the differ-ence, however, between the following two ex-amples:

Diesels will replace internal combustion en-gines withina. 6 yearsb. 10 yearsc. 20 yearsd. 25 years

According to the June 1982 issue of DieselWeekly, diesels will replace internal combus-tion engines withina. 6 yearsb. 10 yearsc. 20 yearsd. 25 years

In the first example, students must essentiallyguess what your opinion would be so that they canagree with it. In the second, you have identified aspecific authority as the source of the opinion, sothat students can use their knowledge of this au-thority's opinion as the basis for their answer.

Don't end your stem with a give-away. Some-times, the last word in the stem can give the answeraway if you are not careful in writing your re-sponses. In the following example, notice that thestem ends with the word an.

A physician who specializes in the structure,functions, and diseases 01 the eye is an1. hematologist2. optician3. optometrist4. ophthamologist

Many students will be able to rule out the first re-sponse just on the basis of their knowledge of theEnglish languageyou don't say "an hematologist."One way to remedy this error is to end the stem atthe word is and to include the word a or an, as ap-propriate, in each response.

All responses for a single stem should use thesame grammatical form. This rule is closely re-lated to the preceding one. If your stem is an incom-plete statement, each response should be stated ina form that correctly completes the statement in thestem. In the following example, students could againuse their knowledge of the English language to ruleout the second response.

Farmers rotate their crops in order to1. spread out the work load2. ease of marketing3. conserve the soil4. balance their diet

Once again, the stem does not combine with thesecond response to form a grammatically correctsentence in English.

All responses should be about the same length.Item analysis has repeatedly shown that the correctresponse tends to be longer than the distracters.Often, the correct answer must be qualified with ex-tra words and phrases just in order to make it true.If your answers are usually longer than your distrac-ters, however, test-wise students are provided witha clue to the correct response.

All distracters should be plausible. You might,for example, be writing a multiple-choice item to teststudents' knowledge of the name of an inventor (e.g.,The sewing machine was invented by). If so, Chris-topher Columbus is not a plausible answer, at leastfor secondary or postsecondary students. You couldexpect any teenager or adult to be able to rule outColumbus on the basis of general knowledge. Youwouldbe presenting, again, a clue to the correct re-sponse. An excellent source of plausible distractersis the incorrect answers students have given to re-lated completion items in the past.

Avoid using negative statements in the stem.Logically, it might well seem that there should be noproblem in the following example:

Which of the following woods is not a hard-wood?1. maple2. oak3. cherry4. pine

24 25

Page 27: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

The questioc, does not appear difficultfour woodsare listed, one of them is not a hardwood, so whichone is it? Research shows consistently, however, thatstudents do less well on multiple-choice items thathave a negative statement in the stem. Apparently,the small word not is often simply overlooked in thepressure of the testing situation.

Furthermore, most experts agree that it is less thanideal to emphasize negative learning. Why not, theysay, rewrite the previous example so that the stemast Which of the following woods is a softwood?Now, the item emphasizes the positive fact that pinebelongs to the softwoods.

The same experts also agree, however, that it maysometimes be necessary to test for negative knowl-edge (e.g., Which of the following things should youNOT do when using a radial arm saw?). When thisis so, you should at least underline the negative wordor write it in all capital letters, to make it stand outas much as possible.

Finally, in avoiding negatives, you should neverwrite a stem containing a double negative. espe-cially a combination of not and a negative adjective(e.g., not impossible or not unlikely). This, again, onlyserves to introduce irrelevant difficulty into the item.Such an item may test students' skill in using theEnglish language more than their knowledge oftechnical content.

Avoid any pattern of response. Some teacherstend to put the correct response in the middle of thelist, apparently because they feel that putting it inthe first or last position will make the correctness ofthe answer too obvious. Test-wise students, how-ever, will notice this pattern and use it to rule outincorrect responses.

Likewise, students are likely to notice any otherpattern you might use. It is best to go back to eachitem after you have completed writing all items andto distribute all responses randomly. In that way, youcan ensure that there will be no pattern of re-sponses for students to use as a clue.

Don t give clues to one item in another. An ob-vious example of this would be the following twostems:

The catalytic converter, in common use since1975, was invented byWhen did the catalytic converter first come intocommon use?

The example probably seems extreme, especiallysince the two items are right next to each other.However, you can very easily write two such itemsand not notice it yourself when you have written sev-eral other items in between. It is best to review allitems together to ensure that none of them containclues to any of the others.

Avoid all of the above and none of the aboveresponses. Using these responses can, once more,provide clues to experienced test-takers. Studentscan assume that, if they find two responses that theyknow to be correct, the answer has to be all of theabove. On the other hand, if students can find oneanswer that they know is correct, they can automat-ically rule out none of the above. They may, there-fore, select the correct response not through knowl-edge, but through the process of elimination.

Fit the item to the objective it covers. If yourobjective is for students to apply knowledge, thenyour multiple-choice item should actually test theirability to do that. The item must present a new sit-uation in which students are to apply knowledge. Ifyou use the same example on the test that you usedin a classroom discussion, all you are testing is stu-dents' recall of your earlier example. You are nottesting students' ability to Jse information in a situ-ation different from the original learning context.

Use four or five responses in each item. Themore responses there are in an item, the less likelyit is that students will be able to get the correct an-swer by guessing or by process of elimination. Fouror five items seems to be the best number. The ele-ment of pure chance is reduced, yet all the re-sponses ci:in still be read fairly quickly.

Keep all multiple-choice items together. If yourtest contains more than one kind of item, you shouldkeep all the items of one kind together. This keepsto a minimum the number of times students have toget the right mindset for the type of item. They don'thave to switch back and forth from a true-false itemto a multiple-choice item to a completion item andback to another multiple-choice item. If you mix yourtypes of items all together this way, you are alsotesting students' ability to switch rapidly from onemode of reasoning and answering to another.

Give clear, simple, complete directions. Yourstudents are very likely to know that there are differ-ent ways to answer multiple-choice items. You mightwant them to choose the one correct answer, theone best answer of several that are all correct tovarying degrees, all possible correct answers nomatter how many, and so on. Consequently, it is im-portant that you state clearly and completely howstudents are to respond.

You should also be sure to tell students how tomark their answers. Depending on how you plan toscore the finished tests, you may want students tomark their answer sheet in different ways. For ex-ample, if tests are to be computer-scored, studentswill need to be told to use the computer answer formand a #2 pencil. Or you can ask students to placethe correct letter or number in a particular blank, tomark an X through each correct response, or to circlethe correct response, and so on.

25Co

Page 28: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

,f .aria1,10t t hdAL,t1 007411:

Litt, 1, '1411 ,110,,,,p/1(11! ;1,0i W c. t ictr Wt1C11111/f

14 1 t.,` 47 tfylf 14 If lttr711,: A ' C at cr t: 4t4..14.t /Vt.'

.1 t.4 1.,11 t,p10 11111,t tI(NVIA El? A!ly t!V xil. 1v t: a'rr ioltoent>ii 1.0 rItl.,04)Itt

tevr,i't st ttl.attl ht, Vrt,r,tr. I. * Arht twls!, Illoy

an o;/,...Yi,lci 01 111,04 10C C

C

PVC

p*, oilt

ti., tet

* Skt Ming y itu.) ;3114100h rrt AtI1 thorn orOGA:lit, if Vl'.1.1

, cl ,3,11p,v vott ;V.- -+IliiiWqrtost:j1/ the rltirttycl! Iritcrii1011 y,t,fr Mny t he ;I Mum.

c. A thr, 100 wall a hiiI'n

, L,3' c.t . L nt t rt.pon'40 V1/4/1111n V0t1r-* r h ',1.itiorit 41'0,

o 1*iosoonskt !,liovvino thiough a hole should bei .1 with :;t1 of tosponse is wrong

A trotl:'ti may not work, howoyor, if you ask stu-[tont, to circle oach cotroct response In that case,thou ttliciht IV bigger than the holes in the

which wouid Yl)li to thatk ca correcttut:ipott-,r) az.; itit7(1tred

A toad to easiest to use students use anansvvw shoot with columns of numbers or letters forMA1.1t10 ,answers F=or this mason, some teachersalway:, put such columns in one margin or tho other.

t,,dents mad. their answers in this column, and not..n the choices actually listed in the item, as follows:

2 3 4 Murtie verlocit ordinarily ex-pressed hr1. yards per second2. yards pee minute3. foot per minute4. foot per second

26

a

I

Page 29: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

SAMPLE 3

DIRECTIONS FOR MULTIPLE-CHOICE ITEMS

Selecting a single correct response:Directions: Each of the items below is followed by four possible responses. For each item, only one ofthe responses is correct; the others are incorrect. Select the one correct response for each item. Indi-cate your answer by placing an X over the number of the correct response.

EXAMPLE: The commonly accepted industrywide standard for beginning -level typists is1. 40 words per minute, with a minimum of five errors2. 40 words per minute, error-free

X 45 words per minute, with a maximum of three errors4. 45 words per minute, error-free

Selecting a single best response::.;rections: Each of the items below is followed by four possible responses. For each item, any or allof the responses could be correct. Select the response that is best for each item. Indicate your answerby placing an k over the nurr.!29r of the best response.

EXAMPLE: You have just typed a letter for your supervisor, Mrs. Staunton. She gives it back to youand says that she has decided that she doesn't want to include the last paragraph in theletter after all. What would you do to remove that last paragraph:1. use correction tape2. cut and paste

X retype the letter4. use correction fluid

Selecting all correct responses:Directions: Each of the items below is followed by four possible responses. For each item, any or allof the responses could be correct. Select all the correct responses for each item. Indicate your answerby placing an X over the number of each correct response for each item.

EXAMPLE: You can thicken gravy withX cornstarch2. cornmealX wheat flour4. baking soda

27 23

Page 30: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

Select one or more student performance objectives in your occupational spe-cialty that express or require cognitive learning and that lend themselves tothe use of multiple-choice items to measure their achievement. (If you needhelp in identifying student performance objectives that require or expresscognitive performance, check with your resource person.)

Construct five multiple-choice items to measure achievement of the studentperformance objectives you selected. Number each item for easy referenceduring feedback. Include directions and a scoring key.

You may wish to ask one or more of your peers who have taken or are takingthis module to critique the multiple-choice items you have constructed. Dis-cuss any suggested changes and then make any necessary revisions.

After you have constructed your multiple-choice items, use the Checklist forMultiple-Choice Items, p. 29, to evaluate your work.

28

29

I

I

Page 31: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

CHECKLIST FOR MULTIPLE-CHOICE ITEMSDirections: Place an X in the NO, PARTIAL, or FULL box to indicate thateach of the following performance components was not accomplished, par-tially accomplished, or fully accomplished. If, because of special circum-stances, a performance component was not applicable, or impossible to exe-cute, place an X in the N/A box.

Name

Date

Resource Person

LEVEL OF PERFORMANCE

?° 4'D kThe multiple-choice items meet the following criteria:

1. A significant question or problem is stated in the stem of each item I i El El

2. All responses for each item are grammatically correct

3. All ElElElt he response choices within each item are approximately the same Ellength

4. An authority is quoted when the item contains opinion

5. There are no ambiguous statements in the stems or responses

6. All distracters for each item are plausible I I

7. The items are written at the language level of the students

8. The items do not contain clues to the correct responses for other items

9. The items are constructed to measure the level of knowledge speci-fied in the objectives E El El El

10. Stems do not contain negative items

11. Correct responses are randomly distributed

12. Each item contains four or five responses

13. Clear, sufficient directions are provided

14. An appropriate scoring key is provided

Level of Performance: All items must receive FULL or N/A responses. If any item receives a NO or PARTIAL

response, review the materials in the information sheet, Multiple-Choice Items, pp. 22-27, revise your itemsaccordingly, or check with your resuurce person if necessary.

'11 ,

29

Page 32: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

NOTES

3031

I

4

I

Page 33: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

Learning Experience IIIOVERVIEW

Activity

Activity

2

Activity

3

After completing the required reading, construct a matching item to measurethe achievement of selected cognitive student performance objectives.

You will be reading the information sheet, Matching Items, pp. 32-34.

You will be selecting one or more student performance objectives that requirestudents to recall or recognize information.

You will be constructing a matching item to measure achievement of the stu-dent performance objectives you select.

You may wish to ask peers to give suggestions for the improvement of thematching item you develop.

You will be evaluating your competency in constructing a matching item, usingthe Checklist for Matching Item, p. 37.

3132

Page 34: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

Another familiar and popular form of objective testing is the matching item.For information on what to do and what to avoid in constructing matchingitems, read the following information sheet.

MATCHING ITEMSLike multiple-choice items, matching items are a

form of selection item. The matching item consistsof two lists of elements. One list (usually the one onthe left) is made up of premises, while the othercontains responses. The students' task is to matcheach premise with the appropriate response on thebasis of a stated association or relationship be-tween the two elements.

Uses of Matching Items

There are many kinds of associations or relation-ships that can be used as the basis for a matchingitem. Students could be asked to match inventionswith their inventors, tools or pieces of equipment withtheir uses, terms with their definitions, principles withexamples of their application, and so on. Since youwill probably want to test students' knowledge of suchassociations or relationships in your owr, occupa-tional area, you will find matching items to be useful.

Furthermore, matching items can be constructedusing real objects or materials, pictures, drawings,or models. Instead of using words as your premises,you could use pictures for students to match re-sponses with. You could, for example, give horticul-ture students pictures of different leaves to matchwith the names of the trees they come from.

Or, you could use one large, clear drawing or dia-gram of a radial arm saw, with arrows drawn to in-dicate different parts of the saw. Students would thenmatch each part indicated by an arrow with the c r-rect name from the list of responses.

You can also construct matching items with threeparts instead of two. You could use a three-partmatching item for terminology in health care occu-pations. The list of premises r ould consist of Latinor Greek terminology. The first list of responses couldbe literal translations of the premises, and the sec-ond list of responses could be the actual definitionsof the terms. It is not recommended that you attemptto use matching items with more than three parts(one premise and two response:).

Matching items are really only appropriate, how-ever, for measuring student achievement at the low-est cognitive levelknowledge. Experts in testinggenerally agree that matching items are not a validmeasure of any of the higher cognitive levels.

Constructing Matching Items

While matching items may be familiar and popu-lar, it is very important to take care in constructingthem. When constructed carefully, matching itemscan be valid, reliable, and useful in evaluting studentknowledge. The following are guidelines for writinggood matching items.

Use a reasonable number of both premises andresponses. Neither of these two lists should be eithertoo long or too short. If there are too few elements,it becomes easier for students to guess correct an-swersespecially the last oneby process ofelimination. (If there are only three items in each list,you can get them all right even if you only know two.)On the other hand, students have to spend too muchtime reading the list over and over if there are toomany elements in the list.

Unfortunately, there is no single answer concern-ing how long the lists should be. Almost all the ex-perts agree that five is the fewest number of ele-ments you should put in a list. The maximum numberrecommended varies from ten to fifteen. It would ap-pear safest to keep to the middle ground and haveapproximately ten elements in each listcertainlyno fewer than five and no more than fifteen.

All premises or responses in one list shouldbe homogeneous. Stated simply, this means thatall the elements in a list, whether premises or re-sponses, should be the same kind of thing. You might,for example, want students to match two differentsets of things: technical terms with their definitions,and tools with their uses. This would best be doneby constructing two different matching items. In one,students could match terms with definitions. In theother, they could match tools with uses.

3233

Page 35: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

The whole matching item should be on a singlepage. Students should not have to flip back and forthfrom one page to the next to scan all the elementsin either list. This would probably confuse studentsand take too much time. If necessary, you shouldleave the bottom of one page blank and start yourmatching item at the top of a fresh page in order tofit it all on one page.

List enough responses so that some are leftover. If you have exactly the same number of prem-ises and responses, and each response is used onlyonce, many students will get the last one correct au-tomatically, by process of elimination. This reducesthe validity of the item, of course.

You can avoid this problem in one of three ways.The first is to use more responses than premisesfor example, six premises and nine responses. Thenstudents still have four possible responses left whenthey get to the last premise. Listing two or three moreresponses than premises is sufficient.

The second way to avoid the problem is to use asingle response as the correct answer for more thanone premise. You might, for example, have tenpremises and ten responses, with one response beingthe correct answer for two premises and one re-sponse not used at all. If you do so, you must tellstudents that a response can be used once, morethan once, or not at all.

One final variation is possible. You can state inyour directions that a response can be used once,more than once, or not at all. However, you can thenlist the same number of responses as premises, witheach response being used once.

This is a legitimate variation, since it does fit yourdirections for using responses. It can be effective ineliminating guessing by process of elimination un-der one conditionthat your matching tests do usea single response more than once often enough thatstudents know that this is a real possibility. If yousimply recite the formula for every matching item butnever really use responses that way, students willsoon learn that you don't really mean it.

List your responses in some logical order. Ifresponses are listed in a logical order, students canfind the response they are looking for more easilyand quickly. If the responses are dates, they couldbe listed in chronological order. If they are names orterms, they could be in alphabetical order. Otherpossible logical orders might be increasing or de-creasing size or importance.

The premises can be long, but the responsesshould be short. When students are actually takingthe test, they usually read one premise and then scanthe list of responses looking for the correct one. Asa result, they end up reading and rereading the listof responses. To save time, the responses that theywill be reading over and over should be short.

Avoid matching patterns. It is surprisingly easyto place your responses so that they fo. :n a patternin relation to your premises. You might do this with-out realizing it. For example, you might automati-cally put a response in the bottom half of the listwhen its premise is in the top half. Or, if the premiseis toward the middle you might tend to put the ap-propriate response in the middle also. Other pat-terns are also possible, and all should be avoided.Otherwise, some students will notice your patternsand find clues to correct responses.

One way to avoid matching patterns is to reviewthe entire matching item after you have finished writ-ing it. Look for patterns in the way elements match.If you find such patterns, you should use a proce-dure to randomize the order of the responses. Youmight roll dice, pick numbered slips of paper out ofa hat, number the sides of a pencil and roll it, orperhaps use the randomization function of a calcu-lator. Listing your responses in some logical order,as described earlier, may also serve to randomizethem.

Provide clear, simple, and complete direc-tions. You should not assume that your students willunderstand what to do with a matching item as soonas they see it. Nor should you assume that studentswill automatically know what the association or re-lationship between the premises and responses is.Students should also be told how to mark their an-swers.

Thus, in your directions, you should tell studentswhat the relationship in the item isfor example,that they are to match breeds of swine with theiridentifying characteristicsand exactly how theyshould mark their answers. Sample 4 presents amatching item with directions for its use.

Prepare a scoring key. A scoring key alwaysmakes the job of scoring easier and helps to elimi-nate errors in scoring. You can just write the num-bers and letters of the premises and the correct re-sponses down the side of a piece of paper. Then,for quick and accurate scoring, you can lay this keynext to the blanks in which students mark their re-ponses.

33 34

Page 36: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

SAMPLE 4

MATCHING ITEM AND DIRECTIONS

Matching Test:The Typewriter Keyboard

Directions: The list on the left describes the functions of different keys on the typewriter keyboard. The liston the right contains the names of different keys. Match the names of the keys with their correct functionsby writing the letter of the correct response in the blank to the left of each function. Use each responseonly once; the first item is completed as an example.

b 0.

Functions

Removes a letter, number, or symbol already printed

Moves the typing element quickly across the keyboard toa position of your choosing

Moves the typing element one space to the left withoutprinting anything

Prints capital letters (A) instead of lower-cased letters (a)

Moves the typing element one space to the right withoutprinting anything

Moves the typing element to the left margin and rolls thepaper up to the next line

Prints whole words or sentences in capital letters

a.

b.

c.

d.

e.

f.

g.

h.

Keys

Back space

Correction key

Margin release

Return

Shift

Shift lock

Space bar

Tab

1.

2.

3.

4.

5.

6.

35

34

I

Page 37: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

Activity

2

Select one or more cognitive student performance objectives in your occu-pational specialty that require students to recall or recognize information.

Construct a matching item to measure achievement of the student perfor-mance objectives you selected. Include directions and a scoring key.

You may wish to ask one or more of your peers who have taken or are takingthis module to critique the matching item you have constructed. Discuss anysuggested changes and then make any necessary revisions.

After you have constructed your matching item, use the Checklist for Match-ing Item, p. 37, to evaluate your work.

35 30

Page 38: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

NOTES

36 3 7

I

I

Page 39: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

CHECKLIST FOR MATCHING ITEMDirections: Place an X in the NO, PARTIAL, or FULL box to indicate thateach of the following performance components was not accomplished, par-tially accomplished, or fully accomplished. If, because of special circum-stances, a performance component was not applicable, or impossible to exe-cute, place an X in the N/A box.

Name

Date

Resource Person

LEVEL OF PERFORMANCE

'gr. 43 4.* kJThe matching item meets the following criteria:

1. Only homogeneous premises and responses are grouped within theitem

2. There are not less than 5 and not more than 15 responses in the item

3. Premises are arranged for maximum clarity and convenience for thestudent

4. The responses are logically (e.g., alphabetically, chronologically) ar-ranged

5. The directions clearly indicate the basis for matching and the methodfor marking answers

6. Real materials, pictures, drawings, or models are used when practical

7. Perfect one-to-one matching in the arrangement of premises and re-sponses is avoided

8. All the premises and responses are on one page

9. There %re more responses than premises, or single responses can bematched with more than one premise

10. Correct answers are randomly distributed

11. Responses are shorter than premises

12. An appropriate scoring key is provided

Level of Performance: All items must receive FULL or N/A responses. If any item receives a NC orPARTIAL response, review the material in the information sheet, Matching Items, pp. 32-34, revise youritem accordingly, or check with your resource person if necessary.

Page 40: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

NOTES

38 3j

I

4

I

Page 41: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

Learning Experience IVOVERVIEW

Activity

Activity

3

After completing the required reading, construct five completion items to

measure the achievement of selected cognitive student performance objec-

tives.

You will be reading the information sheet, Completion Items, pp. 40-41.

You will be selecting onn or more student performance objectives that

require students to recall information.

You will be constructing five completion items to measure achievement of the

student performanceobjectives you select.

You may wish to ask peers to give suggestions for the improvement of the

completion items you develop.

You will be evaluating your competency in constructing completion items,

using the Checklist for Completion Items, p. 43.

39 40

Page 42: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

Comp: er on items can be used to great advantage in evaluating studentKnowleoge For information on how to construct high-quality completion items,read ttie tol!ow:ng information sheer.

COMPLETION ITEMSCompletion Items are supply itemsIlia!' is, stu-

dents must supply their own answers in5.te?.1 ofchoosing from given possible answers. Cohipictionitems can appear in two forms. One form is simply

sentence with a blank in place of onr, or more'hissing words. as in the following example:

The unit used to moasure electrical retdstanceis called the

Another form consists of a complete question re-Quinng a short answer (hence the name short-answerrest ometirnes u,ed for a test using these items).Students wale their answer in a blank provided atthe ervi of the question, as in the foilowilig example:

What we the parallel beams that support theroof of a structure called'

Uses of Completion ItemsCompletion items test students' recall of knowl-

edge Hence they are different from selection itemset, 9 multiplechoice and matching) in which stu-dents need orgy to be able to recognize the correctarviwet In this respect, completion items are similartcI et.oiay or oral items because all of these formsrequire sturienh to supply answers from their ownmorality possibility of guessing thus is mini-mirtert

(....,orrip4stion items can be used to test student re-ot many offerent kinds of factual information.

I troy can be used to evaluate students' knowledgerit terminology, classifications, methodol-

unr,l,,als, principles, and so on. Furthermore.tcy can bo u.ied in any content area However,i.;.,rnolittion item., are most appropriate hr use at the

level of the cognitrie don Ain.

Constructing Completion ItemsCare %tiould b4 taken in constructing completion

items it Lin iti9rn t not clearly stated, students mayinisinter ;get it and answer incorrectly, even though!hely might know the correct answer. On the otherharIl. a poorly wnlItIn OM may contain clues that a

..tudent can tP,0 in determining the correctfiuth 'J1)001 be ilVOic'od

40

Ir

The following are suggestions for ccitstructingcompletion items.

Use your own words. Many authorities feel thatthe main weakness of the completion item is that itonly measures recall of rote learning, thus encour-aging lower-level cognitive activity at the expense ofhigher-level activity. Lifting textbook quotes verba-tim to use as completion items encourages rotelearning even more.

Test Only for significant bits of knowledge. Youwill need to use your own judgment to determinewhat knowledge is significant in your own area. Usingthat judgniunt, you can ensure that your completionitems focus on significant knowledge in the area.Students might truly need to know that a particulartechnological development occurred in your field in1962. It is unlikely, however, that students need toknow that this development took place on aWednesday morning.

Have each item clearly call for oneand onlyone--answer. Perhaps the most common mistaketeachers make in writing completion items is in notbeing clear enough. Consider the following ex-ample:

The cathode ray tube with fluorescent screenwas first introduced by K.F. Braun in

How many possible answers are there to this item?The tube in question was first introduced by Braun:In 1897? In the laboratory? In Germany? Studentscould legitimately give any one of these three an-swers, And what if a cagey student wrote the an-swer "in modern times"is that right or wrong?

The item could be saved by rephrasing it. Stu-dents could be asked specifically in what year or inwhat country, for example, the cathode ray tube wasintroduced by Braun, as in the following examples:

In what year was the cathode ray tube withfluorescent screen first introduced?In what country was the cathode ray tube withfluorescent screen first introduced?

Don't go overboard with your blanks. Thisguideline is closely related to the previous one. If

Page 43: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

you put too many blanks in one item, there won't beenough significant information left in the item for stu-dents to know how to answer it. The following is anexample of too many blanks spoiling the broth:

In the entersthe

There really is no way for students to know howto complete this item. However, you could delete thefirst blank, replacing it with the word it stands for,and improve the item as follows:

In the lungs, enters the

The item is now clear enough that students shouldbe able to supply the correct answers (i.e., oxygenenters the bloodstream). If you have any questionconcerning how many blanks you can put in a givenitem, remember that fewer is usually better. If nec-essary, you should write several items, each with oneblank, to replace a single item with several blanksthat is impossible to understand.

Do not turn completion items into essay items.As stated previously, another name for the comple-tion item is the short-answer item. Completion itemsdiffer from essay items mainly in the length of theanswer required. You should ensure that your com-pletion items are not really essay items (even mini-essay items) in disguise. This point applies particu-larly to completion items in the form of questions.

Keep blanks in all items the same length. Inpractice, this means that all blanks should be longenoujh for the longest answer to be used in the test.You should never leave a short blank for a short an-swer, medium blank for a medium answer, and soon. Students will quickly understand that there is aclue to the answer in the length of the blank if youdo.

Don't give grammatical clues to the answer.Consider the following example:

The remove waste substancesfrom the blood

Many students will be able to use their knowledgeof the English language to help answer this item.

The verb remove in the item ir plural (he, she, or itremoves; they remove). Any student who realizesthis will also realize that a plural answer is requiredto be the subject of the verb. Otherwise, the sen-tence would be grammatically incorrect. This ex-ample could be rewritten as follows:

lottiste substances are removed from the bloodby the

There are no grammatical clues in this version.since we use the with both singular and plural nounsin English. You should also avoid having either a oran immediately before the blank. The indefinite ar-ticle a tells students that the next word starts with aconsonant, and an indicates that the next word startswith a vowel.

Place the blanks for easy scoring. It 's helpfulto put all the blanks in one column, in ea' ler the leftor right margin. Students actually write their an-swers in this column of blanks. A short blank in th.,middle of items can indicate which word or phrasestudents are to supply. The following is an example:

Explosive gas is used as 3an energy source in the

(3) engine.

Prepare a scoring key. This can make scoring alot easier, especially if you put all the blanks in acolumn. List the correct answers to the items as youwrite them. Then, as you score the test, you mayfind other answers that are acceptable in students'papers. You should add these to your key as well.

Assign the same number of points to eachblank. Each blank should test student recall of asignificant bit of knowledge. Each answer should thushave the same weight, since each item should be ofcomparable importance and difficulty. This also makescomputation of students' scores on the test easier.

Keep all completion items together. This guide-line always applies, no matter what type of item isinvolved. Students should not have to switch fromone type of item to anotherfrom one mode of rea-soning and answering to anotherany more thannecessary.

Page 44: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

Activity

2

Select one or more cognitive student performance objectives in your occu-pational specialty that require students to recall information.

Construct five completion items to measure achievement of the student per-formance objectives you selected. Number each item for easy reference dur-ing feedback. Include directions and a scoring key.

You may wish to ask one or more of your peers who have taken or are takingthis module to critique the completion items you have constructed. Discussany suggested changes and then make any necessary revisions.

After you have constructed your completion items, use the Checklist forCompletion Items, p. 43, to evaluate your work.

4243

Page 45: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

CHECKLIST FOR COMPLETION ITEMS

Directions: Place an X in the NO, PARTIAL, or FULL box to indicate that

each of the following performance components was not accomplished, par-

tially accomplished, or fully accomplished. If, because of special circum-

stan,:e.s, a performance component was not applicable, or impossible to exe-

cute place an X in the N/A box.

Name

Date

Resource Person

LEVELJOF PERFORMANCE

'. ze

The completion items meet the following criteria:1. Only significant words are omitted in incomplete sentence items El

2. Enough clues are left in the incomplete sentence items so that therequired responses make sense to the students El El

3. There are no grammatical clues to reveal the correct responses El El

4. All the blanks are the same length GI El

5. All omitted phrases or words are of comparable importance

6. Any question asked is specific enough to evoke the correct type of

responseEl El

7. Direct quotes from a textbook are avoided

8. All answers are a word, numeral, symbol, or short phrase

9. All answers have equal point value

10. An appropriate scoring key is provided

Level of Performance: All items must receive FULL or N/A responses. If any item received a NO or

PARTIAL response, review the material in the information sheet, Completion Items, pp. 40-41, revise your

items accordingly, or check with your resource person if necessary.

Page 46: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

NOTES

44 45

Page 47: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

Learning Experience VOVERVIEW

EnablingObjective

After completing the required reading, construct ten true-false items to mea-sure the achievement of selected cognitive student performance objectives.

You will be reading the information sheet, True-False Items, pp. 46-48.

Activity

2

You will be selecting one or more student performance objectives that requirestudent knowledge or comprehension of information.

Activity

3

You will t.onstructing ten true-false items to measure achievement of thestudent performance objectives you select.

You may wish to ask peers to give suggestions for the improvement of thetrue-false items you develop.

You will be evaluating your competency in constructing true-false items, usingthe Checklist for True-False Items, p. 51.

454 0

Page 48: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

Activity True-false items have perhaps the most definite advantages and disadvan-tages of the objective forms of testing. For information on how to constructtrue-false items, read the following information sheet.

TRUE-FALSE ITEMSTrue -false items are also called alternate-re-

sponse items. They are selection items in whichstudents must choose one of two given answers.Most commonly, students are asked to judge whetherstatements are true or false. Students can also begiven a list of questions to which they must answeryes or no.

Uses of True-False ItemsTrue-false items can measure student achieve-

ment at the knowledge and comprehension levels ofthe cognitive domain. Furthermore, students cananswer true-false items rather quickly, so they canbe used to sample a wide range and large amountof content. Finally, true-false items can be scoredcompletely objectively.

Against these advantages, however, you mustweigh one large disadvantage. Since students haveonly two answers to choose from, their chances ofguessing correctly c re 50-50. Because of this, true-false items are less ;pliable. They do not discrimi-nate as well between students who know the mate-rial and those who don't.

Constructing True-False ItemsYou can minimize the effects of guessing by fol-

lowing some basic guidelines in writing true-falseitems. You can avoid giving unnecessary clues thatstudents can use in guessing.

Keep items simple. The longer and more com-plicated the item is, the more you are testing stu-dents' reading ability instead of their technicalknowledge. Longer items also give students moreopportunities to "read into" the statement an inter-pretation you didn't intend. Simple items can be moreclearly true or false.

Furthermore, many teachers tend to write trueitems that are consistently longer than false items.This is because, often, a statement must be quali-fied or modified to be sure it really is true. Experi-enced test-takers soon realize that a long item islikely to be true, thereby improving their chances ofguessing correctly.

Make each item entirely true or entirely false.Sometimes teachers write items that are partly true,party false. Consider the following example:

T F The rip saw is best suited to cuttingalong the grain because of the angleof the grip to the blade.

The student who knows his or her rip saws will re-alize that part of this statement is truerip saws areused to cut along the grain. However, the studentwill also know that this has nothing to do with grip.Is this statement true for the cut or false for the grip?

Don't make items false on a technicality. Closelyrelated to the last guideline, this rule means that youshould test for significant bits of knowledge. Con-sider the following example:

T F Pasteur first inoculated a human, ayoung child, with his treatment of ra-bies on July 7, 1885.

The answer is falsethe inoculation occurred on July6. The item, then, becomes trivial. It is difficult toimagine why students would need to know the exactday of the inoculation. It would be equally trivial tochange the year to 1886. You will need to use yourown knowledge of technical content to determinewhen details are significant and when they are not.

Avoid qualitative terms in your items. When youare writing items, don't use words that express in-definite quantities (e.g., several, some, many). Thesemake the statement open to different interpreta-tions, as in the following example:

T F Several different thickeners can be usedin making gravy.

If several means two or three, this item would betrue. But if several means five or six, the item wouldbe false. You could improve the item by using aspecific number instead of the word several.

Quote authorities for opinions. You might usetrue-false items to test students' knowledge of opin-ion in your area. It may be significant that estab-lished authorities in the field believe that a techno-logical development will become widespread. But you

4746

Page 49: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

must tell students whose opinion you are quoting,so they can know how to answer. The following ex-ample would tell students what they need to know:

T F The editors of Keyboard and Consolepredict that word processing equip-ment will replace current technology inoffice operations within five years.

Avoid negative items. Negative statements aremore difficult for students to read and consequentlyfavor students with higher reading skills. You shoulda, aid items with negative wordsnot, never, noth-,ng, no, and so on. You should especially avoid doublenegatives, as in the following example:

T F Broken ears are a not uncommon de-fining characteristic of breeds of swine.

If students are unable to decode the double nega-tive not uncommon, it won't matter whether they knowthe information you are testing for. The item couldbe improved by stating it in positive terms, as fol-lows:

T F Broken ears are a common definingcharacteristic of breeds of swine.

Avoid specific determiners. Certain words areusually associated with either true or false state-ments. Statements are usually false if they containthe words always, never, all, or none. On the otherhand, the words sometimes, some, may, and shouldgenerally appear in true statements. If you use thesewords in your items, you are giving students an un-necessary clue to the answer.

Use equal numbers of true and false state-mnrits While you don't need to take this point liter-ally, you should have about the same number of trueand false statements. If you consistently have moreof one kind than the other, students can use this as

u,a in guessing. When you have finished writingall your items, you can count your true and false itemsand add items as necessary to make them roughlyeqt.

Prepare a scoring key. Scoring is very easy ifyou prepare a key. The best would be a stencilacopy r f the test with the correct answers punchedou . Ole T and F columns. You can place this over

ents' papers and see their answers at a glance.u use a stencil, you should have students mark

their answers by placing an X over the T or F. If theycircle their answers, they could make their circlesbigger than the holes in your stencil and you wouldn'tsee them.

Give simple and clear directions. You should al-ways tell students clearly and simply what they are

to do. Your directions should include how studentsare to mark their answers (e.g., place an X over theT or F). The following is an example of directions fortrue-false items:

Directions: Each of the following statementsis either true or false. Read each statement anddecide whether it is true or false. Indicate youranswer by placing an X over the T if the state-ment is true and over the F if the statement isfalse.

Watch out for patterns of answers. If you al-ways use two true items followed by two false items,students will soon recognize this pattern. You shouldreview the answers when you have finished writingyour items. If you notice any pattern in the answers,you can use various procedures to distribute the an-swers randomly (e.g., rolling dice or pulling num-bered slips out of a hat).

Don't quote from the textbook. This practice onlyplaces a premium on rote learning and does not en-courage higher-level cognitive activities. In addition,many textbook statements are ambiguous when re-moved from their original context.

Consider using modified true-false items. Be-cause responding to regular true-false items oftenbecomes a kind of guessing game for students,modified true-false items are often used when rec-ognition and recall of information are being mea-sured. There are many methods for modifying true-false items. The main disadvantage is in the gradingof the items, in that more time is required.

The modification consists of asking students tocorrect or give the right response to any false item.The value in the modification is that the students arerequired not only to recognize a false statement, butto recall the correct response as well.

Two forms of this modification are illustrated insamples 5 and 6. Notice that in both of these modi-fications, the word or words that make the statementtrue or false are underlined. In sample 5, studentsmust supply their own correct response. In sample6, each true-false item provides multiple choices forcorrecting the item if it is false. These are much moredifficult to construct, because you have to list theplausible choices for each true-false item.

In the form illustrated in sample 6, grading or giv-ing of points for the responses could be a problem.One suggestion is to give each item, whether trueor false, a value of 2. Recognition of a tri state-ment is rewarded with 2 points; a value of 1 is givenfor recognizing a false statement; and a value rfis given for recalling the correct response.

47 48

Page 50: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

SAMPLE 5

MODIFIED TRUE-FALSE ITEM

Directions: Circle the T or F in the column to the left of each true (T) or false (F) statement.Then, if the statement is false, correct the statement in the space provided under each state-ment.

T F 1. The horizontal threads in fabric are called the warp threads.

Correct statement:

SAMPLE 6

MODIFIED TRUE-FALSE ITEM

Directions: Circle the T or F in the column to the left of each true (T) or false (F) statement.Then, if the statement is false, select the correct alternative and place the letter in the spaceprovided in the left-hand column.

T F 1. The primary colors are blue, green, and yellow.

a. green, orange, and redb. green, red, and yellowc. violet, yellow, and redd. orange, blue, and violete. yellow, green, and violet

48

40

Page 51: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

Activity

2

Select one or more cognitive student performance objectives in your occu-pational specialty that require student knowledge or comprehension of infor-mation.

Construct ten true-false items to measure achievement of the student perfor-mance objectives you selected. Number each item for easy reference duringfeedback. Include directions and a scoring key.

You may wish to ask one or more of your peers who have taken or are takingthis module to critique the true-false items you have constructed. Discussany suggested changes and then make any necessary revisions.

After you have constructed your true-false items, use the Checklist for True-False Items, p. 51, to evaluate your work.

Page 52: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

NOTES

50 5i

Page 53: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

CHECKLIST FOR TRUE-FALSE ITEMSDirections: Place an X in the NO, PARTIAL, or FULL box to indicate thateach of the following performance components was not accomplished, par-tially accomplished, or fully accomplished. If, because of special circum-stances, a performance con.onent was not applicable, or impossibie to exe-cute, place an X in the N/A box.

Name

Date

Resource Person

The true-false items meet the following criteria:1. All statements are entirely true or entirely false

2. No trivial details are used to make any of the statements false

3. The statements are concise without more elaboration than is neces-sary to give clear meaning

4. The statements have not been quoted from a textbook

5. The statements avoid using specific determiners that signal answers

6. The items are stated in positive rather than negative terms

7. An authority is quoted when the item contains opinion

8. A pattern of responses is avoided in the items

9. The responses for the group of items are written in a manner thatmakes scoring easy (e.g., in a column to the left)

10. Statements avoid qualitative terms (e.g., several, many)

11. Numbers of true and false items are roughly equal

12. Clear and simple directions are provided

13. An appropriate scoring key is provided

LEVEL OF PERFORMANCE

Level of Performance: All items must receive FULL or N/A responses. If any item receives a NO orPARTIAL response, review the material in the information sheet, True-False Items, pp. 46-48, revise youritems accordingly, or check with your resource person if necessary.

51 52

Page 54: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

NOTES

5253

Page 55: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

Activity

1

Activity

2

Learning Experience VIOVERVIEW

After completing the required reading, construct three essay items and twooral items to measure the achievement of selected cognitive student per-formance objectives.

You will be reading the information sheet, Essay and Oral Items, pp. 54-55.

You will be selecting one or more student performance objectives that requirestudents to apply, analyze, synthesize, or evaluate information.

You will be constructing three essay items and two oral items to measureachievement of the student performance objectives you select.

You may wish to ask peers to give suggestions for the improvement of theessay and oral items you develop.

You will be evaluating your competency in constructing essay and oral items,using the Checklist for Essay and Orai Items, p. 57.

5354

Page 56: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

Essay and oral torus have distinct advantages and disadvantaqec. Properlyconstructed, they can be excellent evaluation devices. For information onwriting high quality essay and oral items, read the following information sheet.

ESSAY AND ORAL ITEMSand -oral items, like completion items, are

supply 'toms Completion items, however, requireNtudents to supply the own knowledge one pieceat a tune Essay or oral sterns require students to

anu 1,upply numerous pieces of "noviledge.Vurthermore, they require students to organize andpresent this knowledge in a logical ant uader-standable mariner

Moth essay and oral items can be divided into twokin(1%. depending on the kind of response they callfor The extended-response Item allows studentsconsiderable freedom in structuring and presentingtheir own answers. Tho following item, for example,could bo answered in a number of d;fferent ways:

Discuss the stand of the American MedicalAssociation on socialized modicine.

A restricted-response Item, on the other hand,limits the content and form of the esponse in some-ow Both the item and the response it calls for arem(.._n more specitic. as in the following example:

Describe the criticisms made by the AmericanMedical Association against socialized medi-cine. kidsent one example to support eachcnbcism and one example to ounter eachcriticism.

Uses of rssay and Oral ItemsEssay and oral items can be used to test student

performance at any level sn the cognitive domain.They are most useful in evaluating performance onhiqh.:evel instr_ctional objectives requiring the ap-diication. analysis, synthesis, or evaluation of infor-mation

Compared to objective items, there are some-olsrovantages" in using e ay and oral items.Sconng of students' answers to essay and oral itemsis more subiectIve and, thus, can be more unreli-able Even a restricted-response item allows stu-dents some latitude in responding, and scoring mustallow equal credit for different responses of equalmerit Thus, It is critical that you prepare and utilizea structured. yet flexible. answer key in scoring re-sponses. so that your ratings are as objective andfair as possible.

Answering essay and oral items and scoring theanswers also demand a good deal of time. Oral tests

must be conducted and score( individually, whichmay require large amounts of teacher and studenttime. Teachers can often expect to spend severalhours scoring essay items for a whole classunlikeobjective items, which can usually be answered andscored in a matter of minutes.

Finally, these items test more than just students'knowledge of technical content. They also test stu-dents' ability to communicate, orally or in writing.Students must be able not only to recall and apply,analyze, synthesize, or evaluate technical knowl-edge. They must also be able to communicate theirknowledge and their treatment of it to you.

However, when the student performance objec-tives call for higher levels of cognition and when abilityto organize and communicate information is impor-tant for occupational success, then essay and oralitems should be used. The time involved in scoringwill be wail worth it. Short-answer tests are usefuland have their place in evaluation, but only throughessay and oral items can you test, and challenge,students' ability to thinkto take what they know,analyze it, synthesize it, organize it, and communi-cate their ideas to others.

Constructing Essay and Oral ItemsThere are several guidelines that you can follow

to minimize any limitations of essay and oral items.

Develop clear, precise items. Your essay and oralitems should always communicate your intent tostudents with no room for misinterpretation. In prac-tice, restricted-response items are usually prefer-able. Each item should tell students, as clearly aspossible, what information is required in the answerand how that information should be presented:*

Clear, precise items allow less opportunity for stu-dents to misunderstand an item. They also reducethe opportunities for students to evade respondingto an item by speaking or writing at length in gener-alities. Scoring also becomes easier when the re-quired content and form of the response are clearlyindicated in the item.

Require students to answer all items. Teacherssometimes make some essay or oral items optional.

4 to gain skill in presenting questions orally to students, you may wishto refer to Module C-12, Employ Oral Questioning Techniques.

54 55

I

Page 57: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

Students might be directed, for example, to respondto any one of three items. However, when differentstudents answer different items, it is almost impos-sible to get comparable scores because studentshave not been performing comparable tasks. Withoptional items, you end up testing different groupsof students on different bodies of information anddifferent ways of treating those different bouies ofinformation.

Use a larger number of shorter items. Sincestudents take more time in responding to essay ororal items, you cannot sample as wide a range ofknowledge as with objective items. The best way tomake your sample as wide as possible is to use alarger number of items that require shorter answers.Your essay or oral items will then evaluate a widervariety of instructional objectives. -

Develop items for higher-level objectives andcontent. Given the nature of essay and oral items,you should use them to measure only appropriate,high-level instructional achievement. Objective itemsshould be used to measure the lower cognitive lc':els (i.e., knowledge and comprehension)a task theycan accomplish accurately and efficiently. The useof each type of itemsubjective and objective- -should be restricted to the levels of accomplisrimentfor which each is most suited.

Provide full and clear directions. In your direc-tions, you should tell students exactly what their taskisfor example, to answer each of the three itemsprovided on a written essay test. You should tell stu-dents what the point value of each item is and howmuch time is available for each item or for the wholetest. Students will then know approximately how muchtime to spend on each item. Students must also betold if their scores will be affected by mechanicalfactors (e.g., spelling and handwriting on essay items,grammar on essay or oral items).

Give students advance notice and practicetests. You should give your students warning of up-coming essay or oral tests. This provides them withan opportunity to review the material to be coveredfrom the appropriate perspective. You probably knowfrom your own experiences that preparing for an es-say test requires the use of different study skills thanpreparing for objective tests.

Furthermore, students should be given practicetests. Some students may never have been re-quired to iespond to essay or oral items before. Theymay have had no experience in organizing and pre-senting a body of information, as these items re-quire. They may not be aware that they will need tospend some of the time allotted for the test in plan-ning their answers. Students should have the chanceto practice answering oral or essay items before beingevaluated on their performance of these skills.

Don't use open-book tests. Allowing students touse their books doesn't really help them to answeritemsunless you are also testing their ability to lo-cate appropriate information in reference books.Remember that an essay or oral item should be usedto evaluate student ability at the higher cognitivelevels. At these levels, it matters somewhat less whatknowledge students supply. More important is howthey can apply that knowledge in answering the es-say or oral item. Studies have shown that studentsdon't do any better or worse tl Ian expected whenthey are allowed to use their books.

Prepare a structured key fo- scoring. The sub-jectivity of scoring can be minimized if you write astructured scoring keya model answer to the es-say or oral item, usually in outline form. As you writeyour own answer to the item, you can assign spe-cific point values to specific aspects of the answer.Students' answers can then be compared to the keyto assign point values to their answers.

It is a good idea to have your scoring key re-viewed by a colleague who is also knowledgeablein your occupational area. He or she may identifyeNther information to include in your model answer.Likewise, you may find information in students' an-swers that is correct but that you did not anticipate.Students might include facts not mentioned in yourkey, or they might organize and present their factsin a way you hadn't thought of. Any such informationshL,'Ild be added to your key as well.

Read all students' answers to one essay itemat one sitting. The best way to score answers toessay items consistently is to read all students' an-swers to one item at one sitting. In other words, don'tread one student's entire test before reading anoth-er's. You can review your key for the item you arescoring and concentrate on that item. If possible, youshould not even know whose answer you are read-ing. Studies have shown that the "halo effect" is realteachers tend to score answers higher for studentswho are known to be more capable.

Score mechanical factors separately. Expertsdisagree on whether such factors as handwriting,spelling, or grammar should be considered at all inscoring students' answers. Some are all for it; someare equally against it. You will need to use your ownjudgment to determine whether to score students'answers for these factors. If these skills are impor-tant for success in the occupation, then they shouldat some point and in some way, be evaluated.

If you do score for these factors, however, youshould assign a score separate from that for tech-nical content. Regardless of how important skill incommunicating may be in your occupational area,the technical content of student:- answers shouldbe scored on its own merit.

555b

Page 58: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

Select one or more student performance objectives in your occupational spe-Activity cialty that require students to use the higher cognitive levels (application,

2analysis, synthesis, evaluation).

Construct three essay items and two oral items to measure achievement ofthe student performance objectives you selected. Number each item for easyreference during feedback. Include directions and a scoring key.

You may wish to ask one or more of your peers who have taken or are takingthis module to critique the essay and oral items you have constructed. Dis-cuss any suggested changes and then make any necessary revisions.

After you have constructed your essay and oral items, use the Checklist forEssay and Oral Items, p. 57, to evaluate your work.

56

5r

Page 59: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

IICHECKLIST FOR ESSAY AND ORAL ITEMSDirections: Place an X in the NO, PARTIAL, or FULL box to indicate thateach of the following performance components was not accomplished, par-tially accomplished, or fully accomplished. If, because of special circum-stances, a performance component was not applicable, or impossible to exe-cute, place an X in the N/A box.

Name

Date

Resource Person

LEVEL OF PERFORMANCE

s,,... v

4) AIv ..

4The essay and oral items meet the following criteria:1. The language in each item is precise and unambiguous El OD {12. The problems or tasks require students to use the higher levels of

cognition CI 0 El El3. Open books are not allowed DO Ell4. Directions clearly state th' task, point value of items, and time allowed

for answering {1 El 0 El5. The items clarify any additional directions needed beyond the general

set of directions

III6. Students are required to answer all items

7. A structured scoring key is provided DODOLevel of Performance: All items must receive FULL or N/A responses. If any item receives a NO orPARTIAL response, review the material in the information sheet, Essay and Oral Items, pp. 54-55, reviseyour items accordingly, or check with your resource person if necessary.

557

58

I

Page 60: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

NOTES

5358

Page 61: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

Learning Experience VIIOVERVIEW

EnablingObjective

After completing the required reading, construct one case study or problem-solving item to measure the achievement of a selected cognitive studentperformance objective.

You will be reading the information sheet, Case Studies and Problem-SolvingItems, pp. 60-61.

Activity

2

You will be selecting one student performance objective that requires stu-dents to apply, analyze, synthesize, or evaluate information.

Activity

3

You will be constructing one case study or problem-solving item to measureachievement of the student performance objective you select.

You may wish to ask peers to give suggestions to improve the clarity of the

case study or problem-solving item you develop.

You will be evaluating your competency in constructing a case study or prob-lem-solving item, using the Checklist for Case Studies and Problem-Solving

Items, p. 63.

59 60

Page 62: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

Activity The case study or problem-solving technique is often used as a method ofinstruction. For information on how to use these techniques in evaluatingstudent knowledge, read the following information sheet.

CASE STUDIES AND PROBLEM-SOLVING ITEMSCase studies and problem-solving items present

students with a situation to which they must re-spond. These items differ from other types of testitems in the amount of detail they contain. Objectiveitems are usually shortone sentence. Essay or oralitems are usually no longer than three or four sen-tences. Some case studies or problem-solving itemscan be much longer; each item might be a page ormore in length.

Case studies and problem-solving items are usu-ally presented to students in written form. You might,for example, write a case study describing how aworker observed safety regulations on the job. Stu-dents could read the case study and critique theworker's safety practices. Students could also begiven a table, graph, or diagram and asked to an-swer items on the basis of it.

But such items do not necessarily have to be inwritten form. Students might be asked to react .ocase studies or problem-solving items presented onvideotape or audiotape.

Students are asked to use the information pre-sented in the case study or problem-solving item toanalyze a situation or solve a problem. Often, otherknowledge is used as well. Students in health careoccupations, for instance, could be given a series ofproblem-solving items, each of which describes apatient's symptoms in a particular emergency con-dition (e.g., choking on a piece of food). Studentscould be asked to identify the condition and de-scribe the correct emergency procedures to follow.

Students can also be asked to respond to the sit-uafon or solve the problem in a number of differentways. They might answer essay or oral items in whichthey evaluate methods, materials or solutions. True-false items could be used to evaluate student com-prehension of the data presented in a case study.Students' ability to troubleshoot a malfunctioningmachine could be evaluated using multiple-choiceitems.

Uses of Case Studies andProblem-Solving Items

The case study or problem-solving statement couldask students to predict trends or to make infer-ences of consequences (high-level types of know-ing). For example, assume you are teaching stu-dents the factors to consider in selecting anoccupation. For evaluation purposes you might wantto design a case study in which students would haveto predict from the data presented which jobs will beavailable in the next two years.

You can also use case studies and problem-solv-ing items when the objective is to teach the appli-cation of principles, generalizations, and concepts.For example, an instructor could ask the studentswhat automotive repair principles they would use tosolve a list of problem situations faced in an autobody shop. This type of item requires students toapply what they have learned.

Some objectives stress the ability of students toanalyzeto break down material into its parts, de-tect relationships of the parts, and recognize the or-ganizational principles of the structure as a whole.These require a higher level of knowing, arid casestudies and problem-solving items are frequently usedfor evaluating student performance at this level.

For example, if a teacher wanted students in of-fice machine repair to be able to troubleshoot (ana-lyze) a defective calculating machine, he or she coulddevelop a problem-solving item listing all the symp-toms of a malfunctioning calculator and then askstudents to describe the probable cause of thetrouble.

The case study or problem-solving item is also anexcellent means of evaluating students' ability tojudge (evaluate or appraise) ideas, methods, ma-terials, solutions, and so on. For example, achieve-ment of objectives involving the purchasing of goods

Page 63: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

S

and services can be evaluated through case stud-ies, in that such objectives require students to makevalue judgments.

Constructing Case Studies andProblem-Solving Items

Several guidelines can help you to write clear casestudies and problem-solving items. In general, youritems should be practical and realistic and containenough detail for students to perform the task re-quired in relation to the case study or problem.

Describe a realistic situation. A marketing anddistributive education instructor, for example, couldwrite a case study describing hpw a retail clerk waitedon customers and ask students to critique the clerk'scustomer service. The clerk's performance shouldbe realisticneither all good nor all bad, since peopleare neither one nor the other. Also, mistakes in per-formance should not be obvious. Don't, for example,have the retail clerk chase a customer out of thestore.

Give students all the detail they will need forthe task. For example, a case study might list thesymptoms of a malfunctioning carburetor. Studentscould be asked to identify probable causes of thetrouble. They could even be asked to describe howthey would verify their diagnoses. To complete thistask successfully, they would need to know all therelevant detail from reading the case study. If youexpect students to identify a particular c Ise for aproblem, be sure they have enough information tocome to that conclusion.

You can still hold your students responsible forother knowledge in the task, of coy if a case studyinvolves the application of Ohm:_ Law to an elec-tronic circuit design, the instructor could legitimatelyexpect students to know what Ohm's Law is and how

61

to apply it (assuming they have studied those thingsalready). That, in fact, would be one of the thingsthe instructor would be testing in this item. He or shewould have to be sure, however, to supply in thecase study all the information students would needon the circuit design.

Clearly state the task to be performed. Stu-dents should be told clearly what they are to do withthe case study or problem once they have read it. If

there are objective items for students to answer, youneed to tell them to answer the items using the in-formation in the case study or problem statement. Ifstudents are to write longer, essay-type responses(e.g., analyzing, critiquing, evaluating), you need totell them how detailed their answers should be.

Follow good item-writing guidelines. As statedpreviously, you might ask students to answer true-false items about a case study, multiple-choice itemsabout a circuit diagram, or essay items evaluating aworker's safety practices. In each case, you shouldfollow all the guidelines for writing that type of item.Specifically, you should provide directions for an-swering the items. The type of item you use shouldbe chosen to fit the learning activity being evalu-ated.

Write items in terms students can understand.Your case studies or problem-solving items shouldnot penalize students with lower reading or oralcommunication skills. You should use technical termsthat students know or define those terms in youritems.

Use appropriate scoring keys and proce-dures. You should prepare scoring keys (e.g., oLt-lines of model answers) for your case studies orproblem-solving items. Likewise, you should useappropriate procedures for scoring student re-sponses to the items you have used.

62

Page 64: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

Activity

2

Select a student performance objective in your occupational specialty thatrequires students to apply, analyze, synthesize, or evaluate information.

Construct one case study or problem-solving item to measure achievementActivity of the student performance objective you selected. Include directions and a

3scoring key.

You may wish to ask one or more of your peers who have taken or are takingthis module to critique the case study or problem-solving item you have con-structed. Discuss any suggested changes and then make any necessaryrevisions.

After you have constructed your case study or problem-solving item, use theChecklist for Case Studies and Problem-Solving Items, p. 63, to evaluateyour work.

62

63

Page 65: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

CHECKLIST FOR CASE STUDIES ANDPROBLEM-SOLVING ITEMS

Directions: Place an X in the NO, PARTIAL, or FULL box to indicate thateach of the following performance components was not accomplished, par-tially accomplished, or fully accomplished. If, because of special circum-stances, a performance component was not applicable, or impossible to exe-cute, place an X in the N/A box.

Name

Date

Resource Person

LEVEL OF PERFORMANCE

.41;*

41. eThe case study or problem-solving item meets the following criteria:

1. The situation developed is based on the student performance objec-tive El El

2. The situation is practical and realistic El3. The situation is described in sufficient detail for the task

4. The problem or case study is worded at the students' comprehensionlevel El El El El

5. The task requ?res students to apply, analyze, synthesize, or evaluatelearning El

6. The item contains a clear statement of what students are to do

7. The item indicates the level of detail required in the solution

8. Test items concerning the case study or problem follow good item-writing guidelines El

9. Item directions state point value:- and time limits

10. Appropriate scoring keys are provided

Level of Performance: All items must receive FULL or N/A responses. If any item receives a NO orPARTIAL response, review the material in the information sheet, Case Studies and Problem-Solving Items,pp. 60-61, revise your items accordingly, or check with your resource person if necessary.

63 64

Page 66: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

NOTES

64 65

Page 67: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

Learning Experience VIIIFINAL EXPERIENCE

Activity

1

In an actual teachins-,1 situation,* assess student cognitive (knowledge) per-

formance.

As you fulfill you;- teach:rq duties, determine when the nature of thestudent perforrnance obi;;A;t;ves requires the assessment of studentknowledge. Based on that decision, assess student cognitive perfor-mance. This will include

determining The appropriate types of test items to use to niersure stu-dents' achievement of unit and lesson objectivesconstructing 4t least one unit test designed to measure student progresstoward the objectivesadministering the unit test to students

NOTE: As you complete each of the above activities, document your actions(in writing, on tape, through a log) for assessment purposes.

Arrange to have your resource person review your test and other documen-tation (e.g., audiotape of your directions to students prior to administering thetest).

Your total competency will be assessed by your resource person, using theTeacher Performance Assessment Form, pp. 67-68.

Based upon the criteria specified in this assessment instrument, your re-source person will determine whether you are competent in assessing stu-dent cognitive (knowledge) performance.

For a definition of 'actual teaching sluation; see the inside sack cover.

665

66

Page 68: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

NOTES

66s7

Page 69: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

S

TEACHER PERFORMANCE ASSESSMENT FORMAssess Student Performance: Knowledge (D-2)

Directions: Indicate the level of the teacher's accomplishment by placingan X in the appropriate box under the LEVEL OF PERFORMANCE heading.If, because of special circumstances, a performance component was notapplicable, or impossible to execute, place an X in the N/A box.

Name

Date

Resource Person

In developing a unit test to assess student cognitive per-formance, the teacher:

1. identified objectives in the cognitive (knowledge) area ....

2. identified types of test items appropriate for assessing thelevel of knowledge implied in the performance objectives

3. developed each test item according to the guidelines forconstructing that type of item

4. constructed a test that was:a. valid

b. reliable

c. objective

d. discriminating

e. comprehensive

f. easy to use

5. grouped together all test items of like type

6. grouped together within each type the items relating to thesame objective

7. included no more than three or four different types of testitems

8. sequenced like items logically

9. included clear and concise directions for each different typeof test item

10. constructed the test so that recording responses was simpleand consistent

11: numbered items consecutively from the beginning to theend of the test

12. developed an appropriate key for sccring

6768

LEVEL OF PERFORMANCE

o..oo 4. b ..oir o$) 4°o ...4. o 0,.9 0-k° 1/47

11

O 00000O E10000O 000E10O r10000O 0E1F100O 000E10O 00000O 00000

00E1O 0E1E1

I-1

1-1

O 000O 0E10D C CIO

Page 70: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

In administering the test, the teacher:13 arranged the physical environment so it was suitable for

testtaking

14 explained the use of the test to students in a nonthreaten-ing way

15 informed students concerning how the test woJld be gradedprior to their taking the test

16 gave specific instructions for taking the test

17 had all needed testing materials and supplies rearly for dis-tribution

18 indicated approximate time needed to complete the test bysections and as a whole

19 allowed sufficient time for the test to be completed by all ormost students

20 answered students' questions during the test in a mannerthat did not disturb the total group

Level of Performance: All items must receive N/A, GOOD, or EXCELLENT responses. If any item re-ceives a NONE, POOR, or FAIR response, the teacher and resource person should meet to determinewhat additional activities the teacher needs to complete in order to reach competency in the weak area(s).

68 UPU 9871111E

Page 71: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

ABOUT USING THE NATIONAL CENTER'SPBTE MODULES

0

OrganizationEach module is designed to help you gain competency in aparticular skill area considered important to teaching suc-cess. A module is made up of a series of learning experi-ences, some providing background information, some pro-viding practice experiences, and others combining thesetwo functions. Completing these experiences should en-able you to achieve the terminal objective in the finallearning experience. The final experic, ice in each modulealways requires you to demonstrate the skill in an actualteaching situation when you are an intern, a student teach-er, an inservice teacher, or occupational trainer.

ProceduresModules are designed to allow you to individualize yourteacher education program. You need to take only thosemodules covering skills that you do not already possess.Similarly, you need not complete any learning experiencewithin a module if you already have the skill needed tocomplete it. Thereiore, before taking any module, youshould carefully review (1) the introduction, (2) the objec-tives listed on p. 4, (3) the overviews preceding each learn-ing experience, and (4) the final experience. After compar-ing your present needs and competencies with the informa-tion you have read in these sections, you should be ready tomake one of the fo/lowing decisions:

That you do not have the competencies indicated andshould complete the entire moduleThat you are competent in one or more of the enablingobjectives leading to the final learning experience and,thus, can omit those learning experiencesThat you are already competent in this area and areready to complete the final learning experience inorder to "test out"That the module is inappropriate to your needs at thistime

When you are ready to complete the final learning experi-ence and have access to an actual teaching situation,make the necessary arrangements with your resource per-son. If you do not complete the final experience success-fully, meet with your resource person and arrange to (1)repeat the experience or (2) complete (or review) previoussections of the module or other related activities suggestedby your resource person before attempting to repeat thefinal experience.Options for recycling are also available in each of thelearning experiences preceding the final experience. Anytime you do not meet the minimum level of performancerequired to m, t an objective, you and your resource per-son may meet to select activities to help you reach compe-tency. This could involve (1) completing parts of the modulepreviously skipped, (2) repeating activities, (3) reading sup-plementary resources or completing additional activitiessuggested by the resource person, (4) designing your ownlearning experience, or (5) completing some other activitysuggested by-you or your resource person.

TerminologyActual Teaching Situation: A situation in which you areactually working with ar-i responsible for teaching sec-ondary or postsecondary vocational students or other oc-cupational trainees. An intern, a student teacher, an in-service teacher, or other occupational trainer would befunctioning in an actual teaching situation. If you do nothave access to an actual teaching situation when you aretaking the module, you can complete the module up to thefinal learning experience. You would then complete thefinal learning experience later (i.e., when you have accessto an actual teaching situation).Alternate Activity or Feedback: An item that may substi-tute for required items that, due to special circumstances,you are unable to complete.Occupational Specialty: A specific area of preparationwithin a vocational service area (e.g., the service areaTrade and Industrial Education includes occupational spe-cialties such as automobile mechanics, welding, and 6ac-tricity.Optional Activity or Feedback: An item that is not re-quired but that is designed to supplement and tanrich therequired items in a learning experience.Resource Person: The person charge of your educa-tional program (e.g., the professor, instructor, administra tor,instructional supervisor, cooperating/supervising/clat:s-room teacher, or training supervisor who is guiding you incompleting this module).Student: The person who is receiving occupational in-struction in a secondary, postsecondary, or other trainingprogram.Vocational Service Area: A major vocati,nal field: agri-cultural education, business and office education, marketing and distributive education, health occupations e...uca-tion, home economics education, industrial arts education,technical education, or trade and industrial education.You or the Teacher/Instructor: The person who is com-pleting the module.

Levels of Performance for Final AssessmentN/A: The criterion was not met because it was not appli-cable to the situation.None: No attempt VMS made to meet the criterion, al-though it was relevant.Poor: The teacher is unable to perform this skill or has onlyvery limited ability to perform it.Fair: The teacher is unable to perform this skill in an ac-ceptable manner but has some ability to perform it.Good: The teacher is able to perform this skill in an effec-tive manner.Excellent: The teacher is able to perform this skill in a veryeffective manner.

'70

Page 72: 11111 ' 111111 - ERIC · 11111 2.2. L. 11111 ' ... Norman E. Measurement and Evaluation in Teaching. Fourth Edition. New ... pp. 19-41; Gronlund, Measurement and. Eval-uation in Teaching…

Titles of the National Center's Performance-Based Teacher Education Modules

Category A: Program Planning, Development, and Evaluation Category G: School-Community Relations

A-1 Prepare for a Community Survey G-1 Develop a School-Community Relations Plan for Your Vocational ProgramA-2 Conduct a Community Survey G-2 Give Presentations to Promote Your Vocational ProgramA-3 Report the Findings of a Community Survey G-3 Develop Brochures to Promote Your Vocational Program

A-4 Organize an Occupational Advisory Committee G-4 Prepare Displays to Promote Your Vocational ProgramA-5 Maintain an Occupational Advisory Committee G-5 Prepare News Releases and Articles Conceming Your Vocational Program

A-6 Develop Program Goals and Objectives G-6 Arrange for Television and Radio Presentations Conceming Your Vocational

A-7 Conduct an Occupational Analysis Program4-8 Develop a Course of Study G-7 Conduct an Open HouseA-9 Develop Long-Range Program Plans G-8 Work with Members of the Community

A-10 Conduct a Student Follow-Up Study G-9 Work with State and Local Educators

A-11 Evaluate Your Vocational Program G-10 Obtain Feedback about Your Vocational Program

Category B: Instructional Planning Category H: Vocational Student Organization

B-1 Determine Necds and Interests of Students H-1 Develop a Personal Philosophy Concerning Vocational Student

B-2 Develop Student Performance Objectives Organizations

13-3 Develop a Unit of Instruction H-2 Establish a Vocational Student Organization

B-4 Develop a Lesson Plan H-3 Prepare Vocational Student Organization Members for Leadership Roles

13-5 Select Student Instructional Materials H-4 Assist Vocational Student Organization Members in Developing and

8-6 Prepare TeacherMade Instructional Materials Financing a Yearly Program of Activities

Category C: Instructional ExecutionH-5H-6

Supervise Activities of the Vocational Student OrganizationGuide Participation in Vocational Student Organization Contests

C-1C-2

Direct Field TripsConduct Group Discussions, Panel Discussions, and Symposiums

Category I: Professional Role and Development

C-3 Employ Brainstorming, Buzz Group, and Question Box Techniques 1-1 Keep Up to Date ProfessionallyC-4 Direct Students in Instructing Other Students 1-2 Serve Your Teaching ProfessionC-5 Employ Simulation Techniques 1-3 Develop an Active Personal Philosophy of Education

C-6 Guide Student Study 1-4 Serve the School and Community

C-7 Direct Student Laboratory Experience 1-5 Obtain a Suitable Teaching Position

C-8 Direct Students in Applying ProblemSolving Techniques 1-6 Provide Laboratory Experiences for Prospective Teachers

C-9 Employ the Project Method 1-7 Plan the Student Teaching ExperienceC-10 Introduce a Lesson 1-3 Supervise Student TeachersC-11C-12

Summarize a LessonEmploy Oral Questioning Techniques

Category J: Coordination of Cooperative Education

C-13 Employ Reinforcement Techniques J-1 Establish Guidelines for Your Cooperative Vocational Program

C-14 Provide Instruction for Slower and More Capable Learners J-2 Manage the Attendance. Transfers, and Terminations of Co-Op Students

C-15 Present an Illustrated Talk J-3 Enroll Students in Your Co-Op Program

C-16 Demonstrate a Manipulative Skill J-4 Secure Training Stations for Your CoOp Program

C-17 Demonstrate a Concept or Principle J-5 Place Co-Op Students on the Job

C-18 Individualize Instruction J-6 Develop the Training Ability of Onthe-Job Instructors

C-19 Employ the Team Teaching Approach J-7 Coordinate Onthe-Job InstructionC-20 Use Subject Matter Experts to Present Information J-8 Evaluate Co-Op Students' OntheJob Performance

C-21 Prepare Bulletin Boards and Exhibits J-9 Prepare for Students' Related Instruction

.;-22 Present Information with Models, Real Objects, and Flannel Boards J-10 Supervise an Employer-Employee Appreciation Event

C-23C-24

Present Information with Overhead and Opaque MaterialsPresent Information with Filmstrips and Slides

Category K: Implementing Competency-Based Education (CBE)

C-25 Present Information with Films K-1 Prepare Yourself for CBE

C-26 Present Information with Audio Recordings K-2 Organize the Content for a CBE Program

C-27 Present Information with Televised and Videotaped Materials K-3 Organize Your Class and Lab to Install CBE

C-28 Employ Programmed Instruction K-4 Provide Instructional Materials for CBE

C-29 Present Information with the Chalkboard and Flip Chart K-5 Manage the Daily Routines of Your CBE Program

C-30 Provide for Students' Learning Styles K-6 Guide Your Students Through the CBE Frogram

Category D: Instructional EvaluationD-1D-2D-3D-4D-50-6

Establish Student Performance CriteriaAssess Student Performance: KnowledgeAssess Student Performance: AttitudesAssess Student Performance: SkillsDetermine Student Gradesevaluate Your Instructional Effectiveness

Category E: Instructional ManagementE-1 Project Instructional Resource NeedsE-2 Manage Ycar Budgeting and Reporting ResponsibilitiesE-3 Arrange for Improvement of Your Vocational FacilitiesE-4 Maintain a Filing SystemE-5 Provide for Student SafetyE-6 Provide for the First Aid Needs of StudentsE-7 Assist Students in Developing SelfD ;ciplineE-8 Organic a the Vocational LaboratoryE-9 Manage the Vocational LaboratoryE-10 Combat Problems of Student Che :al Use

Category F: GuidanceF-1 Gather Student Data Using Formal DataCollection TechniquesF-2 Gather Student Data Through Personal ContactsF-3 Use Conferences to Help Meet Student NeedsF-4 Provide Information on Educational and Career OpportunitiesF-5 Assist Students in Applying for Employment or Further Education

Category L: Serving Students with Special/Exceptional NeedsL-1 Prepare Yourself to Serve Exceptional StudentsL-2 identify and Diagnose Exceptional StudentsL-3 Plan Instruction for Exceptional StudentsL-4 Provide Appropriate Instructional Materials for Exceptional StudentsL-5 Modify the Leaming Environment for Exceptional StudentsL-6 Promote Peer Acceptance of Exceptional StudentsL-7 Use Instructional Techniques to Meet the Needs of Exceptional StudentsL-8 Improve Your Communication SkillsL-9 Assess the Proc-ess of Exceptional StudentsL-10 Counsel Excep,. anal Students with Personal-Social ProblemsL-11 Assist Exceptional Students in Developing Career Planning SkillsL-12 Prepare Exceptional Students for EmployabilityL-13 Promote Your Vocational Pi ogram with Exceptional Students

Category M: Assisting Students in Improving Their Basic Skills

M-1 Assist Students in Achieving Basic Reading SkillsM-2 Assist Students in Developing Technical Reading SkillsM-3 Assist Students in Improving Their Writing SkillsM-4 Assist Students in Improving Their Oral Communication SkillsM-5 Assist Students in Improving Their Math SkillsM-6 Assist Students in Improving Their Survival Skills

riELATED PUPLICATIONSStudent Guide to Using P tormanceBased Teacher Education MaterialsResource Person Guide to Using PerformanceBased Teacher Education MaterialsGuide to the Implementation of PerformanceBased Teacher EducationPerformance-Based Teacher Education: The State of theArt, General Education and

Vex -Ilion& Education

or information regarding availability and prices of these materials contactAAVIM, American Association for Vocational Instructlaterials, 120 Driftmier Engineering Center, University of Georgia, Athens, Get rgla 30602, (404) 542-2586

71 ISBN 0-89606-'