creating rubrics for...
TRANSCRIPT
Workshop Goals
4
Considering the various rubric options
Creating a Customized Rubric
Evaluating A Rubric for effective assessment
Calibrating rubric scoring for reliable assessment data
Rubric Overview
• Rubrics provide the criteria for classifying products or behaviors into categories that vary along a continuum. – They can be used to classify virtually any product or behavior, such as
essays, research reports, portfolios, works of art, recitals, oral presentations, performances, and group activities.
– Judgments can be self-assessments by students; or judgments can be made by others, such as faculty, other students, supervisors, or external reviewers.
– Rubrics can be used to provide formative feedback to students, to grade students, and/or to assess programs.
• There are two major types of scoring rubrics:– Holistic scoring: one global, holistic score for a product or behavior– Analytic rubrics: separate, holistic scoring of specified characteristics
of a product or behavior.
5
Rubric Options
• Considering the various rubric options
– Nature of the rubric
• Assessment of student – Knowledge
– Performance
• Assigning grades– Feedback for students
• Assessment of curriculum– Feedback for faculty
6
Rubric Options
• Considering the various rubric options (continued):– Nature of the rubric
• Aligned with the course syllabus and student learning outcomes– Designed to assess what is being taught
• User friendly– Can faculty scoring be calibrated to achieve reliable data?
– Can students benefit from the rubric’s feedback?
• Holistic– Summative
• Analytic– Formative feedback
» Quantitative
» Qualitative
7
Rubric Options
• What do you want to assess with your rubric?
– Student
• Knowledge: recall facts, terms, formulas
• Skills: ability to use acquired knowledge
• Performance: ability to produce a performance based product
• Student Learning Outcomes
• Learning Objectives
8
Creating Rubrics
• What do you want to assess with your rubric?
– Student
• Knowledge: recall facts, terms, formulas rather than a rubric, a scantron test may be more suitable
• Performance: ability to produce a performance based product
An analytic rubric can generate a "profile" of specific student strengths and weaknesses
9
Creating Rubrics
– Assigning Grades
• Effective assessment and feedback for students– What are the outcomes and supporting objectives you want to
assess?
– Have the students been given the opportunity to practice these outcomes and objectives in class under the instructors’ supervision?
– Are these outcomes and objectives present in the course syllabus?
10
Creating Rubrics
– Assigning Grades• Effective assessment and feedback for students
Look for meaningful assessment categories, wording and terminology that not only point out weaknesses and deficiencies, but also provide the information the student can use to understand why they received the grade they did as well as to improve their performance in the future. Avoid including in your rubric any outcomes and objectives that have not been included in the syllabus or taught in the class.
11
Creating Rubrics
– Assessing course curriculum
• Effective assessment and feedback for faculty– Is the rubric instructionally aligned with the course
curriculum?
– Does the rubric adequately address the student learning outcomes and supporting objectives assigned to that course?
– Does the rubric provide enough information to allow faculty to pinpoint problems and weaknesses in the curriculum that will need to be adjusted in the future?
12
Creating Rubrics
– Assessing course curriculum
• Effective assessment and feedback for faculty
• For curriculum purposes, a rubric should assess a course’s student learning outcomes and the supporting learning objectives, preferably one student learning outcome and its objectives per rubric instrument.
• A holistic rubric would be given first consideration as they are less time consuming and not designed to provide detailed feedback to students.
13
Creating Rubrics
– Course curriculum
• Syllabus and student learning outcomes are aligned with actual classroom student performance products.
– For rubric produced data to be reliable, the syllabus and its student learning outcomes should be aligned with what is actually being taught in the classroom.
– Are full time faculty and adjuncts teaching the same material and using the same assignments?
– Are the textbooks and syllabi aligned with the curriculum?
14
Course curriculum alignment matrix
I = Introduced, D = Developed & Practiced with Feedback,
M = Demonstrated at the Mastery Level Appropriate for Graduation
15
Course Outcome 1 Outcome 2 Outcome 3 Outcome 4 Outcome 5
1105 I, D I
1305 I D
1307 D D
1405 D D
1407 D
1411 D M
1412 D M
1413
1414 M DM M
Rubric Options
– Course curriculum
• Syllabus and student learning outcomes are aligned with actual student performance
• If there are too many variables between the curriculum, the textbooks, the syllabi, the full time faculty and the adjuncts; rubric produced data will be unreliable.
• It is critical to ensure there are alignment of course and teaching variables, as well as calibration of grading and scoring.
16
Rubric Options
– Course curriculum
• User friendly– Can faculty scoring be calibrated to achieve reliable data?
» Is a wide variations in grading unusual?
– Can students benefit from the rubric’s feedback?
» Should the learning process continue throughout the testing or assessing process?
17
Rubric Options
– Course curriculum
• User friendly– Can faculty scoring be calibrated to achieve reliable data?
– Faculty and adjuncts can be quickly trained to grade or score student work with standardized results
– Can students benefit from the rubric’s feedback?
– Using clear and easily understandable wording in the rubric provide students with helpful feedback that will encourage improved performance
18
Creating Rubrics
– Course curriculum
• Holistic– Summative
– The focus of a score reported using a holistic rubric is on the overall quality, proficiency, or understanding of the specific content and skills. Holistic rubrics can result in a less time consuming scoring process than use of analytic rubrics.
19
Creating Rubrics
– Course curriculum• Analytic
– Student feedback
» Quantitative
» Qualitative
– An analytic rubric articulates levels of performance for each criterion so instructor can assess student performance on each. Provides detailed feedback for the students on their strengths and weaknesses. Most important, it continues the learning process.
20
Creating Rubrics
– Course curriculum• Analytic
– Student feedback» Quantitative» Qualitative
– The various levels of student performance can be defined using either quantitative (i.e., numerical) or qualitative (i.e., descriptive) labels. In some instances, both quantitative and qualitative labels can be utilized. If a rubric contains four levels of proficiency or understanding on a continuum, quantitative labels would typically range from "1" to "4." When using qualitative labels, there is much more flexibility, and can be more creative. A common type of qualitative scale might include these labels: master, expert, apprentice, and novice. Nearly any type of qualitative scale will suffice, provided it "fits" with the task.
21
Creating Rubrics
Rubrics support data driven decision making
– Data-driven decision making (DDDM):
• uses data on function, quantity and quality of inputs
• Examines how students learn to suggest educational solutions
• Based on the assumption that scientific methods can effectively evaluate educational programs, and instructional methods.
– Rubrics provide significant data to support effective DDDM.
22
Creating Rubrics
• Creating a customized rubric• Step 1: Re-examine the learning objectives to be addressed
by the task. This allows you to match your scoring guide with your objectives and actual instruction.
• Step 2: Identify specific observable attributes that you want to see (as well as those you don’t want to see) your students demonstrate in their product, process, or performance. Specify the characteristics, skills, or behaviors that you will be looking for, as well as common mistakes you do not want to see.
• Step 3: Brainstorm characteristics that describe each attribute. Identify ways to describe above average, average, and below average performance for each observable attribute identified in Step 2.
24
Creating Rubrics
• Creating a customized rubric
• Step 4a: For holistic rubrics, write thorough narrative descriptions for excellent work and poor work incorporating each attribute into the description. Describe the highest and lowest levels of performance combining the descriptors for all attributes.
• Step 4b: For analytic rubrics, write thorough narrative descriptions for excellent work and poor work for each individual attribute. Describe the highest and lowest levels of performance using the descriptors for each attribute separately.
25
Creating Rubrics
• Creating a customized rubric• Step 5a: For holistic rubrics, complete the rubric
by describing other levels on the continuum that ranges from excellent to poor work for the collective attributes. Write descriptions for all intermediate levels of performance.
• Step 5b: For analytic rubrics, complete the rubric by describing other levels on the continuum that ranges from excellent to poor work for each attribute. Write descriptions for all intermediate levels of performance for each attribute separately.
26
Creating Rubrics
• Creating a customized rubric• Step 6: Collect samples of student work
that exemplify each level. These will help you score in the future by serving as benchmarks.
• Step 7: Revise the rubric, as necessary. Be prepared to reflect on the effectiveness of the rubric and revise it prior to its next implementation.
27
Creating RubricsTable 2: Template for analytic rubrics
Criteria #1 Beginning 1
Developing2
Accomplished3
Exemplary4
Score
Criteria #2 beginning level of performance
movement toward mastery level of performance
achievement of mastery level of performance
highest level of performance
Criteria #3 beginning level of performance
movement toward mastery level of performance
achievement of mastery level of performance
highest level of performance
Criteria #4 beginning level of performance
movement toward mastery level of performance
achievement of mastery level of performance
highest level of performance
Criteria #5 beginning level of performance
movement toward mastery level of performance
achievement of mastery level of performance
highest level of performance
28
Creating Rubrics
Table 1:Template for Holistic Rubrics
Score Description
5Demonstrates complete understanding of the problem. All requirements of task are included in response.
4Demonstrates considerable understanding of the problem. All requirements of task are included.
3Demonstrates partial understanding of the problem. Most requirements of task are included.
2Demonstrates little understanding of the problem. Many requirements of task are missing.
1 Demonstrates no understanding of the problem.
29
Creating Rubrics
• Assigning scores to the rubric• If a rubric contains four levels of proficiency or understanding on a continuum,
quantitative labels would typically range from "1" to "4." When using qualitative labels, teachers have much more flexibility, and can be more creative. A common type of qualitative scale might include the following labels: master, expert, apprentice, and novice. Nearly any type of qualitative scale will suffice, provided it "fits" with the task.
• One potentially frustrating aspect of scoring student work with rubrics is the issue of somehow converting them to "grades." It is not a good idea to think of rubrics in terms of percentages (Trice, 2000). For example, if a rubric has six levels (or "points"), a score of 3 should not be equated to 50% (an "F" in most letter grading systems). The process of converting rubric scores to grades or categories is more a process of logic than it is a mathematical one. Trice (2000) suggests that in a rubric scoring system, there are typically more scores at the average and above average categories (i.e., equating to grades of "C" or better) than there are below average categories. For instance, if a rubric consisted of nine score categories, the equivalent grades and categories might look like this:
30
Creating Rubrics
31
Table 3: Sample grades and categories
Rubric Score Grade Category
8 A+ Excellent
7 A Excellent
6 B+ Good
5 B Good
4 C+ Fair
3 C Fair
2 U Unsatisfactory
1 U Unsatisfactory
0 Y Unsatisfactory
Hands on Practice
Create a holistic or analytical rubric for a course in your discipline
Use blank rubric forms in your handbook
32
Evaluating A Rubric
• Evaluating A Rubric for effective assessment
– A Rubric for Rubrics: A Tool for Assessing the
Quality and Use of Rubrics in Education
• Downloaded January 22, 2010 from http://webpages.charter.net/bbmullinix/Rubrics/A%20Rubric%20for%20Rubrics.htm
• Dr. Bonnie B. Mullinix © Monmouth University December 2003
34
Evaluating A RubricCriteria 1 Acceptable 3 Acceptable 3 Exemplary Exemplary
Clarity of criteria Criteria being assessed are unclear, inappropriate and/orhave significant overlap
Criteria being assessed can be identified, but are not clearlydifferentiated or areinappropriate
Criteria being assessed are clear, appropriate and distinct
Each criteria is distinct, clearly delineated and fully appropriate for the assignment(s) /course
Distinction betweenLevels
Little/no distinction can be made between levels of achievement
Some distinction between levels is made, but is not totally clear how well
Distinction between levels is apparent
Each level is distinct and progresses in a clear and logicalorder
Distinction betweenLevels
Rubric is not shared with learners
Rubric is shared and provides some idea of the assignment/expectations
Rubric is referenced - used to introduce an assignment/guidelearners
Rubric serves as primary reference point for discussionand guidance for assignments as well as evaluation ofassignment(s),
35
Evaluating A RubricCriteria 1 Acceptable 3 Acceptable 3 Exemplary Exemplary
Reliability of Scoring
Cross-scoring among facultyand/or students often results insignificant differences
Cross-scoring by faculty and/orstudents occasionally producesinconsistent results
There is general agreementbetween different scorers whenusing the rubric (e.g. differs byless than 5-10% or less than ½level)
Cross-scoring of assignmentsusing rubric results in consistentagreement among scorers
Clarity ofExpectations/Guidance to Learners
Rubric is not shared withlearners
Rubric is shared and providessome idea of the assignment/expectations
Rubric is referenced - used to introduce an assignment/ guide learners
Rubric serves as primaryreference point for discussionand guidance for assignments aswell as valuation of assignment(s)
36
Evaluating A RubricCriteria 1 Acceptable 3 Acceptable 3 Exemplary Exemplary
Support ofMetacognition(Awareness of Learning)
Rubric is not shared with learners
Rubric is shared but not discussed/ referenced withrespect to what is being learnedthrough theassignment(s)/course
Rubric is shared and identifiedas a tool for helping learners tounderstand what they are learning through the Assignment/ in the course
Rubric is regularly referencedand used to help learners identify the skills andknowledge they are developingthroughout the course/assignment(s)
Engagement ofLearners in RubricDevelopment/ Use *
Learners are not engaged ineither development or use of therubrics
Learners offered the rubric andmay choose to use it for selfassessment
Learners discuss the design of the rubric and offer Feedback / input and areresponsible for use of rubrics inpeer and/or self-evaluation
Faculty and learners are jointlyresponsible for design of rubricsand learners use them in peerand/or self-evaluation
37
Calibrating a Rubric
• Calibrating rubric scoring for reliable assessment data.– The validity of your rubric and assessment depends in
part on the validation of the rubric scores.– Therefore calibration is an essential process.– Calibration training for a group of instructors who can
then score rubrics and produce valid data is critical.– The process for calibration will be determined by each
discipline or program and may involve paired scoring or open table scoring. Rubric scores are determined by consensus and part of the calibration expert's role is to resolve discrepancies in scoring.
38
Calibrating Rubric Scoring
• Each participant should have a copy of the task, samples of student work, extra rubrics, and scoring sheets.
• Explain procedure (as below)• Ground rules:
– Only one person speaks at a time– Wait to be called on– No side conversations, please– Respect each other’s comments
• Designate a recorder to note any issues, record discussions, and initial and final scores for each box.
39
Calibrating Rubric Scoring
• Examine task for standards, instructor and student directions, prompt, and rubric.
• Discuss and record.• Read first sample of student work only.• Score individually, marking rubric. • When determining final score, consider the preponderance
of evidence to determine where a student’s work falls.• Check the teacher directions page for any accommodations
that might have been made for students with Letters of Accommodation.
• When everyone has finished, ask for a show of hands for an overall score of exceeds, meets, etc.
40
Calibrating Rubric Scoring
• Work box by box, asking for an overall initial score. • Then ask volunteers to support the evaluation, citing
evidence from the student work. • Generally, start with someone who is most supportive of
student work, work towards “below” categories. • This discussion takes quite a bit of time, depending upon
the task and the sample of student work. • At the end of the discussion of each box, take a revote to
determine the final score for that box.• When all boxes have been discussed, ask for a final
overall score.• Use same procedure for sample two.
41
Calibrating Rubric Scoring
• OR (depending upon time and level of expertise) pair score sample two and discuss as needed after an overall vote.
• Read Central scoring calibration notes for comparison/validity purposes.
• Pair score tasks, using score sheets.
• Second scorer should not look at first evaluator’s score until his/her own evaluation is completed.
42
Calibrating Rubric Scoring
• At that point, if both scores agree, record the score on the cover of the task. (Both scorers should also fill out a score sheet).
• If there is a discrepancy of a 3-4 or a 2-3, then a discussion takes place to determine the final score.
• If agreement is reached, one of the scorers changes his/her score sheet to agree with the other score.
43
Glossary of Terms
• Analytic Rubric: An analytic rubric articulates levels of performance for each criterion so the teacher can assess student performance on each criterion.
• Assessment: The systemic process of gathering and analyzing information about student learning to ensure the program and the students are achieving the desired results in relation to the student learning outcomes.
• Assessment plan: a document which states the procedures and processes used to gather and interpret
student learning outcome data.
44
Glossary of Terms
• Authentic Task: An assignment given to students
designed to assess their ability to apply standards-driven knowledge and skills to real-world challenges. A task is considered authentic when 1) students are asked to construct their own responses rather than to select from ones presented; and 2) the task replicates challenges faced in the real world. Good performance on the task should demonstrate, or partly demonstrate, successful completion of one or more standards. The term task is often used synonymously with the term assessment in the field of authentic assessment.
45
Glossary of Terms
• Bloom’s Taxonomy: A classification of learning behaviors or categories used as a way of determining learning progression in a course or program. The revised taxonomy includes lower level skills and high order thinking skills. Bloom’s Taxonomy is used to align Student Learning Outcomes and their subsequent objectives.
• Calibration: Training faculty to score rubrics in a similar fashion to ensure validity of scores and subsequent data.
• Closing the Loop: Primarily regarded as the last step in the assessment process, closing the loop actually starts the process over again if the data is analyzed and the desired results are not achieved. Closing the loop refers specifically to using the data to improve student learning.
46
Glossary of Terms
• Course Goals: Generally phrased, non measureable
statements about what is included and covered in a course.
• Course Guide: a booklet and online resource that helps
students select subjects.
• Curriculum Guide: a practical guide designed to aid
teachers in planning and developing a teaching plan for specific subject areas.
47
Glossary of Terms
• Course Objectives: A subset of student learning outcomes, course objectives are the specific teaching objectives detailing course content and activities.
• Criteria: Characteristics of good performance on a particular task. For example, criteria for a persuasive essay might include well organized, clearly stated, and sufficient support for arguments. (The singular of criteria is criterion.
• Data-driven decision making: A process of making decisions about curriculum and instruction based on the analysis of classroom data, rubric assessment, and standardized test data.
48
Glossary of Terms
• Descriptors: Statements of expected performance at each level of performance for a particular criterion in a rubric - typically found in analytic rubrics. See example and further discussion of descriptors.
• Direct Assessment Method: The assessment is based on an analysis of student behaviors or products in which they demonstrate how well they have mastered learning outcomes.
49
Glossary of Terms
• Indirect Assessment Method: The assessment is based on an analysis of reported perceptions about student mastery of learning outcomes or the learning environment.
• Instructional Alignment: the process of ensuring that Student Learning Outcomes and the subsequent objectives use the same learning behaviors or categories in Bloom’s Taxonomy.
50
Glossary of Terms
• Program Assessment: An on-going, systemic process designed to evaluate and improve student learning by identifying strengths and areas for improvement. The data from the evaluation is used to guide decision making for the program.
• Reliability: The degree to which a measure yields consistent results.
• Rubric: the criteria for classifying products or behaviors into categories which varies along a continuum. Rubrics are used as a way of assessing a Student Learning Outcome. A scoring scale used to evaluate student work. A rubric is composed of at least two criteria by which student work is to be judged on a particular task and at least two levels of performance for each criterion.
51
Glossary of Terms
• Student Learning Outcomes: Student Learning Outcomes are defined as the specific knowledge, skills, or attitude students should be able to demonstrate effectively at the end of a particular course or program. Student Learning Outcomes are measured and provide students, faculty, and staff the ability to assess student learning and instruction. Each course should have four to seven Student Learning Outcomes.
• Validity: The degree to which a certain inference from a test is appropriate and meaningful.
52
Student Learning Cycle
Course Goal
Student Learning Outcomes
Course Objectives
SLO Rubric and Assessment
Program Assessment
53
Bloom’s Taxonomy
First published in 1956, a classification of levels of learning behaviors or categories.
The elements of the original Taxonomy included Knowledge, Comprehension, Application,
Analysis, Synthesis, and Evaluation.
The revised Taxonomy now includes, in the same order, Remembering, Understanding,
Applying, Analyzing, Evaluating, and Creating.
The Taxonomy is often used as a way of determining the progression of learning or
intellectual skills in a program.
More advanced course work, for example, should rely on higher order skills while developmental or
elementary course work could focus on lower level skills.
54
Bloom’s Pyramid
55
Evaluation
Level 6
Synthesis
Level 5
Analysis
Level 4
Application
Level 3
Comprehension
Level 2
Knowledge
Level 1
Bloom's Taxonomy of Cognitive LevelsVerb Table
1 Knowledge 2 Comprehension 3 Application
list
name
identify
show
define
recognize
recall
state
summarize
explain
put into your own
words
interpret
describe
compare
paraphrase
differentiate
demonstrate
visualize
find more
information about
restate
solve
illustrate
calculate
use
interpret
relate
manipulate
apply
classify
modify
put into
practice
4 Analysis 5 Synthesis 6 Evaluation
analyze
organize
deduce
choose
contrast
compare
distinguish
design
hypothesize
support
schematize
write
report
discuss
plan
devise
compare
create
construct
evaluate
choose
estimate
judge
defend
criticize
justify
56
Instructional Alignment
SLO’s are the foundation for teaching and learning district-wide.
Employ the Bloom’s Taxonomy of Cognitive Levels’ verb chart
Align Course Goal, Student Learning Outcomes, Course Objectives,
and Rubric Assessment
A level five SLO should be aligned with level five course objectives
as well as a level five rubric assessment
57
Resource Links• http://www.hccs.edu/hccs/faculty-staff/student-
learning-outcomes--01
58
59
THE CENTER for TEACHING & LEARNING EXCELLENCE
Online Resources: http://hccs.edu/tle