Once the evidence of learning is collected, it needs to be assessed. Consistently applying standards is essential to reliable and usable data. Most commonly, stringent scoring schemas or rubrics are applied to learning artifacts.
A rubric can be defined as a descriptive guideline, a scoring guide or specific pre-established performance criteria in which each level of performance is described to contrast it with the performance at other levels.
Primary Assessment
Primary assessments occur within a course, typically completed by the instructor contemporaneously with the grading process.
Intra-rater Reliability
Intra-rater reliability describes how consistently an individual rater/juror is able to assess evidence. Rather than being indicative of personal inconsistency, poor intra-rater reliability is often a symptom of poorly-designed rubrics.
Juried Assessment
Juried assessment occurs outside of a course setting, typically by collecting student work from multiple courses to be assessed by a committee. The instructors may or may not be jurors on the committee. Also referred to as calibration, norming is a process by which raters/jurors assess the same evidence to level-set expectations prior to assessment of the evidence. Revision, clarification, and interpretation of the rubric can occur during a norming session.
Inter-rater Reliability
Inter-rater reliability describes how consistently a group of raters/jurors is able to assess evidence, or the degree of agreement among raters. Inter-rater reliability is often measured by comparing the assessment results of the same artifacts by two or more independent raters. Poor inter-rater reliability can be indicative of a poorly-designed rubric, but may also be minimized with norming.