← Blog
Traditional tests mostly check memory. Performance assessment asks students to use what they know—solve a problem, make a thing, argue a claim, run a test, explain the why. It’s closer to real work: engineers prototype, scientists model and measure, artists perform, historians work from sources. The point is transfer—can someone apply ideas in a new situation with actual constraints?
Real tasks build the kinds of skills people need for college, work, and community life—reasoning, communicating, collaborating. When there’s a real audience and purpose, effort goes up. Students get why quality matters and how content powers decisions.
It’s also fairer. Timed tests tend to reward speed or test tricks. Performance tasks open more doors: talks, briefs, prototypes, models, performances, reflections. With clear supports and culturally responsive contexts, more students can show what they know—not just the fastest readers.
Think like a coach across a season: quick tasks test transfer, projects build depth, exhibitions add pressure and pride, portfolios show growth.
Start with the end. Name three to five actions students must do to show mastery (e.g., model a system with justified assumptions, make an evidence-based claim, communicate for a specific audience). Decide what counts as convincing evidence—what you need to see or hear.
Use RAFT to frame it:
Map supports in advance: exemplars, sentence frames, graphic organizers, curated datasets, bilingual glossaries. Add two or three non-negotiables tied to validity (quantify uncertainty, cite two sources, name one limitation). Pilot with a small group, collect questions, tweak, then launch.
Rubrics turn complex work into teachable parts. Use analytic rubrics for criterion-level feedback (reasoning, accuracy, communication) and holistic rubrics for quick overall judgments. Four levels usually balance clarity with nuance. Write observable descriptors: “Integrates two data sources to justify the claim and quantifies the relationship,” not “strong reasoning.”
Share rubrics early. Study exemplars with students. Do quick co-scoring with a colleague to align interpretations. Early on, a student-facing single-point rubric keeps focus without overload. Add a light workflow criterion (timeliness, documentation) so real-world habits show up without overshadowing thinking.
Launch by unpacking the prompt and rubric, then examine exemplars. Teach mini-lessons on tricky bits (reading error bars, audience-aware intros, quick prototype tests). Use checklists to make invisible steps visible. Build checkpoints—proposal, outline, draft, peer review, rehearsal—with short, criterion-tied feedback at each step. In teams, assign rotating roles and require individual evidence.
Plan access from the start: UDL-aligned options, multiple representations, language supports. Keep constraints realistic—limited materials, noisy real data, time budgets. A quick “pre-flight” conference helps: students explain how their plan hits each criterion before they go further.
Make feedback timely, specific, and tied to criteria: “Claim is clear; quantify the trend in Figure 2 to strengthen reasoning.” Use targeted peer-review checklists so comments stay on-task. Protect class time for revision or feedback turns into background noise. Require a brief response-to-feedback note on final submissions, naming criteria targeted and changes made; it speeds grading and builds metacognition.
Check validity first: does the task actually elicit the construct? If you’re assessing argumentation with data, don’t bury it under excessive reading. Reliability grows from clear rubrics, anchors at each level, and quick calibration cycles during the unit. For fairness, minimize construct-irrelevant barriers and audit contexts for hidden assumptions. Do a fast “bias and burden” check with a colleague and adjust before launch.
Performance assessment works best as a system, not a one-off. Map task types across a course or grade band to build independence. Portfolios curate artifacts over time with reflective notes tied to competencies; schedule defenses or conferences. Exhibitions create real accountability and pride. Report learning with standards-based snapshots and brief narrative comments; criterion-level evidence is far more useful than a single score. Add a few short, standardized micro-tasks across the year to complement long projects.
Traditional tests are efficient for recall and procedures. Performance tasks center application, reasoning, and communication in realistic conditions. Project-Based Learning is an instructional approach; performance assessment is how you evaluate results. They pair well: a PBL unit can culminate in a performance task with a clear rubric and public audience. A balanced mix—quizzes for retrieval, on-demand tasks for transfer, exhibitions/portfolios for the full portrait—beats any single measure.
Pick tools that make this doable: rubric platforms with criterion scoring and comments, portfolio tools with tagging and reflection prompts, peer-review workflows, version history. Build a small exemplar bank each term—samples at each level plus short commentaries explaining why they fit. Over time, that library becomes your best PD for new staff and a fast way to calibrate and norm with students.
Context & Purpose: Short real-world setup and why it matters.
RAFT:
Evidence Targets: 3–5 criteria tied to standards (evidence-based reasoning, modeling, precision, communication, collaboration).
Rubric Skeleton: Four levels with concise, observable descriptors; include anchors.
Resources & Supports: Texts, datasets, tools, mini-lessons, language/access options.
Deliverables & Checkpoints: Proposal, draft, peer review, rehearsal, final product, reflection on how feedback shaped revisions.
Elementary: Design-a-solution challenges (shade structures), reading-to-writing responses, science journals with simple models; lots of visuals and rehearsal.
Middle School: Data-argument briefs in science and social studies, community proposals in ELA/civics, multi-week projects with clearly staged deliverables and peer review.
High School: Portfolio defenses, capstones, civic exhibitions, lab practicums, juried arts critiques; add pro norms like deadlines, documentation, and public audiences.
CTE/Arts: Client briefs, prototypes with test logs, performances, curated galleries with artist statements or engineering notebooks; score process and product.
Cross-Curricular: STEAM showcases, humanities-science collabs, short design sprints on local issues.
Let the rubric-level data drive next steps. In class, target mini-lessons to the criteria that dipped. As a team, share tasks, co-calibrate, and grow your exemplar library. For students, turn assessment into agency: set criterion-based goals, track growth across artifacts, reflect and plan next moves. Program-wide, keep an eye on task quality, scoring reliability, and equitable outcomes.
Start small: one on-demand task, four-criterion rubric, a 20-minute co-scoring chat, and a revision pass. Then scale. The goal isn’t perfection—it’s giving students a fair, meaningful way to show what they can really do.