Resources · Writeiq framework
The Integrated Writing Assessment Framework is the pedagogical spine of Writeiq. This page explains what it is in plain English, how the four achievement bands work, and the research evidence behind each decision — including which calls are research-backed and which are considered design choices we will validate through standard-setting work.
IWAF stands for the Integrated Writing Assessment Framework. It is a criterion-referenced writing assessment framework developed by Edsthetic that structures the way Writeiq marks, gives feedback on, and reports student writing across six writing modes (narrative, persuasive, analytical, primary narrative, recount, primary persuasive) and nine curriculum frameworks (Victorian Curriculum 2.0, Australian Curriculum v9, NESA, QCAA, WASCSA, IB MYP, IB PYP, VCE, and the UK National Curriculum).
The framework is the intellectual property of Edsthetic and is listed in our terms of service. “Integrated” has a specific meaning here: the framework integrates three things that are usually treated separately — criterion-level descriptors, logic gates that catch scoring inconsistencies, and actionable feedback language for students, teachers, and leaders. Each criterion has one entry in the framework that produces all three outputs.
IWAF 3.0 is the current version. It launched in April 2026 with recalibrated achievement band thresholds, a reorganised criterion structure (the framework now uses CR-01 to CR-11 canonical codes rather than the older theme labels), and expanded feedback content authored by our writing specialist Mikhaila Picone.
Most writing assessment produces a score and stops. The teacher still has to work out what the score means, which skills to teach, and what feedback to give. We built IWAF to close that loop.
IWAF 3.0 assigns each student one of four achievement bands based on their total score as a percentage of the task maximum. The four bands, and their thresholds, are:
The four-point structure matches the shape most teachers already think in, and matches the four-point achievement labels used across every major Australian curriculum framework (at standard, above standard, below standard, well below standard — each with local variant terminology). We picked four bands deliberately rather than three, five, or a continuous score, because four lets a leader look at a heat map and instantly see where the work is.
We want to be clear about which of the four threshold values are research-derived and which are design decisions. This matters because we are building a product schools rely on, and overstating research support is how educational products lose credibility.
| Threshold | Evidence strength | What the evidence says |
|---|---|---|
| 80% Extending | Strong | Bloom’s mastery learning framework uses 80% as the mastery criterion (Bloom, 1968; Guskey, 2007). Rosenshine’s research on effective instruction identifies 80% as the threshold for moving from guided to independent practice (Rosenshine, 2012). AERO’s synthesis of Australian evidence-based teaching practice uses 80% consistently. This threshold has strong empirical grounding. |
| 56% Consolidating | Moderate | The boundary between “can do with support” and “can do reliably alone” sits around the middle of the score range in most standards-referenced assessment literature (Popham, 2008; Sadler, 2005). 56% (rather than exactly 60%) reflects criterion-referenced scale calibration where consistency across a whole piece, not just strong moments, determines the threshold. |
| 31% Developing | Design choice | The boundary between Emerging and Developing is currently a design decision, not a research-derived value. It reflects our judgement about where “starting to show the skill” sits for writing specifically. We will validate this threshold through an ACER Body of Work standard-setting study in Term 3 2026 and adjust if needed. |
A narrative is marked out of 50. An analytical response is marked out of 46. A primary persuasive piece is marked out of 51. A percentage normalises across these different maxima so a 70% in narrative and a 70% in analytical reflect comparable developmental standing. This also lets schools compare growth across tasks of different types and across year levels.
IWAF 3.0 draws on five strands of published research. We list the primary sources so you can read them yourselves; the page on curriculum framework alignment covers how IWAF bands map to Victorian Curriculum, Australian Curriculum, NESA, QCAA, WASCSA, MYP, PYP, VCE, and UK descriptors.
The four IWAF achievement bands are designed to align with the Multi-Tiered System of Supports (MTSS) framework — the approach the Australian Education Research Organisation (AERO) recommends for organising literacy and numeracy intervention in Australian schools. This alignment is deliberate, not coincidental.
MTSS is a three-tier prevention framework that sorts students into layers of support based on how they respond to instruction. It emerged from decades of Response to Intervention (RTI) and Positive Behavioural Interventions and Supports (PBIS) research in the United States, and is now endorsed in Australia by AERO, the Australian Government Department of Education, and widely implemented across Catholic, independent, and government school sectors.
The three tiers are:
| Tier | Expected share | What it means |
|---|---|---|
| Tier 1 | ~80% | Universal / core instruction. High-quality, evidence-based classroom teaching that meets the needs of most students without additional intervention. When Tier 1 is strong, around 80% of students progress adequately with no further support. |
| Tier 2 | ~15% | Targeted small-group support. Supplementary intervention for students who are not responding adequately to Tier 1. Usually delivered in small groups, 2–3 times weekly, with regular progress monitoring (at least monthly). |
| Tier 3 | ~1-5% | Intensive individualised intervention. For students who show minimal response to Tier 2. Individualised, delivered 4–5 times weekly in smaller groups or 1:1, with frequent (weekly) progress monitoring. |
These tier proportions are canonical across the MTSS literature. Fuchs (2010) and the American Institutes for Research’s National Center on Intensive Intervention state them explicitly. AERO’s 2024 explainer cites the same distribution. A school that finds more than 20% of students needing Tier 2 support is generally considered to have a Tier 1 instruction problem, not a student problem.
When Mikhaila Picone recalibrated the IWAF 3.0 band thresholds in April 2026, the proportions fell out almost exactly onto the MTSS distribution:
| IWAF band | Threshold | MTSS correspondence |
|---|---|---|
| Extending | 80%+ | Tier 1. Student is thriving in core instruction. Continue universal teaching; consider extension tasks, exemplar status, or peer mentoring roles. |
| Consolidating | 56–79% | Tier 1. Student is on track with core instruction. No additional intervention required — keep teaching at pace, continue to monitor progress. |
| Developing | 31–55% | Tier 2 signal. Student is showing partial or inconsistent skill application. Consider targeted small-group support for the specific criteria they’re struggling with. Writeiq’s Teach Next panel identifies those priority criteria automatically. |
| Emerging | 0–30% | Tier 3 signal. Student requires intensive individualised support. A sustained Emerging pattern across multiple tasks, despite adjustments, is a strong indicator that intensive intervention is warranted. |
Writeiq does not make tier placement decisions. Tier decisions are made by school intervention teams — literacy leads, learning support coordinators, psychologists, classroom teachers — using multiple sources of evidence over time. Writeiq contributes one important piece of that evidence picture: consistent, criterion-level, curriculum-aligned data on how each student’s writing is developing.
A student showing Emerging across two or three Writeiq assessments on different tasks is a flag worth discussing. A student moving from Developing to Consolidating after targeted teaching is evidence that the intervention is working. A student who stays Emerging despite adjustments is evidence of limited response that the school’s intervention team can act on.
The Nationally Consistent Collection of Data on School Students with Disability (NCCD) requires schools to document the adjustments they provide to students with disability, and to retain evidence of those adjustments. The NCCD Portal explicitly lists “documentation of ongoing learning needs that have a limited response to targeted intervention over time” among the valid categories of evidence supporting inclusion decisions (nccd.edu.au, What evidence is the NCCD based upon?).
Writeiq data can contribute to this evidence picture in three specific ways:
The MTSS framework and its application to Australian schools are grounded in a concentrated body of local research:
We think it matters to say clearly what we are not claiming. Five things in particular:
A rubric is a scoring tool. A framework describes a whole way of doing something. IWAF includes the rubric descriptors and the logic gates that check scoring consistency, the band structure that interprets the scores, the feedback language for students and teachers, the curriculum framework mappings, and the pedagogical sequence for teaching to the identified priority. Calling IWAF a rubric would understate what it actually does.
Three substantive changes. First, band thresholds recalibrated from 40/60/80 to 31/56/80. The old thresholds let too many students sit in Developing when their writing showed clearer foundational gaps. Second, the criterion structure was reorganised from dimension-keyed (sentence, paragraph, craft…) to criterion-keyed (CR-01 Audience and Purpose, CR-02 Text Structure, CR-03 Ideas, and so on). This makes the framework easier to reason about and easier to align to curriculum descriptors. Third, the feedback content library was rewritten against new authoring rules around positive framing, concrete next steps, and no self-diagnosis.
The four IWAF bands align with the three-tier MTSS framework endorsed by AERO and the Australian Government. Extending and Consolidating capture the roughly 80% of students supported successfully by Tier 1 core instruction. Developing corresponds to the roughly 15% typically needing Tier 2 targeted small-group support. Emerging corresponds to the roughly 1–5% needing Tier 3 intensive individualised intervention.
This alignment is deliberate — the 80% Extending threshold draws on the same mastery-learning research base (Bloom, Rosenshine, AERO) that underpins the MTSS tier distribution. See the full MTSS section above for the tier-by-tier correspondence and source references.
Writeiq does not make tier placement decisions. Tier decisions are made by school intervention teams using multiple sources of evidence. Writeiq contributes one important piece of that picture: consistent criterion-level writing data.
Yes, in the same way any systematically collected assessment and intervention data can contribute to NCCD evidence. The NCCD Portal explicitly lists “documentation of ongoing learning needs that have a limited response to targeted intervention over time” and “evidence of interventions provided over time, with monitoring of the effectiveness of the intervention” as valid evidence categories.
Writeiq generates exactly this kind of evidence: dated, criterion-level writing assessments showing where a student sits, how they respond to teaching, and whether that response is adequate. Teacher score adjustments are logged with reason text, capturing NCCD-relevant accommodations (reader, extra time, and so on) alongside the student’s work.
What Writeiq does not do: diagnose disability, determine NCCD category, or set level of adjustment. Those remain the professional judgement of teachers, learning support staff, and specialist assessors. For broader NCCD allocation, evidence accumulation, and 3-sheet acquittal export, see Edsthetic’s companion product Allocateiq.
Not yet. IWAF 3.0 is designed for the conventions of English writing: narrative, persuasive, and analytical text response. Applying English criteria to a science investigation or a humanities source analysis would produce misleading feedback — those faculties assess different qualities (methodology, evidence, reasoning) that the English rubric does not measure.
We have an IWAF 4.0 roadmap item for bespoke, faculty-designed criterion sets. That work is scoped for a future major update so the rubrics are built properly with faculty leads rather than adapted on the fly. If you’d like to be part of that design work, email hello@edsthetic.com.au.
Open the submission, click Adjust next to any criterion score, and change it. You will be prompted to confirm the change, enter the revised score, and add a reason. The original Writeiq score is retained with a strikethrough, your adjustment is flagged, and the reason is logged. This is the moderation pathway — Writeiq is designed to give you a defensible starting point, not a final ruling.
Yes. We have planned an ACER Body of Work standard-setting study for Term 3 2026. This is the standard Australian method for validating criterion-referenced assessment thresholds empirically. We will publish the results, including any threshold adjustments the study recommends.
We also welcome independent review from universities, research centres, and state education departments. If you are in a position to help validate the framework, contact hello@edsthetic.com.au.
IWAF 3.0 has three primary modes — Primary Narrative, Recount, and Primary Persuasive — each with criteria pitched at Years 3–6 expectations. The framework is calibrated to Derewianka and Jones’ primary metalanguage (2016) and aligned to AERO’s School Writing Instruction Framework (2025). Primary feedback uses simpler language and shorter sentences, but holds the same four-band structure so schools can track growth into secondary without breaking the data.
Criteria are what gets scored. Logic gates are consistency rules the marker must pass before a score is returned. For example: if a student’s writing has very little punctuation, a gate caps the Sentence Structure score because a reader cannot judge sentence structure without punctuation. If a student uses bullet points, a gate caps the Cohesion score because cohesion is assessed across continuous prose. Gates do not add information; they catch the kinds of inconsistency that undermine trust in a score.
Many gates surface a framing sentence to the student: “Your writing has very little punctuation. Before anything else, make sure every sentence has a capital letter at the start and a full stop at the end.” The gate caps the score and tells the student why.
IWAF is authored by Mikhaila Picone (writing specialist and co-founder of Edsthetic) with technical implementation by Ashwin Pillai (co-founder, Deputy Principal for Pedagogy at Sacred Heart College Geelong). The framework draws on Mikhaila’s work across primary and secondary English teaching, and on published research from AERO, Rosenshine, Hattie, Fisher and Frey, Martin and Rose, Derewianka and Jones, and others cited in the References section below.
The research literature IWAF 3.0 is grounded in. This is a working list, not an exhaustive bibliography.
If you are evaluating Writeiq and want to talk through how IWAF 3.0 would apply to your school’s context — including validation, moderation, or how the framework maps to your reporting language — email us at hello@edsthetic.com.au.