Resources · Writeiq framework

What is IWAF 3.0?

The Integrated Writing Assessment Framework is the pedagogical spine of Writeiq. This page explains what it is in plain English, how the four achievement bands work, and the research evidence behind each decision — including which calls are research-backed and which are considered design choices we will validate through standard-setting work.

What is IWAF

IWAF stands for the Integrated Writing Assessment Framework. It is a criterion-referenced writing assessment framework developed by Edsthetic that structures the way Writeiq marks, gives feedback on, and reports student writing across six writing modes (narrative, persuasive, analytical, primary narrative, recount, primary persuasive) and nine curriculum frameworks (Victorian Curriculum 2.0, Australian Curriculum v9, NESA, QCAA, WASCSA, IB MYP, IB PYP, VCE, and the UK National Curriculum).

The framework is the intellectual property of Edsthetic and is listed in our terms of service. “Integrated” has a specific meaning here: the framework integrates three things that are usually treated separately — criterion-level descriptors, logic gates that catch scoring inconsistencies, and actionable feedback language for students, teachers, and leaders. Each criterion has one entry in the framework that produces all three outputs.

IWAF 3.0 is the current version. It launched in April 2026 with recalibrated achievement band thresholds, a reorganised criterion structure (the framework now uses CR-01 to CR-11 canonical codes rather than the older theme labels), and expanded feedback content authored by our writing specialist Mikhaila Picone.

Plain-English definition: IWAF is how Writeiq decides what a piece of writing is strong at, what needs work, and how to say both of those things in a way a student, teacher, or principal can act on.

Why we built it

Most writing assessment produces a score and stops. The teacher still has to work out what the score means, which skills to teach, and what feedback to give. We built IWAF to close that loop.

Problem one · Scores without direction
A number out of 50 tells a teacher how a student did on the task. It does not tell them what to teach next. IWAF assigns each student a band and a priority criterion — the single skill that, if taught, would most lift their band.
Problem two · Feedback that does not differentiate
Generic “try to use better vocabulary” feedback helps no one. IWAF’s feedback content is band-differentiated: a student sitting in Emerging gets a different next-step sentence to a student sitting in Consolidating, even on the same criterion. A Year 7 writing at Extending hears different language to a Year 12 writing at Extending. The framework does this structurally, not through prompt tricks.
Problem three · Inconsistency between teachers and schools
Two English teachers marking the same piece often disagree by five marks. IWAF’s 26 logic gates act as a consistency layer: rules like “vocabulary cannot outrun sentence structure by more than 20 per cent” catch the kinds of inconsistency that moderation meetings spend hours resolving.
Problem four · Writing assessment that does not help leaders
A literacy leader needs whole-school signal, not a pile of individual reports. IWAF produces criterion-level data that aggregates cleanly to class, year level, and school views — the same evidence a teacher uses to plan a lesson is what the head of school uses to plan a term.

How the bands work

IWAF 3.0 assigns each student one of four achievement bands based on their total score as a percentage of the task maximum. The four bands, and their thresholds, are:

Emerging
0–30%
Foundational capability. Requires targeted explicit teaching.
Developing
31–55%
Partial, inconsistent application. Ready for guided practice.
Consolidating
56–79%
Reliable, developing command. Ready for independent tasks.
Extending
80%+
Sophisticated, confident, purposeful application.

The four-point structure matches the shape most teachers already think in, and matches the four-point achievement labels used across every major Australian curriculum framework (at standard, above standard, below standard, well below standard — each with local variant terminology). We picked four bands deliberately rather than three, five, or a continuous score, because four lets a leader look at a heat map and instantly see where the work is.

The important honesty about thresholds

We want to be clear about which of the four threshold values are research-derived and which are design decisions. This matters because we are building a product schools rely on, and overstating research support is how educational products lose credibility.

Threshold Evidence strength What the evidence says
80% Extending Strong Bloom’s mastery learning framework uses 80% as the mastery criterion (Bloom, 1968; Guskey, 2007). Rosenshine’s research on effective instruction identifies 80% as the threshold for moving from guided to independent practice (Rosenshine, 2012). AERO’s synthesis of Australian evidence-based teaching practice uses 80% consistently. This threshold has strong empirical grounding.
56% Consolidating Moderate The boundary between “can do with support” and “can do reliably alone” sits around the middle of the score range in most standards-referenced assessment literature (Popham, 2008; Sadler, 2005). 56% (rather than exactly 60%) reflects criterion-referenced scale calibration where consistency across a whole piece, not just strong moments, determines the threshold.
31% Developing Design choice The boundary between Emerging and Developing is currently a design decision, not a research-derived value. It reflects our judgement about where “starting to show the skill” sits for writing specifically. We will validate this threshold through an ACER Body of Work standard-setting study in Term 3 2026 and adjust if needed.
We will not claim the Developing threshold is research-derived until the standard-setting study is complete. Any school asking us about the evidence base for the 31% boundary will get this honest answer. Our commitment is to validate the calibration empirically and publish the result.

Why percentages and not raw scores

A narrative is marked out of 50. An analytical response is marked out of 46. A primary persuasive piece is marked out of 51. A percentage normalises across these different maxima so a 70% in narrative and a 70% in analytical reflect comparable developmental standing. This also lets schools compare growth across tasks of different types and across year levels.

The research evidence base

IWAF 3.0 draws on five strands of published research. We list the primary sources so you can read them yourselves; the page on curriculum framework alignment covers how IWAF bands map to Victorian Curriculum, Australian Curriculum, NESA, QCAA, WASCSA, MYP, PYP, VCE, and UK descriptors.

1. Gradual Release of Responsibility (GRR)
The “I Do · We Do · You Do” model (Pearson & Gallagher, 1983; Fisher & Frey, 2008) structures how IWAF’s teacher-facing feedback is delivered. Every criterion has explicit modelling strategies, guided practice prompts, and independent application tasks. This is how Writeiq lesson plans move from teacher-led demonstration to student independence.
2. Principles of Instruction
Rosenshine’s Principles of Instruction (2012) and the work of Anita Archer and Charles Hughes on explicit teaching ground the structure of IWAF’s feedback loop: clear modelling, frequent low-stakes checking, structured practice, and systematic review. The 26 logic gates are a direct implementation of Rosenshine’s principle of error-catching during guided practice.
3. Feedback that moves learning forward
Hattie and Timperley’s feedback model (2007) identifies three questions effective feedback must answer: Where am I going? How am I going? Where to next? Every Writeiq student feedback block is structured to answer all three. The criterion score addresses “Where am I going”. The band descriptor addresses “How am I going”. The Next Step sentence addresses “Where to next”.
4. Genre pedagogy
The Sydney School genre pedagogy tradition (Martin & Rose, 2008; Derewianka & Jones, 2016) informs IWAF’s six writing modes and the criterion structure within each. A narrative is assessed against orientation-complication-resolution structure; an analytical response against claim-evidence-effect; a recount against orientation-events-reorientation. These are not arbitrary categories — they are the genre-specific features Australian curriculum assessment has drawn on for two decades.
5. Mastery learning and standard-setting
Bloom’s mastery learning framework (1968) and Guskey’s subsequent research (2007) inform our Extending threshold and the broader commitment that most students can reach high standards given appropriate instruction. The ACER Body of Work standard-setting methodology (Cizek, 2012) is the approach we will use in 2026 to empirically validate IWAF 3.0 band boundaries.

MTSS and NCCD alignment

The four IWAF achievement bands are designed to align with the Multi-Tiered System of Supports (MTSS) framework — the approach the Australian Education Research Organisation (AERO) recommends for organising literacy and numeracy intervention in Australian schools. This alignment is deliberate, not coincidental.

What MTSS is

MTSS is a three-tier prevention framework that sorts students into layers of support based on how they respond to instruction. It emerged from decades of Response to Intervention (RTI) and Positive Behavioural Interventions and Supports (PBIS) research in the United States, and is now endorsed in Australia by AERO, the Australian Government Department of Education, and widely implemented across Catholic, independent, and government school sectors.

The three tiers are:

Tier Expected share What it means
Tier 1 ~80% Universal / core instruction. High-quality, evidence-based classroom teaching that meets the needs of most students without additional intervention. When Tier 1 is strong, around 80% of students progress adequately with no further support.
Tier 2 ~15% Targeted small-group support. Supplementary intervention for students who are not responding adequately to Tier 1. Usually delivered in small groups, 2–3 times weekly, with regular progress monitoring (at least monthly).
Tier 3 ~1-5% Intensive individualised intervention. For students who show minimal response to Tier 2. Individualised, delivered 4–5 times weekly in smaller groups or 1:1, with frequent (weekly) progress monitoring.

These tier proportions are canonical across the MTSS literature. Fuchs (2010) and the American Institutes for Research’s National Center on Intensive Intervention state them explicitly. AERO’s 2024 explainer cites the same distribution. A school that finds more than 20% of students needing Tier 2 support is generally considered to have a Tier 1 instruction problem, not a student problem.

How IWAF bands map to the tiers

When Mikhaila Picone recalibrated the IWAF 3.0 band thresholds in April 2026, the proportions fell out almost exactly onto the MTSS distribution:

IWAF band Threshold MTSS correspondence
Extending 80%+ Tier 1. Student is thriving in core instruction. Continue universal teaching; consider extension tasks, exemplar status, or peer mentoring roles.
Consolidating 56–79% Tier 1. Student is on track with core instruction. No additional intervention required — keep teaching at pace, continue to monitor progress.
Developing 31–55% Tier 2 signal. Student is showing partial or inconsistent skill application. Consider targeted small-group support for the specific criteria they’re struggling with. Writeiq’s Teach Next panel identifies those priority criteria automatically.
Emerging 0–30% Tier 3 signal. Student requires intensive individualised support. A sustained Emerging pattern across multiple tasks, despite adjustments, is a strong indicator that intensive intervention is warranted.
The alignment is by design. The 80% / 56% / 31% thresholds weren’t chosen to fit MTSS — they were chosen because 80% has strong research grounding as a mastery threshold (Bloom; Rosenshine; AERO). The fact that the resulting band distribution aligns with the canonical ~80% / ~15% / ~5% tier distribution reflects the same underlying theory, not a coincidence.

What this means in practice

Writeiq does not make tier placement decisions. Tier decisions are made by school intervention teams — literacy leads, learning support coordinators, psychologists, classroom teachers — using multiple sources of evidence over time. Writeiq contributes one important piece of that evidence picture: consistent, criterion-level, curriculum-aligned data on how each student’s writing is developing.

A student showing Emerging across two or three Writeiq assessments on different tasks is a flag worth discussing. A student moving from Developing to Consolidating after targeted teaching is evidence that the intervention is working. A student who stays Emerging despite adjustments is evidence of limited response that the school’s intervention team can act on.

IWAF and NCCD evidence

The Nationally Consistent Collection of Data on School Students with Disability (NCCD) requires schools to document the adjustments they provide to students with disability, and to retain evidence of those adjustments. The NCCD Portal explicitly lists “documentation of ongoing learning needs that have a limited response to targeted intervention over time” among the valid categories of evidence supporting inclusion decisions (nccd.edu.au, What evidence is the NCCD based upon?).

Writeiq data can contribute to this evidence picture in three specific ways:

1. Sustained non-response documentation
When a student remains Emerging across multiple Writeiq assessments despite exposure to Tier 1 and Tier 2 adjustments, Writeiq’s submission history provides dated, criterion-level documentation of that pattern. This is exactly the kind of evidence NCCD coordinators and learning support teams use when making supplementary or above adjustment decisions.
2. Adjustment-linked evidence
Teacher score adjustments in Writeiq are logged with reason text (required on every override). If a teacher adjusts a score because the student received a reader, extra time, or another NCCD-relevant adjustment, that documentation is captured alongside the evidence of the student’s work. Both the original score and the adjusted score are retained, with the reason attached.
3. Progress monitoring for fidelity of intervention
NCCD expects schools to monitor the effectiveness of the adjustments they provide. Writeiq’s growth tracking across tasks produces the kind of longitudinal criterion-level data that makes progress or non-progress visible. A student moving Emerging → Developing after 10 weeks of Tier 2 support is evidence the intervention is working; a student staying Emerging is evidence a change is needed.
What Writeiq does not do: Writeiq does not diagnose disability. It does not determine NCCD category or level of adjustment. It does not replace specialist assessment by speech pathologists, educational psychologists, or paediatricians. It is one source of evidence among several that schools use to make informed decisions. Edsthetic’s companion product Allocateiq is the dedicated NCCD allocation and evidence platform; Writeiq contributes writing-specific data into that broader picture.

Key Australian references

The MTSS framework and its application to Australian schools are grounded in a concentrated body of local research:

What we do not claim

We think it matters to say clearly what we are not claiming. Five things in particular:

We do not claim IWAF is a research instrument
IWAF 3.0 is a classroom assessment tool grounded in the research literature cited above. It is not a peer-reviewed research instrument and we have not yet published validation data. Our plan is to run an ACER Body of Work standard-setting study in Term 3 2026 and submit calibration data to a peer-reviewed education journal in 2027.
We do not claim IWAF diagnoses disability or determines NCCD inclusion
A band placement is not a diagnosis. Writeiq data contributes evidence that supports decisions made by school intervention teams, specialist assessors, and NCCD coordinators. Determining whether a student has disability as defined in the Disability Discrimination Act 1992, what their NCCD category is, and what level of adjustment applies is always the work of trained school and specialist staff using multiple evidence sources.
We do not claim IWAF is an MTSS screener on its own
Universal screening in MTSS typically uses brief, standardised curriculum-based measures (e.g. DIBELS, YARC, CBMs) alongside broader data. Writeiq produces criterion-level writing assessment that complements those screeners; it is not a replacement for them. Schools should continue to use validated universal screeners as designed.
We do not claim IWAF replaces moderation
Writeiq produces a score and feedback. Teachers should continue to moderate important assessment decisions — especially for reporting and grading. Every Writeiq score can be adjusted by the teacher, with the reason logged and the original score retained. The adjust-score flow is designed for moderation, not around it.
We do not claim the Developing threshold is research-derived
As stated above, the 31% Emerging-to-Developing boundary is currently a design decision. We will say this honestly to any school that asks, and we will adjust it if the Term 3 2026 standard-setting study tells us to.

IWAF 3.0 FAQs

Why is IWAF called a framework rather than a rubric?

A rubric is a scoring tool. A framework describes a whole way of doing something. IWAF includes the rubric descriptors and the logic gates that check scoring consistency, the band structure that interprets the scores, the feedback language for students and teachers, the curriculum framework mappings, and the pedagogical sequence for teaching to the identified priority. Calling IWAF a rubric would understate what it actually does.

What changed between IWAF 2.4 and IWAF 3.0?

Three substantive changes. First, band thresholds recalibrated from 40/60/80 to 31/56/80. The old thresholds let too many students sit in Developing when their writing showed clearer foundational gaps. Second, the criterion structure was reorganised from dimension-keyed (sentence, paragraph, craft…) to criterion-keyed (CR-01 Audience and Purpose, CR-02 Text Structure, CR-03 Ideas, and so on). This makes the framework easier to reason about and easier to align to curriculum descriptors. Third, the feedback content library was rewritten against new authoring rules around positive framing, concrete next steps, and no self-diagnosis.

How does IWAF align with MTSS and Response to Intervention?

The four IWAF bands align with the three-tier MTSS framework endorsed by AERO and the Australian Government. Extending and Consolidating capture the roughly 80% of students supported successfully by Tier 1 core instruction. Developing corresponds to the roughly 15% typically needing Tier 2 targeted small-group support. Emerging corresponds to the roughly 1–5% needing Tier 3 intensive individualised intervention.

This alignment is deliberate — the 80% Extending threshold draws on the same mastery-learning research base (Bloom, Rosenshine, AERO) that underpins the MTSS tier distribution. See the full MTSS section above for the tier-by-tier correspondence and source references.

Writeiq does not make tier placement decisions. Tier decisions are made by school intervention teams using multiple sources of evidence. Writeiq contributes one important piece of that picture: consistent criterion-level writing data.

Can IWAF data be used as NCCD evidence?

Yes, in the same way any systematically collected assessment and intervention data can contribute to NCCD evidence. The NCCD Portal explicitly lists “documentation of ongoing learning needs that have a limited response to targeted intervention over time” and “evidence of interventions provided over time, with monitoring of the effectiveness of the intervention” as valid evidence categories.

Writeiq generates exactly this kind of evidence: dated, criterion-level writing assessments showing where a student sits, how they respond to teaching, and whether that response is adequate. Teacher score adjustments are logged with reason text, capturing NCCD-relevant accommodations (reader, extra time, and so on) alongside the student’s work.

What Writeiq does not do: diagnose disability, determine NCCD category, or set level of adjustment. Those remain the professional judgement of teachers, learning support staff, and specialist assessors. For broader NCCD allocation, evidence accumulation, and 3-sheet acquittal export, see Edsthetic’s companion product Allocateiq.

Can I use IWAF for subjects other than English?

Not yet. IWAF 3.0 is designed for the conventions of English writing: narrative, persuasive, and analytical text response. Applying English criteria to a science investigation or a humanities source analysis would produce misleading feedback — those faculties assess different qualities (methodology, evidence, reasoning) that the English rubric does not measure.

We have an IWAF 4.0 roadmap item for bespoke, faculty-designed criterion sets. That work is scoped for a future major update so the rubrics are built properly with faculty leads rather than adapted on the fly. If you’d like to be part of that design work, email hello@edsthetic.com.au.

What if I disagree with how a piece was banded?

Open the submission, click Adjust next to any criterion score, and change it. You will be prompted to confirm the change, enter the revised score, and add a reason. The original Writeiq score is retained with a strikethrough, your adjustment is flagged, and the reason is logged. This is the moderation pathway — Writeiq is designed to give you a defensible starting point, not a final ruling.

Will IWAF be validated independently?

Yes. We have planned an ACER Body of Work standard-setting study for Term 3 2026. This is the standard Australian method for validating criterion-referenced assessment thresholds empirically. We will publish the results, including any threshold adjustments the study recommends.

We also welcome independent review from universities, research centres, and state education departments. If you are in a position to help validate the framework, contact hello@edsthetic.com.au.

What about younger students: Year 3 to Year 6?

IWAF 3.0 has three primary modes — Primary Narrative, Recount, and Primary Persuasive — each with criteria pitched at Years 3–6 expectations. The framework is calibrated to Derewianka and Jones’ primary metalanguage (2016) and aligned to AERO’s School Writing Instruction Framework (2025). Primary feedback uses simpler language and shorter sentences, but holds the same four-band structure so schools can track growth into secondary without breaking the data.

How are the 26 logic gates different from the criteria?

Criteria are what gets scored. Logic gates are consistency rules the marker must pass before a score is returned. For example: if a student’s writing has very little punctuation, a gate caps the Sentence Structure score because a reader cannot judge sentence structure without punctuation. If a student uses bullet points, a gate caps the Cohesion score because cohesion is assessed across continuous prose. Gates do not add information; they catch the kinds of inconsistency that undermine trust in a score.

Many gates surface a framing sentence to the student: “Your writing has very little punctuation. Before anything else, make sure every sentence has a capital letter at the start and a full stop at the end.” The gate caps the score and tells the student why.

Who wrote IWAF 3.0?

IWAF is authored by Mikhaila Picone (writing specialist and co-founder of Edsthetic) with technical implementation by Ashwin Pillai (co-founder, Deputy Principal for Pedagogy at Sacred Heart College Geelong). The framework draws on Mikhaila’s work across primary and secondary English teaching, and on published research from AERO, Rosenshine, Hattie, Fisher and Frey, Martin and Rose, Derewianka and Jones, and others cited in the References section below.

Primary references

The research literature IWAF 3.0 is grounded in. This is a working list, not an exhaustive bibliography.

Questions about IWAF for your school?

If you are evaluating Writeiq and want to talk through how IWAF 3.0 would apply to your school’s context — including validation, moderation, or how the framework maps to your reporting language — email us at hello@edsthetic.com.au.