Evaluation Has Outgrown the Old Playbook
For most of the 3012-d era, evaluation technology had a relatively simple job: schedule observations, collect rubric scores, store SLO documentation, and produce an annual report. Districts that invested in a solid platform got real efficiency gains. But the underlying workflow was predictable enough that even a well-organized spreadsheet could theoretically handle it.
That era is ending.
The STEPS framework under 3012-e introduces a level of complexity that genuinely requires purpose-built technology. Multi-year observation cycles with different schedules for different educators. Per-standard scoring across seven NYS Teaching Standards — or ten leadership standards for principals. Evidence collected from multiple measures simultaneously. Overall ratings generated at the end of cycles that may span two or three years, not just one. And all of it needs to feed into NYSED reporting in a compliant format.
This isn’t a software update. For many districts, it’s a fundamentally different workflow — and the platforms built for 3012-d weren’t designed with any of it in mind.
What Evaluation Technology Should Actually Do
Good evaluation technology doesn’t just store data. It actively supports the people doing the work — building administrators, HR directors, department chairs, and the educators being evaluated. Here’s what that looks like in practice:
- Scheduling that reflects reality. Under 3012-e, tenured educators may be on one-, two-, or three-year cycles. Probationary educators require annual cycles. Educators rated Level 1 or Level 2 must receive a Teacher Support Plan. Whether that plan also requires an annual evaluation cycle, regardless of tenure status, is determined at the local level through the TSP design process and collective bargaining — meaning requirements will vary from district to district. A platform needs to track all of this simultaneously, flag who is due for observation and when, and make it easy for a evaluating administrators managing a staff of sixty to stay on top of every individual’s cycle status without a spreadsheet running alongside it.
- Standards-based scoring, not just rubric scoring. Under 3012-d, a platform needed to capture rubric domain scores and combine them into an observation sub-score. Under 3012-e, evidence from observations and additional measures needs to map to specific NYS Teaching Standards — and each standard gets its own 1–4 score. That’s a different data architecture, and platforms that weren’t built for it will struggle to retrofit it cleanly.
- Support for the full range of additional measures. Districts choosing from the NYSED sample measures — portfolios, surveys, goal-setting, peer observation, student outcome data, LEA-developed measures — need a platform that can manage evidence collection and scoring for all of them, not just classroom observations. Districts are also free to invent their own measures! A system that handles observations well but treats everything else as an afterthought isn’t ready for 3012-e.
- Cumulative rating generation across multi-year cycles. This is one of the most technically demanding requirements. When a tenured educator is on a three-year cycle, their overall rating isn’t generated annually — it’s generated at the end of the cycle, drawing on evidence accumulated across all three years. The platform needs to hold that evidence, track its accumulation, and produce a compliant overall rating and NYSED report at the right moment.
- Meaningful feedback, not just documentation. The best evaluation systems don’t just record what happened — they help administrators and educators understand what the data means. Standards-based scoring creates a genuinely richer picture of educator practice than 3012-d’s composite score ever did. A good platform surfaces that picture in a way that drives professional growth conversations, not just compliance paperwork.
Worth knowingThe shift to standards-based scoring under 3012-e means evaluation data becomes significantly more granular. Instead of one composite score, each educator receives a score on each NYS Teaching Standard. A good platform turns that data into a genuinely useful picture of where educators are strong and where they need support — but only if it was designed to handle that level of detail.
Making Observations Faster and More Useful
Even before 3012-e, classroom observation was the most time-intensive part of the evaluation process for administrators. Under 3012-d, every educator required a full observation cycle every year — and principals in larger buildings were spending a disproportionate share of their instructional leadership time on observation paperwork rather than the conversations that matter. The shift to multi-year cycles helps, but technology can do even more to reduce that burden without sacrificing the quality of the feedback.
Two features in particular make a meaningful difference in observation efficiency:
Stored comments. Experienced evaluators use a relatively consistent vocabulary when describing what they observe — strong questioning technique, effective use of wait time, clear learning targets, evidence of differentiation. A platform that lets administrators build and store a library of their own frequently used comments, searchable and insertable with a click, dramatically reduces the time spent writing up observation notes. The evaluator is still making the professional judgment — they’re just not retyping the same well-crafted sentence for the fortieth time. Good platforms let both the district and individual evaluators maintain their own libraries, so the language reflects local priorities and rubric alignment.
Pushing comments directly into rubric indicators. The traditional observation workflow involves taking notes during the classroom visit, then separately scoring rubric domains, then writing a narrative summary — three distinct steps that often cover the same ground. A better approach lets the evaluator capture a comment and push it directly to the rubric indicator it evidences, in the moment. Walk into a classroom, observe a strong example of student-led discussion, tap the relevant indicator, add the comment — it’s simultaneously an observation note and rubric evidence. By the time the observation is over, the rubric is substantially populated. The write-up becomes a synthesis of evidence already captured, not a reconstruction of memory.
Together these features don’t just save time — they improve the quality and specificity of feedback. Comments tied directly to rubric indicators are more actionable for teachers than general narrative summaries. Evidence captured in the moment is more accurate than evidence reconstructed an hour later. And administrators who spend less time on paperwork have more capacity for the pre- and post-observation conversations that are where real professional growth actually happens.
The observation efficiency testWhen evaluating any platform, ask to see the observation workflow from start to finish — not just the reporting dashboard. How many clicks does it take to record an observation note and attach it to a rubric indicator? Is there a comment library? Can the evaluator work on a tablet in the classroom, or only at a desk afterward? The answers reveal whether the platform was designed around how evaluators actually work, or just around what compliance requires.
Where Most Platforms Fall Short
The uncomfortable reality is that most evaluation platforms in use across New York districts today were designed for the 3012-d world. They were built around a specific, state-prescribed workflow: collect observation scores, collect SLO scores, apply the matrix, generate the HEDI rating, submit to NYSED. That workflow is well-understood, and platforms that execute it well have earned their place.
But 3012-e breaks most of those assumptions:
- The fixed annual cycle becomes a locally determined multi-year cycle
- The two-component scoring model becomes a per-standard model with locally determined weighting
- The state-prescribed SLO becomes a locally selected menu of additional measures
- The annual overall rating becomes a cycle-end rating that may be one, two, or three years out
A platform that was engineered around the old structure doesn’t automatically flex to accommodate the new one. Multi-year cycle scheduling, in particular, requires a fundamentally different data model — one where the unit of evaluation is a cycle, not a school year. Districts that try to force 3012-e workflows into a 3012-d platform will find themselves doing in spreadsheets what their software should be doing automatically.
The Technology Decision Comes Earlier Than Districts Think
Here is the sequencing mistake most districts are making: they’re planning to figure out their technology after they’ve bargained their STEPS plan.
That’s backwards — and it creates real risk.
The measures your district selects, the cycle lengths you design, the weighting methodology you negotiate — all of these decisions have technology implications. If you bargain a STEPS plan that includes peer observation as an additional measure, and your platform doesn’t support peer observation workflows, you’ve created a compliance problem before you’ve even started. If you negotiate a three-year cycle with annual per-standard scoring but no annual overall rating, your platform needs to be able to hold three years of evidence and generate the final rating correctly. If it can’t, your plan is unimplementable as written.
The right sequence is:
- Understand the 3012-e requirements
- Assess whether your current platform can support what you want to design
- Make the technology decision — upgrade, replace, or confirm your current platform is ready
- Then bargain your STEPS plan knowing your technology can execute it
Districts that wait until after bargaining to discover their platform isn’t ready will face a difficult choice: renegotiate the plan, switch technologies at the last minute or implement it manually. None of these is a good option.
Questions to Ask Any Evaluation Platform Vendor
When assessing whether your current platform — or a prospective new one — is genuinely ready for 3012-e, these are the questions that matter:
Not just annually — can it track individual educators on 1-, 2-, or 3-year cycles simultaneously, flag who is due when, and manage exceptions for probationary staff and those at the two lower rating levels?
Can evidence from observations, portfolios, surveys, goal-setting, and other measures all map to specific NYS Teaching Standards — and generate a 1–4 score per standard, not just a composite?
Does it accumulate evidence over a multi-year cycle and produce a compliant overall rating and NYSED report at cycle completion — not just at the end of each school year?
Observations are table stakes. Can it also manage peer observation workflows, portfolio evidence, survey data, goal-setting documentation, student outcome data, and locally developed measures?
Many districts are looking to consolidate evaluation systems. A platform that handles only certified staff leaves transportation, nutrition, clerical, technology, nursing, and security staff without a comparable solution.
There’s a meaningful difference between a platform that helps administrators conduct better, more efficient observations and one that simply serves as a filing cabinet for completed forms. Ask to see the observation workflow in action, not just the reporting dashboard.
The Bottom Line
Technology doesn’t make evaluation meaningful on its own — thoughtful design, honest feedback, and genuine professional growth conversations do that. But the right technology removes the friction that gets in the way of all of it. Scheduling burden, data management complexity, reporting compliance — these are problems that good platforms solve quietly in the background, freeing administrators to focus on the work that actually matters.
As New York districts move toward 3012-e, the technology question isn’t a back-burner item. It’s a foundational planning decision that shapes what kind of STEPS plan you can realistically design and implement. The districts that get this right will assess their platforms early, make informed decisions, and bargain plans they can actually execute. The ones that don’t will discover the gap too late.
NY State Education STEPS (3012-e) Guidance Documents →
Read our plain-language guide to 3012-e APPR changes →
RightPath™ was built for exactly this moment
Our evaluation platform supports multi-year observation cycle scheduling, per-standard scoring across all selected measures, overall rating generation at cycle end, the full range of NYSED sample additional measures, and complete STEPS-compliant NYSED reporting. We also support world-class evaluation of non-educator staff — transportation, nursing, clerical, nutrition, technology, security, and more.
We’ve been supporting NY districts through every evolution of APPR. We’re ready for STEPS — and we’re ready to help you get there.

