Overall performance levels can be helpful, when used well. When there are repeated assessments, overall performance can help teachers quickly and easily understand student growth. Overall performance can also help inform decision-making for special programming, especially when comparing data to local and national peers.
Still, when the 2020 revisions of the USNS were originally released, overall performance levels were intentionally not included. I still wonder if they sometimes tend to work against, rather than in favor of the goals of the USNS. When teachers focus on the results of the individual tasks, instructional responses for all students become apparent.
The overall performance levels for the USNS are both criterion and norm referenced. Because each of the tasks of the USNS align with specific skills – or learning criteria – the overall scores published in the PDF are also criterion referenced. Norm references for each of the assessments can be seen in the Single Assessment Report using the Performance view. The largest norm reference provided is the Global Cohort, an aggregation of results collected in Forefront across the country and internationally. Forefront also enables districts to see district level norms.
There are 4 overall performance levels for the USNS:
The overall score for each student is found using the mean of the performance levels for each of the tasks. Overall scores that round to 3 (≥2.5) indicate Level 3 – Proficient. If the results average to a 2 (1.5 – 2.49), then the overall performance level is Level 2 (yellow). The lowest set of results, those that average to a 1 (≤1.49), are broken into two performance levels: the lower half of the set indicates Level 0 – Well Below Basic; and the upper half of the set indicates Level 1 – Below Basic.
Why the reluctance to create overall performance categories in the first place? First and foremost, I do not like categorizing and labeling students. A primary purpose of screening assessments in an MTSS is to identify students who might be at risk of not performing well. However, when the focus becomes content that needs to be learned, it leads to thinking about how to instruct all students who might need help in specific areas. Targeted instructional groups then focus around learning needs, rather than a list of student names.
When averaging scores together, teachers lose valuable information. In reality each student is unique. No two assessments are ever the same, and to focus on the average tends to suggest a homogeneity of that group that really does not exist.
Imagine the “overall” as a portrait instead of a number. Take these profiles for example:

student 1 student 2
The balance of colors on these two student wheels from Forefront (using the Number Sense Lens) is quite similar, when one just thinks of the balance of green, yellow, and orange. Overall, their scores land in Level 2 (yellow). However, the distribution of performance is quite different. While student 1 has quite a bit of yellow and orange in the upper right quadrant — which represents Numerals, Words, and Sequences — student 2 appears to be doing generally better in that area. Student 2 has more yellow and orange is in the lower right quadrant which represents Addition and Subtraction within 20, and more green in Numerals, Words and Sequences, an important asset.
Every student wheel is different. Rather than sorting students based on their overall performance, it helps us to recognize each student as a complex individual with strengths and areas where they need to grow.
While there are legitimate purposes for calculating overall performance levels of students, including tracking growth and for MTSS/RtI programming purposes, the results of the assessments that are most helpful to teachers and interventionists are the class’s results on each of the tasks. Create your small group/differentiated instruction around the topics assessed on the USNS, rather than around students who performed below basic. This can result in more diverse and dynamic grouping, improved efficiency and collaboration, and less labeling.
David Woodward is an educator with more than 25 years of experience as a classroom teacher and district leader in math education. He founded Forefront Education to help educators better understand student learning with meaningful assessments and support standards-based grading that is automatic, accurate, and reliable. David recently retired as a math coach at Boulder Valley School District in Colorado in June 2020. He leads the Universal Screeners for Number Sense project (formerly known as the BVSD K-5 Math Screeners). Read more.
Our team and tools help schools implement standards-based grading, streamline assessment systems, and use meaningful data to drive decision-making.
Recorded April 2, 2025
Forefront Education and math leaders from Alabaster City and Cambridge Public Schools led a virtual panel discussion on implementing the USNS.
District and building-level leaders shared their experiences, including how they navigated challenges and leveraged USNS data to support student learning.
Forefront is the only assessment data solution optimized for classroom assessment results, leveraging these results to fuel instruction, PLCs, and grading. Elevate meaningful assessment data district-wide to transform how you understand and communicate about student learning across your schools.
Copyright © 2025 Forefront Education, Inc. All Rights Reserved.