Smarter Balanced Interim Assessment Blocks (IABs) for ELA and math

Interim Assessment Blocks are available for ELA and math. These can be further divided into two types: IABs and Focused IABs (FIABs). IABs assess 3-8 academic targets, while FIABs assess 1-3 academic targets. For this user guide, both types of Interim Assessment Blocks are referred to as IABs. There are several ways to view student IAB results in CERS.

IAB/IA Dashboard

The IAB/IA Dashboard provides a summary of IAB/IA assessments taken over the course of a school year for a group of students. The dashboard also provides an easy way to select these assessments to generate a detailed assessment report. Selecting the [View IAB/IA Dashboard] button on the Assigned Groups panel shown in figure 1 displays this dashboard, shown in figure 2.

View IAB Dashboard button

Figure 1. View IAB/IA Dashboard button

Graphical user interface, IAB Dashboard

Figure 2. IAB/IA Dashboard

The IAB/IA Dashboard displays a summary of IAB/IA results based on the drop-downs provided. It identifies the total number of students from the selected group, how many participated in each IAB/IA, and basic performance statistics for each assessment. One or more of these assessments can then be selected and used to generate more detailed reports.

  1. Group drop-down: Contains all the groups from the Assigned Groups and My Groups panels. This drop-down defaults to a group belonging to the panel from which the [View IAB/IA Dashboard] button was selected. It may be changed to any group from either Assigned Groups or My Groups.
  2. School Year drop-down: Allows a user to select from available school years
  3. Subject drop-down: Can be set to limit the displayed IABs/IAs to a single content area or set to All
  4. IAB/IA selection indicator: A checkmark indicates that the IAB/IA card is selected. Multiple IABs/IAs may be selected at the same time.
  5. [View Assessments] button: Selecting this button generates assessment reports for all the selected IABs/IAs. It is disabled until at least one IAB/IA is selected.

IAB Results in Group View

The group view provides an aggregate and individual display of a selected assessment for an assigned group, a custom group, or a school. Refer to the View Student Assessment Results by School and Grade section or the View Student Group Assessment Results section for details.

Once a user selects an IAB to display in the group view, the Results panel for an Interim Assessment Block (IAB) is shown in figure 3.

Graphical user interface, IAB Results in Roster View

Figure 3. IAB Results in Group View

Each IAB Results page displays the following information and elements:

  1. Name and grade of the assessment
  2. Testing session(s) from which results are displayed. These sessions appear once a test is completed (and scored). Please note that testing sessions appear here for interim assessments only.

  3. Group Aggregate panel displays aggregate data for the selected groups of students
  4. Select a results view drop-down menu offers the following options for display results:
    1. [Results By Student] (default view)
    2. [Results By Item]
    3. [Key/Distractor Analysis] (for selected-response items)
    4. [Writing Trait Scores] (for ELA Performance Tasks only)
  5. Results View table
  6. [Instructional Resources] button that displays related instructional resource links to Smarter Balanced Tools for Teachers website resources or a user’s LEA-provided resources. This button is available for IABs, FIABs, CAST IAs, and ELPAC IAs.
  7. Display value as field: Contains toggles allowing a user to change the display to show the student score distribution in percentages or numbers of students in each reporting category
  8. [Collapse All] button: Hides the Results View table for all the displayed assessments and toggles to an [Expand All] button that displays the results again
  9. [Export CSV] button: Gives the option to download the results in CSV format

IAB Group Aggregate Results

The IABs report one overall performance score. For Overall scores, the Student Score Distribution reports three reporting categories.

Graphical user interface, Overall Group Results for an IAB

Figure 4. Overall Group Results for an IAB

The Group Aggregate panel for IABs displays the following information and elements:

  1. Average Scale Score: Average scale score for the selected group of students and an error band based on the Standard Error of the Mean
  2. Number of student test results from which the scores are calculated
  3. Student Score Distribution: Score distribution for the performance levels of the overall score
  4. [Show Results] button: Displays the Results View table for the displayed assessment and toggles to a [Hide Results] button that can hide the table

IAB Results By Student

When the Results By Student view is active, the IAB results page shows a table with each row displaying a student’s assessment results. The table has sortable columns and additional information icons, as shown in figure 5:

Graphical user interface, IAB Results By Student

Figure 5. IAB Results By Student

The table includes the following information for each student:

  1. Student: Student’s full name with a Context Menu three-dot icon [three dots icon] allows a user to access a pop-up menu that allows further selections of student details to be made (refer to IAB Student Options section).
  2. Date: Date student completed assessment
  3. Session: Testing session identifier
  4. Enrolled Grade: Student’s enrolled grade at the time of the assessment
  5. School: Student’s assigned school at the time of the assessment
  6. Status: The assessment status, which includes whether the assessment administration (Manner of Administration) was Standardized or Nonstandardized, or if the assessment was marked Partial or Complete (Completeness).
  7. Reporting Category: The score reporting category, which is Below Standard, Near Standard, or Above Standard as well as a link, indicated with the [Instructional Resources] book icon [], to available instructional resources.
  8. Scale Score/Error Band: The student’s scale score on the assessment and the error band based on the Standard Error of Measurement (SEM) associated with that student’s score.

IAB Student Options

When the Context Menu three-dot icon [] next to a student’s name is selected, the Student Options pop-up menu, shown in figure 6, appears.

Student Options view

Figure 6. Student Options context menu

The menu options are:

  • [Student’s Responses]: This displays the student’s responses to each item on the assessment. Refer to the Student Responses section for details
  • [Student’s Test History]: This displays the test history of the student. Refer to the Student Test History Report sections for details
  • [Print Student’s Full IAB Report]: This links to a printable Student Report in PDF format. The selections already known in the context of the IAB results, such as the assessment type, subject, and school year, are auto-filled. Refer to the Printable Reports section for details.

IAB Results By Item

When the Results By Item view is active, the IAB results panel shows a results table with an aggregate view of how the students performed on each item. The table has sortable columns and additional information icons, as shown in figure 7:

Figure 7. IAB Results By Item

The data shown in the Results By Item view indicates how the students performed overall on each assessment item, as described next.

  1. Item #: Number of the item in the order it was presented to the student
  2. Claim/Target: Topic area and particular knowledge and skills tested by the item
  3. Item Difficulty: Item difficulty (easy, moderate, or difficult) based on the item's difficulty for students in a reference population
  4. Standard: Content standard(s) that define what students should know, understand, and can do
  5. Full Credit: Percentage or number of students in the selected group who earned the maximum number of points for the item
  6. 0, 1, 2, 3, etc.: Percentage or number of students who earned each possible score-point level on the item

Users may sort by each of these categories of information to quickly identify items of interest. Additionally, an [Export] button is available to export the report results to a CSV file.

IAB Results By Item Details

Selecting an assessment item’s drop-down arrow icon [] expands the row to reveal additional details about that item, as shown in figure 8.

Screen Shot of IAB Results By Item additional details

Figure 8. IAB Results By Item additional details

Student Points and Responses

The [Student Points and Responses] tab lists each student in the selected session(s) and the points earned by the student for each item, as previously shown. It displays each student’s name, test date and session identifier, enrolled grade, and school. The report includes the number of points the student earned on the item, the maximum points available for the item, and the correctness of the student’s response, which is the ratio of the student’s points to the maximum points. For example, a student who earned 1 point out of a maximum of 2 would have a correctness value of 0.50.

Selecting the drop-down arrow icon [] next to a student’s name displays the assessment item the student viewed when the student took the assessment and the student’s response. The item number is shown at the top-left for the item being viewed. This view is read-only; there are no user-editable fields. An example is shown in figure 9.

Student Scores and Responses

Figure 9. Student Points and Responses tab

Item Viewer

The [Item Viewer] tab displays the assessment item as it appeared on the test but does not show the student’s response. The item number is shown at the top-left for the item being viewed. This view is read-only; there are no user-editable fields. An example Item Viewer panel is shown in figure 10.

Figure 10. Item Viewer tab

Rubric and Exemplar

The [Rubric and Exemplar] tab displays the answer key, or an exemplar, and the rubric for the assessment item, as shown in figure 11.

Rubric and Exemplar

Figure 11. Rubric and Exemplar tab

The Exemplar panel provides the correct answer and, in some cases, multiple correct answers. For selected response test items (e.g., multiple-choice, multiple-select, or evidence-based selective response items), a Key (answer key) is shown.

The Rubric panel can also describe the point values associated with specific responses, as shown in figure 12.

the Rubric

Figure 12. Rubric

Item Information

The [Item Information] tab displays a detailed description of the assessment item (figure 13).

Item Information tab

Figure 13. Item Information tab

Elements shown in the Item Information view may include:

  • Claim—Major topic area of the test item content
  • Target—Particular knowledge and skills measured by the test item
  • Depth of Knowledge (DOK)/Performance Level Descriptor—Level of cognitive demand this test item requires; includes a link to the DOK matrix
  • Item Difficulty—One of three possible values: Easy, Moderate, Difficult
  • Standard—Content standard(s) that define what students should know, understand, and can do
  • Calculator—For mathematics and science questions, this element will indicate whether or not a calculator was available to students completing this item
  • Supporting Documentation—Link to the Interpretive Guide

Key/Distractor Analysis

The Key/Distractor Analysis view allows the user to evaluate whether there may have been common areas of misunderstanding among students who chose incorrect answers, called “distractors,” in selected-response (multiple choice and multiple select) items for interim assessments.

When the Key/Distractor Analysis view is active, the Results panel shows an aggregate view similar to that of the Results By Item table, but with the percent distribution of student responses by answer option, as shown in figure 14.

Key / Distractor Analysis results

Figure 14. Key/Distractor Analysis results

The Key/Distractor Analysis table appears as described in the IAB Results by Item section, with the following difference:

  1. [Item drop-down arrow] icon: Selecting an item’s down-arrow icon [arrow icon] expands the row to reveal additional details. Refer to the IAB Results by Item Details section.
  2. Credit Columns: The columns indicating students’ score-point levels are replaced by columns that indicate the percentage or number of students who selected each answer choice. The correct answers are highlighted in green.
  3. [Export] button: Exporting is available to export the results of the report to a CSV file

IAB Writing Trait Scores

Writing trait scores are displayed for ELA Performance Task IABs that include a writing extended response (essay) item.

When the Writing Trait Scores view is active, the Results panel shows a Writing Trait Scores Item table with an aggregate view of how the students performed on the writing extended response item. Only items with writing trait information are listed, as shown in figure 15.

Writing Trait Scores tables

Figure 15. Writing Trait Scores view

The Writing Trait Scores Item table is similar to the Results By Item table, with the following differences:

  1. Writing Purpose: A new column displays the purpose of the writing tasks (argumentative, explanatory, informational, narrative, or opinion)
  2. [Item drop-down arrow] icon: Selecting an item’s drop-down arrow icon [ drop down arrow icon]expands the row to allow viewing of the item and additional information about the item (refer to the IAB Results by Item Details section).
  3. Points Distribution table:
    1. Each row lists the category of writing performance or writing trait (Evidence/Elaboration, Organization/Purpose, and Conventions); the average and maximum points for that category; and the percentage or number of students who earned each number of points for that category.
    2. The final row is for the transformed points calculated from all categories (refer to the Interpretive Guide for an explanation of how transformed points are calculated for Writing Extended Response items).
  4. [Export] button: Exporting is available to export the report results—both the Results By Item table and the Points Distribution table combined—to a CSV file.