iFAB (Individualized Formative Assessment of Behavior)
DBR-SIS: Academic Engagement
Summary
Individualized Formative Assessment of Behavior (iFAB) is a mobile-enabled web-based system for behavioral progress monitoring of elementary students in grades K -3, that is flexible, feasible, and psychometrically-sound. Using iFAB, teachers or other school staff can generate assessment plans, enter data, record events (e.g., change of intervention plan), and review auto-generated time series charts of student progress. Developed from the ground up for the purpose of formative behavior assessment, iFAB offers the ability to assess both academic enablers (i.e., Academic Engagement, Organization Skills, Interpersonal Skills) and problem behaviors that commonly occur in school settings (Disruptive, Oppositional, Interpersonal Conflict, Anxious-Depressed, and Social Withdrawal). The system offers considerable flexibility in that ratings may be completed daily (direct behavior rating method) or weekly (formative behavior rating method). Users also have the choice of rating student behavior using single-item scales or multi-item scales. This submission applies to the DBR rating method, the single item scale, and the Academic Engagement behavior construct.
- Where to Obtain:
- Self
- r.volpe@neu.edu
- 617-702-6220
- https://www.ifabonline.com/
- Initial Cost:
- Free
- Replacement Cost:
- Free
- Included in Cost:
- At this point, we are making the tools available at no cost.
- NA
- Training Requirements:
- Training not required
- Qualified Administrators:
- No minimum qualifications specified.
- Access to Technical Support:
- Assessment Format:
-
- Rating scale
- Scoring Time:
-
- Scoring is automatic OR
- 0 minutes per
- Scores Generated:
-
- Raw score
- Composite scores
- Administration Time:
-
- 3 minutes per student
- Scoring Method:
-
- Automatically (computer-scored)
- Technology Requirements:
-
- Computer or tablet
- Internet connection
Tool Information
Descriptive Information
- Please provide a description of your tool:
- Individualized Formative Assessment of Behavior (iFAB) is a mobile-enabled web-based system for behavioral progress monitoring of elementary students in grades K -3, that is flexible, feasible, and psychometrically-sound. Using iFAB, teachers or other school staff can generate assessment plans, enter data, record events (e.g., change of intervention plan), and review auto-generated time series charts of student progress. Developed from the ground up for the purpose of formative behavior assessment, iFAB offers the ability to assess both academic enablers (i.e., Academic Engagement, Organization Skills, Interpersonal Skills) and problem behaviors that commonly occur in school settings (Disruptive, Oppositional, Interpersonal Conflict, Anxious-Depressed, and Social Withdrawal). The system offers considerable flexibility in that ratings may be completed daily (direct behavior rating method) or weekly (formative behavior rating method). Users also have the choice of rating student behavior using single-item scales or multi-item scales. This submission applies to the DBR rating method, the single item scale, and the Academic Engagement behavior construct.
- Is your tool designed to measure progress towards an end-of-year goal (e.g., oral reading fluency) or progress towards a short-term skill (e.g., letter naming fluency)?
-
ACADEMIC ONLY: What dimensions does the tool assess?
- BEHAVIOR ONLY: Please identify which broad domain(s)/construct(s) are measured by your tool and define each sub-domain or sub-construct.
- Academic Engagement
- BEHAVIOR ONLY: Which category of behaviors does your tool target?
Both
Acquisition and Cost Information
Administration
Training & Scoring
Training
- Is training for the administrator required?
- No
- Describe the time required for administrator training, if applicable:
- We have purposely evaluated the measures as administered with no training.
- Please describe the minimum qualifications an administrator must possess.
- We made a conscious choice to design formative behavior assessment tools that required no training. As such, no training was involved in our psychometric studies.
-
No minimum qualifications
- Are training manuals and materials available?
- No
- Are training manuals/materials field-tested?
- No
- Are training manuals/materials included in cost of tools?
- If No, please describe training costs:
- NA
- Can users obtain ongoing professional and technical support?
- Yes
- If Yes, please describe how users can obtain support:
Scoring
- Please describe the scoring structure. Provide relevant details such as the scoring format, the number of items overall, the number of items per subscale, what the cluster/composite score comprises, and how raw scores are calculated.
- Single item scales follow a DBR format and generate scores for each item. The system also generates two composites for single item scales (academic enablers and problem behaviors).
- Do you provide basis for calculating slope (e.g., amount of improvement per unit in time)?
- No
- ACADEMIC ONLY: Do you provide benchmarks for the slopes?
- ACADEMIC ONLY: Do you provide percentile ranks for the slopes?
- Describe the tool’s approach to progress monitoring, behavior samples, test format, and/or scoring practices, including steps taken to ensure that it is appropriate for use with culturally and linguistically diverse populations and students with disabilities.
- Assessment plans can be tailored to each individual student in that teachers/consultants select relevant scales for monitoring. They can decide which of the four methods to employ for progress monitoring. Our psychometric studies were conducted on diverse groups of students in both regular education and special education. We are currently seeking funding to explore the measurement invariance of the iFAB measures.
Levels of Performance and Usability
- Date
- Size
- Male
- Female
- Unknown
- Eligible for free or reduced-price lunch
- Other SES Indicators
- White, Non-Hispanic
- Black, Non-Hispanic
- Hispanic
- American Indian/Alaska Native
- Asian/Pacific Islander
- Other
- Unknown
- Disability classification (Please describe)
- First language (Please describe)
- Language proficiency status (Please describe)
Performance Level
Reliability
| Age / Grade Informant |
Grades K-3
Teacher |
|---|---|
| Rating |
|
Convincing evidence
Partially convincing evidence
Unconvincing evidence
Data unavailable- *Offer a justification for each type of reliability reported, given the type and purpose of the tool.
- Temporal stability and dependability across occasions are important to examine in formative behavioral assessment as inconsistency of ratings across measurement occasions is a common source of error. For formative measures we want to know the extent to which changes in scores can be attributable to change in the target students behavior as opposed to time-related error.
- *Describe the sample(s), including size and characteristics, for each reliability analysis conducted.
- Dependability (from Volpe et al., 2023): Ninety-one K–5 educators (i.e., teachers, paraprofessionals, and intervention specialists) from elementary schools located in the Northeast, Midwest, and West of the United States participated in the study. Educators were mainly female (n = 86), White (n = 87), and their age ranged from 21 to 62 (M = 39.00, SD = 10.40) years. On average, they had 12.70 (SD = 7.87) years of experience. Each educator rated one student in their classroom. The largest number of participating students were classified as male (n = 61) and their age ranged from 5 to 10 (M = 7.36, SD = 1.39) years. The majority of students were White (n = 60), followed by Black (n = 17), Hispanic (n = 10), and biracial (n = 4). Approximately 45% of the students received special education services (n = 41). Once all student participants were identified, educators were randomly assigned to one of the following four conditions: (a) DBR-SIS, (b) FBRM-SIS, (c) DBR-MIS, or (d) FBRM-MIS. MTMM (test-retest): A total of 68 elementary school teachers were recruited from schools in Boston and the surrounding metro area. The mean age of teachers was 37.82 (range 25-67) and their mean years of experience teaching was 11.49 (range 1 to 40).
- *Describe the analysis procedures for each reported type of reliability.
- As part of our initial development of the scales comprising the iFAB, we conducted a series of exploratory factor analyses for one of our four assessment methods (DBR-MIS). In our three published studies (Briesch et al., 2022; Daniels, Briesch, & Volpe, 2021; Volpe, Chaffee, Yeung, & Briesch, 2019) we reported both coefficient alpha and coefficient omega. We used these indices as indicators of item homogeneity and item quality. We examined the dependability of all four methods across all scales in a recently published series of G studies (Volpe, Matta, Briesch, & Owens, 2023). Here we conducted a set of parallel G studies to investigate differences in dependability across our four rating methods. A single-facet design was used with students (s) as the object of measurement and occasion (o) as the facet. The design was fully crossed with all students evaluated on five occasions (s x o). Time related variance has been well-documented as a source of error in formative behavioral assessment and in this study, we were interested in the extent to which a single week of data collection might generate scores with acceptable levels of dependability. We were particularly interested in making comparisons across methods and constructs. We are currently preparing several manuscripts to examine the criterion-related validity of the iFAB measures. We are using a multi-trait multi-method (MTMM) approach, wherein we examine associations with the iFAB measures and a group of criterion measures. In the diagonal of the MTMM matrix are coefficients of stability for each measure. We report the 1-week stability coefficients for each scale.
*In the table(s) below, report the results of the reliability analyses described above (e.g., model-based evidence, internal consistency or inter-rater reliability coefficients). Include detail about the type of reliability data, statistic generated, and sample size and demographic information.
| Type of | Subscale | Subgroup | Informant | Age / Grade | Test or Criterion | n (sample/ examinees) |
n (raters) |
Median Coefficient | 95% Confidence Interval Lower Bound |
95% Confidence Interval Upper Bound |
|---|
- Results from other forms of reliability analysis not compatible with above table format:
- Scale Informant Grade N Rater N Participant Dependability Coefficient Number of Measurement Occasions to Criterion Academic Engagement (DBR-SIS) Teacher K-5 18 18 0.89 3 daily rating to reach >0.80 coefficient Volpe, R. J., Matta, M., & Briesch, A. M. (2023). Formative behavioral assessment across eight Constructs: Dependability of direct behavior ratings and formative rating measures. Journal of School Psychology. https://doi.org/10.1016/j.jsp.2023.101251
- Manual cites other published reliability studies:
- No
- Provide citations for additional published studies.
- Do you have reliability data that are disaggregated by gender, race/ethnicity, or other subgroups (e.g., English language learners, students with disabilities)?
- No
If yes, fill in data for each subgroup with disaggregated reliability data.
| Type of | Subscale | Subgroup | Informant | Age / Grade | Test or Criterion | n (sample/ examinees) |
n (raters) |
Median Coefficient | 95% Confidence Interval Lower Bound |
95% Confidence Interval Upper Bound |
|---|
- Results from other forms of reliability analysis not compatible with above table format:
- Manual cites other published reliability studies:
- No
- Provide citations for additional published studies.
Validity
| Age / Grade Informant |
Grades K-3
Teacher |
|---|---|
| Rating |
|
Convincing evidence
Partially convincing evidence
Unconvincing evidence
Data unavailable- *Describe each criterion measure used and explain why each measure is appropriate, given the type and purpose of the tool.
- Engagement: The iFab Engagement scales were designed to measure student academic engagement. The ACES Engagement subscale is an established measure of student academic engagement, we hypothesize strong positive correlations with that construct. As the inattentive symptoms of ADHD are associated with low levels of academic engagement, we hypothesize the ADHDSC-4 Inattention subscale and the BASC-3 Attention Problems subscale (both established measures of the inattentive symptoms of ADHD to demonstrate strong negative correlations with the iFab Engagement scales.
- *Describe the sample(s), including size and characteristics, for each validity analysis conducted.
- Elementary school teachers were recruited from schools in Boston and the surrounding metro area. To avoid range restriction in this correlational design, a general sample of students was selected (as opposed to selecting at-risk students alone). Specifically, teachers were asked to complete a packet of rating forms for one male and one female student selected at random. The overall sample was stratified by gender and grade level in order to ensure equal proportions of male and female students in grades K-1 and 2-3. Each of the two students was assigned to one of the two following conditions: • Daily condition: Teachers completed the 8 DBR scales using both SIS and MIS methods each day. Both the order of DBR formats (SIS and MIS) and scales (e.g., Disruptive, Social Withdrawal) was randomized. Both the order and format of scales was randomized. Each daily rating completed over the course of 1-week was averaged for a composite of ratings for each scale. • Weekly condition: Teachers completed the 8 DBR scales using both SIS and MIS methods at the end of each week. Both the order of DBR formats (SIS and MIS) and scales (e.g., Disruptive, Social Withdrawal) was randomized. Teachers were also asked to complete a set of established rating measures, including ACES, ADHD Symptom Checklist-IV, BASC-3, BRIEF-2 (Plan/Organize and Organization of Materials subscales), CDI-2, and SAS-TR, at the end of the week (i.e., on the weekend). Description of Students and Teachers in the Daily Condition: A total of 71 Kindergarten – third-grade teachers completed ratings for the study. Teacher raters were between 25 and 67 years of age (M = 37.96; SD = 10.28) and had between 1- and 40-years teaching experience (M = 11.64; SD = 9.0). All teachers were female, with the overwhelming majority of them identifying as White. One teacher identified as Hispanic. Description of Students in the Daily Condition: There were 72 students in the Daily Condition (34 boys, 33 girls). Sex data were missing for 5 students. Kindergarten students comprised 33.3% of the sample, first-graders 26.4%, second-graders 16.7% and third-graders 16.7% of the sample. Grade data were missing for 5 cases. Approximately 68% of the sample was White, 13.9% Black, 1.4% Hispanic, 2.8 Asian, and 5.6% other. In regard to ethnicity, 20.8% of students were Hispanic. A total of 11 students were receiving some kind of special education services or were in the process of being evaluated for special education eligibility. Student participants were between 5- and 9-years of age (M = 6.7; SD = 1.23). According to the current NCES statistics on school-aged student demographics (White = 44%, Hispanic/Latino = 28%, Black non-Hispanic = 15%, Asian = 6%, two or more races = 5%, American Indian/Alaskan Native = 1%, and Native Hawaiian/Pacific Islander < 1%). White students were overrepresented in our sample. While Hispanic students were underrepresented, and Asian students were somewhat underrepresented, the representation of Black students was similar to national estimates.
- *Describe the analysis procedures for each reported type of validity.
- To examine concurrent validity, we conducted bi-variate correlations between each iFAB measure and criterion measures. We performed bootstrapping on 1,000 bootstrap samples setting the desired confidence level to 95%, the resulting output provided correlation coefficients in addition to the lower and upper bounds of the 95% confidence intervals.
*In the table below, report the results of the validity analyses described above (e.g., concurrent or predictive validity, evidence based on response processes, evidence based on internal structure, evidence based on relations to other variables, and/or evidence based on consequences of testing), and the criterion measures.
| Type of | Subscale | Subgroup | Informant | Age / Grade | Test or Criterion | n (sample/ examinees) |
n (raters) |
Median Coefficient | 95% Confidence Interval Lower Bound |
95% Confidence Interval Upper Bound |
|---|
- Results from other forms of validity analysis not compatible with above table format:
- Manual cites other published reliability studies:
- Provide citations for additional published studies.
- Describe the degree to which the provided data support the validity of the tool.
- Do you have validity data that are disaggregated by gender, race/ethnicity, or other subgroups (e.g., English language learners, students with disabilities)?
If yes, fill in data for each subgroup with disaggregated validity data.
| Type of | Subscale | Subgroup | Informant | Age / Grade | Test or Criterion | n (sample/ examinees) |
n (raters) |
Median Coefficient | 95% Confidence Interval Lower Bound |
95% Confidence Interval Upper Bound |
|---|
- Results from other forms of validity analysis not compatible with above table format:
- Manual cites other published reliability studies:
- Provide citations for additional published studies.
Bias Analysis
| Age / Grade: Informant |
Grades K-3
Teacher |
|---|---|
| Rating | Not Provided |
- Have you conducted additional analyses related to the extent to which your tool is or is not biased against subgroups (e.g., race/ethnicity, gender, socioeconomic status, students with disabilities, English language learners)? Examples might include Differential Item Functioning (DIF) or invariance testing in multiple-group confirmatory factor models.
- No
- If yes,
- a. Describe the method used to determine the presence or absence of bias:
- b. Describe the subgroups for which bias analyses were conducted:
- c. Describe the results of the bias analyses conducted, including data and interpretative statements. Include magnitude of effect (if available) if bias has been identified.
Growth Standards
Sensitivity to Behavior Change
| Age / Grade: Informant |
Grades K-3
Teacher |
|---|---|
| Rating |
|
Convincing evidence
Partially convincing evidence
Unconvincing evidence
Data unavailable- Describe evidence that the monitoring system produces data that are sensitive to detect incremental change (e.g., small behavior change in a short period of time such as every 20 days, or more frequently depending on the purpose of the construct). Evidence should be drawn from samples targeting the specific population that would benefit from intervention. Include in this example a hypothetical illustration (with narrative and/or graphics) of how these data could be used to monitor student performance frequently enough and with enough sensitivity to accurately assess change:
Reliability (Intensive Population): Reliability for Students in Need of Intensive Intervention
| Age / Grade Informant |
Grades K-3
Teacher |
|---|---|
| Rating |
|
Convincing evidence
Partially convincing evidence
Unconvincing evidence
Data unavailable- Offer a justification for each type of reliability reported, given the type and purpose of the tool:
- Describe the sample(s), including size and characteristics, for each reliability analysis conducted:
- Describe the analysis procedures for each reported type of reliability:
In the table(s) below, report the results of the reliability analyses described above (e.g., model-based evidence, internal consistency or inter-rater reliability coefficients). Report results by age range or grade level (if relevant) and include detail about the type of reliability data, statistic generated, and sample size and demographic information.
| Type of | Subscale | Subgroup | Informant | Age / Grade | Test or Criterion | n (sample/ examinees) |
n (raters) |
Median Coefficient | 95% Confidence Interval Lower Bound |
95% Confidence Interval Upper Bound |
|---|
- Results from other forms of reliability analysis not compatible with above table format:
- Manual cites other published reliability studies:
- No
- Provide citations for additional published studies.
- Do you have reliability data that are disaggregated by gender, race/ethnicity, or other subgroups (e.g., English language learners, students with disabilities)?
- No
- If yes, fill in data for each subgroup with disaggregated reliability data.
Type of Subscale Subgroup Informant Age / Grade Test or Criterion n
(sample/
examinees)n
(raters)Median Coefficient 95% Confidence Interval
Lower Bound95% Confidence Interval
Upper Bound
- Results from other forms of reliability analysis not compatible with above table format:
- Manual cites other published reliability studies:
- No
- Provide citations for additional published studies.
Validity (Intensive Population): Validity for Students in Need of Intensive Intervention
| Age / Grade Informant |
Grades K-3
Teacher |
|---|---|
| Rating |
|
Convincing evidence
Partially convincing evidence
Unconvincing evidence
Data unavailable- Describe each criterion measure used and explain why each measure is appropriate, given the type and purpose of the tool.
- Describe the sample(s), including size and characteristics, for each validity analysis conducted.
- Describe the analysis procedures for each reported type of validity.
- In the table(s) below, report the results of the validity analyses described above (e.g., concurrent or predictive validity, evidence based on response processes, evidence based on internal structure, evidence based on relations to other variables, and/or evidence based on consequences of testing), and the criterion measures.
Type of Subscale Subgroup Informant Age / Grade Test or Criterion n
(sample/
examinees)n
(raters)Median Coefficient 95% Confidence Interval
Lower Bound95% Confidence Interval
Upper Bound
- Results from other forms of validity analysis not compatible with above table format:
- Manual cites other published reliability studies:
- No
- Provide citations for additional published studies.
- Describe the degree to which the provided data support the validity of the tool.
- Do you have validity data that are disaggregated by gender, race/ethnicity, or other subgroups (e.g., English language learners, students with disabilities)?
- If yes, fill in data for each subgroup with disaggregated validity data.
Type of Subscale Subgroup Informant Age / Grade Test or Criterion n
(sample/
examinees)n
(raters)Median Coefficient 95% Confidence Interval
Lower Bound95% Confidence Interval
Upper Bound
- Results from other forms of validity analysis not compatible with above table format:
- Manual cites other published reliability studies:
- No
- Provide citations for additional published studies.
Decision Rules: Data to Support Intervention Change
| Age / Grade: Informant |
Grades K-3
Teacher |
|---|---|
| Rating |
|
Convincing evidence
Partially convincing evidence
Unconvincing evidence
Data unavailable- Are validated decision rules for when changes to the intervention need to be made specified in your manual or published materials?
- No
- If yes, specify the decision rules:
-
What is the evidentiary basis for these decision rules?
Decision Rules: Data to Support Intervention Selection
| Age / Grade: Informant |
Grades K-3
Teacher |
|---|---|
| Rating |
|
Convincing evidence
Partially convincing evidence
Unconvincing evidence
Data unavailable- Are validated decision rules for what intervention(s) to select specified in your manual or published materials?
- No
- If yes, specify the decision rules:
-
What is the evidentiary basis for these decision rules?
Data Collection Practices
Most tools and programs evaluated by the NCII are branded products which have been submitted by the companies, organizations, or individuals that disseminate these products. These entities supply the textual information shown above, but not the ratings accompanying the text. NCII administrators and members of our Technical Review Committees have reviewed the content on this page, but NCII cannot guarantee that this information is free from error or reflective of recent changes to the product. Tools and programs have the opportunity to be updated annually or upon request.

