iFAB (Individualized Formative Assessment of Behavior)
FBRM-MIS: Interpersonal Skills
Summary
Individualized Formative Assessment of Behavior (iFAB) is a mobile-enabled web-based system for behavioral progress monitoring of elementary students in grades K -3, that is flexible, feasible, and psychometrically-sound. Using iFAB, teachers or other school staff can generate assessment plans, enter data, record events (e.g., change of intervention plan), and review auto-generated time series charts of student progress. Developed from the ground up for the purpose of formative behavior assessment, iFAB offers the ability to assess both academic enablers (i.e., Academic Engagement, Organization Skills, Interpersonal Skills) and problem behaviors that commonly occur in school settings (Disruptive, Oppositional, Interpersonal Conflict, Anxious-Depressed, and Social Withdrawal). The system offers considerable flexibility in that ratings may be completed daily (direct behavior rating method) or weekly (formative behavior rating method). Users also have the choice of rating student behavior using single-item scales or multi-item scales. This submission applies to the FBRM rating method, the multi-item scale, and the Interpersonal Skills behavior construct.
- Where to Obtain:
- Self
- r.volpe@neu.edu
- 617-702-6220
- https://www.ifabonline.com/
- Initial Cost:
- Free
- Replacement Cost:
- Free
- Included in Cost:
- At this point, we are making the tools available at no cost.
- NA
- Training Requirements:
- Training not required
- Qualified Administrators:
- No minimum qualifications specified.
- Access to Technical Support:
- Assessment Format:
-
- Rating scale
- Scoring Time:
-
- Scoring is automatic OR
- 0 minutes per
- Scores Generated:
-
- Raw score
- Composite scores
- Administration Time:
-
- 3 minutes per student
- Scoring Method:
-
- Automatically (computer-scored)
- Technology Requirements:
-
- Computer or tablet
- Internet connection
Tool Information
Descriptive Information
- Please provide a description of your tool:
- Individualized Formative Assessment of Behavior (iFAB) is a mobile-enabled web-based system for behavioral progress monitoring of elementary students in grades K -3, that is flexible, feasible, and psychometrically-sound. Using iFAB, teachers or other school staff can generate assessment plans, enter data, record events (e.g., change of intervention plan), and review auto-generated time series charts of student progress. Developed from the ground up for the purpose of formative behavior assessment, iFAB offers the ability to assess both academic enablers (i.e., Academic Engagement, Organization Skills, Interpersonal Skills) and problem behaviors that commonly occur in school settings (Disruptive, Oppositional, Interpersonal Conflict, Anxious-Depressed, and Social Withdrawal). The system offers considerable flexibility in that ratings may be completed daily (direct behavior rating method) or weekly (formative behavior rating method). Users also have the choice of rating student behavior using single-item scales or multi-item scales. This submission applies to the FBRM rating method, the multi-item scale, and the Interpersonal Skills behavior construct.
- Is your tool designed to measure progress towards an end-of-year goal (e.g., oral reading fluency) or progress towards a short-term skill (e.g., letter naming fluency)?
-
ACADEMIC ONLY: What dimensions does the tool assess?
- BEHAVIOR ONLY: Please identify which broad domain(s)/construct(s) are measured by your tool and define each sub-domain or sub-construct.
- Interpersonal Skills
- BEHAVIOR ONLY: Which category of behaviors does your tool target?
Both
Acquisition and Cost Information
Administration
Training & Scoring
Training
- Is training for the administrator required?
- No
- Describe the time required for administrator training, if applicable:
- We have purposely evaluated the measures as administered with no training.
- Please describe the minimum qualifications an administrator must possess.
- We made a conscious choice to design formative behavior assessment tools that required no training. As such, no training was involved in our psychometric studies.
-
No minimum qualifications
- Are training manuals and materials available?
- No
- Are training manuals/materials field-tested?
- No
- Are training manuals/materials included in cost of tools?
- If No, please describe training costs:
- NA
- Can users obtain ongoing professional and technical support?
- Yes
- If Yes, please describe how users can obtain support:
Scoring
- Please describe the scoring structure. Provide relevant details such as the scoring format, the number of items overall, the number of items per subscale, what the cluster/composite score comprises, and how raw scores are calculated.
- Multi-item scales generate scale composites and also allow users to track scores by item as they tend to overlap with specific target behaviors. Broad composites for academic enablers and problem behaviors also are generated.
- Do you provide basis for calculating slope (e.g., amount of improvement per unit in time)?
- No
- ACADEMIC ONLY: Do you provide benchmarks for the slopes?
- ACADEMIC ONLY: Do you provide percentile ranks for the slopes?
- Describe the tool’s approach to progress monitoring, behavior samples, test format, and/or scoring practices, including steps taken to ensure that it is appropriate for use with culturally and linguistically diverse populations and students with disabilities.
- Assessment plans can be tailored to each individual student in that teachers/consultants select relevant scales for monitoring. They can decide which of the four methods to employ for progress monitoring. Our psychometric studies were conducted on diverse groups of students in both regular education and special education. We are currently seeking funding to explore the measurement invariance of the iFAB measures.
Levels of Performance and Usability
- Date
- Size
- Male
- Female
- Unknown
- Eligible for free or reduced-price lunch
- Other SES Indicators
- White, Non-Hispanic
- Black, Non-Hispanic
- Hispanic
- American Indian/Alaska Native
- Asian/Pacific Islander
- Other
- Unknown
- Disability classification (Please describe)
- First language (Please describe)
- Language proficiency status (Please describe)
Performance Level
Reliability
| Age / Grade Informant |
Grades K-3
Teacher |
|---|---|
| Rating |
|
Convincing evidence
Partially convincing evidence
Unconvincing evidence
Data unavailable- *Offer a justification for each type of reliability reported, given the type and purpose of the tool.
- Internal consistency: This is a relevant indicator of item homogeneity and item quality for our two multi-item methods (DBR-MIS and FBRM-MIS). This is not specific to formative assessment measures, but is an important consideration for any multi-item scale or test generating a composite score. Temporal stability and dependability across occasions are important to examine in formative behavioral assessment as inconsistency of ratings across measurement occasions is a common source of error. For formative measures we want to know the extent to which changes in scores can be attributable to change in the target students behavior as opposed to time-related error.
- *Describe the sample(s), including size and characteristics, for each reliability analysis conducted.
- Internal Consistency A (from Volpe et al., 2019): Participants were general and special education teachers from 35 public school districts across 13 states. Investigators e-mailed principals and school psychologists from partner network schools to provide information regarding the study. Subsequent to administrator approval, interested teachers in participating schools were contacted directly by investigators and provided with a brief description of the purpose and procedures of the study. A total of 307 K-3 teachers each completed ratings for one randomly selected student in their class. Teachers were primarily female (95.8%) with a wide range of teaching experience. The student sample was comprised of 187 (60.9%) males and 120 (39.1%) females, and students were nearly evenly distributed across grade levels. The composition of students by race and ethnicity was as follows: 67.1% White, 13% Black, 15.0% Hispanic, 3.3% Asian, 1.0% American/Alaska Native, and 5.2% Unknown. Approximately 35% of students were receiving special education services at the time of data collection. Dependability (from Volpe et al., 2023): Ninety-one K–5 educators (i.e., teachers, paraprofessionals, and intervention specialists) from elementary schools located in the Northeast, Midwest, and West of the United States participated in the study. Educators were mainly female (n = 86), White (n = 87), and their age ranged from 21 to 62 (M = 39.00, SD = 10.40) years. On average, they had 12.70 (SD = 7.87) years of experience. Each educator rated one student in their classroom. The largest number of participating students were classified as male (n = 61) and their age ranged from 5 to 10 (M = 7.36, SD = 1.39) years. The majority of students were White (n = 60), followed by Black (n = 17), Hispanic (n = 10), and biracial (n = 4). Approximately 45% of the students received special education services (n = 41). Once all student participants were identified, educators were randomly assigned to one of the following four conditions: (a) DBR-SIS, (b) FBRM-SIS, (c) DBR-MIS, or (d) FBRM-MIS. For the multi-item scales (DBR-MIS and FBRM-MIS), we also report internal consistency. Cronbach’s alpha was calculated on the first ratings completed by teachers for each student. For DBR-MIS, the first rating occurred after one school day at the beginning of the week (typically, on Monday). For FBRM-MIS, the first rating occurred at the end of one school week (typically, on Friday). MTMM (test-retest): A total of 68 elementary school teachers were recruited from schools in Boston and the surrounding metro area. The mean age of teachers was 37.82 (range 25-67) and their mean years of experience teaching was 11.49 (range 1 to 40).
- *Describe the analysis procedures for each reported type of reliability.
- As part of our initial development of the scales comprising the iFAB, we conducted a series of exploratory factor analyses for one of our four assessment methods (DBR-MIS). In our three published studies (Briesch et al., 2022; Daniels, Briesch, & Volpe, 2021; Volpe, Chaffee, Yeung, & Briesch, 2019) we reported both coefficient alpha and coefficient omega. We used these indices as indicators of item homogeneity and item quality. We examined the dependability of all four methods across all scales in a recently published series of G studies (Volpe, Matta, Briesch, & Owens, 2023). Here we conducted a set of parallel G studies to investigate differences in dependability across our four rating methods. A single-facet design was used with students (s) as the object of measurement and occasion (o) as the facet. The design was fully crossed with all students evaluated on five occasions (s x o). Time related variance has been well-documented as a source of error in formative behavioral assessment and in this study, we were interested in the extent to which a single week of data collection might generate scores with acceptable levels of dependability. We were particularly interested in making comparisons across methods and constructs. We are currently preparing several manuscripts to examine the criterion-related validity of the iFAB measures. We are using a multi-trait multi-method (MTMM) approach, wherein we examine associations with the iFAB measures and a group of criterion measures. In the diagonal of the MTMM matrix are coefficients of stability for each measure. We report the 1-week stability coefficients for each scale.
*In the table(s) below, report the results of the reliability analyses described above (e.g., model-based evidence, internal consistency or inter-rater reliability coefficients). Include detail about the type of reliability data, statistic generated, and sample size and demographic information.
| Type of | Subscale | Subgroup | Informant | Age / Grade | Test or Criterion | n (sample/ examinees) |
n (raters) |
Median Coefficient | 95% Confidence Interval Lower Bound |
95% Confidence Interval Upper Bound |
|---|
- Results from other forms of reliability analysis not compatible with above table format:
- Scale Informant Grade N Rater N Participant Dependability Coefficient Number of Measurement Occasions to Criterion Interpersonal Skills (FBRM-MIS) Teacher K-5 27 27 0.98 3 weekly rating to reach >0.80 coefficient Volpe, R. J., Matta, M., & Briesch, A. M. (2023). Formative behavioral assessment across eight Constructs: Dependability of direct behavior ratings and formative rating measures. Journal of School Psychology. https://doi.org/10.1016/j.jsp.2023.101251
- Manual cites other published reliability studies:
- No
- Provide citations for additional published studies.
- Do you have reliability data that are disaggregated by gender, race/ethnicity, or other subgroups (e.g., English language learners, students with disabilities)?
- No
If yes, fill in data for each subgroup with disaggregated reliability data.
| Type of | Subscale | Subgroup | Informant | Age / Grade | Test or Criterion | n (sample/ examinees) |
n (raters) |
Median Coefficient | 95% Confidence Interval Lower Bound |
95% Confidence Interval Upper Bound |
|---|
- Results from other forms of reliability analysis not compatible with above table format:
- Manual cites other published reliability studies:
- No
- Provide citations for additional published studies.
Validity
| Age / Grade Informant |
Grades K-3
Teacher |
|---|---|
| Rating |
|
Convincing evidence
Partially convincing evidence
Unconvincing evidence
Data unavailable- *Describe each criterion measure used and explain why each measure is appropriate, given the type and purpose of the tool.
- Interpersonal Skills: The iFab Interpersonal Skills scales were designed to measure student social skills. Both the ACES IS and BASC-3 social skills scales are established measures of student social skills in school settings.
- *Describe the sample(s), including size and characteristics, for each validity analysis conducted.
- Elementary school teachers were recruited from schools in Boston and the surrounding metro area. To avoid range restriction in this correlational design, a general sample of students was selected (as opposed to selecting at-risk students alone). Specifically, teachers were asked to complete a packet of rating forms for one male and one female student selected at random. The overall sample was stratified by gender and grade level in order to ensure equal proportions of male and female students in grades K-1 and 2-3. Each of the two students was assigned to one of the two following conditions: • Daily condition: Teachers completed the 8 DBR scales using both SIS and MIS methods each day. Both the order of DBR formats (SIS and MIS) and scales (e.g., Disruptive, Social Withdrawal) was randomized. Both the order and format of scales was randomized. Each daily rating completed over the course of 1-week was averaged for a composite of ratings for each scale. • Weekly condition: Teachers completed the 8 DBR scales using both SIS and MIS methods at the end of each week. Both the order of DBR formats (SIS and MIS) and scales (e.g., Disruptive, Social Withdrawal) was randomized. Teachers were also asked to complete a set of established rating measures, including ACES, ADHD Symptom Checklist-IV, BASC-3, BRIEF-2 (Plan/Organize and Organization of Materials subscales), CDI-2, and SAS-TR, at the end of the week (i.e., on the weekend). Description of Students in the Weekly Condition: There were 71 students in the Weekly Condition (33 boys, 34 girls). Sex data were missing for 4 students. Kindergarten students comprised 33.8% of the sample, first-graders 26.8%, second-graders 16.9% and third-graders 16.9% of the sample. Grade data were missing for 4 cases. Approximately 62% of the sample was White, 16.9% Black, 2.8% Hispanic, 4.2 Asian, and 4.2% other. In regard to ethnicity, 19.7% of students were Hispanic. A total of 13 students were receiving some kind of special education services or were in the process of being evaluated for special education eligibility. Student participants were between 5- and 9-years of age (M = 6.58; SD = 1.28). According to the current NCES statistics on school-aged student demographics (White = 44%, Hispanic/Latino = 28%, Black non-Hispanic = 15%, Asian = 6%, two or more races = 5%, American Indian/Alaskan Native = 1%, and Native Hawaiian/Pacific Islander < 1%). White students were overrepresented in our sample. While Hispanic students were underrepresented, and Asian students were somewhat underrepresented, the representation of Black students was similar to national estimates.
- *Describe the analysis procedures for each reported type of validity.
- To examine concurrent validity, we conducted bi-variate correlations between each iFAB measure and criterion measures. We performed bootstrapping on 1,000 bootstrap samples setting the desired confidence level to 95%, the resulting output provided correlation coefficients in addition to the lower and upper bounds of the 95% confidence intervals.
*In the table below, report the results of the validity analyses described above (e.g., concurrent or predictive validity, evidence based on response processes, evidence based on internal structure, evidence based on relations to other variables, and/or evidence based on consequences of testing), and the criterion measures.
| Type of | Subscale | Subgroup | Informant | Age / Grade | Test or Criterion | n (sample/ examinees) |
n (raters) |
Median Coefficient | 95% Confidence Interval Lower Bound |
95% Confidence Interval Upper Bound |
|---|
- Results from other forms of validity analysis not compatible with above table format:
- Manual cites other published reliability studies:
- Provide citations for additional published studies.
- Describe the degree to which the provided data support the validity of the tool.
- Do you have validity data that are disaggregated by gender, race/ethnicity, or other subgroups (e.g., English language learners, students with disabilities)?
If yes, fill in data for each subgroup with disaggregated validity data.
| Type of | Subscale | Subgroup | Informant | Age / Grade | Test or Criterion | n (sample/ examinees) |
n (raters) |
Median Coefficient | 95% Confidence Interval Lower Bound |
95% Confidence Interval Upper Bound |
|---|
- Results from other forms of validity analysis not compatible with above table format:
- Manual cites other published reliability studies:
- Provide citations for additional published studies.
Bias Analysis
| Age / Grade: Informant |
Grades K-3
Teacher |
|---|---|
| Rating | Not Provided |
- Have you conducted additional analyses related to the extent to which your tool is or is not biased against subgroups (e.g., race/ethnicity, gender, socioeconomic status, students with disabilities, English language learners)? Examples might include Differential Item Functioning (DIF) or invariance testing in multiple-group confirmatory factor models.
- No
- If yes,
- a. Describe the method used to determine the presence or absence of bias:
- b. Describe the subgroups for which bias analyses were conducted:
- c. Describe the results of the bias analyses conducted, including data and interpretative statements. Include magnitude of effect (if available) if bias has been identified.
Growth Standards
Sensitivity to Behavior Change
| Age / Grade: Informant |
Grades K-3
Teacher |
|---|---|
| Rating |
|
Convincing evidence
Partially convincing evidence
Unconvincing evidence
Data unavailable- Describe evidence that the monitoring system produces data that are sensitive to detect incremental change (e.g., small behavior change in a short period of time such as every 20 days, or more frequently depending on the purpose of the construct). Evidence should be drawn from samples targeting the specific population that would benefit from intervention. Include in this example a hypothetical illustration (with narrative and/or graphics) of how these data could be used to monitor student performance frequently enough and with enough sensitivity to accurately assess change:
Reliability (Intensive Population): Reliability for Students in Need of Intensive Intervention
| Age / Grade Informant |
Grades K-3
Teacher |
|---|---|
| Rating |
|
Convincing evidence
Partially convincing evidence
Unconvincing evidence
Data unavailable- Offer a justification for each type of reliability reported, given the type and purpose of the tool:
- Describe the sample(s), including size and characteristics, for each reliability analysis conducted:
- Describe the analysis procedures for each reported type of reliability:
In the table(s) below, report the results of the reliability analyses described above (e.g., model-based evidence, internal consistency or inter-rater reliability coefficients). Report results by age range or grade level (if relevant) and include detail about the type of reliability data, statistic generated, and sample size and demographic information.
| Type of | Subscale | Subgroup | Informant | Age / Grade | Test or Criterion | n (sample/ examinees) |
n (raters) |
Median Coefficient | 95% Confidence Interval Lower Bound |
95% Confidence Interval Upper Bound |
|---|
- Results from other forms of reliability analysis not compatible with above table format:
- Manual cites other published reliability studies:
- No
- Provide citations for additional published studies.
- Do you have reliability data that are disaggregated by gender, race/ethnicity, or other subgroups (e.g., English language learners, students with disabilities)?
- No
- If yes, fill in data for each subgroup with disaggregated reliability data.
Type of Subscale Subgroup Informant Age / Grade Test or Criterion n
(sample/
examinees)n
(raters)Median Coefficient 95% Confidence Interval
Lower Bound95% Confidence Interval
Upper Bound
- Results from other forms of reliability analysis not compatible with above table format:
- Manual cites other published reliability studies:
- No
- Provide citations for additional published studies.
Validity (Intensive Population): Validity for Students in Need of Intensive Intervention
| Age / Grade Informant |
Grades K-3
Teacher |
|---|---|
| Rating |
|
Convincing evidence
Partially convincing evidence
Unconvincing evidence
Data unavailable- Describe each criterion measure used and explain why each measure is appropriate, given the type and purpose of the tool.
- Describe the sample(s), including size and characteristics, for each validity analysis conducted.
- Describe the analysis procedures for each reported type of validity.
- In the table(s) below, report the results of the validity analyses described above (e.g., concurrent or predictive validity, evidence based on response processes, evidence based on internal structure, evidence based on relations to other variables, and/or evidence based on consequences of testing), and the criterion measures.
Type of Subscale Subgroup Informant Age / Grade Test or Criterion n
(sample/
examinees)n
(raters)Median Coefficient 95% Confidence Interval
Lower Bound95% Confidence Interval
Upper Bound
- Results from other forms of validity analysis not compatible with above table format:
- Manual cites other published reliability studies:
- No
- Provide citations for additional published studies.
- Describe the degree to which the provided data support the validity of the tool.
- Do you have validity data that are disaggregated by gender, race/ethnicity, or other subgroups (e.g., English language learners, students with disabilities)?
- If yes, fill in data for each subgroup with disaggregated validity data.
Type of Subscale Subgroup Informant Age / Grade Test or Criterion n
(sample/
examinees)n
(raters)Median Coefficient 95% Confidence Interval
Lower Bound95% Confidence Interval
Upper Bound
- Results from other forms of validity analysis not compatible with above table format:
- Manual cites other published reliability studies:
- No
- Provide citations for additional published studies.
Decision Rules: Data to Support Intervention Change
| Age / Grade: Informant |
Grades K-3
Teacher |
|---|---|
| Rating |
|
Convincing evidence
Partially convincing evidence
Unconvincing evidence
Data unavailable- Are validated decision rules for when changes to the intervention need to be made specified in your manual or published materials?
- No
- If yes, specify the decision rules:
-
What is the evidentiary basis for these decision rules?
Decision Rules: Data to Support Intervention Selection
| Age / Grade: Informant |
Grades K-3
Teacher |
|---|---|
| Rating |
|
Convincing evidence
Partially convincing evidence
Unconvincing evidence
Data unavailable- Are validated decision rules for what intervention(s) to select specified in your manual or published materials?
- No
- If yes, specify the decision rules:
-
What is the evidentiary basis for these decision rules?
Data Collection Practices
Most tools and programs evaluated by the NCII are branded products which have been submitted by the companies, organizations, or individuals that disseminate these products. These entities supply the textual information shown above, but not the ratings accompanying the text. NCII administrators and members of our Technical Review Committees have reviewed the content on this page, but NCII cannot guarantee that this information is free from error or reflective of recent changes to the product. Tools and programs have the opportunity to be updated annually or upon request.

