Behavior rating scales like the DESSA-Mini strive to measure the frequency of behaviors that indicate age and context-appropriate emotional recognition and management, social awareness and interactions, and responsible decision-making. Behavior rating scales can be used repeatedly, across settings, and with multiple informants to capture an understanding of the child’s behavior over time and relative to a standardized reference group. Yet, behavioral rating scales have been criticized for their potential to incorporate rater bias into assessment scores (Elliot, Frey, & Davies, 2015).
Rater bias is “the presence of substantial and systematic error in ratings of performance or behavior caused by rater attitudes, beliefs, or experiences” (Mason, Gunersel, & Ney, 2014). Previous studies have considered sources of divergence between teacher ratings and a “true” criterion (Podsakoff, MacKenzie, & Podsakoff, 2012). Since SEC does not have a consensus indicator of “true” levels, this paper seeks to understand the variance in DESSA-Mini scores that is accounted for by rater characteristics.
Methods: Data were collected within a diverse California school district (13% Asian, 12% Black, 7% Filipino, 54% Hispanic, 9% White, 4% Other/Multiple/Not Reported). Rater characteristics were collected from (n=63) elementary school classroom teachers. These teachers assessed 1,676 students using the DESSA-Mini. The students were assessed to have similar levels of SEC (M=50.9; SD=11.7) as the national reference group (M=50.00; SD=10.00). Using an unconditional multilevel model, only 16% of the variance in student scores was attributable to characteristics of the teacher raters.
Findings: Teachers were 97% female, 60% White, 98% credentialed, and 16% district residents. Prior to training for the implementation of a new SEL program, 60% of teachers expressed that social emotional competence was “essential” (4 on 0-4 scale) to school success and 77% expressed typically feeling “eager” or “very eager” (3 or 4 on 0-4 scale) to adopt new initiatives. The extent to which teachers, at the end of an in-service training for the implementation of a new SEL curriculum, believed that the SEL program would benefit their students (B=1.90, p<0.10), intended to implement the program fully (B=1.00, p<0.05,), and did not anticipate implementation challenges (B=-2.46, p<0.01) each reliably accounted for shared variance in student assessment scores.
Conclusion and Implications: Identifying these sources of “rater bias” may lead to a) controlling for these sources of variance in research and b) shrinking these sources of variance through high-quality program selection, training, and implementation planning.