Skip to main content

Developmental Assessments

Developmental assessments provide a method for gaining information about young or very low-functioning children’s cognitive abilities, as well as many other areas, including academic skills, motor skills, communication and language skills, social skills, and self-help/adaptive skills. The assessments in this category consist of screening instruments, criterion-referenced measures, rating scales, and norm-referenced measures, some of which can be completed by a teacher or caregiver or through direct interaction with the child being assessed.

View Evaluations

Overview

Developmental assessments provide a method for gaining information about young or very low-functioning children’s cognitive abilities, as well as many other areas, including academic skills, motor skills, communication and language skills, social skills, and self-help/adaptive skills. The assessments in this category consist of screening instruments, criterion-referenced measures, rating scales, and norm-referenced measures, some of which can be completed by a teacher or caregiver or through direct interaction with the child being assessed. If a developmental measure does not address adaptive behavior, a separate measure of adaptive behavior will provide additional important data for the evaluation team to consider in programming decisions. Clinicians may consider obtaining scores from more than one developmental or cognitive test since it cannot be assumed that developmental test scores are interchangeable (Bishop, Guthrie, Coffing, & Lord, 2011). In addition, since testing behavior can affect obtained scores particularly among preschoolers (Akshoomoff, 2006), clinicians administering developmental assessments should be familiar with instrument administration, scoring, and interpretation; moreover, they should be familiar with working with preschoolers, including those with ASD.

Included within this section of the TARGET is summary information about the following instruments for developmental assessment:

  • Ages & Stages Questionnaires, Third Edition (ASQ-3)
  • Battelle Developmental Inventory – Second Edition Normative Update (BDI-2 NU)
  • Bayley Scales of Infant and Toddler Development –Fourth Edition (Bayley-4)
  • Developmental Assessment of Young Children – Second Edition (DAYC-2)
  • Developmental Profile –Third Edition (DP-3)
  • Hawaii Early Learning Profile (HELP)
  • Merrill-Palmer-Revised Scales of Development (M-P-R)
  • Mullen Scales of Early Learning (Mullen/MSEL)
  • Psychoeducational Profile –Third Edition (PEP-3)
  • Transdisciplinary Play-Based Assessment – Second Edition (TPBA-2)

The summary of developmental assessments included in this section is not intended to be all-inclusive. Rather, the assessments were selected based on their prevalence within clinical and academic settings as well as their relevance to children with ASD.

Several developmental assessments have been widely studied in the ASD population and found to be psychometrically sound ways of measuring skills of preschoolers with autism. Construct, convergent and divergent validity of the Mullen was similar in ASD and nonspectrum samples (Swineford, Guthrie, & Thurm, 2015). The Bayley-III has also been identified as a useful developmental assessment to evaluate young children with ASD (Long, Gurka, & Blackman, 2011; Torras-Mana et al., 2016), as convergent validity of the Bayley-III with cognitive assessment measures has been found for those with ASD (Torras-Mana et al., 2016). The PEP-3 has also been found to be a useful indicator of cognitive and language functioning in young children with ASD (De Giacomo et al., 2016), as well as of adaptive functioning (Fu et al., 2012). However, Fulton and D’Entremont (2013) warned that PEP-3 scores varied from other developmental measures’ scores in their study, underscoring the need to include multiple developmental measures to obtain the most accurate picture of an individual’s abilities.

Moreover, some developmental tests are predictive of later cognitive test scores for preschoolers with ASD. For example, Dempsey et al. (2018) found that the Cognitive domain of the M-P-R showed a moderate but positive correlation with later WISC-IV scores. Mullen and DAS-II scores have also been found to be highly correlated, which suggests they measure a similar construct; however, mean differences in scores varied depending on the child’s overall cognitive ability (Farmer, Golden, & Thurm, 2015).

Long and colleagues (2014) found that language skills of preschoolers with ASD were significantly lower than their typically developing peers’, but fewer differences were found between overall cognitive skills of children with and without ASD, as measured on the Bayley-III. Similarly, Dempsey et al. (2018) found that preschoolers with ASD had higher general cognitive scores than receptive language scores on the M-P-R.

Developmental assessments are not designed to measure symptoms of ASD, but ASD symptoms can and do affect performance on developmental instruments. For example, higher autism severity scores were associated with total scores of the BDI-2, as well as predicted scores on all BDI-2 domains (Goldin, Matson, Beighley, & Jang, 2014). The ASQ-3, particularly scores on the Communication Domain, can also alert clinicians to the need for autism-specific follow-up evaluation (Hardy, Haisley, Manning, & Fein, 2015). However, Williams and colleagues (2018) found that the ASQ-3 used in conjunction with the ASQ:SE is more likely to accurately identify subtle early concerns that may indicate ASD.

Change-over-time among young children with ASD has also been investigated with developmental assessments. In a 2014 study, Williams and colleagues found that participants with lower overall developmental scores demonstrated greater gains in their BDI-2 scores than those with average scores at baseline. The authors hypothesized that some of the children who initially showed low scores exhibited regression toward the mean upon subsequent testing; alternately children with greater deficits at the outset may have received more intense and/or variety of services, thus affecting DQ. Regardless of the explanation for this observed effect, there is wide variability in development among young children and clinicians should be mindful of the various reasons that developmental scores may change over time. This is also applicable to comparing child performance on developmental tests during the preschool years to their performance on cognitive tests in the school-age years. For example, even though Mullen and DAS-II scores are highly correlated, DAS-II scores were consistently higher in a study conducted by Farmer, Golden, and Thurm (2015); the researchers underscored that initial evaluations using the Mullen with follow-ups on the DAS-II may produce score “changes” that give the false impression of considerable improvement over time. Use of PEP-3 raw scores and developmental ages may be quite useful for tracking change-over-time, including responsiveness to intervention or treatment effects (Chen et al., 2011).

Some research has suggested that developmental assessments may indicate the need for ASD evaluation in children, or serve as a kind of screening instrument. In a large sample (n = 1,668) of 17-36 month-old children, Sipes and colleagues (2011) identified BDI-2 cut-off scores that optimized sensitivity and specificity if this instrument was used for ASD screening purposes. One study with the M-P-R concluded that this test was able to identify developmental delay but not differentiate autism specifically (Peters, 2014).

As in other domains of evaluation, multifaceted data collection is important in developmental assessment. Ulvand (2014) concluded that direct assessment with the Mullen, as well as indirect assessment through both parent and teacher reports, yielded valid measurement of the language functioning of preschoolers with autism.

Misconceptions

Myth:

Some students are untestable.

Reality:

No student is untestable. It is a matter of finding the most appropriate and suitable approach. For example, when standardized testing procedures are ineffective, administer a test in a non-standardized fashion and collect meaningful qualitative data. Always document when you break standardization guidelines. Alternate sources of data may be used. Examples include developmental measures, observations, informal student assessment, portfolios, and interviews with the teacher/parent/caregiver.

Myth:

Developmental quotients, IQs, and mental ages provide important information about a student.

Reality:

While scores from measures are informative, test results are only one piece of the picture. Do not underestimate the value of qualitative information gathered during the evaluation. Numbers do not define a student.