Abstract
While many standardized assessment measures exist to track child mental health treatment outcomes, the degree to which such tools have been adequately tested for reliability and validity across race, ethnicity, and class is uneven. This paper examines the corpus of published tests of psychometric properties for the ten standardized measures used in U.S. child outpatient care, with focus on breadth of testing across these domains. Our goal is to assist care providers, researchers, and legislators in understanding how cultural mismatch impacts measurement accuracy and how to select tools appropriate to the characteristics of their client populations. We also highlight avenues of needed research for measures that are in common use. The list of measures was compiled from (1) U.S. state Department of Mental Health websites; (2) a survey of California county behavioral health agency directors; and (3) exploratory literature scans of published research. Ten measures met inclusion criteria; for each one a systematic review of psychometrics literature was conducted. Diversity of participant research samples was examined as well as differences in reliability and validity by gender, race or ethnicity, and socio-economic class. All measures showed adequate reliability and validity, however half lacked diverse testing across all three domains and all lacked testing with Asian American/Pacific Islander and Native American children. ASEBA, PSC, and SDQ had the broadest testing.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
While many standardized assessment tools have been developed to track child mental health treatment outcomes on the individual and aggregate level, the degree of testing for reliability and validity across race, ethnicity and class is uneven. Nearly two thirds (62.1%) of children who received treatment through the national Children’s Mental Health Initiative (CMHI) from 2013 to 2017 were racial or ethnic minorities or biracial, and 71.3% came from families at or below the federal poverty threshold (SAMHSA, 2020). For this reason, it is crucial that clinicians have the information to select measures with proven reliability and validity for their clients.
This paper reviews the published psychometrics literature on the most commonly used standardized outcome measures in use for U.S. outpatient community child mental health care, with two aims:
-
1.
To assist clinicians and policy makers in making informed decisions when selecting a standardized outcome measurement system.
-
2.
To highlight needed avenues of additional testing for commonly used measures, to ensure suitability for diverse client populations.
To these ends, this paper presents a summary of the psychometrics tests for each of the investigated measures, focusing on disparities across gender, class, and race and ethnicity, followed by a comparison of the entire corpus to examine which populations are systematically overlooked across studies. We also briefly discuss the importance of taking into account differences in community background (whether these be race or ethnic group, class, or immigrant culture) and highlight specific ways such differences can impact measurement accuracy, as well as recommended readings for clinicians on the topic of culturally congruent care.
The literature scans for this review were conducted as part of a contracted project with the California Department of Health Care Services (DHCS).
Background
Reducing disparities in the access to, and quality of, child mental health care has long been identified as a national priority area (Perou et al., 2013; US Department of Health and Human Services et al., 2000). The U.S. Department of Health and Human Services’ National Quality Strategy envisions that quality improvement is driven by linking recommended care processes to meaningful clinical outcomes, as well as aligning financial incentives to promote effective care (AHRQ, 2016). The Patient Protection and Affordable Care Act was passed in 2010 (42 U.S.C. § 18,001 et seq.). Although California legislated a mandate in 2011 to develop a performance outcome system for children (SB 1009; California Legislative Analyst’s Office, 2011), the state’s development of data infrastructures to monitor quality and detect disparities have considerably lagged behind national recommendations (Gardner & Kelleher, 2017; Glied et al., 2015; Patel et al., 2015; Pincus, 2012; Zima et al., 2013).
In 2016, the California Department of Health Care Services (DHCS) contracted our university to help them address the question, “What is the best statewide approach to evaluate functional status for children/youth that are served by the California public specialty mental health service system?” (DHCS, 2015, p. 6). To assist the CA DHCS in developing their outcome monitoring infrastructure, we created a list of all standardized measures in common use for tracking overall child mental health outcomes in the U.S., and ranked them on nine minimum criteria (Pourat et al., 2017; Zima et al., 2019). In this paper, we will lay out the findings from two rounds of systematic literature scans, and examine the breadth and diversity of psychometric testing on each candidate measure, namely: (1) published evidence for its use as a clinical outcome measure; (2) its psychometric properties and variation among diverse study populations; and (3) whether the existing published evidence adequately included children of diverse genders, classes, ethnicities and races. Details about demographics, sample size, etc., for each of the cited studies can be found in the tables. Results are divided into two sections: first the analyses of individual measures, then an aggregate analysis of the entire corpus for systematic patterns and gaps.
The Discussion section examines the significance of these results, particularly the systematic under-representation of Native Americans and Asian Americans. We also discuss the importance of clinicians’ having an understanding of cultural and linguistic differences across class, race, and ethnicity (in addition to cultural differences for immigrant families), and guide clinicians in recognizing specific types of misreporting that can occur if a measure is not properly developed for a particular population. Finally, we examine the pros and cons of three possible solutions: creating adjusted scoring guides; using internationally developed questionnaires for immigrant children; and developing measures specifically for under-represented populations.
Methods
Compiling a List of Candidate Measures
The first step in this project was to create a pool of all eligible measures in common use in the U.S. that might meet the DHCS’s needs, after which we would research each individual measure more closely. We conducted three initial investigations: (1) an exploratory 5-year literature scan; (2) an environmental scan of U.S. state Department of Mental Health agency websites, and (3) two statewide California surveys. Additionally, (4) DHCS requested information on four measures that had been recommended to them by other sources. Table 1 lists which criteria were met for each measure.
Exploratory Literature Scan
Systematic searches of PubMed, PsycInfo, and Scopus were conducted for peer-reviewed journal articles published between January 2010 and December 2015, with English language abstracts, that examined children ages 0–18 years in U.S. community-based outpatient care.Footnote 1 (Further details on article inclusion and exclusion criteria are in Supplemental Table 1, the 5-year cutoff was deliberately chosen to capture measures in current use). A list was compiled of all the standardized measures that were used in these studies as data collection tools to track outcomes, resulting in approximately 225 child clinical outcome measures in 127 unique articles. Thirty-four of these measures appeared in three or more articles.Footnote 2 Of these, twenty-one were specific to a single diagnosis or condition (e.g. ADHD), three were general health or quality of life measures, and three were excluded for other reasons,Footnote 3 leaving seven candidate measures.
Measures in Common Use (Nationally or in California) or of Interest to the DHCS
An environmental scan of state Department of Mental Health agency websites was conducted to determine which measures were in common use across the U.S. Thirty-five states listed at least one standardized assessment measure, for a total of 36 different measures (Pourat et al., 2016b).
To determine which measures were in common use in California, we conducted a statewide survey of county behavioral health agency directors (56 counties), and a second convenience sample of outpatient clinic staff (21 responses) which yielded seven eligible measures that were used in more than one county (Pourat et al., 2016a). Finally, the DHCS Subject Matter Experts team requested we also investigate four measures that had been recommended to them (three of which also appeared in the county and state lists).
Required Scope of Candidate Measures
To align with the priorities of our DHCS agency partners, the list was narrowed to measures that had the following characteristics: (1) track overall behavioral or emotional health (i.e. not specific to a single diagnosis such as depression); (2) are designed for children 5–16 years; (3) have been normed or tested with children in the U.S.; and (4) produce quantifiable scores that can be used to compare treatment outcomes of different patients, or can be aggregated to compare the quality of care of different service provides.
The final list consisted of all measures meeting those criteria that also met at least one of the following use criteria: (1) appeared in at least 3 studies in the exploratory literature scan (7 measures); (2) was reported by at least 2 California counties (4 measures); or (3) was recommended on at least 2 state DMH webpages (6 measures). One measure on the DHCS interest list did not meet any of the other criteria (TOP); it was also included. This yielded a total of 10 measures flagged for further investigation, listed in Table 1.
Examination of Psychometric Properties and Capacity to Identify Disparities
After the candidate list was compiled, ten systematic literature reviews (one for each measure) were conducted to examine psychometric properties and suitability for diverse communities. For each measure, a Web of Science search was made of published articles with English language abstracts from the measure’s initial development through March 2019, that tested reliability and/or validity with children under 18 years of age for either symptoms or functioning.Footnote 4 (Further details on article inclusion and exclusion criteria are in Supplemental Table 2) Literature reviews and meta-analyses were manually examined for additional citations. Further citations were obtained from measure developers’ or vendors’ webpages (if they existed) as well as from articles recommended for inclusion by a DHCS-selected team of Subject Matter Experts.
Studies that focused on specific demographics (e.g. by ethnicity or socioeconomic status) were included, as well as studies focusing on populations that might be recipients of outpatient care in California (e.g. immigrants or adopted children). Studies that used non-English versions of the measure were included if they used an independent metric to test reliability and validity (i.e. were not simply comparing a translation to the English language original).
We examined the characteristics of the research participants in each study and across the entire corpus, as well as noting reported differences in psychometric properties by race or ethnicity, gender, and class or SES. For ethnicity and gender, the researchers’ own categories were used. To determine class diversity, we looked for: explicit mention of SES or household income, more general class labels (e.g. “working-class," “upper middle class,” “poor”), eligibility for financial services or other aid (e.g. Medicaid or free school lunches), or enrollment in programs specifically designed for low-income families (e.g. HeadStart).
Results for Individual Measures
The final list of candidate measures by selection criteria are summarized in Table 1. The candidate measures were Achenbach System of Empirically Based Assessment (ASEBA); Child and Adolescent Functional Assessment Scale (CAFAS); Child and Adolescent Needs and Strengths (CANS); Children’s Global Assessment Scale (CGAS); Clinical Global Impressions Scale (CGI); Ohio Youth Problem, Functioning and Satisfaction Scales (Ohio); Pediatric Symptom Checklist (PSC); Strengths and Difficulties Questionnaire (SDQ); Treatment Outcome Package (TOP); and Youth Outcome Questionnaire (Y-OQ).
Use as a Clinical Outcome Measure in Community-based Mental Health Programs
Findings from the exploratory literature scan are summarized in Table 2. The three measures most frequently used to track clinical outcomes among children receiving community-based mental health care were ASEBA (21 studies),Footnote 5 CGAS (14 studies), and SDQ (13 studies). Five measures were only used in the U.S. (CAFAS, CANS, Ohio, PSC, and Y-OQ), while four were also used internationally (ASEBA, CGAS, CGI, and SDQ). TOP was added to the list of measures to investigate at the request of DHCS, however it did not appear in the literature scan. One fifth (13 of 57) of the studies combined multiple candidate measures, most frequently CGAS with either SDQ (4 studies) or CGI (3 studies). Other measures used in combination were: CGI (6 of 8 studies), SDQ (6 of 13), PSC (2 of 2), ASEBA (3 of 21), CAFAS (1 of 2), and Ohio (1 of 4). All measures were applied to children with a diverse range of mental health conditions including: general use across psychiatric conditions (18 studies), broad categories such as behavioral or emotional problems (6 studies) or trauma (7 studies); or specific diagnoses such as anxiety (7 studies) or ADHD (5 studies).
All ten measures were designed for wide age ranges and covered at minimum 5–18 years. Despite this, over one quarter of the studies (15 of 57) used a measure for children outside the recommended age range.
Follow-up intervals varied extensively and did not consistently correspond to the measure’s recommended use. One fifth (12) of the studies used archival data from existing clinic records, illustrating the feasibility of using the measure in clinical practice, but also indicating that clinics do not always track their clients’ outcomes at consistent, regular intervals. Similarly, one third (21 studies) administered the measure only at intake and end of treatment or patient discharge, which led to variation in episode of care across their data set.
Sample Diversity and Psychometric Properties of Candidate Measures
The following sections summarize the diversity of study samples and overall psychometrics for each individual measure. Sample characteristics for each study are summarized in Table 3. Measures are listed in order of number of published studies. Most had less than five studies testing reliability and validity with U.S. children, however PSC (23 studies), ASEBA (22) and SDQ (12) were more extensively tested.
Pediatric Symptom Checklist—PSC (n = 23 Articles; 23 Study Samples)
Sample Diversity
Studies included diversity across ethnicity, culture (including immigrant children), class and gender. One third (7 studies) were predominately (over 60%) White, while half were either mixed (5) or predominately African American (3) or Latino (4). Four studies did not list ethnicity. Two thirds of the studies recruited low-income or Medicaid-receiving participants (9) or used a mixed-class sample (6). Three were mostly middle class, and four did not provide information. Two studies focused on foster youth and five focused on Spanish-dominant parents. Three quarters of the community samples (14 of 18) were recruited via pediatric primary care.
Results suggest using a lower clinical cutoff for disadvantaged families (Simonian & Tarnowski, 2001) and children of Latino immigrants (Jutte et al., 2003). Murphy et al. (1996) found that Mexican immigrant parents scored their children slightly higher when answering the PSC orally than when filling out the written form, suggesting that they are more likely to describe problems verbally. Pagano et al. (1996) and Jutte et al. (2003) also validated PSC for Spanish speaking parents.
Validity results for low-income and minority children were also mixed, as discussed below. Gender results were mixed: Leiner et al. (2007) found no significant gender differences for Mexican families while Boothroyd and Armstrong (2010) found small to moderate gender effect in a mixed-ethnicity sample.
For children in foster care, PSC showed slightly lower test–retest reliability (Jacobson et al., 2019), moderate convergent validity (Parker et al., 2019), and mixed results for discriminant validity (Jacobson et al., 2019; Parker et al., 2019).
Validity for low-income and minority children is mixed: Earlier studies supported the validity of PSC with African American and low-income children (Murphy et al., 1992) and showed comparable validity and reliability compared to middle class children (Jellinek et al., 1986; Murphy & Jellinek, 1988). However, Kostanecka et al. (2008) found PSC-17’s externalizing and attention subscales to have low discriminant validity with their low-income, predominately African American sample.
Psychometric Properties
PSC showed high inter-rater reliability between parent and student (Murphy et al., 1989), significant correlation with parents’ reports of functioning problems (Pagano et al., 1996, 2000), moderate correlation with pediatricians’ (Jellinek et al., 1986, 1988), teachers’ (Pagano et al., 2000) and school counselors’ reports (Murphy & Jellinek, 1988), and moderate to high agreement with other standardized measures including CBCL (Jellinek et al., 1986; Leiner et al., 2007), CGAS (Jellinek et al., 1988; Murphy et al., 1992), SCARED and CDI (Gardner et al., 2007; Parker et al., 2019) and the Diagnostic Interview for Children and Adolescents Parent Report (DICA-P) (Jellinek et al., 1988).
PSC showed high specificity compared to pediatrician ratings: Jellinek et al., (1988, 1995) found that children who had experienced high stress might meet clinical criteria on PSC even when rated as functional by pediatricians. However, they also found an overall trend of pediatricians under-detecting problems when compared to child psychologists, particularly for low-income families (Jellinek et al., 1995).
Construct and discriminant validity were high (Jacobson et al., 2019) as were reliability over time and as an outcome measure (Boothroyd & Armstrong, 2010; Murphy & Jellinek, 1988; Murphy et al., 1992, 2012; Navon et al., 2001). Test–retest reliability was high for PSC-35 (Jellinek et al., 1986; Navon et al., 2001) and PSC-17 (Murphy et al., 2016) as well as the preschool PSC-18 (Sheldrick et al., 2012), and moderate for the 0–18 month Baby PSC (Sheldrick et al., 2013).
Achenbach System of Empirically Based Assessment— ASEBA (n = 23 Articles; 25 Samples)
Sample Diversity
Studies included diversity across class and gender although half of the samples (12 out of 25) had larger proportions of boys. Like PSC, one third were predominately White (9 samples, 8 studies), but African American families (3 studies) and Spanish speaking Puerto Rican families (2 studies, 3 samples) were represented, as well as six studies (5 samples) using ethnically diverse samples. Four studies (5 samples) did not provide ethnicity.
Half the studies (13 studies, 14 samples) included class information. Of these, the community samples were either middle class (3 from 2 studies) or mixed (4 from 5 studies), while the samples of children receiving mental health care were predominately lower income (5 of 7). ASEBA was also tested with families with special needs adopted children (Tharinger et al., 1986) and military families (Jensen et al., 1993, 1996). Two studies validated CBCL for Spanish speaking Puerto Rican parents: Rubio Stipec et al. (1990), who used their own translation, and Bird et al. (1987a).
Konold et al. (2004) found no effect of child/parent gender, but others found that the Attention Problems subscale correlated with internalizing conditions for girls but externalizing ones for boys (Song et al., 1994), and that ASEBA had trouble distinguishing between girls' anxiety versus depression (Ebesutani et al., 2010). An early version of CBCL showed poor factor model fit for African American families (Jastrowski Mano et al., 2009) with a mismatch between parent-reported problems and CBCL’s list of problem items (Lambert et al., 2002). These issues are not mentioned in later studies and may have been fixed.
Psychometric Properties
Inter-rater reliability is high (Reed & Edelbrock, 1983) including between parent and clinician (Dutra et al., 2004), parent and teacher (Konold et al., 2004), parent and child (Ebesutani et al., 2011), and mother and father (Ebesutani et al., 2010; Konold et al., 2004). ASEBA shows strong correlation with relevant variables such as hospitalizations, suicide attempts, personality pathology, and family history (Dutra et al., 2004), as well as academic achievement measures (Hogez & McKay, 1986), and clinical diagnoses and history of mental health service use (Jensen et al., 1996; Rubio-Stipec et al., 1990; Tharinger et al., 1986). Jensen et al. (1996) found that while CBCL was less accurate than a diagnostic interview (DISC), it was “reasonably comparable.” There is also strong correlation with other standardized measures including BASC (Jastrowski Mano et al., 2009) and the DSM III (Tharinger et al., 1986) among others (Jastrowski Mano et al., 2009; Lambert et al., 2002; Nakamura et al., 2009; Salcedo et al., 2018; Tharinger et al., 1986). Although Reed and Edelbrock (1983) found strong correlations between TRF scores and referrals for problem behavior, Nelson et al. (2002) found low correlations with school disciplinary referrals for behavioral problems.
Discriminant validity is supported, although CBCL has higher specificity (true negatives) than sensitivity (true positives) (Rishel et al., 2005). Some early studies found correlations between subscales but they were still distinguishable (Dedrick et al., 1997; Nakamura et al., 2009), and recent research shows subscales are good to fair at distinguishing between the condition targeted by the subscale and other conditions (Ebesutani et al., 2010) although poor at more detailed distinctions such as type of anxiety disorder (Knepley et al., 2019), especially for internalizing conditions (Jensen et al., 1993).
Strengths and Difficulties Questionnaire—SDQ (n = 12 Articles; 13 Samples)
Sample Diversity
Gender balance was even across most samples. Three samples were primarily White, three were ethnically mixed, and three focused on specific ethnic groups: African Americans (Jee et al., 2011), Chinese and Korean immigrants (Yu et al., 2016), and Latinos (Downs et al., 2012). The other four samples (3 studies) did not include information on race or ethnicity. Half the studies did not mention class/income (6 studies, 7 samples), the others were either mixed or predominately low-income families. One study focused on foster youth (Jee et al., 2011).
Two studies examined Spanish dominant families: Downs et al. (2012) compared English and Spanish-speaking preschoolers, while a quarter of Hill and Hughes’ (2007) parent study sample were bilingual or limited English proficiency.Footnote 6 Only two studies recruited children receiving mental health treatment: psychiatric inpatients (Kovacs & Sharp, 2014) and youth in residential care (Mason et al., 2012). The rest used community samples, although two focused on children who were flagged as “at risk” based on SES (Downs et al., 2012) or low literacy (Hill & Hughes, 2007).
Care should be taken to use the culturally-normed cutoff scores available on the developer’s webpage, particularly with immigrant families (Dickey & Blumberg, 2004; Downs et al., 2012) as three studies found cultural differences in the parent-rated Conduct Problems and Peer Problems subscales. In a cross-national study, American parents had several items correlate more strongly with the Hyperactivity and Emotional Problems subscales compared to British parents, suggesting cultural differences in interpretations of child behavior (Dickey & Blumberg, 2004) or measure responses. Yu et al. (2016) found low reliability for Chinese and Korean immigrant parents, and low convergent and discriminant validity on the Hyperactive/Inattentive scale. Finally, Downs et al. (2012) found the Emotional Problems subscale to be suitable for Spanish-speaking U.S. preschool boys, but inadequate for girls.
For foster youth, sensitivity was high compared to CHIPS (93%) when youth and foster parent reports were combined, but lower for each one alone (54% for youth, 71% for foster parents) (Jee et al., 2011).
Psychometric Properties
Internal consistency was high for Total Score and moderate to high for most subscales except for Peers and Conduct and the already mentioned cultural issues (Bourdon et al., 2005; Downs et al., 2012; Yu et al., 2016). SDQ showed moderate to high correlation with CBCL (Kovacs & Sharp, 2014) over time (Mason et al., 2012) and strong correlation with reports of service use (Bourdon et al., 2005) and other variables known to be predictive of mental health problems, e.g. low SES (Bourdon et al., 2005).
Good construct and longitudinal validity (Deutz et al., 2018). Good convergent validity (He et al., 2013; Hill & Hughes, 2007) apart from the cultural issues noted. A longitudinal cohort study of first-graders showed poor discriminant validity (Hill & Hughes, 2007). Test–retest reliability is strong (Downs et al., 2012).
Child and Adolescent Functional Assessment Scale—CAFAS (n = 4 Articles; 3 Samples)
Sample Diversity
Murphy et al. (1999) recruited a predominately Hispanic community sample of English and Spanish speaking low-income pre-school children enrolled in a Head Start program.Footnote 7 Hodges and Wong (1996) recruited predominately White children of army personnel who had been referred for mental health services; they also compared ratings for fictional written patient vignettes. Bates et al. (2006) collected clinician and student ratings of individual questions items.Francis et al. (2012) compared CAFAS and GAF (the adult version of CGAS) for a multi-ethnic sample of adolescents referred for mental health evaluation, but did not present any conclusions regarding the accuracy of either measure.
Psychometric Properties
Inter-rater reliability was moderate to high for written vignettes for the school-age CAFAS (Hodges & Wong, 1996) and for parental reports for the preschool PECFAS (Murphy et al., 1999). There is strong correlation with reported problematic behaviors, poor academic performance, and teacher rating of psychosocial problems (Hodges & Wong, 1996; Murphy et al., 1999). The only construct validity tests were based on graduate student ratings of individual items (Bates et al., 2006). In a comparison with CAFAS, GAF identified roughly equal proportions of functional impairment for youth with externalizing versus internalizing diagnoses, while CAFAS identified twice as many externalizing cases compared to internalizing ones (Francis et al., 2012).
Children’s Global Assessment Scale—CGAS (n = 4 Articles; 5 Samples)
Sample Diversity
One sample was all male, two were at least two thirds male, and two were balanced gender. Two studies (3 samples) included information on ethnicity: Francis et al. (2012) used a multi-ethnic sample to compare the older GAF version with CAFAS (see above) and Bird et al. (1987b) recruited clinic-referred and community samples of Spanish speaking Puerto Rican children. Two studies used predominately working-class samples, the others did not list such information. Green et al. (1994) extracted ratings from clinical records of psychiatric inpatients, they found different patterns from other studies (see below).
Psychometric Properties
CGAS showed consistently high inter-rater reliability (Bird et al., 1987b; Shaffer et al., 1983) even among raters with different types of experience, such as psychiatrists versus nurses (Green et al., 1994). It also showed high discriminant validity between clinic and community populations (Bird et al., 1987b), and between inpatients and outpatients (Shaffer et al., 1983). While Bird et al. (1987b) found high concurrent validity with CBCL, Green et al. (1994) found no significant correlation between CGAS rating and symptomatology measures (the CBCL Behavior Problems subscale) but did find strong correlations with measures of functioning (e.g. the CBCL Activities and School subscales, WISC-R IQ scale, and measures of social relatedness), and these were stronger for higher functioning children. Green and colleagues explain the discrepancy in results by arguing that their study used clinical records, in contrast to more controlled studies (such as Bird’s), which either used homogenous structured data (e.g. written vignettes or videos) or tested with raters who were not involved in the child’s treatment. In both cases, study raters would be focusing on different features than would clinicians and staff in normal practice. Additionally, they add, CBCL is normed on non-clinical samples which, by definition, would have a broader range of scores than the psychiatric inpatient samples that Green’s group looked at. Francis et al. (2012) also found discrepancies between GAF and CAFAS (discussed above) but did not investigate the cause.
Child and Adolescent Needs and Strengths—CANS (n = 3 Articles)
Sample Diversity
The short form was tested with predominately Hispanic children; the other two studies were primarily (67–78%) White. Participants in one study were Medicaid recipients (Anderson et al., 2003), income status for the others was unknown. All studies focused on children receiving mental health care; there were no community comparison samples. Worth noting, Alamdari and Kelber used Y-OQ to test concurrent validity; as we will discuss below, Y-OQ has not been adequately tested with non-White children.
Psychometric Properties
According to Rosanbalm et al. (2016), CANS was not designed using a psychometric approach and was originally intended to be used only for individuals, not aggregated.
Three different variations of CANS were tested: a mental health scale (Anderson et al., 2003), a short form (Alamdari & Kelber, 2016), and a trauma screener (Kisiel et al., 2018). Anderson et al. (2003) found high interrater reliability between caseworkers and researchers on the mental health scale but did not address validity. The short form showed good concurrent validity with Y-OQ’s somatic and behavior dysfunction subscales (Alamdari & Kelber, 2016), while the trauma screener showed good convergent and discriminant validity when compared to TSCC-A and CBCL (Kisiel et al., 2018).
Youth Outcome Questionnaire—Y-OQ (n = 3 Articles; 10 Samples)
Sample Diversity
There was little information on sample characteristics: three of the five community samples were 80% White, two were also predominately middle class. There was no information on ethnicity or class for the others. One community sample was over two thirds female, and four clinic samples from the same study were predominately male; the rest were all equally balanced.
Psychometric Properties
Y-OQ showed high internal consistency across community and clinical samples for Total Score, and moderate to high for subscale scores, as well as moderate to high correlation with CBCL, the Connors Parent Rating Scale (Burlingame et al., 2001), and YSR (Ridge et al., 2009). Burlingame et al. (2001) tested discriminant validity via a sensitivity analysis using a combined sample recruited from clinics, schools, and the general community. Their results suggest Y-OQ can differentiate non-clinical, outpatient, and inpatient children. Test–retest reliability was high when tested on non-clinical samples in all three studies.
Ohio Youth Problem, Functioning and Satisfaction Scales—Ohio (n = 2 Articles; 8 Samples)
Sample Diversity
Both studies provided sparse information: Dowell and Ogles (2008) (2 samples) recruited predominately Caucasian families, while Ogles et al. (2001) (6 samples) provided no information on race or ethnicity. There was no information on SES or class. Three samples were predominately male, three were balanced, two did not include gender information.
Psychometric Properties
Internal consistency and test–retest reliability were high (Dowell & Ogles, 2008; Ogles et al., 2001), with moderate to high correlations with CBCL (parent), Vanderbilt Functioning Index (parent and caseworker), YSR (youth), and with CAFAS, CGAS and the Progress Evaluation Scales (all caseworker reports) (Ogles et al., 2001). CAFAS and CGAS both have fairly little testing with non-White children (CAFAS with Latino preschoolers and CGAS only with Puerto Ricans); both are discussed elsewhere in this paper. Moderate correlation with the BASC was found for a community sample, but no statistically significant correlation for the corresponding service client sample (Dowell & Ogles, 2008).
Treatment Outcome Package—TOP (n = 1 Article; 1 Sample)
Sample Diversity
Community participants were given anonymous packets to fill out and return by mail, no details about ethnicity or class were collected. There were no studies testing TOP with children receiving mental health care.
Psychometric Properties
Moderate to high correlations between TOP’s subscales and equivalent CBCL and SDQ subscales; no information on reliability.
Clinical Global Impressions Scale—CGI (n = 0 Articles)
There were no studies testing CGI on U.S. child populations, even though it is in use as a child mental health measure—it appeared in 8 studies in our exploratory scan—and has been used as a benchmark to test other measures developed for children (e.g. the Obsessive–Compulsive Inventory—Child Version, see McGuire et al., 2019). This gap in testing may be partly because CGI was originally developed for schizophrenia (Guy, 1976) and only later adopted as a general measure of functioning. Regardless of the reason, we mention it in this review because its reliability and validity need to be examined with child populations if it is to be used as a general mental health measure.
Aggregated Results Across Measures
Use of Measures in Research and Clinical Practice
Extracting Data from Clinical Records to Monitor Quality of Care
In the exploratory literature scan, at least 12 of 57 studies pulled data from existing clinical records. While these records confirm feasibility of the measures’ use in usual care practice for community-based outpatient settings, they also reveal that irregular follow-up intervals are common, for example administering only at intake and discharge. This can complicate comparing patients or aggregating results, and can also make it difficult to analyze effectiveness of care for children who have dropped out mid-treatment. Archival records therefore may not be sufficient for state or county level quality monitoring.
Measures that were Used as Benchmarks for Psychometrics Testing
Across the 73 psychometrics articles, five of the candidate measures (ASEBA, CGAS, CAFAS, SDQ and Y-OQ) were used as benchmarks in psychometrics tests of other measures, another indicator of their popularity. The ASEBA package appeared in 15 articles and was used as a benchmark for all the other measures except CANS. CGAS appeared in 8 articles to test CBCL, Ohio, and PSC. The other three each appeared once: CAFAS for Ohio, SDQ for TOP, and Y-OQ for CANS.Footnote 8 All except Y-OQ have some testing with minority (African American or Latino) families, as well as with low income or working-class families.
Breadth and Limitations of Psychometric Testing by Sociodemographic Characteristics
Examining the entire corpus of psychometric studies reveals systematic gaps in the populations with whom the measures are being tested, namely: a tendency toward male-dominated samples in clinic populations; lack of representation of Asian American, Pacific Islander, and Native American children; and lack of examination of differences across social class or SES. For information on the status of each measure, see Table 4.
Testing by Gender
Almost all study samplesFootnote 9 (83 of 89; 93.3%) included participant gender. Of these, over half (48; 57.8%) had a roughly even gender balance,Footnote 10 while over a third (31; 37.3%) were all or mostly male, including five all-male samples. In contrast, only four samples (4.8%) were mostly female. Two thirds (32 of 52; 61.5%) of the community (i.e. “control”) samples were gender balanced, however when examining samples recruited from clinics or other contexts where children had been referred for emotional or behavioral problems, we found that over half of them (20 of 37; 54%) were male-dominant.
Testing by Race and Ethnicity
While only two thirds of samples tracked participant race or ethnicity (61 of 89; 68.5%), most published studies (58 of 72;Footnote 11 80%) provided information for at least one sample. Of the 61 samples for which information was provided, nearly half (27; 44.3%) were predominately White, a third (19; 31.5%) were predominately non-White and a quarter (15; 24.5%) were mixed.
The most frequently listed non-White categories were African American or Black, Hispanic or Latino (sometimes as separate classifications), and mixed-race. In contrast, Asian Americans, Pacific Islanders, and Native Americans were not even mentioned in many of the studies. Only three studies had samples with more than 10% of A/PI-heritage participants: Francis et al. (2012) (17%); Nakamura et al. (2009) (14%); and Yu et al. (2016) (who focused exclusively on Chinese and Korean immigrant parents).
Testing by Class and Socioeconomic Status
Half of the study samples (45 of 89), and nearly half of the published studies (31 of 72, 43%), had no information on participants’ socioeconomic status, class, or household income. Of the 44 samples that did, almost half focused on working class or low-income demographics (21, 47.7%; or 23.6% of total corpus) while over a quarter used mixed-class samples (12, 27.3%; or 13.4% of total corpus). This combination of facts raises some concerns of possible selection or sampling bias, i.e. that the researchers who include information on social class tend to be those who are intentionally recruiting low-income participants or creating mixed class samples. Studies that do not report such information may be more class-homogeneous or less representative, particularly in older studies as health research has suffered from a lack of data collection on economic status or social class (see e.g. Krieger & Fee, 1994). Of course, it is impossible to say for certain since we only have access to what is in the published articles, but caution should be taken when generalizing from these studies.
Testing by Linguistic and Cultural Diversity
The majority (83%) of the 72 studies used the original English version of the measure. Eleven studies (for 5 measures) included Spanish-speaking families, and one study focused on Chinese and Korean-speaking immigrant parents. Table 4 presents the breakdown by measure, and the individual studies are cited in Table 3. For this review, we did not examine studies that compared how well a translation matched the original English-language measure because these would not provide information about the accuracy and quality of the measure itself. However, we encourage clinicians working with limited English proficiency families to seek out such studies (e.g. Stolk et al.’s, 2017 literature review on translations of SDQ into languages spoken by refugee families).
Other Participant Characteristics
Most of the measures were tested on both community samples and samples of children diagnosed with, or receiving care for, a mental or behavioral health condition. The exceptions were CANS, which was only tested on mental health care clients, TOP, which was only tested on an anonymous community sample, and CGI, which, as mentioned, was not tested on children at all. Some studies examined more specific populations such as special needs adopted children (Tyson et al., 2011) or children of military personnel (Hodges & Wong, 1996), details on these can be found in Table 3.
Discussion
This discussion will address two issues that emerged in our study: the specific absence of Native American and Asian American families in psychometrics testing, and the more general lack of diversity testing for several measures in popular use. We will examine how cultural and linguistic differences (including those emerging from racial or ethnic, class, or geographic differences) can lead to errors on standardized measures, discuss the pros and cons of possible solutions (including norming of cutoff scores), and provide two examples of measures that were intentionally designed for Native American youth.
Degree of Testing of Psychometric Properties Across Diverse Populations
As discussed above, published evidence supporting reliability and validity were found for all measures except CGI, however the number of studies and the diversity of participants varied widely.
The number of publications for each measure varied widely, however this number does not indicate a measure’s quality or the thoroughness of the testing. Once the initial validity and reliability tests have been published, there is little incentive to publish replicated results unless a significantly different new version is released or unless the author can provide something new that was not in the original publication, such as applying it in a different setting or with a different population.Footnote 12 While the two measures with the most publications (ASEBA and PSC, with over 20 studies each) did have testing across a more diverse population, overall breadth of testing is more important than number of articles, and here many of the measures fell short.
Three measures (PSC, SDQ and ASEBA) were tested across diverse genders, across SES or class, and with both Latino and African American as well as White children, however none were adequately tested with Native American or US-born Asian American children. Of these three measures, only ASEBA and PSC had breadth of testing across both community populations and diagnosed (or in-care) children.
Table 4 presents a comparison of gaps in testing for each measure. The popularity of CANS in particular (used in over half of California counties, see Pourat et al., 2016a), suggest that in some cases measures might be used in clinical practice without knowing whether they are suitability to the client population.
Underrepresented Populations: Asian Americans, Native Americans, and Immigrants
Information on Asian Americans, Pacific Islanders, and Native Americans was heavily lacking for all measures. Only three studies had over 15% Asian-identifying participants; two did not provide any data about ethnic differences, and the third (Yu et al., 2016) focused on Chinese and Korean immigrants, not US-born Asian Americans. Native Americans were not even examined.
These absences are problematic as both populations are highly vulnerable. American Indian/Alaska Native adolescents have the highest rates of depression of all ethnic groups (American Psychiatric Association, 2017). Both Native American and Asian American adolescents have rates of suicide ideation and suicide attempts that exceed those of White adolescents (US Dept. of Health & Human Services: Office of Minority Health, 2018a, b). There is also evidence of differences in symptom reports for parents of Asian and Pacific Islander ethnicity compared to non-minority parents (Okamura et al., 2016) even when child self-reports are similar, indicating a need for measurement tools tailored to these groups.
Immigrant families were also under-represented. There were a few studies involving Mexican immigrants or parents who spoke Spanish, but none with immigrants from other language groups. Currently one out of ten U.S. inhabitants (12.9%) and nearly one third of California inhabitants (27.2%) are foreign born, and that number is predicted to rise in the future (Trevelyan et al., 2016; U.S. Census Bureau, 2016), therefore this population needs further attention.
Lack of Evidence Does Not Equal Low Quality
A brief caveat: the lack of robust testing does not indicate a measure’s lack of suitability for diverse client populations, simply that further testing is required. We encourage interested researchers to fill in the gaps discussed here, and we encourage developers to broadly pilot their measures across gender, race and ethnicity, and social class. In the next section, we briefly highlight ways a standardized measure may fall short when administered with populations for whom it was not designed.
How Culture, Class, or Language Impact the Accuracy of Standardized Measures
It is important for care providers to be aware of how parents’ background, upbringing, and even language,Footnote 13 can impact their responses on standardized measures and interviews, and how unintentional misreporting or miscommunication about a child’s status can cause outsized effects on children and families. Cultural and linguistic differences are not limited to immigrant parents, we can find different cultural values and dialects across classes, racial or ethnic groups, and even geographic communities within the U.S.
Misdiagnoses (“false positives”) due to unclear reporting put financial burden and emotional distress on a family in addition to lost time (Au-Yeung et al., 2018; Baker & Bell, 1999). Failure to diagnose a problem (“false negatives”)—for example because questions were not understood or symptoms were not described in a way that the clinician recognized (Bailey et al., 2009; Shen et al., 2018)—can also lead to inappropriate treatment, or even denial of services if a standardized diagnostic tool does not accurately show a child’s level of need.
To give a severe example, measurement tools that assume Euro-American childrearing practices have led to First Nations parents being judged unfit (and children removed from their homes), because they are designed to assess a two-parent nuclear family rather than an extended family of multiple primary caregivers (Choate & McKenzie, 2015).
The importance of accurate reporting becomes heightened as primary care providers are being given increased responsibility in diagnosing, referring, and sometimes even treating mental illness issues themselves (Glazier et al., 2015), especially with low income families (Hodgkinson et al., 2017). While pediatricians are adapting to meet this need (Foy et al., 2019), they do not have the same level of training as a specialist and must rely more heavily on standardized measurement tools.
Specific Issues When Administering Measures
Details of racial, ethnic and cultural bias in test design have been extensively discussed elsewhere (see e.g. Reynolds et al., 2021 for a historical overview) and there is also a body of literature on how to develop and adapt measures to different patient populations.
Here we would like to briefly point out some specific misreport issues, so that providers can be aware of and take these into account when administering standardized measures:
Parents may have different benchmarks of severity when asked to rank a behavior as “problematic” or “burdensome.” Their ratings may be affected by cultural attitudes toward certain behavioral problems (Heiervang et al., 2008) or by the level and type of caregiving support available. In such cases, a teacher report may also be of assistance.
Parent/youth may be unfamiliar with the questionnaire structure. Parents and youth of different cultural backgrounds or education levels may not understand or be comfortable with certain types of standardized questions, such as rating an emotion on a numerical scale (Lee et al., 2002) or grid-formatted questions (Ware et al., 1995). Some red flags to check for are skipped answers, especially if they are all the same type of question (Ware et al., 1995), or whether a parent often marks the “neutral” midpoint (e.g. 3 on a scale of 1–5) when asked for level of agreement/disagreement or how positively/negatively they feel about something (Lee et al., 2002).
Differences in word nuance or meaning can occur when a translation does not convey the nuance of the original language, as well as when parent and clinician (or measure developer) speak different geographic or class dialects (Epstein and et al., 2015; Leplège et al., 1998). Similarly, the same diagnostic term may have different presentations in different cultures (Haroz et al., 2017; Jani & Deforge, 2014), and folk illness categories may not translate neatly into biomedical illness categories (Flores et al., 2002).
The diagnostic tool may not ask about (or code for) the parent’s concerns. Connected to the previous point, social environment and community values affect children’s symptom presentation and what behaviors parents consider problematic. For example, Lambert et al. (2002) (discussed in the ASEBA section above) listed over 20 concerns that appeared in clinical case notes with African American parents but could not be coded in CBCL’s schema. They hypothesized that many of these reflect the African American community’s higher valuing of community support, mutual respect, and education.
Parent report may differ from child’s self-report. ASEBA, SDQ, Ohio and Y-OQ all have a child version (typically for ages 11+) to supplement the parent report. While it is unsurprising that parent/teacher reports may differ from self-reports, the level of discrepancy can vary across ethnicity and across class (Ha et al., 1998; Okamura et al., 2016).
Clinician comprehension and unconscious bias will affect interpretation of parent concerns. Over half of the measures in our reviewFootnote 14 were designed to be completed by a clinician or social worker rather than the parent. In such cases, the accuracy of the tool is dependent on the provider being able to understand parental concerns and translate them into the standardized categories and ratings on the form. Many of the parent-report issues we discussed above also come into play in interviews with providers: perception of behavior severity, different word nuances, even parents’ level of familiarity with clinical interviews and how to describe mental health symptoms (Probst et al., 2007).
Clinician’s own biases also affect the accuracy of the report. Even well-intentioned providers may hold false beliefs or unconscious stereotypes about racial or ethnic groups (Bailey et al., 2019; Hoffman et al., 2016). Such beliefs may, on the one hand, cause providers to rank an item as less/more severe, or a symptom as less/more abnormal, for similarly presenting patients of different races. On the other hand, unawareness of real differences in cultural, social, and religious upbringing may hinder providers’ ability to recognize key symptoms when described in nonmainstream ways. For example folk explanations involving blood temperature (Elderkin-Thompson et al., 2001) or possession by evil spirits (Malgady et al., 1987) may be perceived as irrelevant or, even worse, as symptoms of psychosis.
Addressing Clinician Bias and Developing Cultural Awareness
Providing culturally appropriate care goes beyond merely avoiding racist or classist stereotypes. Because biases and gaps in cultural knowledge are often unconscious, the care provider must actively work at being “anti-racist” and at educating oneself about social differences (Cénat, 2020). Here are some helpful starting points for clinicians and social workers interested in this topic: Cénat (2020) gives a timely guide on how to address the needs of Black patients within the larger social context of racial violence and police brutality. McGregor et al.’s (1998) holistic approach to Hawai’ian health assessment incorporates family, community, spirituality, and relationship to the local ecology. Thyer (2015) critiques the DSM while McQuaide (1999) discusses its pros and cons in social work. Finally, Rohe (1985) reveals the impact of urban planning: how the physical layout of a city or community can benefit or impair its residents’ mental health.
Possible Solutions for Adjusting Measures to Different Populations—and Their Down Sides
Norms Cutoffs and Item Weighting
A common fix is to adjust clinical cutoffs, or even individual item weights, in accordance with the baseline of a community. For example, if cultural ideas of appropriate child behavior lead to consistently higher/lower ratings of impulsiveness or aggression, these items could be re-weighted in the final score.
However, there is strong danger in this solution. Using race- or class-based scoring sheets without understanding the underlying reasons for the reported differences runs the risk of reifying harmful stereotypes about populations (e.g. as innately more aggressive or with lower self-control) rather than understanding and addressing real social inequalities (Gasquoine, 2009).
That said, SDQ and ASEBA both have publicly available data on mean scores for community populations in various countries. ASEBA’s website uses three “norm groups,” (with some countries falling into different groups for different questionnaires e.g. CBCL versus YSR). It also has a Module with Multicultural Options (MMO) supplement that allows clinicians to compare a child’s score with the grading scales for multiple countries (e.g. looking at both the immigrant parent’s home country and the family’s country of residence). SDQ’s webpage lists means by age and gender, as well as a searchable database of published research studies.
Using Internationally Developed Questionnaires for Immigrant Families
Providers working with immigrant families may be asking whether they should look for measures validated in the family’s country of origin. Unfortunately, this can be equally problematic, as immigrant children are not raised in the same social environment as their cousins abroad. For example, Shen et al. (2017) developed and piloted a set of scales to measure anxiety, depression, and school problems in Chinese middle school students, but because their tool is designed for children immersed in China’s intense examination-oriented education system, it is not useable for Chinese-immigrant adolescents attending U.S. schools.
Creating Mental Health Measures for Minority Youth, Two Examples for Native American Youth
Finally, we offer two examples of mental health measures that were designed for, and tested with, Native American and First Nations youth. While these do not fit our original criteria (one is outside our age range, the other does not provide an overall mental health score), they may be of use as supplements to clinicians. Additionally, the details of their development may serve as models for researchers.
The Life Trajectory Interview for Youth was designed to “address gaps in our understanding of the links between large-scale structural conditions and social processes and individual outcomes such as mental health.” It was developed and pilot tested on a 60% Anglo/40% Cherokee sample of young adults (aged 19–24) living in western North Carolina. The measure covers four domains: life course milestones, life course barriers, social affordances, and material goods (Brown et al., 2006).
The Cultural Connectedness Scale (CCS) was designed for and piloted with Canadian First Nations, Métis, and Inuit youth (Snowshoe et al., 2015). Although not itself a mental health scale, its developers showed a link between cultural connectedness and mental wellness for these population (Snowshoe et al., 2017).
Use of Appropriate Benchmarks When Testing or Developing Measures
As noted earlier, several of the measures we reviewed (particularly CBCL) were used to test validity or correlations in psychometric tests of others measures. CBCL was the most widely used and it is a well-established and broadly tested measure. While we have discussed the importance of using diverse and representative study samples, it is also important for measure developers and researchers to beware of hidden biases or disparities within the benchmarks being used with these populations. For example, several of the studies reviewed here tested for correlations between a measure’s report of behavioral problems and poor academic performance (e.g. Hill & Hughes, 2007; Hodges & Wong, 1996) or school disciplinary referrals (e.g. Hill & Hughes, 2007; Hodges & Wong, 1996; Hogez & McKay, 1986; Murphy et al., 1999; Nelson et al., 2002; Reed & Edelbrock, 1983). Children of low SES families and minority children are at disproportionate risk of both of these, particularly African American boys (Heath, 1982; Skiba et al., 2002).
Scope and Limitations of this Review
Because the original goal of this project was to find candidate measures suitable for use in California DHCS outpatient child mental health care, we narrowed the scope of our review in the following ways: (1) The initial five-year exploratory search focused on English-language articles and only public or community clinics; this may have under-represented some measures’ popularity internationally or in other systems of care. (2) We only examined psychometrics tests on U.S. populations and (3) excluded those focusing on treatments or conditions which would not be covered in DHCS outpatient settings (e.g. substance abuse or skills training for autism). (4) Because we were interested in the reliability and validity of the measure itself, we did not examine studies that only compared the fidelity of a translation to the English version, but we did include non-English translations that were tested against independent metrics.
As noted, sample representativeness was examined for gender, race and ethnicity, and class/SES, as these were the three variables reported in most studies. While other participant characteristics such as adoption or type of family (e.g. foster families or same-sex parents) were not systematically compared, they are listed in Table 3 and in the sections focusing on individual measures.
Conclusion
By laying out the available and missing information about these ten clinical outcome measures, our goal has been to assist practitioners, researchers, and legislators in selecting appropriate standardized measures that have been tested and normed on samples that resemble their own client populations. In the process, we discovered that some popular measures lacked breadth of testing on diverse patient and community populations. Therefore, a second emerging goal of this paper has been to give clinicians insight into how cultural and linguistic differences (including those between racial, ethnic, and class groups) can impact measurement reports.
Testing with Asian American, Pacific Islander, and Native American families should be a high priority, as well as comparisons across classes. Further testing with immigrant families of various backgrounds is also needed. Because measures such as CANS, Y-OQ, and Ohio are already in wide clinical use, patient records may yield a good source of data, although we have discussed some of the caveats of using them.
Finally, researchers who are in the process of creating or adapting their own measures are encouraged to include these under-examined populations starting from the earliest stages of development and pilot testing.
Notes
U.K. and Australian CAMHS (Child and Adolescent Mental Health Services) and CYMHS (Child and Youth Mental Health Services) programs were included as a type of community mental health care.
This count does not include measures of treatment alliance, level of service use, or parenting outcomes. Qualitative measures such as semi-structured interviews and goal-tracking tools are also not counted since they cannot be aggregated or compared across patients.
These were: HoNOSCA (Health of the Nations Outcome Scale for Children and Adolescents), which has not been not validated for U.S. children; CIS (Columbia Impairment Scale), which is only for children 9 years and older, and GAF (Global Assessment of Functioning, formerly GAS) which was dropped in favor of the child-specific C-GAS.
Searches used the topic search string: TS = ([full name of measure] OR [abbreviation]) AND TS = (psychom* OR reliability OR validity).
All but one of these studies used the parent report Child Behavior Check-List (CBCL); 4 studies combined the CBCL with one of the other reports in the Achenbach package: either the Youth Self Report or the Teacher Report Form.
These parents were mailed both Spanish and English versions of SDQ but the authors did not report how many parents chose the Spanish version.
Head Start is a program run by the U.S. Department of Health and Human Services. It funds local community programs for low income families with children under 5 years old, in order to promote school readiness, children’s development, and family wellbeing.
Several studies included multiple different participant samples with different demographics. See Table 3 for further details.
“Roughly balanced” was defined as no more than 55% of a single gender (including any missing data).
One study did not conduct any tests with families (Bates et al., 2006).
Patients with other communicative or social stressors also face similar risks, see e.g. Au-Yeung et al. (2018) on mental illness misdiagnosis with autistic patients.
CAFAS, CANS, CGAS and CGI are provider-only. Ohio and Y-OQ have provider, parent, and child versions.
References
Accomazzo, S., Israel, N., & Romney, S. (2015). Resources, exposure to trauma, and the behavioral health of youth receiving public system services. Journal of Child & Family Studies, 24(11), 3180–3191.
Achenbach, T. M., & Edelbrock, C. S. (1983). Manual for the Child Behavior Checklist and revised child behavior profile. University of Vermont.
AHRQ. (2016). The National Quality Strategy fact sheet (publication 14(17)-M006-EF). Agency for Healthcare Research and Quality.
Alamdari, G., & Kelber, M. S. (2016). The child and adolescent needs and strengths as an outcome measure in community mental health: Factor analysis and a validation of the short form. Community Mental Health Journal, 52(8), 1118–1122.
American Psychiatric Association. (2017). Mental health disparities: American Indians and Alaska Natives. Retrieved from https://www.psychiatry.org/psychiatrists/cultural-competency/education/mental-health-facts
Anderson, R. L., Lyons, J. S., Giles, D. M., Price, J. A., & Estle, G. (2003). Reliability of the Child and Adolescent Needs and Strengths Mental Health (CANS-MH) Scale. Journal of Child & Family Studies, 12(3), 279–289.
Au-Yeung, S. K., Bradley, L., Robertson, A. E., Shaw, R., Baron-Cohen, S., & Cassidy, S. (2018). Experience of mental health diagnosis and perceived misdiagnosis in autistic, possibly autistic and non-autistic adults. Autism, 23(6), 1508–1518.
Bailey, R. K., Blackmon, H. L., & Stevens, F. L. (2009). Major depressive disorder in the African American population: Meeting the challenges of stigma, misdiagnosis, and treatment disparities. Journal of the National Medical Association, 101(11), 1084–1089.
Bailey, R. K., Mokonogho, J., & Kumar, A. (2019). Racial and ethnic differences in depression: Current perspectives. Neuropsychiatric Disease and Treatment, 15, 603–609.
Baker, F. M., & Bell, C. C. (1999). Issues in the psychiatric treatment of African Americans. Psychiatric Services, 50(3), 362–368.
Balottin, U., Ferri, M., Racca, M., Rossi, M., Rossi, G., Beghi, E., Chiappedi, M., & Termine, C. (2014). Psychotherapy versus usual care in pediatric migraine and tension-type headache: A single-blind controlled pilot study. Italian Journal of Pediatrics, 40(1), 6.
Bates, M. P., Furlong, M. J., & Green, J. G. (2006). Are CAFAS subscales and item weights valid? A preliminary investigation of the Child and Adolescent Functional Assessment Scale. Administration and Policy in Mental Health and Mental Health Services Research, 33(6), 682–695.
Baxter, E. E., Alexander, P. C., Kraus, D. R., Bentley, J. H., Boswell, J. F., & Castonguay, L. G. (2016). Concurrent validation of the Treatment Outcome Package (TOP) for children and adolescents. Journal of Child and Family Studies, 25(8), 1–8.
Bird, H. R., Canino, G., Gould, M. S., Ribera, J., Rubio-Stipec, M., Woodbury, M., Huertas-Goldman, S., & Sesman, M. (1987a). Use of the Child Behavior Checklist as a screening instrument for epidemiological research in child psychiatry: Results of a pilot study. Journal of the American Academy of Child & Adolescent Psychiatry, 26(2), 207–213.
Bird, H. R., Canino, G., Rubio-Stipec, M., & Ribera, J. C. (1987b). Further measures of the psychometric properties of the Children’s Global Assessment Scale. Archives of General Psychiatry, 44(9), 821–824.
Boothroyd, R. A., & Armstrong, M. (2010). An examination of the psychometric properties of the Pediatric Symptom Checklist with children enrolled in Medicaid. Journal of Emotional and Behavioral Disorders, 18(2), 113–126.
Bourdon, K. H., Goodman, R., Rae, D. S., Simpson, G., & Koretz, D. S. (2005). The Strengths and Difficulties Questionnaire: U.S. normative data and psychometric properties. Journal of the American Academy of Child and Adolescent Psychiatry, 44(6), 557–564.
Brown, R. A., Worthman, C. M., Costello, E. J., & Erkanli, A. (2006). The Life Trajectory Interview for Youth (LTI-Y): Method development and psychometric properties of an instrument to assess life-course models and achievement. International Journal of Methods in Psychiatric Research, 15(4), 192–206.
Bruns, E. J., Pullmann, M. D., Sather, A., Brinson, R. D., & Ramey, M. (2015). Effectiveness of wraparound versus case management for children and adolescents: Results of a randomized study. Administration and Policy in Mental Health and Mental Health Services Research, 42(3), 309–322.
Burlingame, G. M., Mosier, J. I., Wells, M. G., Atkin, Q. G., Lambert, M. J., Whoolery, M., & Latkowski, M. (2001). Tracking the influence of mental health treatment: The development of the Youth Outcome Questionnaire. Clinical Psychology & Psychotherapy, 8(5), 361–379.
California Legislative Analyst’s Office. (2011). Mental health realignment. Retrieved January 16, 2017, from http://www.lao.ca.gov/handouts/Health/2011/Mental_Health_1_26_11.pdf
California Senate Bill No. 1009: Health and Human Services (2012). Retrieved from http://leginfo.legislature.ca.gov/faces/billNavClient.xhtml?bill_id=201120120SB1009
Cannon, J. A. N., Warren, J. S., Nelson, P. L., & Burlingame, G. M. (2010). Change trajectories for the Youth Outcome Questionnaire self-report: Identifying youth at risk for treatment failure. Journal of Clinical Child and Adolescent Psychology, 39(3), 289–301.
Cantos, A. L., & Gries, L. T. (2010). Therapy outcome with children in foster care: A longitudinal study. Child and Adolescent Social Work Journal, 27(2), 133–149.
Cénat, J. M. (2020). How to provide anti-racist mental health care. The Lancet, 7(11), 929–931.
Choate, P. W., & McKenzie, A. (2015). Psychometrics in parenting capacity assessments—A problem for First Nations parents. First Peoples Child & Family Review, 10(2), 31–43.
Clark, T. C., Johnson, E. A., Kekus, M., Newman, J., Patel, P. S., Fleming, T., & Robinson, E. (2014). Facilitating access to effective and appropriate care for youth with mild to moderate mental health concerns in New Zealand. Journal of Child & Adolescent Psychiatric Nursing, 27(4), 190–200.
Clarke, G., McGlinchey, E. L., Hein, K., Gullion, C. M., Dickerson, J. F., Leo, M. C., & Harvey, A. G. (2015). Cognitive-behavioral treatment of insomnia and depression in adolescents: A pilot randomized trial. Behaviour Research and Therapy, 69, 111–118.
Cohen, J. A., Mannarino, A. P., & Iyengar, S. (2011). Community treatment of posttraumatic stress disorder for children exposed to intimate partner violence: A randomized controlled trial. Archives of Pediatric and Adolescent Medicine, 165(1), 16–21.
Cook, M. N., Crisostomo, P. S., Simpson, T. S., Williams, J. D., & Wamboldt, M. Z. (2014). Effectiveness of an intensive outpatient program for disruptive children: Initial findings. Community Mental Health Journal, 50(2), 164–171.
Coren, E., Thomae, M., Hutchfield, J., & Iredale, W. (2013). Report on the implementation and results of an outcomes-focused evaluation of child sexual abuse interventions in the UK. Child Abuse Review, 22(1), 44–59.
Creswell, C., Hentges, F., Parkinson, M., Sheffield, P., Willetts, L., & Cooper, P. (2010). Feasibility of guided cognitive behaviour therapy (CBT) self-help for childhood anxiety disorders in primary care. Mental Health in Family Medicine, 7(1), 49–57.
De Souza, M. A. M., Salum, G. A., Jarros, R. B., Isolan, L., Davis, R., Knijnik, D., Manfro, G. G., & Heldt, E. (2013). Cognitive-behavioral group therapy for youths with anxiety disorders in the community: Effectiveness in low and middle income countries. Behavioral & Cognitive Psychotherapy, 41(3), 255–264.
Dedrick, R. F., Greenbaum, P. E., Friedman, R. M., Wetherington, C. M., & Knoff, H. M. (1997). Testing the structure of the Child Behavior Checklist/4-18 using confirmatory factor analysis. Education and Psychological Measurement, 57(2), 306–313.
Deutz, M. H. F., Shi, Q., Vossen, H. G. M., Huijding, J., Prinzie, P., Dekovic, M., van Baar, A. L., & Woltering, S. (2018). Evaluation of the Strengths and Difficulties Questionnaire-Dysregulation Profile (SDQ-DP). Psychological Assessment, 30(9), 1174–1185.
DHCS. (2015). Legislative report: Performance outcomes system plan for Medi-Cal Specialty Mental Health Services for children & youth. California Department of Health Care Services.
Dickey, W. C., & Blumberg, S. J. (2004). Revisiting the factor structure of the Strengths and Difficulties Questionnaire: United States, 2001. Journal of the American Academy of Child and Adolescent Psychiatry, 43(9), 1159–1167.
Dorsey, S., Pullmann, M. D., Berliner, L., Koschmann, E., McKay, M., & Deblinger, E. (2014). Engaging foster parents in treatment: A randomized trial of supplementing trauma-focused cognitive behavioral therapy with evidence-based engagement strategies. Child Abuse and Neglect, 38(9), 1508–1520.
Dour, H. J., Chorpita, B. F., Lee, S., & Weisz, J. R. (2013). Sudden gains as a long-term predictor of treatment improvement among children in community mental health organizations. Behaviour Research and Therapy, 51(9), 564–572.
Dowell, K. A., & Ogles, B. M. (2008). The Ohio Scales Youth Form: Expansion and validation of a self-report outcome measure for young children. Journal of Child & Family Studies, 17(3), 291–305.
Downs, A., Strand, P. S., Heinrichs, N., & Cerna, S. (2012). Use of the teacher version of the Strengths and Difficulties Questionnaire with German and American preschoolers. Early Education and Development, 23(4), 493–516.
Duffy, F., & Skeldon, J. (2014). A CAMHS intensive treatment service: Clinical outcomes in the first year. Clinical Child Psychology, 19(1), 90–99.
Dunleavy, A. M., & Leon, S. C. (2011). Predictors for resolution of antisocial behavior among foster care youth receiving community-based services. Children & Youth Services Review, 33(11), 2347–2354.
Dunn, T. W., Burlingame, G. M., Walbridge, M., Smith, J., & Crum, M. J. (2005). Outcome assessment for children and adolescents: Psychometric validation of the Youth Outcome Questionnaire 30.1 (Y-OQ®-30.1). Clinical Psychology & Psychotherapy, 12(5), 388–401.
Dura-Vila, G., Klasen, H., Makatini, Z., Rahimi, Z., & Hodes, M. (2013). Mental health problems of young refugees: Duration of settlement, risk factors and community-based interventions. Clinical Child Psychology and Psychiatry, 18(4), 604–623.
Dutra, L., Campbell, L., & Westen, D. (2004). Quantifying clinical judgment in the assessment of adolescent psychopathology: Reliability, validity, and factor structure of the Child Behavior Checklist for clinician report. Journal of Clinical Psychology, 60(1), 65–85.
Ebesutani, C., Bernstein, A., Martinez, J. I., Chorpita, B. F., & Weisz, J. R. (2011). The Youth Self Report: Applicability and validity across younger and older youths. Journal of Clinical Child & Adolescent Psychology, 40(2), 338–346.
Ebesutani, C., Bernstein, A., Nakamura, B. J., Chorpita, B. F., Higa-Mcmillan, C. K., & Weisz, J. R. (2010). Concurrent validity of the Child Behavior Checklist DSM-Oriented Scales: Correspondence with DSM diagnoses and comparison to syndrome scales. Journal of Psychopathological and Behavioral Assessment, 32(3), 373–384.
Elderkin-Thompson, V., Silver, R. C., & Waitzkin, H. (2001). When nurses double as interpreters: A study of Spanish-speaking patients in a US primary care setting. Social Sciences and Medicine, 52(9), 1343–1358.
Epstein, J., Santo, R. M., & Guillemin, F. (2015). A review of guidelines for cross-cultural adaptation of questionnaires could not bring out a consensus. Journal of Clinical Epidemiology, 68(4), 435–441.
Eslinger, J. G., Sprang, G., & Otis, M. (2015). Children with multi-trauma histories: Special considerations for care and implications for treatment selection. Journal of Child & Family Studies, 24(9), 2757–2768.
Flores, G., Rabke-Verani, J. B., Pine, W. B., & Sabharwal, A. (2002). The importance of cultural and linguistic issues in the emergency care of children. Pediatric Emergency Care, 18(4), 271–284.
Foa, E. B., McLean, C. P., Capaldi, S., & Rosenfield, D. (2013). Prolonged exposure vs supportive counseling for sexual abuse-related PTSD in adolescent girls: A randomized clinical trial. Original Investigation, 310(24), 2650–2657.
Foreman, D. M., & Morton, S. (2011). Nurse-delivered and doctor-delivered care in an Attention Deficit Hyperactivity Disorder follow-up clinic: A comparative study using propensity score matching. Journal of Advanced Nursing, 67(6), 1341–1348.
Foy, J. M., Green, C. M., Earls, M. F., AAP Committee On Psychosocial Aspects of Child and Family Health, & Mental Health Leadership Work Group. (2019). Mental health competencies for pediatric practice. Pediatrics, 144(5), e20192757.
Francis, S. E., Ebesutani, C., & Chorpita, B. F. (2012). Differences in levels of functional impairment and rates of serious emotional disturbance between youth with internalizing and externalizing disorders when using the CAFAS or GAF to assess functional impairment. Journal of Emotional and Behavioral Disorders, 20(4), 226–240.
Gardner, W., & Kelleher, K. J. (2017). Core quality and outcome measures for pediatric health. JAMA Pediatrics, 171(9), 827–828.
Gardner, W., Lucas, A., Kolko, D., & Campo, J. V. (2007). Comparison of the PSC-17 and alternative mental health screens in an at-risk primary care sample. Journal of the American Academy of Child & Adolescent Psychiatry, 46(5), 611–618.
Gasquoine, P. G. (2009). Race-norming of neuropsychological tests. Neuropsychology Review, 19, 250–262.
Glazier, K., Swing, M., & McGinn, L. K. (2015). Half of obsessive-compulsive disorder cases misdiagnosed: Vignette-based survey of primary care physicians. Journal of Clinical Psychiatry, 76(6), e761–e767.
Glied, S. A., Stein, B. D., McGuire, T. G., Beale, R. R., Duffy, F. F., Shugarman, S., & Goldman, H. H. (2015). Measuring performance in psychiatry: A call to action. Psychiatric Services, 66, 872–878.
Goodman, R. (1997). The Strengths and Difficulties Questionnaire: A research note. Journal of Child Psychology & Psychiatry, 38(5), 581–586.
Green, B., Shirk, S., Hanze, D., & Wanstrath, J. (1994). The Children’s Global Assessment Scale in clinical practice: An empirical evaluation. Journal of the American Academy of Child & Adolescent Psychiatry, 33(8), 1158–1164.
Grip, K. K., Almqvist, K., Axberg, U., & Broberg, A. G. (2013). Children exposed to intimate partner violence and the reported effects of psychosocial interventions. Violence and Victims, 28(4), 635–655.
Grip, K., Almqvist, K., & Broberg, A. G. (2012). Maternal report on child outcome after a community-based program following intimate partner violence. Nordic Journal of Psychiatry, 66(4), 239–247.
Guy, W. (1976). CGI: Clinical global impressions. In ECDEU assessment manual for psychopharmacology (revised) (pp. 218–222). US Department of Health, Education and Welfare.
Ha, E. H., Lee, S. J., Oh, K. J., & Hong, K. E. (1998). Parent-adolescent agreement in the assessment of behavior problems of adolescents: Comparison of factor structures of K-CBCL and YSR. Journal of the Korean Academy of Child and Adolescent Psychiatry, 9(1), 3–12.
Hansen, B., Howe, A., Sutton, P., & Ronan, K. (2015). Impact of client feedback on clinical outcomes for young people using public mental health services: A pilot study. Psychiatry Research, 229(1–2), 617–619.
Haroz, E. E., Ritchey, M., Bass, J. K., Kohrt, B. A., Augustinavicius, J., Michalopoulos, L., Burkey, M. D., & Bolton, P. (2017). How is depression experienced around the world? A systematic review of qualitative literature. Social Science & Medicine, 183, 151–162.
He, J.-P., Burstein, M., Schmitz, A., & Merikangas, K. R. (2013). The Strengths and Difficulties Questionnaire (SDQ): The factor structure and scale validation in U.S. Adolescents. Journal of Abnormal Child Psychology, 41(4), 583–595.
Heath, S. B. (1982). What no bedtime story means: Narrative skills at home and school. Language in Society, 11, 49–76.
Heiervang, E., Goodman, A., & Goodman, R. (2008). The Nordic advantage in child mental health: Separating health differences from reporting style in a cross-cultural comparison of psychopathology. Journal of Child Psychology and Psychiatry, 49(6), 678–685.
Hill, C. R., & Hughes, J. N. (2007). An examination of the convergent and discriminant validity of the Strengths and Difficulties Questionnaire. School Psychology Quarterly, 22(3), 380–406.
Hodges, K. (no date). Child and Adolescent Functional Assessment Scale (CAFAS): Overview of reliability and validity. Retrieved May 27, 2018, from http://www2.fasoutcomes.com/RadControls/Editor/FileManager/Document/FAS611_CAFAS%20Reliability%20and%20Validity%20Rev10.pdf
Hodges, K., & Wong, M. M. (1996). Psychometric characteristics of a multidimensional measure to assess impairment: The Child and Adolescent Functional Assessment Scale. Journal of Child & Family Studies, 5(4), 445–467.
Hodgkinson, S., Godoy, L., Beers, L. S., & Lewin, A. (2017). Improving mental health access for low-income children and families in the primary care setting. Pediatrics, 139(1), e20151175.
Hoffman, K. M., Trawalter, S., Axt, J. R., & Oliver, M. N. (2016). Racial bias in pain assessment and treatment recommendations, and false beliefs about biological differences between blacks and whites. PNAS, 113(16), 4296–4301.
Hogez, R. D., & McKay, V. (1986). Criterion-related validity data for the Child Behavior Checklist-Teacher’s Report Form. Journal of School Psychology, 24(4), 387–393.
Jacobson, J. H., Pullmann, M. D., Parker, E. M., & Kerns, S. E. U. (2019). Measurement based care in child welfare-involved children and youth: Reliability and validity of the PSC-17. Child Psychiatry & Human Development, 50(2), 332–345.
Jani, J. S., & Deforge, B. R. (2014). Contextually appropriate measurement as the basis for culturally appropriate interventions: A case study in Managua, Nicaragua. Social Work in Public Health, 30(2), 157–174.
Jastrowski Mano, K. E., Davies, H. W., Klein-Tasman, B. P., & Adesso, V. J. (2009). Measurement equivalence of the Child Behavior Checklist among parents of African American adolescents. Journal of Child & Family Studies, 18(5), 606–620.
Jee, S. H., Szilagyi, M., Conn, A., Nilsen, W., Toth, S., Baldwin, C. D., & Szilagyi, P. G. (2011). Validating office-based screening for psychosocial strengths and difficulties among youths in foster care. Pediatrics, 127(5), 904–910.
Jellinek, M., Little, M., Murphy, J. M., & Pagano, M. (1995). The Pediatric Symptom Checklist: Support for a role in a managed care environment. Archives of Pediatrics & Adolescent Medicine, 149(7), 740–746.
Jellinek, M. S., Murphy, J. M., & Burns, B. J. (1986). Brief psychosocial screening in outpatient pediatric practice. Journal of Pediatrics, 109, 371–378.
Jellinek, M. S., Murphy, J. M., Robinson, J., Feins, A., Lamb, S., & Fenton, T. (1988). Pediatric Symptom Checklist: Screening school-age children for psychosocial dysfunction. Journal of Pediatrics, 112(2), 201–209.
Jensen, T. K., Holt, T., Ormhaug, S. M., Egeland, K., Granly, L., Hoaas, L. C., Hukkelberg, S. S., Indregard, T., Stormyren, S. D., & Wentzel-Larsen, T. (2014). A randomized effectiveness study comparing trauma-focused cognitive behavioral therapy with therapy as usual for youth. Journal of Clinical Child and Adolescent Psychology, 43(3), 356–369.
Jensen, P. S., Salzberg, A. D., Richters, J. E., & Watanabe, H. K. (1993). Scales, diagnoses, and child psychopathology: I. CBCL and DISC relationships. Journal of the American Academy of Child & Adolescent Psychiatry, 32(2), 397–406.
Jensen, P. S., Watanabe, H. K., Richters, J. E., Roper, M., Hibbs, E. D., Salzberg, A. D., & Liu, S. (1996). Scales, diagnoses, and child psychopathology: II. Comparing the CBCL and the DISC against external validators. Journal of Abnormal Child Psychology, 24(2), 151–168.
Jutte, D. P., Burgos, A., Mendoza, F., Ford, C. B., & Huffman, L. C. (2003). Use of the Pediatric Symptom Checklist in a low-income, Mexican American population. Archives of Pediatrics & Adolescent Medicine, 157(12), 1169–1176.
Kamin, H. S., McCarthy, A. E., Abel, M. R., Jellinek, M. S., Baer, L., & Murphy, J. M. (2015). Using a brief parent-report measure to track outcomes for children and teens with internalizing disorders. Child Psychiatry & Human Development, 46(6), 851–862.
Karpenko, V., & Owens, J. S. (2013). Adolescent psychotherapy outcomes in community mental health: How do symptoms align with target complaints and perceived change? Community Mental Health Journal, 49(5), 540–552.
Kisiel, C., Patterson, N., Torgersen, E., den Dunnen, W., Villa, C., & Fehrenbach, T. (2018). Assessment of the complex effects of trauma across child serving settings: Measurement properties of the CANS-Trauma Comprehensive. Children and Youth Services Review, 86(41), 64–75.
Knepley, M. J., Kendall, P. C., & Carper, M. M. (2019). An analysis of the Child Behavior Checklist Anxiety Problems Scale’s predictive capabilities. Journal of Psychopathology and Behavioral Assessment, 41, 249–256.
Konold, T. R., Walthall, J. C., & Planta, R. C. (2004). The behavior of child behavior ratings: Measurement structure of the Child Behavior Checklist across time, informants, and child gender. Behavioral Disorders, 29(4), 372–383.
Kostanecka, A., Power, T., Clarke, A., Watkins, M., Hausman, C. L., & Blum, N. J. (2008). Behavioral health screening in urban primary care settings: Construct validity of the PSC-17. Journal of Developmental and Behavioral Pediatrics, 29(2), 124–128.
Kovacs, S., & Sharp, C. (2014). Criterion validity of the Strengths and Difficulties Questionnaire (SDQ) with inpatient adolescents. Psychiatry Research, 219(3), 651–657.
Kraus, D. R., Seligman, D. A., & Jordan, J. R. (2005). Validation of a behavioral health treatment outcome and assessment tool designed for naturalistic settings: The Treatment Outcome Package. Journal of Clinical Psychology, 61, 285–315.
Krieger, N., & Fee, E. (1994). Social class: The missing link in U.S. health data. International Journal of Health Services, 24(1), 25–44.
Lambert, M. C., Rowan, G. T., Lyubansky, M., & Russ, C. M. (2002). Do problems of clinic-referred African-American children overlap with the Child Behavior Checklist? Journal of Child & Family Studies, 11(3), 271–285.
Lee, J. W., Jones, P. S., Mineyama, Y., & Zhang, X. E. (2002). Cultural differences in responses to a Likert scale. Research in Nursing & Health, 25, 295–306.
Leiner, M. A., Balcazar, H., Straus, D. C., Shirsat, P., & Handal, G. (2007). Screening Mexicans for psychosocial and behavioral problems during pediatric consultation. Revista de Investigación Clínica, 59(2), 116–123.
Leplège, A., Ecosse, E., Verdier, A., & Perneger, T. V. (1998). The French SF-36 Health Survey: Translation, cultural adaptation and preliminary psychometric evaluation. Journal of Clinical Epidemiology, 51, 1013–1023.
Liber, J. M., Van Widenfelt, B. M., van der Leeden, A. J. M., Goedhart, A. W., Utens, E. M. W. J., & Treffers, P. D. A. (2010). The relation of severity and comorbidity to treatment outcome with cognitive behavioral therapy for childhood anxiety disorders. Journal of Abnormal Child Psychology, 38(5), 683–694.
Liotta, L., Springer, C., Misurell, J. R., Block-Lerner, J., & Brandwein, D. (2015). Interventions for child sexual abuse group treatment for child sexual abuse: Treatment referral and therapeutic outcomes. Journal of Child Sexual Abuse, 24(3), 217–237.
Lundh, A., Forsman, M., Serlachius, E., Lichtenstein, P., & Landén, M. (2013). Outcomes of child psychiatric treatment. Acta Psychiatrica Scandinavica, 128(1), 34–44.
Lyons, J. S. (1999). The child and adolescent needs and strengths for children with mental health challenges and their families. Northwestern University.
Malgady, R. G., Rogler, L. H., & Costantino, G. (1987). Ethnocultural and linguistic bias in mental health evaluation of Hispanics. Professional Psychology: Research and Practice, 27(1), 73–77.
Mason, W. A., Chmelka, M. B., & Thompson, R. W. (2012). Responsiveness of the Strengths and Difficulties Questionnaire (SDQ) in a sample of high-risk youth in residential treatment. Child & Youth Care Forum, 41(5), 479–492.
McCrae, J. S., Barth, R. P., & Guo, S. (2010). Changes in maltreated children’s emotional-behavioral problems following typically provided mental health services. American Journal of Orthopsychiatry, 80(3), 350–361.
McGregor, D. P., Minerbi, L., & Matsuoka, J. (1998). A holistic assessment method of health and well-being for native Hawaiian communities. Pacific Health Dialog, 5(2), 361–369.
McGuire, J. F., Geller, D. A., Murphy, T. K., Small, B. J., Unger, A., Wilhelm, S., & Storch, E. A. (2019). Defining treatment outcomes in pediatric obsessive-compulsive disorder using a self-report scale. Behavior Therapy, 50(2), 314–324.
McQuaide, S. (1999). A social worker’s use of the diagnostic and statistical manual. Families in Society: The Journal of Contemporary Social Services, 80(4), 410–416.
Misurell, J., Springer, C., Acosta, L., Liotta, L., & Kranzler, A. (2014). Game-based cognitive-behavioral therapy individual model (GB-CBT-IM) for child sexual abuse: A preliminary outcome study. Psychological Trauma: Theory, Research, Practice, and Policy, 6(3), 250–258.
Mittler, S., Horesh, N., Maytal, H. R., & Toren, P. (2014). Impact of environmental and personality factors upon adolescents before and after psychotherapeutic intervention. Comprehensive Psychiatry, 55(8), 1791–1802.
Mueller, C. W., Tolman, R., Higa-Mcmillan, C. K., & Daleiden, E. L. (2010). Longitudinal predictors of youth functional improvement in a public mental health system. Journal of Behavioral Health Sciences & Research, 37(3), 350–362.
Murphy, J. M., Bergmann, P., Chiang, C., Sturner, R., Howard, B., Abel, M. R., & Jellinek, M. (2016). The PSC-17: Subscale scores, reliability, and factor structure in a new national sample. Pediatrics, 138(3), e20160038.
Murphy, J. M., Blais, M., Baer, L., McCarthy, A., Kamin, H., Masek, B., & Jellinek, M. (2015). Measuring outcomes in outpatient child psychiatry: Reliable improvement, deterioration, and clinically significant improvement. Clinical Child Psychology & Psychiatry, 20(1), 39–52.
Murphy, J. M., Ichinose, C., Hicks, R. C., Kingdon, D., Crist-Whitzel, J., Jordan, P., Feldman, G., & Jellinek, M. S. (1996). Utility of the Pediatric Symptom Checklist as a psychosocial screen to meet the federal Early and Periodic Screening, Diagnosis, and Treatment (EPSDT) standards: A pilot study. The Journal of Pediatrics, 129(6), 864–869.
Murphy, J. M., & Jellinek, M. (1988). Screening for psychosocial dysfunction in economically disadvantaged and minority-group children: Further validation of the Pediatric Symptom Checklist. American Journal of Orthopsychiatry, 58(3), 450–456.
Murphy, J. M., Jellinek, M., & Milinsky, S. (1989). The Pediatric Symptom Checklist: Validation in the real world of middle school. Journal of Pediatric Psychology, 14(4), 629–639.
Murphy, J. M., Kamin, H., Masek, B., Vogeli, C., Caggiano, R., Sklar, K., Drubner, S., Buonopane, R., Picciotto, M., Gold, J., & Jellinek, M. (2012). Using brief clinician and parent measures to track outcomes in outpatient child psychiatry: Longer term follow-up and comparative effectiveness. Child and Adolescent Mental Health, 17(4), 222–230.
Murphy, J. M., Pagano, M. E., Ramirez, A., Anaya, A. A. Y., Nowlin, C., & Jellinek, M. (1999). Validation of the Preschool and Early Childhood Functional Assessment Scale (PECFAS). Journal of Child & Family Studies, 8(3), 343–356.
Murphy, J. M., Reede, J., Jellinek, M. S., & Bishop, S. J. (1992). Screening for psychosocial dysfunction in inner-city children: Further validation of the Pediatric Symptom Checklist. Journal of the American Academy of Child & Adolescent Psychiatry, 31(6), 1105–1111.
Nakamura, B. J., Ebesutani, C., Bernstein, A., & Chorpita, B. F. (2009). A psychometric analysis of the Child Behavior Checklist DSM-oriented scales. Journal of Psychopathology and Behavioral Assessment, 31(3), 178–189.
Navon, M., Nelson, D., Pagano, M., & Murphy, M. (2001). Use of the Pediatric Symptom Checklist in strategies to improve preventive behavioral health care. Psychiatric Services, 52(6), 800–804.
Nelson, J. R., Benner, G. J., Reid, R. C., Epstein, M. H., & Currin, D. (2002). The convergent validity of office discipline referrals with the CBCL-TRF. Journal of Emotional and Behavioral Disorders, 10(3), 181–188.
Nilsen, T. S., Handegård, B.-H., Eisemann, M., & Kvernmo, S. (2015). Evaluating change in symptomatic and functional level of children and youth with emotional disorders: A naturalistic observation study. European Child & Adolescent Psychiatry, 24(10), 1219–1231.
O’Donnell, K., Dorsey, S., Gong, W., Ostermann, J., Whetten, R., Cohen, J. A., Itemba, D., Manongi, R., & Whetten, K. (2014). Treating maladaptive grief and posttraumatic stress symptoms in orphaned children in Tanzania: Group-based trauma-focused cognitive-behavioral therapy. Journal of Traumatic Stress, 27(6), 664–671.
Ogles, B. M., Melendez, G., Davis, D. C., & Lunnen, K. M. (2001). The Ohio Scales: Practical outcome assessment. Journal of Child & Family Studies, 10, 199–212.
Okamura, K. H., Ebesutani, C., Bloom, R., Higa-McMillan, C. K., Nakamura, B. J., & Chorpita, B. F. (2016). Differences in internalizing symptoms across specific ethnic minority groups: An analysis across Chinese American, Filipino American, Japanese American, Native Hawaiian, and White youth. Journal of Child and Family Studies, 25(11), 3353–3366.
Overbeek, M. M., de Schipper, J. C., Lamers-Winkelman, F., & Schuengel, C. (2014). Risk factors as moderators of recovery during and after interventions for children exposed to interparental violence. American Journal of Orthopsychiatry, 84(3), 295–306.
Owens, J. S., Holdaway, A. S., Serrano, V. J., Himawan, L. K., Watabe, Y., Storer, J., Grant, M., & Andrews, N. (2016). Screening for social, emotional, and behavioral problems at kindergarten entry: The utility of two teacher rating scales. School Mental Health, 8(3), 319–331.
Pagano, M. E., Cassidy, L. J., Little, M., Murphy, M., & Jellinek, M. S. (2000). Identifying psychosocial dysfunction in school-age children: The Pediatric Symptom Checklist as a self-report measure. Journal of School Psychology, 37(2), 91–106.
Pagano, M. E., Murphy, J. M., Pedersen, M., Mosbacher, D., Crist-Whitzel, J., Jordan, P., Rodas, C., & Jellinek, M. S. (1996). Screening for psychosocial problems in 4–5-year-olds during routine EPSDT examinations: Validity and reliability in a Mexican-American sample. Clinical Pediatrics, 35(3), 139–146.
Painter, K. (2012). Outcomes for youth with severe emotional disturbance: A repeated measures longitudinal study of a wraparound approach of service delivery in systems of care. Child and Youth Care Forum, 41(4), 407–425.
Palma, S. M. M., Natale, A. C. M. P., & Calil, H. M. (2015). A 4-year follow-up study of attention-deficit hyperactivity symptoms, comorbidities, and psychostimulant use in a Brazilian sample of children and adolescents with Attention-Deficit/Hyperactivity Disorder. Frontiers in Psychiatry, 6, 135.
Parker, E. M., Jacobson, J., Pullmann, M. D. A., & Kerns, S. E. U. (2019). Identifying psychosocial problems among children and youth in the child welfare system using the PSC-17: Exploring convergent and discriminant validity with multiple informants. Child Psychiatry & Human Development, 50(1), 108–120.
Patel, M. M., Brown, J. D., Croake, S., Lewis, R., Liu, J., Patton, L., Potter, D. E. B., & Scholle, S. H. (2015). The current state of behavioral health quality measures: Where are the gaps? Psychiatric Services, 66(8), 865–871.
Patient Protection and Affordable Care Act, 42 U.S.C. § 18001 et seq. (2010). Retrieved January 6, 2017, from https://www.gpo.gov/fdsys/pkg/PLAW-111publ148/pdf/PLAW-111publ148.pdf
Perou, R., Bitsko, R. H., Blumberg, S. J., Pastor, P., Ghandour, R. M., Gfroerer, J. C., Hedden, S. L., Crosby, A. E., Visser, S. N., Schieve, L. A., Parks, S. E., Hall, J. E., Brody, D., Simile, C., Thompson, W. W., Baio, J., Avenevoli, S., Kogan, M. D., & Huang, L. N. (2013). Mental health surveillance among children: United States, 2005–2011. Morbidity and Mortality Weekly Report, 62(2), 1–35.
Pincus, H. A. (2012). Quality measures: Necessary but not sufficient. Psychiatric Services, 63(6), 523.
Pourat, N., Zima, B., Lee, C., & Marti, F. A. (2016a). California Child Mental Health Performance Outcomes System: County and Provider Survey Results. Los Angeles, CA: UCLA Center for Health Policy Research.
Pourat, N., Zima, B., Lee, C., & Marti, F. A. (2016b). California Child Mental Health Performance Outcomes System: National Review of Tools. Los Angeles, CA: UCLA Center for Health Policy Research.
Pourat, N., Zima, B., Marti, F. A., & Lee, C. (2016c). Standardized Outcome Measures: Basic Information, Literature Scan, Psychometric Properties. Los Angeles, CA: UCLA Center for Health Policy Research.
Pourat, N., Zima, B., Marti, F. A., & Lee, C. (2017). California Child Mental Health Performance Outcomes System: Recommendation Report. Los Angeles, CA: UCLA Center for Health Policy Research. http://healthpolicy.ucla.edu/publications/search/pages/detail.aspx?PubID=1660. Accessed 30 Aug 2021.
Probst, J. C., Laditka, S. B., Moore, C. G., Harun, N., & Powell, M. P. (2007). Race and ethnicity differences in reporting of depressive symptoms. Administration and Policy in Mental Health, 34, 519–529.
Radigan, M., & Wang, R. (2013). Relationships between youth and caregiver strengths and mental health outcomes in community based public mental health services. Community Mental Health Journal, 49(5), 499–506.
Reed, M. L., & Edelbrock, C. (1983). Reliability and validity of the direct observation form of the Child Behavior Checklist. Journal of Abnormal Child Psychology, 11(4), 521–530.
Rescorla, L. A., Ewing, G., Ivanova, M. Y., Aebi, M., Bilenberg, N., Dieleman, G. C., Döpfner, M., Kajokiene, I., Leung, P. W. L., Plück, J., Steinhausen, H.-C., Metzke, C. W., Zukauskiene, R., & Verhulst, F. C. (2017). Parent-adolescent cross-informant agreement in clinically referred samples: Findings from seven societies. Journal of Clinical Child and Adolescent Psychology, 46(1), 74–87.
Reynolds, C., Altmann, R. A., & Allen, D. N. (2021). Chapter 15: The problem of bias in psychological assessment. In Mastering modern psychological testing: Theories and methods, 2nd edition. Springer. Retrieved from https://springerlink.fh-diploma.de/chapter/10.1007/978-3-030-59455-8_15
Ridge, N. W., Warren, J. S., Burlingame, G. M., Wells, M. G., & Tumblin, K. M. (2009). Reliability and validity of the Youth Outcome Questionnaire Self-Report. Journal of Clinical Psychology, 65(10), 1115–1126.
Rishel, C. W., Greeno, C., Marcus, S. C., Shear, M. K., & Anderson, C. (2005). Use of the Child Behavior Checklist as a diagnostic screening tool in community mental health. Research on Social Work Practice, 15(3), 195–203.
Rohe, W. M. (1985). Urban planning and mental health. Journal of Prevention & Intervention in the Community, 4(1–2), 79–110.
Rosanbalm, K. D., Snyder, E. H., Lawrence, C. N., Coleman, K., Frey, J. J., van den Ende, J. B., & Dodge, K. A. (2016). Child wellbeing assessment in child welfare: A review of four measures. Children and Youth Services Review, 68(122), 1–16.
Rothmann, K., Hillmer, J.-M., & Hosser, D. (2014). Evaluation of the musical concentration training with Pepe (MusiKo mit Pepe) for children with attention deficits. Zeitschrift Fur Kinder-Und Jugendpsychiatrie Und Psychotherapie, 42(5), 325–335.
Rubio-Stipec, M., Bird, H., Canino, G., & Gould, M. (1990). The internal consistency and concurrent validity of a Spanish translation of the Child Behavior Checklist. Journal of Abnormal Child Psychology, 18(4), 393–406.
Salcedo, S., Rizvi, S. H., Freeman, L. K., Youngstrom, J. K., Findling, R. L., & Youngstrom, E. A. (2018). Diagnostic efficiency of the CBCL thought problems and DSM-oriented psychotic symptoms scales for pediatric psychotic symptoms. European Child & Adolescent Psychiatry, 27(11), 1491–1498.
Salloum, A., Robst, J., Scheeringa, M. S., Cohen, J. A., Wang, W., Murphy, T. K., Tolin, D. F., & Storch, E. A. (2014). Step one within stepped care trauma-focused cognitive behavioral therapy for young children: A pilot study. Child Psychiatry & Human Development, 45(1), 65–77.
SAMHSA. (2020). The comprehensive community mental health services for children with serious emotional disturbances program: 2017 report to congress [report # PEP20-01-02-001]. U.S. Department of Health and Human Services.
Shaffer, D., Gould, M. S., Brasic, J., Fisher, P., Aluwahlia, S., & Bird, H. (1983). A Children’s Global Assessment Scale (CGAS). Archives of General Psychiatry, 40(11), 1228–1231.
Shapiro, J. P., Youngstrom, J. K., Youngstrom, E. A., & Marcinick, H. F. (2012). Transporting a manualized treatment for children’s disruptive behavior to a community clinic. Journal of Contemporary Psychotherapy, 42(4), 215–225.
Sheldrick, R. C., Benneyan, J. C., Kiss, I. G., Briggs-Gowan, M. J., Copeland, W., & Carter, A. S. (2015). Thresholds and accuracy in screening tools for early detection of psychopathology. Journal of Child Psychology and Psychiatry, 56(9), 936–948.
Sheldrick, R. C., Henson, B. S., Merchant, S., Neger, E. N., Murphy, J. M., & Perrin, E. C. (2012). The Preschool Pediatric Symptom Checklist (PPSC): Development and initial validation of a new social/emotional screening instrument. Academic Pediatrics, 12(5), 456–467.
Sheldrick, R. C., Henson, B. S., Neger, E. N., Merchant, S., Murphy, J. M., & Perrin, E. C. (2013). The Baby Pediatric Symptom Checklist: Development and initial validation of a new social/emotional screening instrument for very young children. Academic Pediatrics, 13(1), 72–80.
Shen, M., Hu, M., & Sun, Z. (2017). Development and validation of brief scales to measure emotional and behavioural problems among Chinese adolescents. BMJ Open, 7(1), e012961.
Shen, H., Zhang, L., Xu, C., Zhu, J., Chen, M., & Fang, Y. (2018). Analysis of misdiagnosis of bipolar disorder in an outpatient setting. Shanghai Archives of Psychiatry, 30(2), 93–101.
Simonian, S. J., & Tarnowski, K. J. (2001). Utility of the Pediatric Symptom Checklist for behavioral screening of disadvantaged children. Child Psychiatry & Human Development, 31(4), 269–278.
Skiba, R. J., Michael, R. S., Nardo, A. C., & Peterson, R. L. (2002). The color of discipline: Sources of racial and gender disproportionality in school punishment. The Urban Review, 34, 317–342.
Snowshoe, A., Crooks, C. V., Tremblay, P. F., Craig, W. M., & Hinson, R. E. (2015). Development of a Cultural Connectedness Scale for First Nations Youth. Psychological Assessment, 27(1), 249–259.
Snowshoe, A., Crooks, C. V., Tremblay, P. F., & Hinson, R. E. (2017). Cultural connectedness and its relation to mental wellness for First Nations Youth. Journal of Primary Prevention, 38(1–2), 67–86.
Song, L., Singh, J., & Singer, M. (1994). The Youth Self-Report inventory: A study of its measurement fidelity. Psychological Assessment, 6(3), 236–245.
Southam-Gerow, M. A., Weisz, J. R., Chu, B. C., McLeod, B. D., Gordis, E. B., & Connor-Smith, J. K. (2010). Does cognitive behavioral therapy for youth anxiety outperform usual care in community clinics? An initial effectiveness test. Journal of the American Academy of Child and Adolescent Psychiatry, 49(10), 1043–1052.
Stefanovics, E. A., Filho, M. V. M., Rosenheck, R. A., & Scivoletto, S. (2014). Functional outcomes of maltreated children and adolescents in a community-based rehabilitation program in Brazil: Six-month improvement and baseline predictors. Child Abuse and Neglect, 38(7), 1231–1237.
Stolk, Y., Kaplan, I., & Szwarc, J. (2017). Review of the strengths and difficulties questionnaire translated into languages spoken by children and adolescents of refugee background. International Journal of Methods in Psychiatric Research, 26, e1568.
Storch, E. A., Salloum, A., King, M. A., Crawford, E. A., Andel, R., McBride, N. M., & Lewin, A. B. (2015). A randomized controlled trial in community mental health centers of computer-assisted cognitive behavioral therapy versus treatment as usual for children with anxiety. Depression & Anxiety, 32(11), 843–852.
Tan, L., & Martin, G. (2015). Taming the adolescent mind: A randomised controlled trial examining clinical efficacy of an adolescent mindfulness-based group programme. Child & Adolescent Mental Health, 20(1), 49–55.
Tharinger, D. J., Laurent, J., & Best, L. R. (1986). Classification of children referred for emotional and behavioral problems: A comparison of PL 94-142 SED criteria, DSM III, and the CBCL system. Journal of School Psychology, 24(2), 111–121.
Thirlwall, K., Cooper, P. J., Karalus, J., Voysey, M., Willetts, L., & Creswell, C. (2013). Treatment of child anxiety disorders via guided parent-delivered cognitive-behavioural therapy: Randomised controlled trial. British Journal of Psychiatry, 203(6), 436–444.
Thyer, B. A. (2015). The DSM-5 definition of mental disorder: Critique and alternatives. In B. Probst (Ed.), Critical thinking in clinical assessment and diagnosis (pp. 45–68). Springer.
Trevelyan, E., Gambino, C., Gryn, T., Larsen, L., Acosta, Y., Grieco, E., Harris, D., & Walters, N. (2016). Characteristics of the U.S. population by generational status: 2013: Current population survey reports. US Census Bureau.
Tsai, M.-H., & Ray, D. C. (2011). Children in therapy: Learning from evaluation of university-based community counseling clinical services. Children & Youth Services Review, 33(6), 901–909.
Tse, Y. J., McCarty, C. A., Stoep, A. V., & Myers, K. M. (2015). Teletherapy delivery of caregiver behavior training for children with attention-deficit hyperactivity disorder. Telemedicine and e-Health, 21(6), 451–458.
Tucker, A. R., Javorski, S., Tracy, J., & Beale, B. (2013). The use of adventure therapy in community-based mental health: Decreases in problem severity among youth clients. Child & Youth Care Forum, 42(2), 155–179.
Tyson, E. H., Teasley, M., & Ryan, S. (2011). Using the Child Behavior Checklist with African American and Caucasian American adopted youth. Journal of Emotional and Behavioral Disorders, 19(1), 17–26.
U.S. Census Bureau. (2016). Selected characteristics of the native and foreign-born populations: 2012–2016 American Community Survey 5-year estimates. Retrieved from https://data.census.gov/cedsci/table?q=ACSST5Y2016.S0501&g=0400000US06&tid=ACSST5Y2016.S0501
US Dept. of Health and Human Services, US Department of Education, & US Department of Justice. (2000). Report of the surgeon general’s conference on children’s mental health: A national action agenda. US Department of Health and Human Services.
US Dept. of Health and Human Services: Office of Minority Health. (2018a). Mental health and American Indians/Alaska Natives. Retrieved March 7, 2019, from https://minorityhealth.hhs.gov/omh/browse.aspx?lvl=4&lvlid=39
US Dept. of Health and Human Services: Office of Minority Health. (2018b). Mental health and Asian Americans. Retrieved March 7, 2019, from https://www.minorityhealth.hhs.gov/omh/browse.aspx?lvl=4&lvlid=54
Vishnevsky, T., Strompolis, M., Reeve, C. L., Kilmer, R. P., & Cook, J. R. (2012). Using latent growth curve modeling to examine changes in mental health outcomes for children enrolled in a system of care. American Journal of Orthopsychiatry, 82(1), 121–128.
Ware, J. E., Jr., Keller, S. D., Gandek, B., Brazier, J. E., & Sullivan, M. (1995). Evaluating translations of health status questionnaires: Methods from the IQOLA Project. International Journal of Technology Assessment in Health Care, 11(3), 525–551.
Warren, J. S., Nelson, P. L., Burlingame, G. M., & Mondragon, S. A. (2012). Predicting patient deterioration in youth mental health services: Community mental health vs. managed care settings. Journal of Clinical Psychology, 68(1), 24–40.
Warren, J. S., Nelson, P. L., Mondragon, S. A., Baldwin, S. A., & Burlingame, G. M. (2010). Youth psychotherapy change trajectories and outcomes in usual care: Community mental health versus managed care settings. Journal of Consulting and Clinical Psychology, 78(2), 144–155.
Warren, J. S., & Salazar, B. C. (2015). Youth self-efficacy domains as predictors of change in routine community mental health services. Psychotherapy Research, 25(5), 583–594.
West, A. E., Weinstein, S. M., Peters, A. T., Katz, A. C., Henry, D. B., Cruz, R. A., & Pavuluri, M. N. (2014). Child- and family-focused cognitive-behavioral therapy for pediatric bipolar disorder: A randomized clinical trial. Journal of the American Academy of Child and Adolescent Psychiatry, 53(11), 1168-1178.e1.
Wolpert, M., Ford, T., Trustam, E., Law, D., Deighton, J., Flannery, H., & Fugard, A. J. B. (2012). Patient-reported outcomes in child and adolescent mental health services (CAMHS): Use of idiographic and standardized measures. Journal of Mental Health, 21(2), 165–173.
Yu, J., Sun, S., & Cheah, C. S. L. (2016). Multitrait-multimethod analysis of the Strengths and Difficulties Questionnaire in young Asian American children. Assessment, 23(5), 603–613.
Zima, B. T., Murphy, J. M., Scholle, S. H., Hoagwood, K. E., Sachdeva, R. C., Mangione-Smith, R., Woods, D., Kamin, H. S., & Jellinek, M. (2013). National quality measures for child mental health care: Background, progress, and next steps. Pediatrics, 131, S38–S49.
Zima, B., Marti, F. A., Lee, C., & Pourat, N. (2019). Selection of a child clinical outcome measure for statewide use in publicly funded outpatient mental health programs. Psychiatric Services, 70(5), 381–388.
Acknowledgements
The authors would like to gratefully acknowledge the strong partnership with the California DHCS Mental Health Services Division and comments from their Subject Matter Experts panel during each stage of this project.
Funding
This study was funded by the California Health Care Foundation in support of the California Department of Health Care Services (CA DHCS).
Author information
Authors and Affiliations
Contributions
All authors contributed to the study conception and design. Literature searches and data analysis were performed by AM. The first draft of the manuscript was written by AM and BZ, and all authors commented on previous versions of the manuscript. All authors read and approved the final manuscript.
Corresponding author
Ethics declarations
Conflict of interest
The authors have no conflicts of interest to declare that are relevant to the content of this article.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Prior presentation: A portion of the data were presented in two reports for the California Department of Health Care Services (DHCS): Standardized Outcome Measures: Basic Information, Literature Scan, Psychometric Properties (Pourat et al., 2016c; not publicly available) and California Child Mental Health Performance Outcomes System: Recommendation Report (Pourat et al., 2017).
Supplementary Information
Below is the link to the electronic supplementary material.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Marti, F.A., Pourat, N., Lee, C. et al. Overview of Ten Child Mental Health Clinical Outcome Measures: Testing of Psychometric Properties with Diverse Client Populations in the U.S.. Adm Policy Ment Health 49, 197–225 (2022). https://doi.org/10.1007/s10488-021-01157-z
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10488-021-01157-z