Predictors of Use of Evidence-Based Practices for Children and Adolescents in Usual Care

11
ORIGINAL ARTICLE Predictors of Use of Evidence-Based Practices for Children and Adolescents in Usual Care Charmaine K. Higa-McMillan Brad J. Nakamura Ashley Morris David S. Jackson Lesley Slavin Ó Springer Science+Business Media New York 2014 Abstract Practice data from 74 therapists providing public mental health services to 519 youth ages 5–19 were exam- ined. Multilevel modeling suggested child and therapist characteristics predicted use of practices derived from the evidence-base (PDEB) and use of practices with minimal evidence support (PMES). Longer episode length predicted greater receipt of PDEB; older youth, males, and youth in out-of-home levels of care were more likely to receive PMES; and youth receiving an evidence-based treatment program were less likely to receive PMES. Professional specialty and theoretical orientation significantly predicted PDEB whereas therapist characteristics did not predict PMES. Implementation implications are discussed. Keywords Usual care Á Evidence-based practices Á Therapists Á Implementation Á Dissemination Á Youth Introduction A growing body of research is dedicated to examining effective ways to disseminate and implement programs with empirical support (e.g., Aarons et al. 2011; Damsch- roder et al. 2009; Proctor et al. 2009). This body of research holds promise to improve mental health services for children and adolescents. A complimentary approach to understanding how to effectively implement new programs is the study of practices already in place in usual care settings (Garland et al. 2010a, b). The examination of usual care practices can provide information about the overlap and discrepancies between programs with demonstrated effectiveness and what is already occurring in usual care. It may also be helpful to identify the predictors of evidence- based practice (EBP) usage in usual care settings so that we can draw on these predictors when developing implemen- tation programs. In one of the first studies to date to compare the prac- tices of usual care clinicians to the evidence-based litera- ture, Garland and colleagues (2010a) found that for youth referred for disruptive behavior problems, clinicians pro- vided an array of treatment strategies, however, all strate- gies were delivered at low intensity. Further, while some strategies that were consistent with the evidence base for disruptive behavior problems (e.g., psychoeducation, problem-solving, positive reinforcement) were delivered frequently, there were also a number of approaches derived from the evidence-base that were rarely observed (e.g., role-play, homework). Consistent with these findings, two recent studies on children receiving public mental health services for traumatic stress and anxiety reported similar findings (Borntrager et al. 2013a; Higa-McMillan et al. 2011, respectively).Both studies found that exposure, the most commonly occurring practice among evidence-based treatment protocols for youth with anxiety and traumatic stress, was reported as being used infrequently. Borntrager et al. (2013a) reported that youth receiving out-of-home services with a diagnosis of post-traumatic stress disorder Parts of this research study were presented in May 2013 at the Seattle Implementation Research Conference in Seattle, WA. C. K. Higa-McMillan (&) Á A. Morris Department of Psychology, University of Hawaii at Hilo, 200 W. Kawili St., Hilo, HI 96720, USA e-mail: [email protected] B. J. Nakamura University of Hawaii at Manoa, Honolulu, HI, USA D. S. Jackson Á L. Slavin Child and Adolescent Mental Health Division, Hawaii Department of Health, Honolulu, HI, USA 123 Adm Policy Ment Health DOI 10.1007/s10488-014-0578-9

Transcript of Predictors of Use of Evidence-Based Practices for Children and Adolescents in Usual Care

ORIGINAL ARTICLE

Predictors of Use of Evidence-Based Practices for Childrenand Adolescents in Usual Care

Charmaine K. Higa-McMillan • Brad J. Nakamura •

Ashley Morris • David S. Jackson • Lesley Slavin

� Springer Science+Business Media New York 2014

Abstract Practice data from 74 therapists providing public

mental health services to 519 youth ages 5–19 were exam-

ined. Multilevel modeling suggested child and therapist

characteristics predicted use of practices derived from the

evidence-base (PDEB) and use of practices with minimal

evidence support (PMES). Longer episode length predicted

greater receipt of PDEB; older youth, males, and youth in

out-of-home levels of care were more likely to receive

PMES; and youth receiving an evidence-based treatment

program were less likely to receive PMES. Professional

specialty and theoretical orientation significantly predicted

PDEB whereas therapist characteristics did not predict

PMES. Implementation implications are discussed.

Keywords Usual care � Evidence-based practices �Therapists � Implementation � Dissemination � Youth

Introduction

A growing body of research is dedicated to examining

effective ways to disseminate and implement programs

with empirical support (e.g., Aarons et al. 2011; Damsch-

roder et al. 2009; Proctor et al. 2009). This body of

research holds promise to improve mental health services

for children and adolescents. A complimentary approach to

understanding how to effectively implement new programs

is the study of practices already in place in usual care

settings (Garland et al. 2010a, b). The examination of usual

care practices can provide information about the overlap

and discrepancies between programs with demonstrated

effectiveness and what is already occurring in usual care. It

may also be helpful to identify the predictors of evidence-

based practice (EBP) usage in usual care settings so that we

can draw on these predictors when developing implemen-

tation programs.

In one of the first studies to date to compare the prac-

tices of usual care clinicians to the evidence-based litera-

ture, Garland and colleagues (2010a) found that for youth

referred for disruptive behavior problems, clinicians pro-

vided an array of treatment strategies, however, all strate-

gies were delivered at low intensity. Further, while some

strategies that were consistent with the evidence base for

disruptive behavior problems (e.g., psychoeducation,

problem-solving, positive reinforcement) were delivered

frequently, there were also a number of approaches derived

from the evidence-base that were rarely observed (e.g.,

role-play, homework). Consistent with these findings, two

recent studies on children receiving public mental health

services for traumatic stress and anxiety reported similar

findings (Borntrager et al. 2013a; Higa-McMillan et al.

2011, respectively).Both studies found that exposure, the

most commonly occurring practice among evidence-based

treatment protocols for youth with anxiety and traumatic

stress, was reported as being used infrequently. Borntrager

et al. (2013a) reported that youth receiving out-of-home

services with a diagnosis of post-traumatic stress disorder

Parts of this research study were presented in May 2013 at the Seattle

Implementation Research Conference in Seattle, WA.

C. K. Higa-McMillan (&) � A. Morris

Department of Psychology, University of Hawaii at Hilo, 200 W.

Kawili St., Hilo, HI 96720, USA

e-mail: [email protected]

B. J. Nakamura

University of Hawaii at Manoa, Honolulu, HI, USA

D. S. Jackson � L. Slavin

Child and Adolescent Mental Health Division, Hawaii

Department of Health, Honolulu, HI, USA

123

Adm Policy Ment Health

DOI 10.1007/s10488-014-0578-9

(PTSD) were less likely than youth without a diagnosis of

PTSD to receive practices derived from the evidence-base

(PDEB). Both studies found that as youth age increased,

clinicians reported greater use of PDEB. On the other hand,

Higa-McMillan et al. (2011) found that youth with a

comorbid diagnosis of a disruptive behavior disorder in

addition to primary anxiety or who were receiving an

evidence-based service that is targeted to reduce DBDs

(i.e., Multisystemic Therapy, Multidimensional Treatment

Foster Care, or Functional Family Therapy) were less

likely to receive PDEB for anxiety.

Equally as important to understanding the practices that

usual care practitioners use (and do not use) are the prac-

titioner characteristics that predict use of EBPs. In a fol-

low-up study of Garland et al.’s (2010a) observational data

of practices in usual care, Brookman-Frazee et al. (2010)

examined child, family and therapist characteristics that

predicted greater use of PDEB. While they found several

child and family characteristics predicted greater use of

PDEB, the only therapist characteristics that predicted

greater use of PDEB were theoretical orientation and

clinical experience. Specifically, they found that therapists

with cognitive-behavioral or behavioral orientations pro-

vided more PDEB than therapists who self-identified as

eclectic or with other orientations and therapists with fewer

months in practice provided more PDEB than those with

more months in practice. Other therapist variables such as

license status, position, and training discipline (i.e., mar-

riage and family therapy, psychology, social work) did not

predict usage patterns in usual care.

In addition to therapist background characteristics, other

therapist- and organizational-level characteristics such as

therapist knowledge of EBPs, therapist attitudes towards

EBPs and organizational support for EBPs may also be

important predictors of EBP implementation in usual care

settings. Indeed, a number of theories in Dissemination and

Implementation Science have identified multiple factors

contributing to the successful implementation of a new

innovation. In particular, Damschroder and colleagues

(2009) proposed the Consolidated Framework for Imple-

mentation Research (CFIR), a comprehensive analysis of

the common factors among existing implementation theo-

ries. The CFIR includes five major domains: the inter-

vention itself, the inner setting, the outer setting, the

process by which implementation is accomplished, and the

individuals involved. With regard to characteristics of

individuals, the CFIR identifies five characteristics com-

mon across the field including knowledge and beliefs about

the intervention, self-efficacy, individual stage of change,

individual identification with organization, and personal

attributes. Most research to date has focused on knowledge

and beliefs about the intervention with less research

examining the latter four variables highlighted in the CFIR.

For example, in a study of 29 school mental health

providers, Stephan and colleagues (2012) examined self-

reported knowledge (none to significant) and frequency of

use (never to frequently) of PDEB for attention deficit-

hyperactivity disorder, disruptive behavior disorders,

depression, and anxiety and compared these to interviewer-

rated quality of treatment for youth depression. They found

that self-rated knowledge scores were significantly corre-

lated with self-reported frequency of use of these practices.

They also found that whereas knowledge of common

practices derived from the evidence base were significantly

correlated with interviewer-rated quality of depression

treatment, frequency of use of these practices was not

significantly correlated with interviewer-rated quality of

treatment for depression.

In another study of 197 children’s mental health care

providers, Jensen-Doss et al. (2009) found that positive

attitudes towards EBPs significantly predicted higher self-

reported usage of such practices and that positive provider

attitudes towards EBPs were predicted by higher ratings of

perceived colleague support, higher ratings of perceived

agency support, higher ratings of the quality of prior EBP

training, and lower ratings of institutional barriers. Simi-

larly, Nelson and Steele (2007) found in a survey of 214

providers across 15 states that positive and negative atti-

tudes mediated the relationship between how open the

clinician perceived their environment to be towards EBPs

and EBP use by the clinician. In other words, training in

EBPs and working in a setting that is open to EBPs may

impact attitudes towards EBPs, thereby increasing EBP

usage. Although these studies provide empirical support for

the theory that positive attitudes towards EBPs predict

greater use of EBPs in practice, they did not examine actual

use of EBPs with youth receiving mental health services

from their provider samples and they did not account for

child characteristics that may likely play a role in the

selection of practices by providers in usual care.

The current investigation attempted to replicate and

extend this body of work by examining predictors of

practices reported by usual care providers in a large

statewide publicly-funded mental health system for chil-

dren and adolescents. Using multilevel modeling (MLM),

this study examined both child- and provider-level vari-

ables that predicted greater use of PDEB as well as prac-

tices minimally found in evidence-based protocols. In

addition to examining child and service characteristics

(age, gender, diagnosis, functional impairment, service

type, and length of service) and provider background

characteristics (years trained, license status, professional

discipline, theoretical orientation), we also examined

characteristics of providers hypothesized to influence

greater use of EBPs including provider knowledge of and

attitudes towards EBPs.

Adm Policy Ment Health

123

Methods

Therapist Participants

Therapist participants were invited to participate at a state

sponsored workshop in EBPs. Of the 397 practitioners who

attended the trainings, 240 (63.3 %) completed one or

more questionnaires from the pre-training survey battery

(see Nakamura et al. 2011a, b). Sixty-three of these prac-

titioners provided mental health services in the schools and

thus their practice data was not available for analysis. Of

the remaining 177 practitioners who were affiliated with

the state of Hawaii’s Child and Adolescent Mental Health

Division (CAMHD), 74 (41.8 %) had at least one practice

measure in the Child and Adolescent Mental Health

Management Information System (CAMHMIS) for a youth

receiving mental health services through CAMHD. Par-

ticipants ranged in age from 25 to 61 (M = 42.4,

SD = 10.8), 64.9 % were female (n = 48), and the pri-

mary ethnicities reported were: White (n = 34; 45.9 %),

Asian (n = 11; 14.9 %), and Hawaiian or Pacific Islander

(n = 5; 6.8 %). Participants reported an average of

6.6 years of clinical training (SD = 5.3), 7.5 years of

clinical experience (SD = 5.5), and 31.1 % (n = 23)

reported holding a state license to practice. Therapist par-

ticipants reported being trained in a variety of different

professional disciplines: marriage and family therapy

(n = 35; 47.3 %), psychology or psychiatry (n = 21,

28.4 %), counseling (n = 12; 16.2 %), social work (n = 5;

6.8 %), and other (n = 1; 1.4 %). When asked about the-

oretical orientation, 86.5 % of participants (n = 64) iden-

tified as Eclectic or selected more than one theoretical

orientation that they identified with and 13.5 % (n = 10)

selected Cognitive-Behavioral Therapy as their primary

theoretical orientation. On average, participants reported

having an active caseload of 7.9 (SD = 4.8) and received

approximately 1.8 h of supervision per week (SD = 1.1).

Youth Participants

Youth were selected for participation if they received

mental health treatment from a contracted provider agency

of the CAMHD between May 2006 and May 2008, had at

least 30 days of available treatment data (mean episode

length = 6.9 months; SD = 5.7 months) and were receiv-

ing treatment for anxiety, attentional, depressive, disruptive

behavior, and/or trauma-related problems. Because youth

were nested within therapists, youth were included only if

their therapist participated in the study. This resulted in 519

youth (64.5 % boys) ages 5 to 19 (M = 14.1; SD = 2.9).

More than half reported a multiracial background

(n = 332; 64.5 %) with other non-mixed races reported at

10.8 % White, 10.0 % Native Hawaiian/Pacific Islander,

5.0 % Asian, and 2.1 % Black/African American. Three-

quarters of youth carried comorbid diagnoses (n = 389;

75.0 %) with primary diagnoses including disruptive

behavior disorders (n = 241; 46.4 %), depressive disorders

(n = 112; 21.6 %), attention-deficit/hyperactivity disorder

(n = 84, 16.2 %), and anxiety or traumatic stress disorders

(n = 82, 15.8 %). Just over half of the youth were

receiving services in in-home settings (n = 300, 57.8 %)

and 22.4 % (n = 116) were receiving services through a

packaged evidenced-based program (i.e., Functional Fam-

ily Therapy, Multisystemic Therapy or Multidimensional

Treatment Foster Care).

Measurement

Child and Adolescent Functional Assessment Scale

(CAFAS; Hodges 1998)

The CAFAS is a 200-item scale that measures a youth’s

level of functional impairment. Based on their experience

with the youth, case managers in the CAMHD review

behavioral descriptions ordered by level of impairment

within eight domains of functioning. The subscales of

school role performance, home role performance, com-

munity role performance, behavior toward others, mood/

emotions, mood/self-harmful behavior, substance use, and

thinking are calculated by scoring the highest level of

impairment (i.e., severe = 30, moderate = 20, mild = 10,

no/minimal = 0) that describes the youth within the

respective domain of items. A total score is calculated by

summing across the eight subscales (i.e., CAFAS Total-8).

The CAFAS has been found to have acceptable internal

consistency across items, inter-rater reliability, stability

across time, and concurrent validity (Hodges and Gust

1995; Hodges and Wong 1996). Of note for the current

study, case managers within the CAMHD are certified for

administration annually. For the purposes of this study, the

CAFAS Total-8 score obtained closest to the treatment

episode start was used to estimate level of functional

impairment at treatment start.

Evidence-Based Practice Attitude Scale (EBPAS; Aarons

2004; Aarons et al. 2010)

The EBPAS is a 15-item well-established measure of cli-

nician attitudes towards EBPs. Participants respond on a

five-point Likert-scale (0 = ‘‘not at all’’ to 4 ‘‘to a very

great extent’’) the extent to which they agree to a particular

statement. Higher mean scores indicate more favorable

attitudes. The EBPAS has demonstrated a four factor

structure and good internal consistency (Aarons 2004;

Aarons et al. 2010). Given that all four scales were highly

correlated in this sample and in order to reduce

Adm Policy Ment Health

123

multicollinearity, we only examined the EBPAS total score

in the MLM. Cronbach’s alpha for the EBPAS total in this

sample was .82, which is slightly higher than that reported

in previous studies (a = .77 in Aarons 2004; a = .76 in

Aarons et al. 2010). The EBPAS total mean in this sample

was 2.93 (SD = .48), which is also slightly higher than that

reported in previous studies (M = 2.30, SD = .45, Aarons

2004; M = 2.33, SD = .45, Aarons et al. 2010).

Knowledge of Evidence Based Services Questionnaire

(KEBSQ; Stumpf et al. 2009)

The KEBSQ is a 40-item measure assessing awareness

knowledge of various evidence-based and non-evidence-

based techniques for youth with anxious/avoidant (A),

depressed/withdrawn (D), disruptive behavior (B), and

attention/hyperactivity (H) problems. Respondents are

asked to circle all problem areas for which a particular type

of practice element is drawn from an evidence-based pro-

tocol. Each individual item is then scored on a scale from

zero to four, with correctly endorsed and omitted responses

per problem area each receiving one point each. As an

example for the present study, exposure has been classified

as being derived from an evidence-based protocol for

anxious/avoidant problems (Chorpita and Daleiden 2007).

In this case, a respondent would get one point for circling

A, one point for not circling D, one point for not circling B,

and one point for not circling H, for a grand total of four

points. In order to differentiate a no-response (e.g., the

participant refused to answer the question) from actively

choosing to indicate that a particular technique is not from

an evidence-based protocol for any of the four problem

areas, participants have the option of circling the letter N

(None) for each item. Total possible scores on the KEBSQ

can range from zero to 160. The KEBSQ has demonstrated

adequate test–retest reliability in a sample of graduate level

and community clinicians (r = .56) and the ability to dis-

criminate between these two populations. Given that the

surveys were administered in 2008 and 2009, the KEBSQ

was scored using definitions of the evidence-base found in

the 2007 CAMHD Biennial Report (Chorpita and Daleiden

2007). Consistent with Stumpf et al.’s (2009) original

study, a technique was considered as derived from an

evidence-based protocol for a particular problem area if

that technique was utilized in 10 % or more of all treatment

protocols receiving Best (Level 1) or Good (Level 2)

Support for that specific problem area. For this study the

KEBSQ total score, which can range from 0 to 160 (40

items ranging from 0 to 4 points per item), was examined.

The KEBSQ total mean score for this sample was 91.3

(SD = 8.8), which was about five points lower than that

reported by Stumpf et al. (2009).

Monthly Treatment and Progress Summary (MTPS;

CAMHD 2008)

The MTPS is a therapist report form designed to measure

treatment targets, clinical progress, and intervention practice

elements. For the purposes of this study, only the practice

elements (PEs; e.g., activity scheduling, catharsis) portion on

the MTPS were examined. On a monthly basis, therapists

identify upto 63 predefined PEs they used in their treatment

with youth clients. The MTPS PEs have demonstrated

acceptable one-month stability estimates (average kap-

pas = .65; Daleiden et al. 2004) and structural validity with

support for three factors corresponding to behavior man-

agement interventions, self-control practices, and family

interventions (Orimoto et al. 2012). Further, in a study

comparing the MTPS PEs and audio recordings from therapy

sessions, 12 of the 63 predefined PEs were determined to be

100 % valid according to clinician-coder agreement utilizing

extensiveness and experiential scales (the other PEs were not

available for coding; Borntrager et al. 2013a, b). As the

MTPS includes PEs that are found in EBPs as well as PEs

that are not included in EBPs, two scores were derived from

the MTPS. Using the CAMHD 2007 Biennial Report

(Chorpita and Daleiden 2007), PDEB were defined as prac-

tices occurring in 30 % or more of level 2 or better treatment

protocols for the problem areas examined in this study (i.e.,

anxiety, attentional problems, depression, disruptive behav-

ior, and traumatic stress). In the CAMHD 2007 Biennial

Report, level 1 treatments are similar to ‘‘well-established’’

treatments as defined by the American Psychological Asso-

ciation Division 12 Task Force for Promotion and Dissemi-

nation of Psychological Procedures (1995) and level 2

treatments are similar to ‘‘probably efficacious’’ treatments

(see Chorpita and Daleiden 2007; Chorpita et al. 2011). This

resulted in 24 PEs (see Table 1) which were summed and

then divided by the total number PEs the youth received over

the course of their entire treatment episode (i.e., PDEB =

Sum of PEs from evidence-based protocols/sum of total

PEs). In order to examine whether therapist characteristics

predicted use of practices with minimal evidence-support

(PMES), four treatment outcome research experts were

consulted to review the remaining list of PEs on the MTPS.

These experts determined that common factors (e.g., sup-

portive listening) and practices that are drawn from evidence-

based protocols for problem areas not included in the study

(e.g., motivational interviewing for substance abuse) should

not included in the PMES score. The remaining PEs were

found to occur in 8 % or less of level 1 or level 2 protocols

for the problem areas in this study. This resulted in ten PEs

(see Table 1) which were summed and then divided by the

total number of PEs the youth received over their treatment

episode (i.e., PMES = Sum of PEs from minimally sup-

ported protocols/sum of total PEs).

Adm Policy Ment Health

123

Therapist Background Questionnaire (TBQ)

The TBQ was developed for this study and assesses basic

demographic information (age, gender, ethnicity/race, ethnic

identity), training and experience information (degrees earned,

state license, professional specialty, theoretical orientation,

years of clinical training, years of clinical experience), and

work setting information (agency name/type, position, clinical

setting, current caseload, hours of supervision per week).

Procedure

The University of Hawai‘i Institutional Review Board

approved this study. Half-day voluntary continuing educa-

tion trainings in EBPs for youth internalizing and external-

izing concerns were held between May 2008 and July 2009

across the state of Hawaii’s four counties. Questionnaires

were administered to attendees prior to workshop partici-

pation. If a practitioner attended more than one training

workshop, his questionnaire from the first training he atten-

ded was utilized for analyses. Prior to any data collection, all

participants underwent standardized Institutional Review

Board-approved notice of privacy and consent procedures.

Youth clients and their legal guardian(s) provided written

informed consent after receiving a complete description of

CAMHD’s notice of privacy and disclosure procedures. The

CAFAS was administered as close to the start of treatment

as possible in conjunction with a diagnostic assessment by a

staff or contracted CAMHD clinician (e.g., staff psycholo-

gist, psychiatrist, community psychologist).

Data Preparation

MTPS scores (PDEB and PMES) as well as predictor vari-

ables from the EBPAS, KEBSQ, and therapist and child

demographics were examined for both statistical outliers and

distribution normality. Standardized z-scores were calcu-

lated for these variables and responses in excess of 3.29

(p \ .001, two-tailed test) were considered outliers (Ta-

bachnick and Fidell 2007). Using this convention, a small

number of outliers were identified for years of training

(n = 1; 1.3 %), years of full-time clinical experience

(n = 2; 2.6 %), typical number of active cases (n = 1;

1.3 %), and episode length (n = 10; 1.9 %).Based on outlier

analysis, skewness, and distribution shape, inverse (reci-

procal) transformations were completed for years training,

years experience, active cases, and episode length. However,

transformations failed to reduce outliers and distributions

remained significantly skewed. As such, scores were chan-

ged for variables with more than one outlier by taking the

next most extreme score and adding one unit (Fields 2009;

Tabachnick and Fidell 2007). For instance, one therapist

identified as having 33 years of training beyond their

undergraduate degree. The next most extreme score was

23 years of training. Thus one unit (1 year) was added to the

second most extreme score (i.e., 23 ? 1 = 24) and this

number replaced the most extreme outlier (i.e., 33).

Missing data were handled using the missing value

analysis module of SPSS 19.0 (SPSS, 2010) and missing

Table 1 MTPS practice use by therapist (N = 74) and youth

(N = 519)

MTPS practices Type of

practice

Therapist

percent use

(N = 74)

Youth

percent

received

(N = 519)

Activity scheduling PDEB 89.2 69.8

Behavioral contracting PDEB 33.8 7.4

Cognitive PDEB 93.2 82.1

Commands PDEB 85.1 79.6

Communication skills PDEB 97.3 91.2

Exposure PDEB 27.0 9.9

Goal setting PDEB 43.2 11.1

Ignoring or DRO PDEB 52.7 24.7

Maintenance/relapse Prevention PDEB 66.2 33.7

Modeling PDEB 87.8 57.8

Parent/teacher monitoring PDEB 82.4 60.1

Parent/teacher praise PDEB 87.8 73.2

Personal safety skills PDEB 16.2 4.3

Problem solving PDEB 97.3 85.8

Psychoeducation child PDEB 79.7 61.9

Psychoeducation parent PDEB 85.1 66.7

Relaxation PDEB 74.3 44.2

Response cost PDEB 51.4 27.0

Self-monitoring PDEB 82.4 51.0

Self-reward/self-praise PDEB 78.4 45.9

Social skills training PDEB 90.5 72.0

Tangible rewards PDEB 86.5 68.3

Therapist praise/rewards PDEB 91.9 76.8

Time out PDEB 67.6 42.4

Catharsis PMES 29.7 13.2

EMDR or tapping PMES 9.5 1.4

Free association PMES 41.9 13.8

Hypnosis PMES 9.5 1.6

Interpretation PMES 40.5 26.1

Line of sight supervision PMES 70.3 44.4

Mentoring PMES 73.0 40.3

Milieu therapy PMES 54.1 31.3

Thought field therapy PMES 10.8 2.9

Twelve step program PMES 35.1 18.5

MTPS monthly treatment and progress summary, PDEB practice

derived from the evidence base, DRO differential reinforcement of

other behavior, PMES practices with minimal evidence support,

EMDR eye movement desensitization and reprocessing

Adm Policy Ment Health

123

data patterns were examined using Little’s missing com-

pletely at random (MCAR) test. Data was missing from the

following variables: years trained (1.9 %), KEBSQ

(3.3 %), EBPAS (3.3 %), and CAFAS (8.1 %) and Little’s

MCAR was significant (v2 = 257.46, df = 52, p \ .001),

suggesting that data was not MCAR. As such, data were

imputed (fully conditional) and all results are reported for

the pooled dataset (average of 20 imputed datasets) except

in instances where multiple imputation procedures did not

yield pooled estimates; for those instances, results from

analyses with non-imputed data are reported.

Data Analytic Strategy

Given the nested nature of the data in this study, MLM using

SPSS version 19 was employed. Youth were defined as the

Level 1 unit of analysis and therapists were defined as the Level

2 unit. Three models were created for each of the two dependent

variables (i.e., PDEB and PMES): (a) null models (i.e., no

predictors models), (b) level 1 models (i.e., child characteris-

tics) and (c) level 2 models (i.e., therapist characteristics).

In order to reduce multicollinearity, highly correlated

predictor variables were examined and one of the pair was

selected for inclusion in final models. In particular, years of

training was selected over years of clinical experience

(r = .59) because the latter variable was more skewed. Also,

supervision hours per week, number of active cases, and

receiving a packaged EBP program were all highly correlated.

Given that we were interested in predicting use of PDEB, the

latter of the three variables was retained in final analyses.

Thus, the final predictor variables in the level 1 models were

(a) child age, (b) child gender, (c) primary diagnosis (anxiety,

trauma, depression, or attention compared to disruptive

behavior), (d) CAFAS Total, (e) receiving services through an

evidence-based program (yes, no), (f) level of care (in-home,

out-of-home), and (g) length of treatment episode. The final

predictor variables in the level 2 models were (a) years trained,

(b) licensed (yes/no), (c) professional specialty (counseling,

marriage and family, other, or psychiatry and psychology

compared to social work), (d) primary theoretical orientation

(Eclectic, CBT), (e) KEBSQ Total, and (f) EBPAS Total.

Alpha was set a priori at .05.

Results

Null Models (No Predictors)

In order to determine if a multilevel could be examined, no

predictors (i.e., null) models were examined. For both

PDEB and PMES, intercepts varied significantly across

therapists (Wald Z Intercept = 3.686, p \ .001 and 4.983,

p \ .001, respectively) and residual parameters varied

significantly within therapists (Wald Z Residual = 14.333,

p \ .001 and 14.845, p \ .001, respectively). Intraclass

correlation coefficients (ICCs) provide estimates of vari-

ance for the dependent variables. The ICCs for PDEB

(.416) and PMES (.553) suggest that 41.6 % of the vari-

ance in PDEB scores and 55.3 % of the variance in PMES

scores is accounted for by youth, service, and therapist

characteristics. Thus, we rejected the null hypothesis and

developed multilevel models to explain the variability in

intercepts between and within therapists.

Level 1 Models (Youth and Service Characteristics)

Results of the level 1 models predicting PDEB and PMES

are presented in Table 2 (significant results are in bold

font). Episode length was the only significant predictor of

PDEB with longer episode lengths predicting greater use of

PDEB (t = 2.874, p = .004). On the other hand, several

youth and service characteristics predicted PMES. Older

youth and males were more likely to receive PMES

(t = 2.769, p = .006 and t = -3.322, p = .001, respec-

tively). Further, youth in out-of-home levels of care were

more likely to receive PMES (t = 6.865, p \ .001)

whereas youth receiving services from an evidence-based

program (i.e., Functional Family Therapy, Multisystemic

Therapy or Multidimensional Treatment Foster Care) were

less likely to receive PMES (t = -2.977, p = .003).

Level 2 Models (Therapist Characteristics)

Results of the level 2 models predicting PDEB and PMES

are presented in Table 3 (significant results are in bold

font). Significant youth and service characteristics from

model 1 were included in model 2 (i.e., episode length for

PDEB and youth age, youth gender, service type, and level

of care for PMES). These characteristics remained signif-

icant in the level 2 models. Therapists who reported having

a Psychology or Psychiatry specialty used significantly

more PDEB than therapists with a Social Work specialty

(t = 2.180, p = .029). No other differences among the

professional specialties were noted. In addition, therapists

who identified with a Behavioral or Cognitive-Behavioral

theoretical orientation used significantly more PDEB than

therapists who identified with an Eclectic theoretical ori-

entation (t = -2.599, p = .009). No significant differ-

ences among therapists were noted for PMES.

Discussion

This research builds on and extends the work to date

examining youth and therapist characteristics that predict

use of PDEB in usual care. Regarding youth and service

Adm Policy Ment Health

123

level characteristics, we found that youth who had longer

episodes of care were more likely to receive PDEB. This

may simply indicate that there was greater opportunity for

use of any practice relative to youth who had shorter

lengths of stay. Related to this, youth in out-of-home levels

of care were more likely to receive PMES. Given that

therapists have greater contact with youth in out-of-home

care, it may be that therapists simply had more contact with

Table 2 Level 1 models predicting PDEB and PMES

Predictor PDEB PMES

b SE t p b SE t p

Youth agea 0.000 0.001 -0.004 0.997 0.002 0.001 2.769 0.006

Youth gendera -0.003 0.009 -0.353 0.724 -0.013 0.004 -3.322 0.001

Primary diagnosisb

Anxiety -0.006 0.017 -0.362 0.718 0.006 0.008 0.844 0.399

ADHD 0.008 0.011 0.693 0.489 0.001 0.006 0.104 0.917

Depression 0.007 0.010 0.684 0.494 -0.009 0.005 -1.900 0.057

Trauma 0.011 0.013 0.817 0.414 0.001 0.006 0.104 0.917

CAFAS total 0.000 0.000 -0.119 0.906 0.000 0.000 1.341 0.180

Service typec 0.000 0.018 0.017 0.987 -0.023 0.008 -2.977 0.003

Level of cared -0.029 0.016 -1.793 0.073 0.047 0.007 6.865 0.000

Episode length 0.000 0.000 2.874 0.004 0.000 0.000 -1.556 0.120

Statistically significant values (p \ 0.05) are given in bold

PDEB practices derived from the evidence base, PMES practices with minimal evidence support, ADHD attention-deficit/hyperactivity disorder

PDEB ICC = .425; PMES ICC = .382a 0=Male; 1=Femaleb Compared to disruptive behavior, 0=non-evidence-based service; 1=evidence-based service (Functional Family Therapy; Multisystemic

Therapy, or Mulidimensional Treatment Foster Care)c 0=In-home level of cared 1=Out-of-home level of care

Table 3 Level 2 models

predicting PDEB and PMES

Statistically significant values

(p \ 0.05) are given in bold

PDEB practices derived from

the evidence base, PMES

practices with minimal evidence

supporta 0=Male; 1=Femaleb 0=Non-evidence-based

service; 1=Evidence-based

service (Functional Family

Therapy; Multisystemic

Therapy, or Mulidimensional

Treatment Foster Care)c 0=In-home level of care;

1=Out-of-home level of cared 0=Not licensed; 1=Licensede Compared to social workf Compared to cognitive-

behavioral

Predictor PDEB PMES

b SE t p b SE t p

Level 1 predictors

Youth age – – – – 0.002 0.001 3.073 0.002

Youth gendera – – – – -0.013 0.004 -3.327 0.001

Service typeb – – – – -0.018 0.009 -2.086 0.037

Level of carec – – – – 0.050 0.008 6.612 0.000

Episode length 0.000 0.000 3.166 0.002 – – – –

Level 2 predictors

Years trained 0.000 0.002 0.002 0.938 0.000 0.001 0.282 0.778

Licensedd -0.031 0.022 -1.411 0.158 0.009 0.010 0.885 0.376

Professional specialtye

Counseling -0.006 0.033 -0.172 0.864 0.011 0.016 0.695 0.487

Marriage and family -0.020 0.027 -0.743 0.457 0.009 0.013 0.687 0.492

Psychology and psychiatry 0.083 0.038 2.180 0.029 -0.001 0.018 -0.048 0.962

Other 0.040 0.037 1.076 0.282 -0.006 0.018 -0.365 0.715

Theoretical orientation

Eclecticf -0.073 0.028 -2.599 0.009 0.013 0.014 0.928 0.353

KEBSQ total 0.000 0.001 0.186 0.852 0.000 0.000 -0.335 0.738

EBPAS total -0.005 0.019 -0.259 0.796 -0.003 0.009 -0.371 0.711

Adm Policy Ment Health

123

youth and thus use more practices in general. We also

found that males were more likely to receive PMES. It is

not clear why this pattern emerged. Additional research is

needed to determine whether there truly are gender dis-

crepancies in practice use. If future research confirms this

finding it will be important to address this gender

inequality. On the positive side, we found that youth

receiving services from an evidence-based program (i.e.,

Functional Family Therapy, Multisystemic Therapy or

Multidimensional Treatment Foster Care) were less likely

to receive PMES, suggesting that these programs are not

including practices that are inconsistent with the evidence-

base. Denneny and Mueller (2012) found similar results in

a study of Multisystemic Therapy in the Hawaii CAMHD

system (2013), suggesting robustness of these findings.

Although we found that older youth were more likely to

receive PMES, Brookman-Frazee et al. (2010) found that

older youth received a higher Child EBP score. However, this

may be an artifact of the different methods employed by the

two studies. For instance, the Brookman-Frazee et al. study

examined youth ages 4–13 whereas the current study inclu-

ded older adolescents (age range = 5–19; mean = 14.1).

The Brookman-Frazee et al. study also examined disruptive

behavior only whereas we included other common diagnoses

in childhood. Further, while the Brookman-Frazee et al.

study employed observational methods to assess practices

and the current study relied on self-reported practices, it is

unclear whether (or how) this methodological artifact may

have resulted in the difference in findings.

Consistent with Brookman-Frazee and colleagues’ study

(2010), we found very few therapist characteristics that

predicted use of PDEB and no characteristics that predicted

PMES. Therapists who identified with a Behavioral or

Cognitive-Behavioral theoretical orientation used signifi-

cantly more PDEB than therapists who identified with an

Eclectic theoretical orientation. It is possible that because

many of the PDEB are consistent with cognitive-behavioral

theory, therapists with a CBT orientation are more likely to

use practices consistent with their orientation. An alterna-

tive but related explanation is that therapists with a CBT

orientation have an epistemology that values science-based

treatments and as such, are more likely to use PDEB with

their youth clients. We also found that therapists who

reported a Psychology or Psychiatry specialty used sig-

nificantly more PDEB than therapists with a Social Work

specialty. This is somewhat concerning give that a large

portion of the mental health workforce is made up of Social

Workers, with Psychologists and Psychiatrists increasingly

playing a larger role in mental health administration and

supervision. Historically, Social Work has focused less on

psychotherapy and the provision of mental health services

and more on connecting individuals in need of services to

those services. Thus there has been less of an emphasis on

training of EBPs in schools of Social Work in the past,

however, there are signs of positive changes in this area

(Barth et al. 2012).

Contrary to expectations, knowledge of EBPs and atti-

tudes towards EBPs did not predict greater use of PDEB

nor did they predict less use of PMES. These findings may

relate to ceiling effects for the attitude measure and

instrumentation issues for the knowledge measure. Con-

cerning EBP attitudes in the current study, it is noteworthy

that the mean total score for the EBPAS was 2.93

(SD = .48), suggesting very positive attitudes towards

EBPs. Indeed, the overall attitude score in the current

sample was greater than those indicated in Aarons’ (2004)

original and Aarons et al. (2010) follow up studies. This

skew in the data, combined with a small standard deviation,

may have resulted in less overall variance to draw from

when examining the relationship between attitudes and

PDEB and PMES scores. This skew may be a result of the

context from which the sample was drawn. As noted ear-

lier, participants were providers who voluntarily partici-

pated in trainings on youth EBPs in Hawaii, a state that has

experienced large system reform with an emphasis on

increased attention towards EBP implementation (Nakam-

ura et al. 2011a, b; Nakamura et al. 2012).

Concerning EBP knowledge, the KEBSQ assesses

knowledge at the awareness (i.e., is a specific technique

part of a larger EBP protocol) rather than the how-to or

implementation (i.e., what are the specific steps associated

with implementing a certain technique) level, and the

relationship between these two types of knowledge are in

need of further investigation (Nakamura et al. 2011a, b).In

addition to the concerns mentioned above, the finding that

attitudes and knowledge did not predict PDEB or PMES

scores speak to the importance of investigating other

individual level, as well as inner and outer context factors

such as organization variables important for implementing

youth EBPs in usual care (e.g., Damschroder et al. 2009).

For instance, in a study of therapists from communities

receiving funding from the Comprehensive Community

Mental Health Services for Children and their Families

Program, Aarons et al. (2009) reported that private agen-

cies (compared to public agencies), organizational support

for EBPs, and attitudes towards EBPs all predicted greater

use of EBPs. When they examined the complex relation-

ships among these variables, they found that the model that

fit their data best was one where the relationship between

EBP use and agency type was mediated by organizational

support for EBPs. Further, while agency type and organi-

zational support predicted more favorable attitudes towards

EBPs, in the final model, attitudes towards EBPs did not

significantly predict greater use of EBPs though there was a

small effect in the hypothesized direction. Aarons and

colleagues (2009) reconciled these contradictory findings

Adm Policy Ment Health

123

by hypothesizing that in the absence of organizational

support of EBPs, attitudes towards EBPs are likely to

predict use of EBPs but when organizations do support

EBPs, individual provider attitudes towards EBPs are not

as important.

Limitations

Although the results of the current study are promising with

regard to continued investigation of predictors of youth ther-

apist practices in usual care, a few limitations in addition to

those mentioned above are noteworthy. First, as mentioned

earlier the current therapists were drawn from a larger sample

of providers that voluntarily signed up to participate in train-

ings on youth EBPs. Along these lines, some caution should be

exercised when interpreting these results and readers are urged

to remember that the provider sample may not be representa-

tive of the broad spectrum of usual care therapists. A second

notable concern surrounds the self-reported nature of the

MTPS, the instrument underlying calculation of the PDEB and

PMES outcome variables. As mentioned above, within the

CAMHD reporting infrastructure, therapists self-report on the

practices they utilized with a youth for any given reporting

month. As such, readers are encouraged to remember the self-

report aspect of the outcome variables of PDEB and PMES, as

well as the potential flaws that typically accompany such

methodology. A third noteworthy caveat concerns instru-

mentation specificity for the assessed constructs of attitudes

and knowledge in the current study. In addition to the concerns

already mentioned above, the level of specificity between the

constructs of attitudes, knowledge, and self-reported PDEB

and PMES practice patterns varied widely. For example,

although the construct of EBP attitude was assessed with a

psychometrically tested instrument, the EBPAS queries about

EBP attitudes in a general sense (as opposed to attitudes about

the 24 specific techniques that make up the PDEB score or the

12 techniques that make up the PMES score). Building upon

this example, the knowledge measure asked about a set of 40

techniques, some of which contributed to the PDEB scores and

some of which contributed to PMES scores. Building upon this

work, forthcoming studies examining attitudes towards,

knowledge of, and self-reported usage patterns of EBPs may

benefit from assessing all three constructs at the same level of

specificity. Indeed, research is emerging that suggests that

therapists’ attitudes towards EBPs can vary as a function of the

type of EBP for which they are queried (Reding et al. 2007).

Conclusion

Given that findings in this study largely replicate previous

findings in different samples, in different contexts, and with

different methodologies, one potential conclusion is that

these findings are relatively robust and that as a field we

should be turning our attention and efforts to other levels as

suggested by theories of implementation science such as

organizational factors, leadership, and the process by which

implementation is accomplished (e.g., Aarons and Som-

merfeld 2012; Damschroder et al. 2009; Proctor et al. 2009).

However, care should be taken to not abandon research on

individual factors just yet, as work in this area will most

likely gain benefit from continued scientific advances related

to the broader EBP dissemination and implementation

movement. For instance, it could be argued that our field has

not yet agreed upon a standardized definition of, or metric

system for EBP, the very construct upon which voluminous

amounts of research are beginning to orbit. It is noteworthy

that large governing bodies playing important roles in dis-

semination and implementation (e.g., American Psycholog-

ical Association, Society of Clinical Child and Adolescent

Psychology, Substance Abuse and Mental Health Services

Administration, American Academy of Pediatrics) all have

slightly varying definitions of this construct. Concerning an

EBP metric system, some groups define EBP at the level of

brand-named treatment manuals, whereas others may utilize

technique commonalities (practice elements) across those

protocols or even broad-based treatment approaches (e.g.,

Cognitive Behavior Therapy). Further complicating these

issues, EBP also varies as a function of youth characteristics

including diagnoses and age, and also changes over time as

the treatment outcome literature continues to mature. As a

result, instrumentation efforts for EBP and related individual

level constructs such as knowledge and attitudes can be

deceivingly complex. Owing to these and other complicating

issues, noise or error variance has the real potential to nullify

important research findings in studies such as this one and

partially account for discrepant findings between various

investigative groups. In the end then, despite our findings

that only a few youth and provider level factors predicted

practice patterns for therapists, it may still be fruitful to

continue with research in this area. Borrowing from the area

of treatment outcome research, the perennial question of

‘‘What works for whom, how and under what circum-

stances?’’ as applied to therapists and EBP implementation

may best be approached through efforts that continue to be

multilevel in nature, that also allow for penetrating media-

tional and mode rational models.

References

Aarons, G. A. (2004). Mental health provider attitudes toward adoption

of evidence-based practice: The evidence-based practice attitude

scale (EBPAS). Mental health services research, 6(2), 61–74.

Aarons, G. A., Glisson, C., Hoagwood, K., Kelleher, K., Landsverk,

J., & Cafri, G. (2010). Psychometric properties and US National

Adm Policy Ment Health

123

norms of the evidence-based practice attitude scale (EBPAS).

Psychological Assessment, 22(2), 356.

Aarons, G. A., Hurlburt, M., & Horwitz, S. M. (2011). Advancing a

conceptual model of evidence-based practice implementation

in public service sectors. Administration and Policy in Mental

Health and Mental Health Services Research, 38(1), 4–23.

Aarons, G. A., & Sommerfeld, D. H. (2012). Leadership, innovation

climate, and attitudes toward evidence-based practice during a

statewide implementation. Journal of the American Academy of

Child and Adolescent Psychiatry, 51(4), 423–431.

Aarons, G. A., Sommerfeld, D. H., & Walrath-Greene, C. M. (2009).

Evidence-based practice implementation: The impact of public

versus private sector organization type on organizational

support, provider attitudes, and adoption of evidence-based

practice. Implementation Science, 4(1), 83.

American Psychological Association Task Force on Promotion and

Dissemination of Psychological Procedures, Division of Clinical

Psychology. (1995). Training in and dissemination of empiri-

cally-validated psychological treatments: Report and recommen-

dations. The Clinical Psychologist, 48, 3–23.

Barth, R. P., Lee, B. R., Lindsey, M. A., Collins, K. S., Strieder, F.,

Chorpita, B. F., et al. (2012). Evidence-based practice at a

crossroads the timely emergence of common elements and common

factors. Research on Social Work Practice, 22(1), 108–119.

Borntrager, C., Chorpita, B. F., Higa-McMillan, C. K., Daleiden, E.

L., & Starace, N. (2013a). Usual care for trauma-exposed youth:

Are clinician-reported therapy techniques evidence-based? Chil-

dren and Youth Services Review, 35(1), 133–141.

Borntrager, C. F., Chorpita, B. F., Orimoto, T., Love, A., & Mueller,

C. W. (2013b). Validity of clinician’s self-reported practice

elements on the monthly treatment and progress summary. The

Journal of Behavioral Health Services and Research. doi:10.

1007/s11414-013-9363-x.

Brookman-Frazee, L., Haine, R., Baker-Ericzen, M., Zoffness, R., &

Garland, A. (2010). Factors associated with use of evidence-

based practice strategies in usual care youth psychotherapy.

Administration and Policy in Mental Health, 37, 254–269.

Child and Adolescent Mental Health Division. (2008). Instructions

and codebook for provider monthly summaries. Honolulu:

Hawaii Department of Health, Child and Adolescent Mental

Health Division.

Chorpita, B.F. & Daleiden, E. L. (2007). 2007 Biennial report:

Effective psychosocial interventions for youth with behavioral

and emotional needs. Child and Adolescent Mental Health

Division, Hawaii Department of Health.

Chorpita, B. F., Daleiden, E. L., Ebesutani, C., Young, J., Becker, K.

D., Nakamura, B. J., et al. (2011). Evidence-based treatments

for children and adolescents: An updated review of indicators

of efficacy and effectiveness. Clinical Psychology, 18(2),

154–172.

Daleiden, E., Lee, J., & Tolman, R. (2004). Child and Adolescent

Mental Health Division: 2004 annual report. Honolulu: Child

and Adolescent Mental Health Division.

Damschroder, L. J., Aron, D. C., Keith, R. E., Kirsh, S. R., Alexander,

J. A., & Lowery, J. C. (2009). Fostering implementation of

health services research findings into practice: A consolidated

framework for advancing implementation science. Implement

Science, 4(1), 50.

Denneny, D. & Mueller, C. (2012). Do empirically supported

packages or their practices predict superior therapy outcomes

for youth with conduct disorders. Poster presented at the Forty-

Sixth Annual Convention of the Association for Behavioral and

Cognitive Therapies, National Harbor, MD.

Field, A. (2009). Discovering statistics using SPSS. CA: Sage

publications.

Garland, A., Brookman-Frazee, L., Hurlburt, M., Accurso, E.,

Zoffness, R., Haine-Schlagel, R., et al. (2010a). Mental

health care for children with disruptive behavior problems: A

view inside therapists’ offices. Psychiatric Services, 61(8),

788–795.

Garland, A. F., Hurlburt, M. S., Brookman-Frazee, L., Taylor, R. M.,

& Accurso, E. C. (2010b). Methodological challenges of

characterizing usual care psychotherapeutic practice. Adminis-

tration and Policy in Mental Health and Mental Health Services

Research, 37(3), 208–220.

Higa-McMillan, C. K., Jackson, D., & Daleiden, E. L. (2011).

The Venn Diagram of usual care and evidence-based care:

Identifying non-overlapping areas to fine-tune dissemination

of evidence-based practices for child anxiety. In C. K.

Higa-McMillan (Chair), Comparing Usual Care and Evidence-

Based Care for Children and Adolescents: Advancing Dissem-

ination in the twentyfirst Century. Symposium presented at the

annual meeting of the Association of Behavioral and Cognitive

Therapies, Toronto.

Hodges, K. (1998). Child and adolescent functional assessment scale

(CAFAS). Ann Arbor: Functional Assessment Systems.

Hodges, K., & Gust, J. (1995). Measures of impairment for children

and adolescents. Journal of Mental Health Administration, 22,

403–413.

Hodges, K., & Wong, M. M. (1996). Psychometric characteristics of a

multidimensional measure to assess impairment: The child and

adolescent functional assessment scale (CAFAS). Journal of

Child and Family Studies, 5, 445–467.

Jensen-Doss, A., Hawley, K. M., Lopez, M., & Osterberg, L. D.

(2009). Using evidence-based treatments: The experiences of

youth providers working under a mandate. Professional Psy-

chology, 40(4), 417.

Nakamura, B. J., Chorpita, B. F., Hirsch, M., Daleiden, E., Slavin, L.,

Amundson, M. J., et al. (2011a). Large-scale implementation of

evidence-based treatments for children 10 years later: hawaii’s

evidence-based services initiative in children’s mental health.

Clinical Psychology, 18(1), 24–35.

Nakamura, B. J., Higa-McMillan, C. K., & Chorpita, B. F. (2012).

Sustaining Hawaii’s evidence-based service system in children’s

mental health. In D. H. Barlow & K. McHugh (Eds.), The

dissemination of evidence-based psychological treatments. New

York: Oxford University Press.

Nakamura, B., Higa-McMillan, C., Okamura, C., & Shimabukuro,

S. (2011b). Knowledge of and attitudes towards evidence-

based practices in community child mental health practi-

tioners. Administration and Policy in Mental Health, 38,

287–300.

Nelson, T. D., & Steele, R. G. (2007). Predictors of practitioner self-

reported use of evidence-based practices: Practitioner training,

clinical setting, and attitudes toward research. Administration

and Policy in Mental Health and Mental Health Services

Research, 34(4), 319–330.

Orimoto, T. E., Higa-McMillan, C. K., Mueller, C. W., & Daleiden,

E. L. (2012). Assessment of therapy practices in community

treatment for children and adolescents. Psychiatric Services,

63(4), 343–350.

Proctor, E. K., Landsverk, J., Aarons, G., Chambers, D., Glisson,

C., & Mittman, B. (2009). Implementation research in

mental health services: An emerging science with conceptual,

methodological, and training challenges. Administration and

Policy in Mental Health and Mental Health Services Research,

36(1), 24–34.

Reding, M. E. J., Chorpita, B. F., Lau, A. S., & Innes-Gomberg, D.

(2007). Providers’ attitudes toward evidence-based practices: Is

it just about providers, or do the practices matter, too?

Adm Policy Ment Health

123

Administration and Policy in Mental Health and Mental Health

Services Research, 34(4), 411–419.

Stephan, S., Westin, A., Lever, N., Medoff, D., Youngstrom, E., &

Weist, M. (2012). Do school-based clinicians’ knowledge and

use of common elements correlate with better treatment

quality? School Mental Health, 4(3), 170–180.

Stumpf, R. E., Higa-McMillan, C. K., & Chorpita, B. F. (2009).

Implementation of evidence-based services for youth assessing

provider knowledge. Behavior Modification, 33(1), 48–65.

Tabachnick, B. G., & Fidell, L. S. (2007). Using Multivariate

Statistics (6th ed.). Needham Heights: Allyn & Bacon.

Adm Policy Ment Health

123