evaluation public health interventions and · process evaluation of the adolescent social action...

29
N PROCESS EVALUATION FOR PUBLIC HEALTH INTERVENTIONS AND RESEARCH Allan Steckler Laura Linnan Editors Foreword by Barbara A. lsrae! :13 l""i**l,ff' - www.russcyuass.uuru

Upload: others

Post on 13-Jul-2020

17 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: EVALUATION PUBLIC HEALTH INTERVENTIONS AND · Process Evaluation of the Adolescent Social Action Program in New Mexico whereas internal evaluators are assumed to have better insight

N

PROCESS EVALUATIONFOR PUBLIC HEALTHINTERVENTIONSAND RESEARCH

Allan Steckler

Laura Linnan

Editors

Foreword by Barbara A. lsrae!

:13 l""i**l,ff'-

www.russcyuass.uuru

Page 2: EVALUATION PUBLIC HEALTH INTERVENTIONS AND · Process Evaluation of the Adolescent Social Action Program in New Mexico whereas internal evaluators are assumed to have better insight

No part of this publication may be reproduced, stored in a retrieva] System, or tlansmitted il any form

o, by ury -aur., electronic, *."f,u.'itut, photocopying recordilg' scannhg' or otherwise' except as

permittedunderSectionl0Torl0softt,etgzoUnitedStatesCoplrightAct,withouteithertheprior'*ritte.

permission of the publisher, or authorization through payment of the app^ropriate Per-copy

feetotheCoplTightClearanceCentellnc',222RosewoodDrive'Danvers'MA01923'9i8-i50-8400'lax978-750-44T0,oronthewebatwww.coplright.com.RequeststothePublisherforpermissionshould be addressed to the Permissions Dep^itm.nt,Jot n Wiley & Sons, Inc , 1 1 1 River Stre et'

Hoboken, NJ 0i030, (201) 748-6011, fa" liOt; 748-6008, e-mail: permcoordinator@wileycom'

Limit of Liability/Disctaimer of warranty: while the publisher and author have used thei best efforts

in preparing this book, they ma'I<e no representations or warranties with respect to the accuracy or

completeness of the contents of this book and specifically disclaim any implied warranties of

merchantability or fitness for a particular p"po" No warranty may be created or.extended by sales

,.p..r"rtutir.r'or witten sales materials. The advice and strategies contained hereil may not be

suitable for your situation. you should consult with a professional where appropriate Neither the

publisher nor author shalt be liable for any loss of profit or any other commercial damages, including

tut not limited to special, incidental, consequential, or other damages'

Jossey-Bassbooksandproductsareavailab]ethroughmostbookstores'TocontactJossey.Bassdirectlycallour customer care Departrnent withir the U.S. ar800-956-7739, ouSide the uSat 317-572-399386 or

fu3l7-572-4002.

Jossey-Bass also publishes its bools in a variety of electronic formats. Some content that appears in print

may not be available in electronic books'

Library of Coagress Cataloging-in-Publication Data

Process evaluation for public health intewentions and research / Allan

Steckle5 Laura Lirtnan, editors ; foreword by Barbara Israel - 1st ed'

P;cm'Includes bibliographicat references and index'

L Pubiichealth-Research-Methodology 2' Health

promotion-Evaluation. 3. Medicine, Preventive-Research-Methodology'

IDNLM: l. He alth Services Research 2' Process Assessment (Health

Cuie) 3. Public Health. \'V 84 3 P963 20021 I Steckler' Allan B II'

Linnan, Laura.

Copgight @ 2002 bYJohn

Published byJossey-Bass

A WileyJmprint

Wiley & Sons, Inc A-ll rights reserued'

989 Market Street, San Francisco, CA 94103-1741 www.iosseybass.com

RA440.B5 .P76 2002

362.1'07'2-dc2l

Printed in the United States of America

FIRST EDITION

HBI'rinnng l0 9 8 7 6 5 + 3 2 I

2002012702

Page 3: EVALUATION PUBLIC HEALTH INTERVENTIONS AND · Process Evaluation of the Adolescent Social Action Program in New Mexico whereas internal evaluators are assumed to have better insight

NCHAPTER FOURl

llllI

ll

llll

Process Evaluation of the Adolescent

Social Action Program in New Mexico

Deborah L. Helitzer and Soo-Jin Yoon

1-r his chapter describe s the process evaluation of a primary prevention pro-

I gru-, the Adolescent Social Action Program (ASAP), which was aimed at

reducing alcohol- and drug-related morbidity and mortality rates among New

Mexico's youth. The program began in 1982 with supervised youth visits to the

Emergency Room of the University of New Mexico (uNM) Hospital. During

these visits, young people interviewed patients who had problems related to al-

cohol and drug use. Over the years since its inception, ASAP was successful in

gaining various funding for its implementation and improvement. A curriculum

was written around the core experience of the patient interviews, and, over time,

ASAP broadened its focus to include tobacco, all types of drugs and substances,

interpersonal violence, gangs, and other issues relevant to young people. With its

evolution, the program's original name, Alcohol and Substance Abuse Preven-

tion, was changed to Adolescent Social Action Program, to reflect its overall phi-

losophy of empowering young people to become capable of prorhoting change

in their communities.

In 1994, ASAP received a five-year research grant (mid- 1994 to mid- 1999)

from the Nationai Institute on Alcohol Abuse and Alcoholism OIIAAA). Nthough

The authors would Like to acknowledge the contributions and.assistance of Nina Wallerstein, RandaII

Starling, andJulie Gri{fin.

Page 4: EVALUATION PUBLIC HEALTH INTERVENTIONS AND · Process Evaluation of the Adolescent Social Action Program in New Mexico whereas internal evaluators are assumed to have better insight

u . Process Evaluation for Public Health lnterventions and Research

previousfundinghadproducedextensiveoutcomeevaluationsfortheprogram,theNIAAAgrantprovidedASAPwithitsfirstopportunitytoconductasystem-atic process .,rulrrutio.. The process evaluation was designed at the beginning of

the NIAAA grant period, utotg with the development of the outcomt t:l:i::t:

instruments, curriculum "t'i'Iot, and planning for program- implementatlon'

Becar1se the quasi-experimental designtf tne NlA'qa-funded outcome evalua-

tion required ASAP to implement its intervention over a period of three and a

half years (seven semest..' btt*te" Spring 1995 and Spring 1998)' the process

evaluation *u, derigred to capture information during this same pcriod'

Process evaluation has been shown to (1) help determine 1i Typt

III error

occurred(Stecklerandothers,lgg2)-thatis,iftherewasthebeliefthatchangesin program outcomes were due to the program when' in fact' there was an ab-

sence of intervention, and (2) ..rhu.rc.ih. understanding of program effects by

linking exposure ,o o,t"orn" (McGraw and others' 1994)' The purpose of this

process evaluation was to document fidelity to the program (tracking students and

documenting participation, attendance' and exposure)' thereby enhancing the

interpretationofthep..g'.*tresults'This.checkforfdetiEl(toensurethatthepro.gram was implementeiu' it was intended) was designed not only for the sum-

mative purpose of a.it'-i"ing if Type III error had occurred but also for the

formative purpose of reducing i" po"iUitlty of such error through the program's

constant monitoring of its imilem..rtutio., as well as appropriate activities to cor-

rect insu{ficient implementaiion. Evidence of program fidelrty enhances investi-

gators, ability to attribute changes in outcomes to tl-r. progrum's intervention' For

the formativ. p,r.por. of fi.r.-"trni.rg the program, this process.evaiuation was

designed to provide immediate feedback to i'ogr^m personnel with data that

wourd have to be, in a timery manner, u.rulyr.d,lrrterpreted, and given back to

the staff members who implemented the program'

The process t"'i;;;; design tt'oL'"J 'lo'*Iy'

The original design was based

on literature ut'ulubit at the tile' and subsequent modifications were made

after the onset of the NIAAA grant. For."u..rpl., fot's gro:p: t::titt*q "lt::

[opt. we re adde d,; *"* #"' ".' iTi :: *: T;J: il i*:'Ji::",T .'; i.1 "';

ilm.::::f:Xilj, I Til:f iiI'l?li ;; ., J, " a o' un gu,a,i o n 1Mit " u" d

Huberman, 1994) with other data sources' These modifications were based on a

prior process t""''^;;;; ';;Jy "o"d'"ttd by Deborah Heiitzer'for an obesitv pre-

vention program io, Nl i". A*"ri.u., school children, in which [wenty-seven sets

of data collection i"ttt'-t't'*t" dt"ttopta' each set havinqfetween three and

five ins*uments, liom lgg+ ,o 1999 (Heliizer and others' 1999)'

Aclassicdistinctioninevaluationiswhethertheevaluatorisinternalorexternaltotheprogram.Externalevaluatorsareassumedtobemoreobjective,

Page 5: EVALUATION PUBLIC HEALTH INTERVENTIONS AND · Process Evaluation of the Adolescent Social Action Program in New Mexico whereas internal evaluators are assumed to have better insight

Process Evaluation of the Adolescent Social Action Program in New Mexico

whereas internal evaluators are assumed to have better insight into program

function and meaning flNeiss, l99B). The Process evaluation of ASAP was con-

ducted, for the most part, internally, which, in this case, involved the handling of

design, instruments, data collection/documentation, and data management, as

well as analysis of the process evaluation-functions considered to be part of

the role of program implementers, which included two co-principal investigators,

a full-time project director, a full-time Program manager, and a small team of part-

time staff members (site coordinators and research assistants). An external con-

sultant was brought in for the process evaluation after the start of the grant, to

increase the objectivity of the evaluation and to attempt to achieve some separa-

tion between program implementation and evaluation'

The Adolescent Social Action Program

Program Summary

The Adolescent Social Action Program had two primary long-term goals. The

first was to prevent morbidity and mortality from risky behaviors-specifically to-

bacco and aicohol abuse-among low-income minority youth in high-risk com-

munities. The second was to empower the young people to become leaders who

were capable of promoting changes in their communities' behaviors, social val-

ues, and environmental policies and norms. The theoretical foundation of the pro-

gram was based on protection motivation theory (Rogers, l983) and Freire's

empowerment dialogue method (Wallerstein and Bernstein, l9BB). Over the

course of sixteen years, ASAP operated in over thirty multiethnic (predominantly

Hispanic and Native American) communities in New Mexico, served nearly six-

teen hundred middle- and high school students, and trained over four hundred

adult facilitators. The program was eventually expanded to serve tlventy-seven

hundred elementary school children through the program's peer education ac-

tivities with younger students. ASAP attracted national and international atten-

tion and received an outstanding program award from the U.S. Department of

Education's Safe and Drug-Free Schools Program for two conseclltive years.

Participating Schools

In 1994, upon receipt of NIAAA funding eight Albuquerque middle schools were

invited to participate in the program and were informed that each of them would

be randomly allocated to either the intervention or the comparison condition.

Hence, four schools were randomly selected to receive the intervention during

85

l,t-

rfnI,

t-

a

;s

v

s

Il

Page 6: EVALUATION PUBLIC HEALTH INTERVENTIONS AND · Process Evaluation of the Adolescent Social Action Program in New Mexico whereas internal evaluators are assumed to have better insight

86'

Process Evaluation for Public Health lnterventions and Research

the research study period, whereas the other four schoois received the delayed

intervention(comparisoncondition).Beforetheonsetoftheresearchactivities,ASAPgainedlnstitutionalReviewBoardapprovalfromtheHumanResearchRe-view Committee at UNM.

Youth Recruitment

Site coordinators, who were part-time paid personnel, had the responsibility of

recruiting young people to participate in ASAP' The ASAP program manager

and site coordinators consulted with the school administration to identify at each

oftheeightschoolsaseventh-gradeteacherwhofacilitatedaccesstohisorherstud.entsanclwaswillingtohelpintherecruitmentprocess.Theseteacherswereprovidedwitha,,o.i,,"ulstipendfortheirtimespentonsuchtasksasarrang-ingmeetingspacesattheschoolforASAPwheneverneeded'collectingconsentforms,u.rdr.*i,dingstudentsofupcomingASAPsessions'ASAPsitecoordina-tors entered the teachers, classrooms to recruit students, and ASAP research

assistants accessed students for data collection (using questionnaires, forms, and

salivasamplesforacotinineassessment]).Thosestudentswhovolunteeredtoparticipate *.r.- .r.oll.d on a first-come, first-served basis' Between eighteen and

tlventy-one students were recruited each semester at each school' site coordinators

were responsible for orienting parents of the students and for coordinating trans-

portationofthestudentstotheprogramsites(eitherUNMhospitalorBernalilioCountY Detention Center)'

ASAP Facilitators

An important part of the ASAP program was the g':lp of adult facilitators who

workedwitntheyoungpeopled"ti"gthesix-weekhospital-detentioncentersessions (a more i.tuitld e*planution of the centers is given later in this chap-

ter).Mostfacilitators*tttt"""'itedthroughUNMclassesandreceiveduniver-sity credit for their involvement, and some facilitators were community members

whovolunteered.(SeetheResultssectionforspecificdemographicinlormationon faciiitators.) A1l facilitators were required to attend a^two-day training course

atthebeginningoftheSemester.D,,.i.,gthesemeste5faci]itatorswererequiredto attend u *..-kly seminar session, whiJh included class readings of current lit-

erature on health issues related to young People and discussions about the pt:Ct:::

of the hospital-detention center sessions. These sessions allowed the opporturury

for facilitators to learn from each other's experiences and share lessons learned

about how to keep the sessions on track'

Page 7: EVALUATION PUBLIC HEALTH INTERVENTIONS AND · Process Evaluation of the Adolescent Social Action Program in New Mexico whereas internal evaluators are assumed to have better insight

process Evaluation of the Adolescent Social Action Program in New Mexico

The Core Curriculum: Hospital-Detention Center Sessions

Students who chose to participate in ASAP were assigned to a small group by

school. ASAP attempted to recruit seven to nine students for each group' Due to

attrition and the need to make up the numbers (for the appropriate sample size

for the outcome evaluation), the number of students per group actually ranged

from two to eleven. Each group of students was assigned two facilitators, who

worked together to implement the core curriculum sessions held once a week for

each group for six consecutive weelss.

The UNM hospital could handle only a moderate number of students with-

out disrupting its normal operations, so the groups were staggered by day and

by waves. In other words, the length of each semester allowed ASAP to schedule

two waves of the six-week curriculum, and each group held a session one night a

week (on a Monday, Tuesday, wednesday, or Thursday). Hence, during a semes-

ter, betvveen five and twelve groups of students went through the curriculum.

The first session of the six-week curriculum was a group-building session, fol-

lowed by three supervised visits to the University-affiliated hospital and one visit

to the county detention center. On these site visits, the students interviewed pa-

tients, their families, and jail residents who were affected by alcohol or substance

abuse problems. After interviews, students and facilitators were debriefed on their

experiences, using the SHOWED model (FIelitzer, Yoon, Wallerstein, and Garcia-

Velarde, 2000) to generate critical thinking and dialogue. This model was derived

from ASAP's theoretical foundation. The acronl.'rn SHOWED stands for S-Whatdid the students see and observe? H-V\4eat is happmingin their stories? O-Howdo the stories relate to ourlives? W-Why is this a problem? E-What would

empower thisperson or us to change? and D-What can we /o to improve our lives

or the lives of others? A variety of other participatory learning exercises, designed

to trigger topics relevant to young people, augmented each session. For example,

for a media iiteracy activiry students constructed a mosaic of cigarette and alco-

hol advertisements from magazines and they discussed how advertisers target

young people. The sixth session was a potluck dinner with family members, and

it provided the opportunity for students to brainstorm about community projects

in which they could further engage after the curriculum sessions.

Social Action Proiects and Booster Sessions

After the core hospital-detention center experience, the site coordinators worked

with the youth groups on social action projects. In alignment with the philosophy

of the program of empowering young people, the student groups were encouraged

E7

Page 8: EVALUATION PUBLIC HEALTH INTERVENTIONS AND · Process Evaluation of the Adolescent Social Action Program in New Mexico whereas internal evaluators are assumed to have better insight

8aProcess Evaluation for Public Health lnterventions and Research

t o c ho o s e t h e ir own p roj e c ts'' l :"i::l:-':'J::i: :,1': I ;:l;?f :: ::'*;

.r,ubtirni"t ASAP clubs at schools to palntmg

' *"ff t:ffiI'liit:

;3::J;, s e s sion was c o nduc te i lrjt1i" T:'":o

or di n at o r s s ix

monthsafterthestudent,receivedthecorecurriculum.ThisboostersesslonCon.sisted of a visit to ' "ot,tt'oom'

where "j;;;;itnessed live hearings and

interviewed j,,o*t' u'i"'':[ll' t" *'l' *tigt't it'to the legal and financiai ramt-

fications of risky b.;;;r] ;ch Js drinkini'""Ja"-"i at"'t'' As in the first

six sessions, 'Ut r<l"tttl-i""f'ata " tt'itty--i""te deilrie{ing that used the

SHOWED model'

Outcome Evaluation

Fortheoutcomeevaluation,aquestionnairewasdevelopedtoassesswhethertheprograminfluencedstudents,attailmentofconstructspertinenttoprotectlonmotivationtheory.Thequestio,,-;-;;;'.ised(l)thre.ar-xudents,percepttonof their Susceptibility and vulnerauri"".", 'i'1,

behaviors, \2) coping-students,

perceptionoftheirself-e{ficaciesandintentionstoengageinself-orsocialpro-tective behaviors, and (3) *ertly ^:;.rirl-r""*r

r.l-utionships and their abil-

ity to exper';:;;i:;;to"iv tnt rttli'g' ^"a thoughts of others' Figure 4'l

illustrates the ASAp program ,h".rr;;;;r, *rri"n is'explained in more detail in

the Theoretical Framework .."tiorl ffi; .frup,., flNallerstein and Sanchez-

Merki, 1994)'

Theoutcomeevaluationquestionnairewasadministeredtoeachstudentinboththeinterventionandthecomparisongroupsatpretest(immediatelybeforeth.y,ttti"d;;;; t"ti"'l"tj' i;*iiu"ty posttest' uttd si* weeks' eight

months, and fifteen months p.*,.,J. i" ^Jdi,io,,, u ii,.". measure of tobacco use

wascollectedinbothgroupstft,ot-tgh'u]i'a-samples^:O:-:*"andfifteenmonthsposttest. ouii.'g th"

"]o,,,,. :f ..Y school Semesters, sixw-three groups of stu-

dents partrcipaied in theproject' ^In;ffi;t'dt"t' who provided basic demo-

graphic it'roi**iot" s+i 1oi'zp;;ii;tte female and 3lB (36'B percent) were

male. sixryfourpercent or tn. r*a.trtJr"ffi.ffi;as "Hispanic" (n = 557)' 23'6

percent ., .'.ffiil);;.,,]; = 20;), il percent as..Amcrican Indian,, (n = 72),

o'o pt'"""t '* "fii'i1' = 57)' ;' i';;;;;"s 'Asian American" (n : iB) and

7'2p""''t;;"o;;;;"1n=62)'G;;;;;;'"ntasesadduotooverl00percentbecause".;;rrof "i'1"r'T::*:#**f.*#:fffi il?:lffi?t,;one sroup' For example' forry-srx:*"':::.::'"":;:;il;;;

betwe e n 76 and 86

and white/Anglo)' Follow-uP rates at fifteen months w

percent t"; ;ffi';;ention and comparison students'

Page 9: EVALUATION PUBLIC HEALTH INTERVENTIONS AND · Process Evaluation of the Adolescent Social Action Program in New Mexico whereas internal evaluators are assumed to have better insight

+o\o\

qJ

NoEUc6

ECoc'o

g

6

qr

o

I

.0J

oo,

oo-o

(!

Oa

6GL''E-c oa3 .E PgO--. F.E na< Y c fi 5'dotr .(! j? E

=-vL.='^-Y'=-g '6ro<6-9 ,iUd +

=-E,voooJq2@n49>\

E9s (,)

o:tr(!: .!!UEcO-oo-E<(!!>ls

'jf, (Jo:z. t-

a'ooFo(!

.9u(u

oLo-

q)

irE(60- T

.@-*FL-.40)PolJd<

P-(,)=^9.-L

xl6E.. eD IqF!P .=.-fr,io-ls\J+ cE or

9 c c'S)r^=P s.;..jj..s 9u9O iOL;

O- qr qJ..fac,3000! lqrEI=YPJ

o -'3EO

= EEtre6 ot !l .9 0.,

-qY=ooJ o-(^o-cE

u

o-

E(!Olo6-

c.)

9€.46u^-!,.s H_.0932LdiE r3!6

oE

o

o

jUJoo

&,oIJJ

G,(,o&eo-

.fUJc,fI

Page 10: EVALUATION PUBLIC HEALTH INTERVENTIONS AND · Process Evaluation of the Adolescent Social Action Program in New Mexico whereas internal evaluators are assumed to have better insight

Process Evaluation for Public Health lnterventions and Research

Theoretical Framework

The hypothesis of the ASAP program was that the experiential aspect of the pro-

gram (for example, patient and inmate interviews)' in combination with the

facilitator-led dijogical method, would lead students to empathize with the patients

and jail residents, i., h...i.rg their stories. students were encouraged to talk about

*f,.ifr.. or how th.y id..rtifi.d with the issues raised in interviews and to critically

analyzethe influences that contributed to risky behaviors. For example, the specific

targeting of young people by the Iiquor and tobacco industries is responsible for

increasing lev.t, oiyo,rti, consumption of alcohol and cigarettes, and young people,

being engaged in such risky behaviors, become negative role models for their friends

and family members. The expectation was that the ability to analyze their own

experienceswouldempowertheseyoungASAPparticipantstotakeaction.TheASAPprogramfocusedonreducingriskybehaviorsandencouraging

socially respo.rsibl. b.hu,riors that would protect others. The curriculum was based

ontwocomplementarytheoreticalperspecti.,esthatwere.woventogetherintheproject curriculum: Freire,s empowerment theory (lNallerstein and Bernstein, 19BB;

wallerstein and sanchez-Merki, 1994) and Rogers's protection motivation theory

(Floyd, Prentice-Dunn, and Rogers, 2000; Rogers' 1983)' Freire's approach was

used in designing the structureJdiulogu. component of the ASAP intervention'

The theories .rr.d in ASAP were hypothesized to provide psychoiogical empow-

erment,includingpersonalefficacy,recoglitionoftheneedforgrouppartlclPa-tion, critical .oriiorrrr.rs, and the willingness to participate in collective action

(Rissel and others, 1996)' Figure 4'I shows how the two theories and the

SHOWED model were expected to move participants to empowerment'

The first theoretical framework was the Freirian dialogue method, which was

used to foster critical thinking about adolescents' perception of alcohol, drug, or

tobacco consumption and the meaning of such behavior to adolescents trpng to

achieve or maintain a certain self-image. For example, a teenage giri feels that

smokingmakesherlookmoregrownup'TheFreirianempowermentdialoguemethod is based on a continuous listening-dialogue-action cycle, in which pro-

gramparticipantsidentifycuestoactionbylisteningtotheirownissuesofemo.tionalandsocialimport,engagingindialoguesabouttheseissues,anddevelopurgstrategiesforaddressingthern.ThesHoWEDmodelwasderivedfromthisthe.oreticalfoundation.Thiscriticalthinkingprocesshasbeenshowntoinfluenceyouthengagementinsuchriskybehavio.sasdrinking,smoking,andunexcusedabsence from school (chassin, Presson, Sherman, and Steinberg, l989;Jacobson,

Atkins, and Hacker, 1983).

Rogers,sprotectionmotivationtheorywasasecondtheoreticaifoundationof the ASAp curriculum. Protection motivation theory hypothesizes that the

Page 11: EVALUATION PUBLIC HEALTH INTERVENTIONS AND · Process Evaluation of the Adolescent Social Action Program in New Mexico whereas internal evaluators are assumed to have better insight

ProcessEvaluationoftheAdolescentSocialActionPrograminNewMexico9l

decision to act is initiated through a range of informational sources and is medi-

ated through a nonlinear cognitive perceptual proce ss (Rogers, l983; Rogers,

Decknel and Mewborn, 1978). A health-seeking response is expected when a

health threat increases one's r.rrlnerability and susceptibiliry when self-eIficacy and

response fficacy $he belief that one's actions can have an effect and that one can

..rpo.rd correctly to a health threat) increase, and when the rewards for engaging

in a maladaptive behavior decline (Rippetoe and Rogers, 1987; Rogers, l983).

The jail and hospital visits were expected to heighten the students' threat appraisal

(or risk perception) of the seriousness of their own susceptibility to the conse-

quences of substance abuse . The dialogue with patients and jail residents was ex-

pected to encourage the students to develop the ability to think through a situation

and make an assessment that would lead to an appropriate behavioral response,

as well as enhance their empathy, knowledge, and intention to change their own

behaviors. It was anticipated that their coping skills and self-eficac2 (personal

responsibility) would be improved by their participation in peer resistance and

decision-making exercises. For example, students were asked to enact scripted

and improvised role-playing exercises in which they responded to common Peer

pressure situations'

Overall Program Results

prior outcome evaluations demonstrated that the curriculum increased the stu-

dents, social skills, competence, critical consciousness, knowledge, and self-efficacy

for their own behaviors, as well as their social responsibility and prosocial behav-

iors, including their seltefficacy related to helping others, their recognition of the

need for group participation, and their willingness to participate in collective action

(Rissel and others, 1996)'

However, the outcome results of the NIAAA-funded study of ASAP showed

no significant di{ferences in alcohol, tobacco, or other substance use among the

yorrrC people-between the intervention groups and the comparison groups

(lryallerstein and Woodail, 2000). In addition, no diflerences in other measures

of threat appraisal or coping appraisal were found. The major pattern observed

in the data was the regular increase over time in the proportior.r' of students

reporting ever having had an alcoholic drink. This percentage increased from 61 '2

percent pretreatment to 64.0 percent posttreatment, to 66.3 percent at six months

posttreatmen t, to 7 4.4 percent at the fifteen-month follow-up' In fact, compared

with the outcomes of other published studies, the ASAP study participants re-

flected higher levels of alcohol initiation and exposure-higher than the national

average for this age group. This was not surprising, given that the risk level of the

study population was higher than the national average. All ethnic groups and both

Page 12: EVALUATION PUBLIC HEALTH INTERVENTIONS AND · Process Evaluation of the Adolescent Social Action Program in New Mexico whereas internal evaluators are assumed to have better insight

92 Process Evaluation for Public Health lnterventions and Research

genders, in both the interuention and the control groups, showed, over the course

of the study, an increased tendency to drink alcohol. The rate for females increased

from 58.6 percent pretreatmen t to 7 5.7 percent at the fifteen-month follow-up'

ffion Design and MethodsJ

TheNlAAA.fundedprocessevaluationoftheASAPprogramwasdesignedtomeasure the foltowing: (1) the fidelity of the implemented curriculum to the de-

signed curriculum, (2-) curriculum implementation consistency across groups of

yorrrg people, (3) the level of exposure and participation of the students, (4) bar-

riersandfacilitatorstoparticipation,(5)competingorinterveninginfluencesonparticipation und e*pos,,.e, (6) the existence of other health-related programs

going on concurrently in both intervention and comparison popuiations, and (7)

the characteristics of facilitators and students in the program' Figure 4'2 lists the

actual evaiuation questions.

The Role of Theory

The criticai theoretical elements previously described were included in the design

of several process evaluation instruments. Examples of theoretical constLucts mea-

sured include the SHOWED model, social analysis, critical thinking, group

process, facilitation style, cognitive dimensions, the continuous listening-dialogr-re-

actio., cycle, cognitive appraisals of threat, and coping abilities'

Process Evaluation Design

The design of the process evaluation was an intervention./comparison group de-

sign with continuous measure'ment throughout the training and the seven semes-

ters of intervention group implementation. Most of the data collection occurred

HGURE 4.2. PROCESS EVALUATION QUESTIONS'

1. Was the curriculum implemented with fidelity to the original design?

2. Was the curriculum implemented similarly in each youth group?

3. What was the level of exposure/participation by students?

4. What were the barriers and facilitators to participation?

5. What were the competing or intervening influences on participation and exposure?

6.Whatwerethecharacteristicsoffacilitatorsandstudentsintheprogram?

Page 13: EVALUATION PUBLIC HEALTH INTERVENTIONS AND · Process Evaluation of the Adolescent Social Action Program in New Mexico whereas internal evaluators are assumed to have better insight

ProcessEvaluationoftheAdolescentSocialActionPrograminNewMexico9,

at the intervention sites, as is appropriate for process evaluation' Some of the instm-

ments were designed to contribute to a formative evaluation, some contributed to

summative evahlation, and others were useful for both types of evaluation. Many

of the instruments were designed to enable triangulation of data' Figure 4'3 shows

the implementation model around which the process evaluation was designed'

Process Evaluation Resources

A small proportion of the grant funds were allocated to the process evaluation,

from which the consuitant was paid for 3 percent of her time for two years' A small

percentage(-5percent)ofthesa]ariesofa]lprogramstaffmemberscoveredtheiriata collection activities. In addition, one staff member conducted the process

evaluation analyses, devoting to this function -50 percent of her tjme for one year'

Data Collection, Sample, and Analyses

Fifteen instruments were used to collect both qualitative and quantitative infor-

mation for the process evaluation data. Figure 4.3 shows the different types of

information gathered and how they were related. The numbers next to each box

correspondwiththeinstrumentslistedinTable4.l,whichpresentsmoredetails

FIGURE 4.3. IMPLEMENTATION MODEL.

FacilitatorCharacteristics

15

Note: The numbers in tni, ,or.t .o|.r"rpond with the list of i,nstruments in Table 4.1 .

Page 14: EVALUATION PUBLIC HEALTH INTERVENTIONS AND · Process Evaluation of the Adolescent Social Action Program in New Mexico whereas internal evaluators are assumed to have better insight

(s9

asEEEE E* i-uc

= i:E=!f a'i, Eiaa*=o =?E-oEE ;fs PEbts.''Ee EsIoiEq sElg pEaasE'

= i+E;;Ea g;iE HgEEEE's rEEE xeE at.EE E

= iE4tr iYfi

i iEiII'ET FE;e itsFssE 6 E!^ o !! A -EpE}\EN

E q : 'E =

!fi*sE E + i=2,7 ;Aa 2.e ^'' y ts==? 9 Ee

=EH.o a^ :otF ri;.oiflj* E6 Eln-2- Fe; z

EiEt ,E eA; EaETEeE..tt=u'ol

l!

Bor- >'7 >E s? i

5=I h {.E9PE =lE F L.:: >+=: e 5

Vvu^,q'EePY a

[a.] Q unHsfiisrgY?E trE g >e0e€ 5

Iaa

a Pnz js .g 8..

EEE*ii;ggEEEg!c.i ,vi + vi

oG'

LLotrct.=

=or,!d

'o0,)

Eoq)o-UoE -lU'E 2Ni!E_>; !.t6=

E.9 .st.+E6l-0.).FYU=ocdo o

I

o-,b0,a(!oo-U=u=oo(J 'tra

'\ I c-o -6 'i.=dc 0(E- ! _ g

E:{ E [Ei c$; E EEEE =

EE; EAil= E.Es3 -^ iuE .yS;2 3-sEEaI EBEu [eE[;9

3sEE E EEie ItE eE3.5oi

aa

oooaA.oGcot!4,

trq,

E)E

!o

o

c)

.>(,tr

0,EEG

co

sCLor

-tIJJ

o

tl

zlrJ

=DG

Zzo

D

lrJvtttlLrJuoGr

Page 15: EVALUATION PUBLIC HEALTH INTERVENTIONS AND · Process Evaluation of the Adolescent Social Action Program in New Mexico whereas internal evaluators are assumed to have better insight

4! !0)9 uot-l>. 6t: .s

'!L

_-da bE Fh^. YE J';f 9E,?9;.? 0Ee! a' o._i=p;= I?: SYL.ii.-CDO F

;;eEg9 E>.I g I9.-c. Z

= b >. r;.9'' <:'":"=-g.

E

(EE>(!<rf2g

UOJU},3 ogE o ?qioJP oEI O tsaYi P.C---=

AE H - ,- IE - gEE:q; EE; &E;uE!q EE= €,ok o.Y < :4 < (! <f t"=a:"a 2A

o E e g r! g = ",EE €=EEt.= E? = s'a

Ea 5 t!E 5giE Ei:E Eg€ a

*a Z EI Z A:;; :E;; iE? Z

:i;' ;EE gIEEE JISEH EE;u=.F-.=t=6r,=(J.t=

9"- c E e E'E E E

gg,era;;g:igEE=EEg:I!fiE= fia fEE= rE= #Ei

lEEl:= ir= Ir :E:= :i= IE=..E.tEt'2c2't!

:Eejs.Y +* E .3dJ:.J

=Etstri =i: EE;E;llF: s;: Escrjoio-

:o

Fo-

il

qlio<oq, o)-nY:ls2i5.0

't)q)

:!cU(.)

!qiilg>6F*3oroFyzo<\.) "'lq-da!

0.)oE(J-uc 6--L-6:-

^7ytd.6YolEa,4.Ya\0P

-e-cu zEPEuEd-gUAoFu* za:

Page 16: EVALUATION PUBLIC HEALTH INTERVENTIONS AND · Process Evaluation of the Adolescent Social Action Program in New Mexico whereas internal evaluators are assumed to have better insight

., lE OJ oE;E E I B! bEEE} : = ^Eg UEl-C*6o '4 Y

EETE: =;EE TEr

Bi,==agli?igiligiggiEEEE r E +

i+ H# E tE a i

Ev t, et;* rr EA = EE '.-E * :f q J7 5Ts -e :: A eE IE= 4 Es ry aE U

ee 6 E; 6 -9t Ie-. = .ex a AE-i=ai g rt':^* !iE*;EsE :EgE EE;Eaor=..=og

c ! E E E Po' E

= Ee EEs.pEEs:i3 ^' = 'E

= ri E"E ; E,E 3 :E 3 5€; EEE= t EE=! i; =Esit*:f faEE=Ba-€! i="i

EE33EE !EEIE EEFEEa i;e3r..5ot5o'

UvO-

8tE:gEEtk 6 e;LL' tr''vi

o,-E=-eE=ojjUy=.\ q).E

B.T

+

!r,jEC tatOsEtr:roct C.

.ti

C6.g=dq->oL -:Zqru.3.8OUc.i

EnE9€ v4 6lEe-aOFor f,il tr =qi; i dSraoJfd!r+ q a ug- 9'_

-=;ao z9'X q.,, € O

3,H gr EE pat 9O.= a\.; oa:\O L uJF J+ (U l-$E.-> 2o=

!o

o

o.>l!L

gqEG

coo

o.or

o)

oo-

GgGco(ln

coE

E

oLUfztrzoU

V1

zlrJ

DE

=zoL

k)

lIIV,tnLUUoE,A

-tUJ

co

Page 17: EVALUATION PUBLIC HEALTH INTERVENTIONS AND · Process Evaluation of the Adolescent Social Action Program in New Mexico whereas internal evaluators are assumed to have better insight

Process Evaluation of the Adolescent Social Action Program in New Mexico

about the instruments-that is, the rationale and purPose of the instruments, the

population sample from which the data were collected, and a brief description of

the analysis methods used. The first three instruments (1, 2, and 3) were used to

collect information on the facilitator characteristics and training, an outcome eva.l-

uation in5trument (4) provided information on student characteristics, two

instruments (5 and 6) examined student participation in and exposure to the

program, four instruments (7, B, 9, and l0) were used to examine school con-

text, and five instruments (11, 12,13,14, and 15) measured the implementation

of the program. As examples, the facilitator checkoff list and observation forms

can be found in Appendixes A and B.

Snrdent comment sheets and facilitator training evaluation forms were anony-

mous and coded by group or semester. The facilitator questionnaire, health-related

student inventory and student questionnaire did not include narnes but were coded

to keep track of the individuals and to allow datasets to be merged by the code.

This iist of codes and names was maintained by ASAP research assistants and was

not accessed by the outcome evaluation data analysts. The health-related proglam

inventory and focus group were coded by school. Facilitator checkoff lists and

log./comment sheets required the facilitators to include their names, as their class

grade depended on their turning in these forms, and they were subsequently coded

by group for analysis purposes. Student attendance information was initiailytracked by student name, which program personnel used to boost attendance

through follow-up, but it was eventually coded to be merged with other student-

Ievel data. Q,uantitative data were entered into an Excel spreadsheet (1\4icrosoft

Applications, 1995) as individual datasets by instrument and were later merged

with the outcome dataset by student, glouP, or school code. Q,ualitative data were

reviewed and analyzed for content where resources allowed. flable 4. I lists which

data were coilected by quantitative versus qualitative methods.)

Results

The information presented in this section describes only the results and use of data

that were completely analyzed. As Table 4. I shows, some data were found to be

unreliable, whereas other data were not analyzed, because of resource constraints.

Further discussion of issues related to the limitations of analyses and the use of

results follows.

Facilitator Characteristics and Training

Because facilitators were so crucial to the program's success, we developed an ex-

tensive analysis of the facilitator characteristics, the training they received, their

Page 18: EVALUATION PUBLIC HEALTH INTERVENTIONS AND · Process Evaluation of the Adolescent Social Action Program in New Mexico whereas internal evaluators are assumed to have better insight

98'

oro..r, Evaluation for Public Health interventions and Research

attendance at the weekly class sessions, and their ability to implement the program

(Helitzer,Yoon,Wallerttei",-'dGarcia-Velarde'2000)'Facilitatorswereprimarilyundergraduate students (Bl percent)' Caucasian (60 percent)' and Hispanic (30 per-

cent), with varied .*pt'it"tt *o'ki"g with teenagers (26 percent having less than

ayear)33percentt,,,it'gonetothreeyears'andl3percenthavingfouryearsor

-, more). These data *ttt t"td only to create a profile of the facilitators for progress

and final rePorts.

Dataontrainingeffectivenessweretriangulatedfromseveralsources:facili.tator training.rulruiio.r, facilitator checkoff list, observation checkoff list, student

comment sheet, and facilitator log'/comment sheet' Training evaluations revealed

that 94 percent ol' the facilitators were confident that the preprogram two-day

training workshop prepared them well for the program' However' the observation

dataandfacilitatoruelt'epottdatasuggestthatthefacilitatorswerenotcompe-tent to implement all of the facfitation methods used in the program and that they

were inconsistent i; their implementation of role-modeling behaviors. Findings

revealedthatfacilitatorsshowedinconsistentandlowerthandesirabieimple-mentation,especiallybythelater(fifthandsixth)sessions'\{trenthesedatasug-gested that specific .or,.rr, or theoretical constructs were not fully grasped by the

facilitators, the training director addressed these issues in the weekly class sessions

for the benefit of othei facilitators, whose sessions were not yet complete'

Exposure: Attendance and lmplementation

Studentattendancescoresbygroupforthesetofsixsessionsrangedfrom43tol00percent.Thiswasdeterminedbycalculatingthetotalpossibleattendanceforeachgroup(thenumberofstudentsinthegroupmultipliedbysix-thenumberofcurriculumsessions)andthendeterminingthemeanattendanceattainedbyallstudentsioth.g.o,,p.Informationonattendancewasdesignedinsuchawaythat program f...""""a had immediate access to the information for constant

monitoring'Thisenabledprogrampersonneltotroubleshootwhenattendancewasdecliningandto.,,,,..thateffortswerebeingmadetoretainstudentsforthesix sessions.

Informationaboutimplementationofthecurriculumwasavailablefromthefacilitatorcheckofflist,thestudentcommentsheet,theobservationcheckofflist'theparticipantcontactdocumentationsheet'andthefacilitatorlog'Thedatashowedthatimplementation,aspreviouslynoted,wasinconsistent.Implemen.tationscoreswerecalculatedbysession,byactivity,""dPlgroup,basedonthe0-to3.pointscoringSystemortr'"r^"ititatorcheckofflist(seeAppendixA)andbytotalingthesescores.Tocheckwhethercertainsessionswereless

Page 19: EVALUATION PUBLIC HEALTH INTERVENTIONS AND · Process Evaluation of the Adolescent Social Action Program in New Mexico whereas internal evaluators are assumed to have better insight

Process Evaluation of the Adolescent Social Action Program in New Mexico

implemented than others, actual scores from the checkoff list for each session

were totaled across aU sixty-three groups and checked against the total possible

implementation score. There was a decline in implementation scores as ses-

sions progressed over the six-week period, but this pattern was consistently

repeated from semester to semester. Repeated elements like the interviews were

highly implemented, but other, more difficult and less frequently implemented

skills, such as role-playing, were often left out. Content / acdvity areas that received

high implementation scores were those activities that were scripted versus

improvised, repeated in several sessions, expected to be discussed in the semi-

nars, and accompanied by visual materials. Poorly covered content areas (in

which coverage was 50 percent or less) tended to be activities that were more

abstract and more time-consuming and which required more skill on the part

of the facilitators. For example, though the patient/inmate interviews were

well implemented, facilitators often failed to follow through with the next, more

diificult and abstract, task of using the SHOWED model as a tool for engag-

ing the young people in critical thinking and dialogue (Helitzer, Yoon,

Walierstein, and Garcia-Velarde, 2000).

Finally, a score for program exposure was calculated for each student. Ex-

posure to each session was calculated by multiplying the student group's fidelity

score for each session (l : low, 2 : medium, 3 : high) by the student atten-

dance score for that session (0 : not attended, I : attended). The program ex-

posure score for the student was then calculated by totaling the exposure scores of

six core sessions. Hence, the formula for program exposure was: Student Program

Exposure : SUM[(Session Fidelity) X (attendance)]. The results revealed that

there was inconsistent exposure to the curriculum across all of the students. With

a possible score of 0 to lB, students'program exposure scores ranged from 2 to

l7(x:10.6,n:403).

Use of Process Evaluation Results

Some results from the process evaluation were used to inform ASAP about the

process of program implementation, whereas other results were not available until

after the completion of the implementation. For example, when initial data analy-

ses suggested that the SHOWED model was being inconsistently implemented,

quizzes were added to the weekly class sessions for facilitators. A final imple-

mentation score for each group (low, medium, or high) was calculated and could

have been used for summative evaluation purposes for a dose response analysis,

because one of the hypotheses was that there might be a positive dose-response

relationship between exposure, participation, implementation, and outcome. Data

Page 20: EVALUATION PUBLIC HEALTH INTERVENTIONS AND · Process Evaluation of the Adolescent Social Action Program in New Mexico whereas internal evaluators are assumed to have better insight

. Process Evaluation for Public Health lnterventions and Research

on student participation (engagement) were not arralyzed" The observation form

wasContinuallyrevisedovertwosemestersinanattempttoimprovelowinterraterreliability. Despite these efforts, when all of the data were analyzed' we found that

observerswhohadpreviouslybeenfacilitatorsweremorecriticalofstudenten-gagementthanthosewhohadneverfacilitatedasession.Asaresult,itwasdiffi-I.,i, ,g,uggr.gate data that was not comparable across groups'

Anotherhypothesiswasthatstudentcharacteristicsmighthaveinfluencedoutcomes. fo, .*u*pt., 't'Jt"t"

school performance or previous exposure to al-

cohol,drugs,o.tobaccomayhaveaffectedtheirfutureriskofnegativeoutcomes.Outcome data suggested that the program had no overall effect on alcohol or

tobaccoconsumption,andadditio,,ulu,'uly.e,werenotconducted.otherinfor-mation sources, such as the participant contact sheet, facilitator log, and student

comment sheet, were t'ot"td, because of the lack of resources to embark on the

iabor-intensive qualitative ana-lyses that these instruments required' Information

on the school context (such as that provided by the barriers focus group, the health-

related program i"".;;; ^ra,r"" health-related student inventory) was not used

by the program, b..ur,.'tht outcome data suggested that the program had not

achieved its intended effect'

After the end of tt" gt^"t period' Helitzer' Yoon' Wallerstein' and Garcia-

Velarde (2000) describ.d ih. ,.iutionship between training and implementation'

At the time, a new grant submission was pending approval' It was thought that

thenewgrantwouldbeabletotakeadvantageofthisinformation.However,the grant was never awarded, which render.dlh. process evaluation results less

useful than anticiPated'

cess Evaluation

Lnssots 1. Get more out of kss data'

Ourexperienceisthatprocessevaluationdatacaneasilybecomeunmanageable.Ifanevaluatorishighlydetail-oriented,heorshewillhavetheinclinationtodoc-ument

"".ry urf..,3f the project. Howeveq it is unethical and a waste of resources

to collect data that will not be used in some way. Unless an evaluator has a plan for

datacollectiot',u"^frti',anduse'itishighlylikelythatheorshewillcollectmorethan is needed. It is a valuable exercise to-fina ways to overlap purposes for tools and

forformativeandsummativeapplications.Themostfrequendyaskedornecessaryquestions, i.rcltding bo ththe whaher and the whst questionE can be formulated in

advance'Earlyprocessevaluations'""hu"ho"fottt"CAICHandPAIrfWAYSprojectsltvt"Ctu*andothers'1994;Helitzerandothers'i999)'hadhundredsof

Page 21: EVALUATION PUBLIC HEALTH INTERVENTIONS AND · Process Evaluation of the Adolescent Social Action Program in New Mexico whereas internal evaluators are assumed to have better insight

Process Evaluation of the Adolescent Social Action Program in New Mexico

instruments, but this is not realistic for most programs. Deciding what are

the most important questions to answer will help the process evaluation be more

efficient.

LrssoN 2. Ltse mixed methods and triangulate.

In addition to being e{frcient with the collection of data, there is a need for data

from di{Ierent instruments, data sources, samples, and types of data. Quaiitativemethods, such as in-depth interviews and focus groups, are labor-intensive and

perceived as subjective data because of the relationship between the data col-

Iector and the data. Furthermore, qualitative data tend to be seen as less valu-

able than quantitative data. Sometimes, this distinction is worded in terms of"soft" (qualitative) versus "hard" (quantitative) data. For these reasons, qualita-

tive data can be seen as iess desirable to include in evaluations than quantita-

tive methods that are seen as more objective and Iess labor-intensive to collect

and ana)yze (Flelitzer and others, 1999). This debate and these biases have edsted

far too long in both primary and evaluation research. Using both qualitative

and quantitative methods, we can better answer both the what and the why ques-

tions. Using mixed methods also provides more opportunities for triangulating

data.

LsssoN 3. [Jse the information on tirne to improue the program.

Process evaluation staff members should be included in project implementation

discussions on a reguiar basis. In this way, these staff members can learn about

the implementation process and can provide frequent feedback for midcourse cor-

rections. This means that program staff people should exPect implementation fail-

ure as a normal part of project implementation. If it is expected, then they can

avoid being defensive about why it is happening, and as a result they can devote

time and resources to fine-tuning and making changes based on feedback. We need

to think about evaluation's purpose as program improvement rather than as

a thumbs-up or thumbs-down judgment as to whether the program is effective.

We need t6 create a cuiture of learning organizations as they relate.to program

imPlementation in Public health.

LBSSON 4. Deuote adequate attention and resources to process eualuation.

Although evaluation has more recently been receiving a part of routine pro-

gram funds, most of these funds and attention are still focused on outcome eval-

uation. The evidence suggests that most programs do not attain the outcomes they

lol

Page 22: EVALUATION PUBLIC HEALTH INTERVENTIONS AND · Process Evaluation of the Adolescent Social Action Program in New Mexico whereas internal evaluators are assumed to have better insight

Process Evaluation for Public Health lnterventions and Research

lo2

aredesignedtoattain,andforthisreasonitisimportantthatmoreattentionbefocused on p.o..,' evaluation' Previously' Process -evfuatiol

was thought of as

"bean counting" because it routinely Jot"*t*td information on' for example'

how many ""it*t

*t* htld' u' *til u' how many trainees were trained' Process

evaluationhasthepotentialtoprovldemuchmoreinformationifitisproperlyplanned and if sufficient resources are provided for it'

LBSSOTq 5. [Jse process eualuation to answer wlry, not just whether'

Programimplementationfailureistobeexpectedandp]annedfor.Also,proglamsunder real-lif-e conditions ,u..ty ,rro* tt;;. ,yp., or results that are published

inresearchstudies.Thisisthedifferencebetweenprogramfficac1andProgram

ffictiaenesstC"*''"aLewis'1986)'Becauseofthis'itismoreimportanttolookatwhyprogramsdonotdemonstratethesameintensityofintendedoutcomes.Programsthataredesignedonpaperandimplementedinreallifefacechallengesthatoltenarenotanticipatedo,pru.,.,.afo,.Fo,exampie,ifteachersimple.mentonlypartsofacurriculum,wecouldaskthemwhytheychosethepartstheydid.However,itwouldbefarmoreusefirltohavetheoriginalcurriculumdesignersindicatewhichpartsofth..,,,i.,t,marelinkedtothetheoreticaiframeworkof the program'

Ltion than others will.Some participants may benefit more from an Interven

Process evaluation data can be used to examine whether a dose-response rela-

tionshipexistsbetweenparticipationandoutcomechanges.Ifsucharelationshipexists, it may explain why some p^'il;;";t show increased changes as compared

withothers.Inaddition,othercharacteristicsoftheparticipants,suchasage,sex,and ethnicity, can be examined for relationshipt' , , - ^

Finally,programsmayworkbetterundercertain^conditionsthanothers.A retrospective analysis can be conducted to examine participants t:-Ii":

outcomes *"tt iut'otuble to see if individual' family' school' communrty' or

program conditions were different for those participants whose outcomes

were not as positive. A strength oir."..$ evaluation-is its ability to provide

more depth ii irrror*",io, u.ra the'oppo.trrniry to build hypothe ses for future

research'

LBssoN 6. Theorlt, theoryt, theory'

Bothastrengthandaweaknessoftheprocesseva]uationdescribedinthischap-ter was lr, .J"r*rg nature, reflecting new knowledge and expertise' The process

evaluationofthisstudylackedatheoreticalframeworkthatmighthaveexam.inedinacomprehensivemannertheassumptionsuponwhichtheprogramwas

Page 23: EVALUATION PUBLIC HEALTH INTERVENTIONS AND · Process Evaluation of the Adolescent Social Action Program in New Mexico whereas internal evaluators are assumed to have better insight

arch

rbelasple,

:ess

erly

tms

hed'am

cok

1es.

L$es

rle-h.yIers

ork

ProcessEvaluationoftheAdolescentSocialActionPrograminNewMexicolo3

based. However, we now know, in 2002, more about process evaluation' which

enables us to be self-reflective. We recognize thal a potential for process evalua-

tion is the examination of both the implementation and the program theory, as

described by Weiss (1998) as the program's "theories of change." Articulating a

program,s "hu.rg.

theories helps us understand the assumptions on which the in-

t.*..rtio., is based, and it is essential for developing appropriate evaluation ques-

tions and evaluation design. Program theories of change help evaluators plan

what data to collect and from whom, as well as what types of short- and long-

term effects might be expected from the intervention. Using weiss's theories of

change upp.ou.h, the process evaluation can focus on detailing and systemati-

.u11y io"rr*.nting each step of the implementation. In this way, the process eva]-

uation can be designed to examine some of the theoretical assumptions of the

program.' "Fig,r.., 4.4 and4.5 illustrate how the theory of change model couid have been

,rs.d fo. the ASAP program. we inserted the instrument numbers into the figure

to demonstrate the comprehensiveness of the original design in terms of program

function but not in terms of theory Process evaluation instruments were designed

to measure whatbut not wlry. They were designed to document the implementa-

tion but not to iook at the context and reasons behind the variable levels of

FIGURE 4.4. THEORY OF CHANGE MODEL FOR ASAP

FACILITATORS.

lmplementatlon TheotY Program TheolY

iill.:la-

hipred

;ex,

lrs.

cmor

:Ies

ide

ilre

Program recruits facilitators --------= Facilitators sign up (1 )

prosramconductsraciritatorrraininstli1i',:,:,i:[::;:l"J-"tJili]l!u,,r,irr,,

/ & self-confidence required to

/ imPlement curriculum (2)

program conducts weekly classes to, I--'---Facilitators attend weekly class (3)

iuiiiltuto6 to enhance knowledge & Facilitators identify areas of concern &

skills, troubleshoot, & provide fledback have enhanced skilE-& support

+

*Facilitators completely implementcurriculum (11, 12, 1 3, 14)

[overlap here with Figure 4.5]lp-ess

m-

Ias I . nrrbers in this model correspond with the list of instruments in Table 4'1'

Page 24: EVALUATION PUBLIC HEALTH INTERVENTIONS AND · Process Evaluation of the Adolescent Social Action Program in New Mexico whereas internal evaluators are assumed to have better insight

l04 Process Evaluation for Public Health lnterventions and Research

FIGURE 4.5. THEORY OF CHANGE MODEL FOR ASAP

CU RRICULUM I MPLEMENTATION.

tmplementation TheorY

*Facilitators completely implementcurriculum (11, 12, 13, 14)

[overlap here with Figure 4.4]

Program TheorY

Programgainscooperation/supportschoolteachersallowtimeinclassroomsfrom schoot administratiJ;';i;;"rr / to, recruitment; schools have low number

/ of comPeting Programs (8' 9' 10)

/Site coordinators recruit students fronL---7 Students gain parental permission & sign up

classrooms ./Site coordinators make efforts to boostl----- Students attend 6-week curriculum sessions

attendance with remind"e;;' .;u;, """"'

,/ {s' s' l)transportation, etc. /

y',Patients/inmate5areavailableforinterviews

(11, 12,1 3, 14) I

Students particiPate fullY (12)

IStudents determine a social action proiect at

,r/ t" end of 6-week curriculum

,/Site coordinators provide support to L-----istudents attend, participate in, & complete

tiro""t group foi sociur u.iiln proiects / social action proiect (5' 1 5)

(6,15) /Site coordinators provide booster session

/ ' students attend & participate in booster

& contacts & invite students at 6-month session (6' 15)

postcurriculum; they.also provide Itransportation &faciritatesession (6, 1s)

il:::H$:i:x,l.J:i:;:L?l:;?::.,& coping skills) for at least 15 months (4)

I

+

Students decrease in risky behaviors (e'g''

smoking, alcohol, & drug consumption) for

at least 15 months (4)

ith the list of instruments in Table 4'1 '

Page 25: EVALUATION PUBLIC HEALTH INTERVENTIONS AND · Process Evaluation of the Adolescent Social Action Program in New Mexico whereas internal evaluators are assumed to have better insight

ProcessEvaluationoftheAdolescentSocialActionPrograminNewMexicolo5

implementation. Many of the instruments measured the same component, and

triangulation was desirable, but given limited resources, it would have been more

effective to have more instruments or sections of instruments to measure a greater

proportion of theoretical components'

Using Weiss's theoretical framework (1998) has been helpfui to us in our

process eiraluation efforts. We recently developed a methodology for working

.to..ty with investigators and program designers in the early stages of planning

and program design. We use implementation and program theory logic models'

suclras those shown in Figures 4.4 and 4'5, as a framework to

. Articulate the program and implementation theory

. Identify where the theory components can be found in the implementation

. Identify which components are critical to program success

. Articulate what assumptions are held about the relationship between theory

and outcome. Identify which program components require tracking

In this way we can be more efficient about collecting data. Included in this

planning process are schedules for pilot testing and feedback, as well as discussions

^bort tirirg for feedback, in order to make the feedback most useful for mid-

course corrections.

In summary, the lessons learned from this study are important for the future

practice of process evaluation. First, evaluators need to gain from investigators

u.rd p.og.^- staff membe rs supportfor and.'interest in the process evaluation' Their

,,rppt.t fr.tps ensure the timely and accurate collection of process evaluation mea-

,r.., ^rd it helps ensure that the evaluation continues despite preliminary outcome

data findings. Se.o.rd, evaluators working with staff members and investigators

need to build their capacitltfor understandingthe value of process evaluation' Such

a process might include discussions early on in the evaluation and implementa-

tion process io help everyone involved in the project understand the important

and interconnected role that all phases of evaluation-f61621ive, process, and

outcome-have in a full and insighdui undelstanding of interventions. Evaluations

are often broken into the familiar parts of formative, process, and outcqme, but for

evaluation to be most effective, greater emphasis must be placed on the fact that

these parts are anintegrated whole evaluation. Such an understanding will go a

Io.rg *uy to*ard resolving the often inadequate resources that are directed toward

the"design, collection, and analysis of process as we[ as formative evaluation data'

Fina]ly, plannizg is essential in successful evaluations, including the process com-

ponent;and inour experience it is extremely helpful for program designers and

evaluators alike'

Page 26: EVALUATION PUBLIC HEALTH INTERVENTIONS AND · Process Evaluation of the Adolescent Social Action Program in New Mexico whereas internal evaluators are assumed to have better insight

Process Evaluation for Public Health lnterventions and Research

r06

Conclusions

Theprocessevaluationdescribedinthischapterrepresen-tsanambitiousunder.taking for il1e arTlounr oI- resources it *as a-llor't"d , an-d it reflects the state of knOwl-

edgeandexperienceaboutprocessevaluationthatexistedinthemid-1990s.Itis

_, through these types of projects that we can be self-reflective and ta'ke away lessons

that can i*p.o* the knowledge base from which future process evaluations will

be designed'r/Vhat follows in this concluding section is a list and discussion of ( 1) the stre ngths

of this process evaluation and (2) the limitations and opportunities for learning.

Strengths

The existmce oif a process eaaluat'ion.It is significant, given the level of knowledge about

process evaiuation at the time, that some funding for process evaluation was pro-

vided by the NIAAA and that the principal investigators had the wisdom to plan

for and allocate these resources to hire a process evaluation consultant. In addi-

tion, the design represented a praiseworthy attempt to be conceptually thorough.

Program personnel did an exceilent and thoroughjob of collecting and managing

the voluminous amount of data. Response rates were extremely high (see Table 4. 1)

for most instruments. Data were organized and easy to access for analysis'

The multipurpose design of instrumatts. A second strength of this process evalua-

tion is that instruments were designed to provide both formative and summative

data. For example, the attendance logs were intended to be used to provide in-

formation on participation for formative purposes, allowing quick response to im-

prove poor attendance by site coordinators. Attendance data also could have been

used to calculate the dose response rate for each student, correlating this rate with

the alcohol use behavior outcome measures'

Tiangulation. A third strength of this process evaluation is the ability to tri-

angulate data from different instruments, data sources, samples, and types of data'

Someofthetriangralationwasaccomplishedbyusingbothqualitativeandquan-titative data. other triangulation was undertaken by using two or more qpes of

informants. For example, the observation checklist was designed to estimate the

fidelity of facilitators io the curriculum by an independent observer during one

session. Facilitator checklist data and student comment sheets for the same session

were compared with observer data to determine whether facilitators were accu-

rately representing their coverage of curriculum components. It was the ability to

triangulate the data from the checklists, comment sheets, and observations that

enabled facilitators to determine that midcourse corrections were needed'

Page 27: EVALUATION PUBLIC HEALTH INTERVENTIONS AND · Process Evaluation of the Adolescent Social Action Program in New Mexico whereas internal evaluators are assumed to have better insight

Process Evaluation of the Adolescent Social Action Program in New Mexico 707

Limitations and Opportunities for Learning

Insuficient planningfor the anafusis process. Despite the ambitious design and allo-

cation of some resources to process evaluation, the program did not dedicate

sufficient attention to planning for the analysis process so that it would pro-

vide useful and timely feedback to the program. This could have been because

human resources were mainly focused on the collection and implementation of

outcome data. For example, the process evaluation consultant was not brought

on board until well into the grant period. When the consultant was brought on,

data were already being collected without a plan for the management and anaiy-

sis of the data. No one anticipated how much data there would be, how long

the data analysis would take, or how labor-intensive it would be. Therefore,

much of the process evaluation data were not analyzed and synthesized in a

time frame that would have enabled the program to make corrective changes

to improve the program's implementation. For example, if the data regarding

facilitator implementation had been analyze d throughout the proje ct rather

than.just at the end, the training process could have been improved to increase

hdelity to the curriculum (Helitzeq Yoon, Wallerstein, and Garcia-Velarde,

2000).

Instfficimt understanding, interest, and appreciatioz. Although the process evaluation

design was thoughtful, there was only partial appreciation for the value of process

evaluation and the contribution that its results could make. This lack of knowl-

edge about the various ways the process evaluation could be used resulted in in-

su{hcient attention being paid to it during the course of the project. For example,

weekiy meetings of the project staff often did not include discussions of process

evaluation data, and rarely were requests made for data to be analyzed for a

specific purpose. Also, the process evaluation expert was paid only 3 percent of

her salary to work on the project, which provided insu{ficient time for the thought-

ful incorporation of process evaluation resuits'

Process eualuation data were not being used to answer the question WhSt?When the

outcome data suggested that there were no intervention effects, process evaluation

data could have been used to understand whether or wlgt there was program fail-

ure and whether or wlgt the theoretical assumptions about the curriculum were not

sound. Fbr example, once we realized that the facilitators were impleriiinting some

parts of the curriculum more consistently than other parts, we could have used the

qualitative data from facilitator logs to try to understand why the facilitators seemed

to have more dilficulty with components requiring role-playing behavior than with

those involving the introduction and review of content areas'

Limited use of qualitatiue data. Most of the qualitative data were not analyzed.

These data were a rich source of information (for example, participant contact

Page 28: EVALUATION PUBLIC HEALTH INTERVENTIONS AND · Process Evaluation of the Adolescent Social Action Program in New Mexico whereas internal evaluators are assumed to have better insight

l08 Process Evaluation for Public Health lnterventions and Research

documentation sheets were the logs of site coordinators working on social action

projects) and could have provided Lsight into unintended outcomes. Howeveq the

..uil y of qualitativ. dutu i, that it can be unwieldy and that it requires a long time

to unulyrc properly. Qualitative analysis requires not only adequate time and

human resources devoled to it but also personnel who possess the appropriate

research skills. In the case of ASAR the human resources devoted to research were

not adequate to be extended to most of the qualitative analyses'

References

Chassin,L.,Presson,C.C.,Sherman,S'J',andSteinberg'L'AdolescentSmokelessTo-bacco IJse: Future Research Needs.,, NCI Monograph,I989, 4 101_105.

Floyd, D. L., Prentice-Dunn, S', and Rogers, R' W' 'A Meta-Analysis of Research on Protec-

tion Motivatio nTheory.'; Jo'urnal d Appltud Slcial Pycholog' 2000' 30(2)' 407429 '

Green, L. W., and Lewis, E M. Measuremmt and Eualuation in Health Education and Health Promo-

hoz. Mountain View, Calif': Mayfield, 1986'

Helitzer,D.L.,Yoon,S.,Walle.stein,N.,andGarcia-Velarde,L...TheRoleofProcessEva].uation in the Training of Facilitators for an Adolescent Health Education Program'" Jour-

nal of School Health,2000,7(4)' l4l-l+7 '

Helitzeq D. L., and others. "Process Evaluation in a Multisite, Primary obe sity-Prevention

Trial in American Indian schoolchildren." American JournaL of clinical JVutiion, 1999' 69

(supp.), B165-8245.

Jacobson, M., Atkins, R., and Hackeq G. The Boo<e Merchants: The Inebiating of Amrica.,Nash-

ington, D.C.: Center for Science in the Public Interest' 1983'

McGraw, S. A., and others. ..Design of Process Evaluation Within the Child and Adolescent

Trial lor cardiovascular Health (CAICH)." Health Educat;ion @Larterll' 1994, 2 (supp )'

s5-26.Microsoft Applications, Microsofi Excel, 1995' Software'

Miles, M. B., and Huberman, A.M. @Lalitatioe Data Anallsis: An Expanded sourcebook (2nd ed')'

Thousand Oaks, Calif.: Sage, 1994'

Rippetoe,PA.,andRogers,R.W..,Eff....ofComponentsofProtection.MotivationThe.ory on Adaptiv. ariMdaduptive coping with a Health Threat." Journal of Personaliry and

Socinl Ps2cholog, l9B7 , 52(3),596-604'Rissel,C.E.,andothers...Empowerment,Alcohol,Eighth-GradestudentsandHealth

Promotion." Jo urnal of Alcohol and Drug Edueation' 1996' 41(2)' 1 05- 1 i 9'

Rogers, R. w ,,6ogniti* "rra

physiologrial processes in Attitudinal Change: A Revised

Theory of protJction-Motivarion.,, InJ. Cacioppo and R. Petty (eds.), Social Ps,choph,sioL

ogr. New York: Guilford Press, 1983'

Rogers, R. W. ..Changing Health.Related Attitudes and Behavior: The Role of Preventive

Health Psychol"gy." h R' McG\m,J' Maddox, C' Stoltenbery and R'J' Harvey (eds')'

Interfaces in Pslchitng. Lubbock: Texas Tech University Press' 1994'

Rog..r, R. W, Decknll C.W', and Mewborn, C' R 'An Expectancy-Value Theory

ApproachtotheLong-TermModificationofsmokingBehaior.,'JournalofClinicaLPslchologlt, 197 B, 3 4, 5 62-5 66'

Page 29: EVALUATION PUBLIC HEALTH INTERVENTIONS AND · Process Evaluation of the Adolescent Social Action Program in New Mexico whereas internal evaluators are assumed to have better insight

process Evaluation of the Adolescent social Action Program in New Mexico 109

Steckler, A., and others. "Toward Integrating Qualitative and Quantitative Methods: An

Introduction." Health Education @LarterLy, 1992, 19, 1-9'

Wallerstein, N., and Bernstein, E. "Empowerment Edr-rcation: Fleire's Ideas Adapted to

Health Education." Health Education @Larterl2, i9BB, 15(4), 379-39+'

Wallerstein, N., and Sanchez-Merki, V "Freirian Praxis in Health Education: Research

Results lrom an Adolescent Prevention Program." Health Education Research, 1994, 9(1),

105 I 18.

Wallerstein, N., and Woodall, G. ASAP Projut Finat Reportfor NIAAA. AJbuquerque: University

of New Mexico, 2000.

Weiss, C. H. EuaLuation. Englewood Cli{Is, NJ.: Prentice Ha1l, 1998'