developmental disabilities program independent evaluation (ddpie) project ucedd meeting –...

Post on 21-Dec-2015

216 Views

Category:

Documents

0 Downloads

Preview:

Click to see full reader

TRANSCRIPT

Developmental Disabilities Program Independent Evaluation (DDPIE) Project

UCEDD Meeting – Technical Assistance Institute May 31, 2007

Lynn Elinson, Ph.D.Project Director

Developmental Disabilities Program Independent Evaluation (DDPIE) Project

Also known as “ADD Independent Evaluation”

Purpose of PowerPoint

To understand the background and progress of the ADD independent evaluation

To obtain a background and context for giving feedback on ADD independent evaluation materials

PowerPoint Outline

1. Background of ADD Independent Evaluation A. Purpose of the DDPIE Project B. Challenges

2. Research design3. Project implementation

A. Overview B. Project activitiesC. Evaluation toolsD. Validation

4. Seeking individualized input5. Progress and timing

1. Background

A. Purpose of the DDPIE Project

Demonstrate impact of DD Network programs on:– Individuals – Families– Service providers– State systems

Provide feedback to ADD to help improve the effectiveness of its programs and policies

Promote positive achievements of DD Network programs by “storytelling”

Promote accountability to the public

Why the independent evaluation?

In 2003 ADD conducted a Program Assessment Rating Tool (PART) self-assessment under OMB guidance.

PART is a series of questions designed to provide a consistent approach to rating programs across the Federal Government.

PART has four parts: (1) Program Purpose & Design; (2) Strategic Planning; (3) Program Management; and (4) Program Results.

PART 4 asks whether an agency has conducted an independent evaluation of sufficient scope and quality to indicate that the program is effective and achieving results?

ADD answered “no” which lowered overall score.

Challenges

Each UCEDD program is unique.

Challenge is to develop performance standards that: are relevant to all UCEDD programs; capture the differences among the programs

(variability); and will be useful to ADD in demonstrating impact.

2. Research design

Design Considerations

PART prefers experimental or quasi-experimental research designs

The structure of the ADD programs does not lend itself to conducting randomized trials or pre- and post-tests.

Research Design: Standards-Based Evaluation

NOT a randomized control trial or quasi-experimental design

IS a standards-based evaluation to:- Set national standards- Determine levels that characterize extent to which national standards are being met - Determine impact DD Network programs (and collaboration among programs) are having on people with developmental disabilities, family members, State systems, and services providers

Reporting at national level

Data will be collected on individual programs and rolled up to national level.

Independent evaluation will NOT be comparing programs to one another

Independent evaluation will NOT replace MTARS, which is specific to individual programs.

2 Types of Standards

Evidence-based Consensus-based Performance standards for DDPIE are

consensus-based Performance standards will be developed for

each DD Network program and collaboration among the three DD Network programs

Key assumptions for designing performance standards

State programs vary on their level of performance across the standards.

Consistently high performance across the standards is related to better outcomes.

Consistently low performance across the standards is related to poor outcomes.

Research design: seeks input and participation from stakeholders

Seeks input from: Project Advisory Panel DD Network Program Working Groups All State programs Validation Panels The public

Role of Advisory Panel

To provide balance, impartiality, and expertise

To provide advice on: DDPIE process Benchmarks, indicators, performance standards, and

performance levels Data collection protocols Pilot study Synthesis of findings and recommendations

Composition of Advisory Panel

Self-advocates Family members Representatives from 3 programs – Richard

Carroll from Arizona UCEDD Child/disability advocates Evaluation expert Federal representative (for PAIMI evaluation)

Working Groups

4 Working Groups (P&A, UCEDD, DD Council, Collaboration)

Process: In-person and telephone meetings Role:

- To assist Westat in understanding programs - To provide feedback on benchmarks, indicators, performance standards

UCEDD Working Group members

Carl Calkins Kansas City, MO

Tawara Goode Washington, DC

Gloria Krahn* Portland, OR

David Mank Bloomington, IN

Fred Orelove* Richmond, VA

Fred Palmer Memphis, TN

Lucille Zeph Orono, ME

*Collaboration Working Group

3. Project implementation

A. Overview

Phases of DDPIE Project

DDPIE will be conducted in 2 phases.- Phase 1 – development and testing of evaluation tools (measurement matrices and data collection protocols)- Phase 2 – full-scale evaluation

Westat was contracted by ADD to implement Phase 1. - Project began September 30, 2005- End of contract – September 29, 2008

Phase 2 will be funded upon completion of Phase 1.

B. Project activities

Steps in Phase I

Construct evaluation tools (measurement matrices and data collection protocols) that contain performance standards and performance levels

Conduct Pilot Study to test evaluation tools (measurement matrices and data collection protocols)

Revise evaluation tools

C. Evaluation tools

2 types of evaluation tools

Measurement matrices, which include:

- Key functions, benchmarks, indicators, performance standards

- Performance levels

Data collection protocols

Definitions of key terms in measurement matrices

Key functions Benchmarks Indicators Performance standards

- Outcome performance standards

- Program performance standards

Logic model/format for measurement matrices

Ben

chm

arks

Indi

cato

rsP

erfo

rman

ce S

tand

ards

Key Functions

Key Functions

Groups of activities carried out by DD Network programs

Cover all aspects of program activity 5 UCEDD key functions 1st four key functions identified by Working Group

(core functions in DD Act) Governance and Management – Relevant to other

four key functions Benchmarks, indicators, and performance standards

are being developed for all key functions.

UCEDD Key Functions

A. Interdisciplinary pre-service training and continuing education

B. Conduct of basic and/or applied research

C. Provision of community servicesD. Dissemination of informationE. Governance and management

Benchmarks

Broad, general statements Set bar for meeting expected

outcome(s) of each key function About 20 UCEDD benchmarks 3-4 benchmarks for each key function

Indicators

Identify what gets measured to determine extent to which benchmarks and performance standards are being met

4 types of indicators: outcome, output, process, structural

Will guide the development of data collection instruments

Performance standards

Criterion-referenced (measurable)Consensus-based 2 types:

- Outcome performance standards

- Program performance standards

Outcome performance standards

Linked to expected outcomes of each key function

Answer the questions:- Were the expected outcomes

met? - To what extent?

Program performance standards

What the program should achieve, have, and do to effectively:- meet the principles and goals of the

DD Act; and - have an impact on people with developmental disabilities, family members, State systems, service providers

Program performance standards (continued)

Linked to the structures, processes, and outputs of UCEDD program

Answers the questions:

- What structures should be in place to carry out UCEDD network key functions? What should they be like?

- What processes should be used? What should they be like?

- What should the UCEDD network produce? What should products be like? To what extent should they be produced (e.g., how often, how many)?

D. Validation

Overview of validation

There is no “gold standard” for an effective UCEDD, so another approach needs to be used to identify performance standards.

The ADD independent evaluation uses a consensus approach. This implies participation in the process and validation from a

wide variety of stakeholders. There will be several opportunities for validation throughout the

development of performance standards. Stakeholders hold a variety of perspectives and, therefore, may

not always agree with one another.

Validation approach for DDPIE project

Consists of obtaining input, feedback, and consensus Consists of validating measurement matrices (indicators and

performance standards) and data collection instruments Is a multi-step process Provides validation opportunities to several types of

stakeholders (e.g., consumers, family members, program representatives, advocates, evaluation experts)

Provides opportunities for validation at different points in the process

Opportunities for validation

Working Group process Advisory Panel meetings State programs (at TA meetings, by

telephone, in writing) Validation Panel process OMB process Pre-test and pilot study

Validation Panels

There will be 4 Validation Panels (UCEDDs, P&As, DD Councils, Collaboration).

Process

- Telephone call orientation

- “Paper” approach (not face-to-face) – accommodation will be provided

- Opportunity for discussion by telephone

Criteria for Validation Panel selection

Stakeholder groups (e.g., people with developmental disabilities, family members, advocates, programs, service providers)

Researchers

Criteria for Validation Panel selection (continued)

Understands consumer needsUnderstands DD Network programsDiverse composition (gender,

race/ethnicity)Mix of junior and senior program staffUrban and rural representation

Focus of Validation Panel process

Will achieve consensus Formal process Builds in objective methodology (e.g., criteria

for eliminating and accepting indicators and performance standards)

OMB approval process is another form of validation

OMB approval process results from the Paperwork Reduction Act

Act is administered by Office of Management and Budget (OMB)

Purpose of Act is to ensure that information collected from the public minimizes burden and maximizes public utility

All Federal agencies must comply

OMB approval process (continued)

When contemplating data collection from the public, Federal agencies must seek approval from OMB.

Must submit an OMB package consisting of description of study and data collection effort, an estimate of burden, and data collection instruments.

Approval process consists of making data collection instruments available for public comment in the Federal Register.

ADD will be submitting an OMB package; all interested parties will have opportunity to comment during public comment period.

Pre-test and Pilot Study – additional form of validation

Data collection protocols will be pre-tested in one state.

A pilot study will be conducted in up to 4 states.

Pilot study states will be chosen randomly. Pilot study will test reliability and validity of

measurement matrices and feasibility of data collection.

4. Seeking individualized input

Opportunities for individualized input

UCEDD TA meeting (May 31, 2007)

- Distribution of draft benchmarks, indicators, and a few examples of performance standards

- Small group discussions facilitated by AUCD

Telephone meetings scheduled in June and July In writing

Small Group Discussions at UCEDD Technical Assistance Meeting (May 31, 2007)

Westat will:- Distribute draft performance standards on UCEDD

Network and Collaboration- Review organization of materials- Describe feedback process for individual UCEDD

programs- Answer questions on process for feedback

UCEDD programs will: - Continue to meet in small groups to discuss the materials (facilitated by AUCD)- Report out in a large group on first impressions

Type of Input Sought

Benchmarks and indicators: Are they the concepts that need to be addressed?

Benchmarks and performance standards: Do they represent what the programs should be achieving/should have/should do in order to be effective in meeting the principles and goals of the DD Act and have an impact on people with developmental disabilities, families, State systems, and service providers?

Indicators: Which seem the most important and feasible to measure? Which could be eliminated?

If not these, then what?

5. Progress and Timing

Progress to Date

Meetings with ADD, head of national associations, TA contractors – November, 2006

Site visit to programs in one state – December, 2006 Review of background materials (provided by ADD; Working

Groups; national websites; other) – October, 2005 – February, 2007

Meetings with Working Groups – March, 2006 – September, 2006

Meetings with Advisory Panel - March, 2006, October, 2006, March, 2007

Synthesis of all information by Westat – September, 2006 to February, 2007

Draft benchmarks, indicators, performance standards – February, 2007

Upcoming DDPIE Project Milestones

Feedback from UCEDD Working Group April – May, 2007

UCEDD TA meeting May 31, 2007

Feedback from all UCEDD programs June - July, 2007

UCEDD Validation Panel Sept. – Dec., 2007

DD Council Validation Panel Oct. – Jan., 2008

P&A Validation Panel Nov. – Feb., 2008

Collaboration Validation Panel Feb. – April, 2008

DDPIE Project Milestones (continued)

Data collection instruments June, 2008

Measurement matrices July, 2008

Final report (with evaluation tools) Sept., 2008

OMB Comment Period

Pilot Study New contract

top related