fea survey about fcps’ teacher performance evaluation ... · fcps purchased the tpep system from...

25
FEA Survey about FCPS’ Teacher Performance Evaluation Program (TPEP) June 2016

Upload: others

Post on 17-Mar-2020

2 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: FEA Survey about FCPS’ Teacher Performance Evaluation ... · FCPS purchased the TPEP system from James H. Stronge, the author of the Guidelines. It has not been significantly revised

FEA Survey about FCPS’ Teacher Performance Evaluation Program (TPEP)

June 2016

Page 2: FEA Survey about FCPS’ Teacher Performance Evaluation ... · FCPS purchased the TPEP system from James H. Stronge, the author of the Guidelines. It has not been significantly revised

2

Contents Executive summary ................................................................................................................................................................. 3

Introduction and background ................................................................................................................................................. 3

Respondent demographics ..................................................................................................................................................... 4

Respondents’ experience with summative TPEP evaluations ................................................................................................ 5

SMARTR goals ..................................................................................................................................................................... 5

Student surveys ................................................................................................................................................................... 6

Observations ....................................................................................................................................................................... 7

TPEP time requirements ....................................................................................................................................................... 12

Self-Assessment ................................................................................................................................................................ 13

SMARTR goals ................................................................................................................................................................... 13

Document log .................................................................................................................................................................... 14

Structured interview ......................................................................................................................................................... 14

Other TPEP activities ......................................................................................................................................................... 15

Respondents’ attitude toward TPEP ..................................................................................................................................... 15

Comments ............................................................................................................................................................................. 17

Summary ............................................................................................................................................................................... 17

Recommendations ................................................................................................................................................................ 18

FCPS and FEA ..................................................................................................................................................................... 18

Virginia Department of Education and VEA ...................................................................................................................... 18

Appendix: Respondents’ comments: selected responses .................................................................................................... 20

Page 3: FEA Survey about FCPS’ Teacher Performance Evaluation ... · FCPS purchased the TPEP system from James H. Stronge, the author of the Guidelines. It has not been significantly revised

3

Executive summary Since the 2012 adoption of the Teacher Performance Evaluation Program (TPEP) used to evaluate teacher scale

employees, the Fairfax Education Association has heard from many members about problems with it. We conducted a

survey of members in May, 2016 in an effort to learn about our members’ experience with TPEP, how successful it has

been, and whether any revisions to the system are indicated.

We found that the system is complex and time-consuming, both for educators being evaluated and for the

administrators who have to do the evaluations.

The mean amount of time that it took a respondent to complete the activities for a summative evaluation was

24.1 hours; the median was 15.8 hours.

Compared to the previous evaluation system (which took 2-3 hours), it adds approximately two days of work to

an educator’s workload.

The additional time demands on administrators are likely responsible for failures in making enough observations,

observing complete lessons, and giving educators useful feedback:

The TPEP manual requires two observation data points; 14% of summative evaluations did not meet the

requirement.

52% report that they did not have a formal observation that lasted an entire lesson.

44.4% of teachers reported having formal observation of non-instructional activities.

Only 45% report receiving observation feedback that includes specific and useful suggestions or guidance for

improvement.

The system is not regarded as useful, worth the time spent, or reliable.

Only 28.8% agree that TPEP provides useful feedback.

Only 16% agree that time spent on TPEP is time well spent.

Only 19.9% agree that TPEP has inter-rater reliability, and that outcomes are independent of who their evaluator

is.

FEA recommends that FCPS reconvene the Evaluation Taskforce to revise the TPEP system to reduce the complexity and

time required. We recommend that the VA Department of Education revise the Guidelines for Uniform Performance

Standards and Evaluation Criteria for Teachers, Principals and Superintendents to allow further simplification of

evaluations and to eliminate or greatly reduce the reliance on Value Added Measures of student progress such as FCPS’

SMARTR goals.

Introduction and background The Code of Virginia (§ 22.1-253.13:5 B.) mandates evaluations for educators “consistent with the Guidelines for

Uniform Performance Standards and Evaluation Criteria for Teachers, Principals and Superintendents” (hereafter the

Guidelines) which is promulgated by the Virginia Board of Education. The Code also requires that “evaluations shall

include student academic progress as a significant component”, known as a Value Added Measure (VAM), which was

added to the law to comply with eligibility requirements for a federal Race to the Top (RTTP) grant (which Virginia did

not ultimately receive).

The Code (§ 22.1 295. C.) requires that teachers with continuing contract status have formal (known as “summative” in

FCPS) evaluation every three years (or more frequently if deemed necessary by the principal), and informal evaluations

(known as “formative” in FCPS) in the other years. Provisional educators are evaluated formally every year until they

achieve continuing contract status.

Page 4: FEA Survey about FCPS’ Teacher Performance Evaluation ... · FCPS purchased the TPEP system from James H. Stronge, the author of the Guidelines. It has not been significantly revised

4

The latest revisions to the Guidelines became effective July 1, 2012. The seven Performance Standards found in the

current TPEP system are mandated by the current Guidelines.

In 2011-12, FCPS convened a Teacher Evaluation Task Force to draft recommendations for a new evaluation tool that

complied with the revised Guidelines. FEA members on the Task Force argued that Standard 7: Academic Performance

should be weighted equally with the other six Standards, or at most at 20% of the total evaluation points. The VA

Department of Education required a 40% weight.

FCPS purchased the TPEP system from James H. Stronge, the author of the Guidelines. It has not been significantly

revised since it was adopted in 2012.

The Fairfax Education Association (FEA) has lobbied FCPS over the years since the introduction of TPEP, asking that FCPS

study the validity and reliability of the new evaluation system. We have been especially concerned with inter-rater

reliability, which is the degree to who which an evaluation rating is independent of the person doing the evaluation. No

such study has been undertaken.

Respondent demographics A total of 426 people responded to the survey. Six respondents identified themselves as administrators. They were not

asked about their experience with being evaluated with TPEP, but were asked about their thoughts about TPEP and

invited to make comments.

We asked respondents about the type of work they do in order to parse the data appropriately (see below). “Other

employee evaluated with TPEP” includes counselors, librarians, psychologists, etc.

ES, 54.8%

HS, 28.0%

MS, 13.9%

Other, 3.3%

Respondents by level

Page 5: FEA Survey about FCPS’ Teacher Performance Evaluation ... · FCPS purchased the TPEP system from James H. Stronge, the author of the Guidelines. It has not been significantly revised

5

Respondents’ experience with summative TPEP evaluations The survey asked several questions about how various parts of the TPEP process played out in the respondents’ most

recent summative evaluations.

SMARTR goals According to the TPEP manual,

Goal setting begins with identifying where students are in relation to what is expected of them. Then, in

collaboration with their evaluator, teachers set specific, measurable goals based on both the demands of

curriculum and the needs of the students. (emphasis added)

Over the past few years, FEA has heard complaints from members that administrators had dictated goals, and that the

development of goals had not been collaborative. While this would seem to violate the procedure set forth in the

Manual, when FEA contacted HR about this problem, we were told that principals were free to require goals, even if

they were not reasonable or attainable.

We wanted to quantify this problem, so we asked how goals are set:

Administrator, 1.4%

Other employee evaluated with

TPEP, 11.8%

SBTS, 2.1%

Teacher, 84.7%

Respondents by job type

Page 6: FEA Survey about FCPS’ Teacher Performance Evaluation ... · FCPS purchased the TPEP system from James H. Stronge, the author of the Guidelines. It has not been significantly revised

6

18.1% of respondents reported that goals are primarily the work of administrators, and fully 7.7% (1 out of 13) indicated

that administrators are fully setting the goals.

We also found that 29.6% of teachers get little or no input from administrators. Adding those whose goals were set by

administrators alone and those set by teachers alone, it seems that 36.3% of goals are being set without significant

collaboration between teacher and administrator.

Student surveys Student surveys can be used as a data source by middle and high school teachers; they can be used voluntarily or

required by the evaluator. We asked MS and HS teachers about their use of student surveys, and found that 71% do not

use surveys. Of those using surveys, the majority do so voluntarily.

Decided by me or my team with little/no

input from administrators, 29.6%

Decided by me or my team with some input from administrators, 53.3%

Decided by administrators

with some input from me or my

team, 9.4%

Decided by administrators with little/no

input from me or my team,

7.7%

Determination of SMARTR goals

Voluntarily used a student survey

18%

Required to use a student survey

11%

Did not use student surveys

71%

MS and HS teachers' use of student surveys

Page 7: FEA Survey about FCPS’ Teacher Performance Evaluation ... · FCPS purchased the TPEP system from James H. Stronge, the author of the Guidelines. It has not been significantly revised

7

Observations We asked several questions about observations. To measure the number of observations, we asked separately about the

number of formal and informal observations, broken down further between observations of instruction and

observations of other activities. We added the numbers together to get information on the total number of

observations.

We were surprised to learn that 5.1% of respondents received summative evaluations without ever being observed

during the process. An additional 8.9% had only one observation.

The TPEP manual requires a minimum of two observation Data Sources; yet 14% of summative evaluations do not meet

this requirement. The mean number of total observations was 5.4 and the median was 4.

Looking just at formal observations, 10.9% had no formal observations at all. The mean was 3.3 and the median was 2.

5.1%

8.9%

15.2% 15.5%

13.0%

8.7%8.2%

4.8%3.9%3.9%

2.7%1.7%1.9%1.4%

0.7%1.0%

3.4%

0.0%

2.0%

4.0%

6.0%

8.0%

10.0%

12.0%

14.0%

16.0%

18.0%

0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 > 15

% o

f al

l res

po

nd

ents

Total number of observations (formal & informal)

Total observations - all respondents

10.9%

21.0%

23.2%

15.2%

9.7%

7.5%

5.3%

1.4% 1.4%0.5% 0.5%

3.4%

0.0%

5.0%

10.0%

15.0%

20.0%

25.0%

0 1 2 3 4 5 6 7 8 9 10 > 10

% o

f al

l res

po

nd

ents

Number of formal observations

Total formal observations - all respondents

Page 8: FEA Survey about FCPS’ Teacher Performance Evaluation ... · FCPS purchased the TPEP system from James H. Stronge, the author of the Guidelines. It has not been significantly revised

8

Looking at the number of informal observations that resulted in written feedback, 32.2% had no informal observations.

The mean was 2.1 and the median was 2.

Some FEA members have reported being observed by multiple observers, so we asked about that.

32.2%

17.2%

21.5%

10.9%

6.8%

1.7%3.6%

1.5% 1.7% 1.0% 0.7% 1.2%

0.0%

5.0%

10.0%

15.0%

20.0%

25.0%

30.0%

35.0%

0 1 2 3 4 5 6 7 8 9 10 > 10

% o

f al

l res

po

nd

ents

Number of informal observations

Total informal observations - all responses

none4%

1 evaluator66%

2 evaluators23%

3 evaluators5%

4 evaluators1%

5 evaluators1%

Number of evaluators making observations

Page 9: FEA Survey about FCPS’ Teacher Performance Evaluation ... · FCPS purchased the TPEP system from James H. Stronge, the author of the Guidelines. It has not been significantly revised

9

One of the goals of the TPEP process is to provide educators with feedback to guide improvement. We asked about

whether observation feedback included useful guidance and/or suggestions for improvement. Only 45% reported

receiving guidance and suggestions that was specific enough to be useful.

We also analyzed the observation data in greater detail for teachers only, in order to get a feel for the number of

observations made of instruction and non- instructional activities.

We found that 12.3% of teachers reported having no formal observation of instruction as part of their formative TPEP

evaluation.

I did not receive any suggestions

or guidance at all12%

I did receive specific and

useful suggestions/

guidance45%

I did receive suggestions/

guidance, but it wasn’t specific enough to be

useful19%

I didn’t receive any feedback on

areas for improvement

24%

Useful feedback from observations

Page 10: FEA Survey about FCPS’ Teacher Performance Evaluation ... · FCPS purchased the TPEP system from James H. Stronge, the author of the Guidelines. It has not been significantly revised

10

The TPEP manual states that

Although there is no specified duration of formal observations, it is highly recommended that the evaluator

remain for the amount of time necessary to observe a complete lesson with a lesson transition.

A number of FEA members have reported formal observations of parts of lessons and receiving feedback questioning

missing elements that were, in fact, present and would have been observed if the observer had only witnessed the

entire lesson. We wanted to learn the extent of this problem, so we asked a follow-up question about whether formal

observations of instruction lasted the entire lesson.

We found that 52% of teachers did not have any formal observations that lasted an entire lesson.

Since teachers’ main job is instruction, we have been surprised at the number of FEA member teachers who have

reported having formal observations of non- instructional activities, such as their participation in CLT meetings. We

asked about that, and we were surprised that 44.4% of teachers reported having formal observations of non-

instructional activities.

12.3%

31.9%29.7%

16.2%

3.1% 2.5% 3.1%

0.0% 0.3% 0.0% 0.3% 0.6%

0.0%

5.0%

10.0%

15.0%

20.0%

25.0%

30.0%

35.0%

0 1 2 3 4 5 6 7 8 9 10 > 10

% o

f te

ach

ers

Number of observations

Formal observations of instruction - teachers only

All of them did, 20.9%

Some of them did, 27.0%

None of them did, 52.0%

Did formal observation(s) of instruction last the entire lesson(s)?

Page 11: FEA Survey about FCPS’ Teacher Performance Evaluation ... · FCPS purchased the TPEP system from James H. Stronge, the author of the Guidelines. It has not been significantly revised

11

We also asked about the number of informal observations (with written feedback – otherwise they don’t count as a data

source). We were surprised that fully 33.6% of teachers reported no informal or mini observations of instruction during

their latest summative year.

55.6%

16.1%13.1%

5.6%3.1% 1.7% 0.3% 0.0% 0.8% 0.0% 0.6%

3.3%

0.0%

10.0%

20.0%

30.0%

40.0%

50.0%

60.0%

0 1 2 3 4 5 6 7 8 9 10 > 10

% o

f te

ach

ers

Number observations

Formal observations of non-instructional activities - teachers only

33.6%

22.3% 21.2%

12.4%

4.2%2.3% 2.5%

0.0% 0.0% 0.6% 0.3% 0.6%0.0%

5.0%

10.0%

15.0%

20.0%

25.0%

30.0%

35.0%

40.0%

0 1 2 3 4 5 6 7 8 9 10 >10

% o

f te

ach

ers

Number of observations

Informal observations of instruction - teachers only

Page 12: FEA Survey about FCPS’ Teacher Performance Evaluation ... · FCPS purchased the TPEP system from James H. Stronge, the author of the Guidelines. It has not been significantly revised

12

The majority of teachers (70.3%) reported no informal observations of non-instructional activities.

TPEP time requirements The new TPEP evaluation system takes considerably more time on the part of teachers than the previous system did and

is the time requirement is the main complaint of FEA members. We asked about the time required for various

components (see below) and added each respondent’s times to generate their overall time spent.

The mean amount of time spent completing the activities for a summative evaluation was 24.1 hours, and the median

was 15.8 hours. FEA members report that the previous system took about 2 hours. The current TPEP system has added

approximately two days of work to the teachers’ workload.

70.3%

14.7%8.2%

4.5%1.1% 0.6% 0.3% 0.0% 0.0% 0.0% 0.3% 0.0%

0.0%

10.0%

20.0%

30.0%

40.0%

50.0%

60.0%

70.0%

80.0%

0 1 2 3 4 5 6 7 8 9 10 >10

% o

f te

ach

ers

Number of observations

Informal observations of non-instructional activities - teachers only

0.0%

2.0%

4.0%

6.0%

8.0%

10.0%

12.0%

1 3 5 7 9 11 13 15 17 19 21 23 25 27 29 31 33 35 37 39 >40

% o

f re

spo

nd

ents

Number of hours spent on completing TPEP requirements

Total time spent on TPEP activities

Page 13: FEA Survey about FCPS’ Teacher Performance Evaluation ... · FCPS purchased the TPEP system from James H. Stronge, the author of the Guidelines. It has not been significantly revised

13

Self-Assessment The mean amount of time spent on self-assessment was 2.8 hours and the median was 2.

SMARTR goals The mean amount of time spent on developing SMARTR goals and analyzing data was 5.0 hours; the median was 3.

0.5%

36.0%

31.6%

13.5%

6.8%4.8%

2.9%0.2% 0.7% 0.2%

2.7%

0.0%

5.0%

10.0%

15.0%

20.0%

25.0%

30.0%

35.0%

40.0%

0 1 orless

2 3 4 5 6 7 8 9 10 ormore

% o

f re

spo

nd

ents

Number of hours spent completing the Self-Assessment form

Amount of time spent on Self-Assessment form

0.2%

14.6%

22.4%

16.8%

14.1%

10.0%

5.6%

1.2%3.4%

0.0%

11.5%

0.0%

5.0%

10.0%

15.0%

20.0%

25.0%

0 1 orless

2 3 4 5 6 7 8 9 10 ormore

% o

f re

po

nd

ents

Number of hours spent writing SMARTR goals and analyzing data

Amount of time spent on SMARTR goals

Page 14: FEA Survey about FCPS’ Teacher Performance Evaluation ... · FCPS purchased the TPEP system from James H. Stronge, the author of the Guidelines. It has not been significantly revised

14

Document log The mean amount of time spent on preparing the document log was 7.9 hours and the median was 4.

Structured interview The mean amount of time spent answering questions for the structured interview was 3.0 hours and the median was 2.

0.7%

9.4%

15.0% 15.2%12.8%

11.1%

6.3%

0.5%

4.6%

24.4%

0.0%

5.0%

10.0%

15.0%

20.0%

25.0%

30.0%

0 1 orless

2 3 4 5 6 7 8 10 ormore

% o

f re

spo

nd

ents

Number of hours spent preparing the document log

Amount of time spent on the document log

15.1%

30.2% 29.2%

10.5%

4.6% 4.1%2.4%

0.0% 0.7% 0.0%3.2%

0.0%

5.0%

10.0%

15.0%

20.0%

25.0%

30.0%

35.0%

0 1 orless

2 3 4 5 6 7 8 9 10 ormore

% o

f re

spo

nd

ents

Number of hours spent answering structured interview questions

Amount of time spent on structured interview

Page 15: FEA Survey about FCPS’ Teacher Performance Evaluation ... · FCPS purchased the TPEP system from James H. Stronge, the author of the Guidelines. It has not been significantly revised

15

Other TPEP activities The mean amount of time spent on other TPEP activities was 6.0 hours and the median was 3.

Respondents’ attitude toward TPEP We asked all respondents to give feedback on TPEP using Likert scale responses to positive statements about the

system. Administrators’ responses are included here.

The first statement was “TPEP provides useful feedback which encourages professional growth and leads to improved

student success.” Only 28.8% agree. 71.2% disagree. Among administrators, 50% agree and 50% disagree.

8.7%

15.1%

22.3%

11.4%

6.4%

12.4%

2.7%

0.0%

3.5%

0.2%

17.3%

0.0%

5.0%

10.0%

15.0%

20.0%

25.0%

0 1 orless

2 3 4 5 6 7 8 9 10 ormore

% o

f re

spo

nd

ents

Number of hours spent on other TPEP activities

Amount of time spent on other TPEP activities

Strongly agree, 1.4%

Agree, 27.4%

Disagree, 42.9%

Strongly disagree, 28.3%

TPEP provides useful feedback . . .

Page 16: FEA Survey about FCPS’ Teacher Performance Evaluation ... · FCPS purchased the TPEP system from James H. Stronge, the author of the Guidelines. It has not been significantly revised

16

The second statement was “Time spent completing TPEP activities is time well spent.” Only 16% agree. Fully 84%

disagree. Among administrators, 33% agree, while 67% disagree.

FEA members have reported has very different experiences depending on who their evaluator was. We asked whether

respondents agree that “I am confident in the inter-rater reliability of TPEP; evaluators use similar standards, and

evaluation outcomes do not depend on who the evaluator is.” Only 19.9% agree. 80.1% disagree. Among administrators,

50% agree and 50% disagree.

Strongly agree, 0.2%

Agree, 15.8%

Disagree, 41.8%

Strongly disagree, 42.1%

Time spent on TPEP activities is time well spent

Strongly agree, 1.7%

Agree, 18.2%

Disagree, 37.9%

Strongly disagree, 42.2%

Confidence in TPEP inter-rater reliability

Page 17: FEA Survey about FCPS’ Teacher Performance Evaluation ... · FCPS purchased the TPEP system from James H. Stronge, the author of the Guidelines. It has not been significantly revised

17

Comments We asked respondents two open ended questions: “How could the current TPEP system be improved?” and “What other

comments do you have about TPEP?” These questions generated hundreds of responses, most of which complained

about the complexity of TPEP and the time required. A sample of responses is included in the Appendix.

Summary The current TPEP evaluation system is complex and time-consuming, both for educators being evaluated and for the

administrators who have to do the evaluations.

The mean amount of time that it took a respondent to complete the activities for a summative evaluation was

24.1 hours; the median was 15.8 hours.

Compared to the previous evaluation system (which took 2-3 hours), it adds approximately two days of work to

an educator’s workload.

The additional time demands on administrators are likely responsible for failures in making enough observations,

observing complete lessons, and giving educators useful feedback:

The TPEP manual requires two observation data points; 14% of summative evaluations did not meet the

requirement.

52% report that they did not have a formal observation that lasted an entire lesson.

44.4% of teachers reported having formal observation of non-instructional activities.

Only 45% report receiving observation feedback that includes specific and useful suggestions or guidance for

improvement.

The system is not regarded as useful, worth the time spent, or being reliable.

Only 28.8% agree that TPEP provides useful feedback.

Only 16% agree that time spent on TPEP is time well spent.

Only 19.9% agree that TPEP has inter-rater reliability, and that outcomes are independent of who their evaluator

is.

The VA Department of Education’s Guidelines cite a number of reasons for implementing changes to the systems

prevalent prior to 2012:

. . . flaws in the current teacher evaluation process include:

• problems with the evaluation instruments themselves (e.g., subjectivity, low validity);

• issues related to time and resources;

• a tendency to focus on paperwork routines rather than improving instruction;

• an absence of standard protocols and practices in teacher practices;

• an absence of meaningful and timely feedback to teachers;

• inadequate administrator training;

• a lack of time to perform adequate evaluations;

• a lack of impact; and

Page 18: FEA Survey about FCPS’ Teacher Performance Evaluation ... · FCPS purchased the TPEP system from James H. Stronge, the author of the Guidelines. It has not been significantly revised

18

• a lack of constructive criticism on the evaluation that can be used to improve professional

practice and often are based on sparse evidence.

It is not clear that any of these problems have been ameliorated by the current TPEP system. If anything, the problems

of time and paperwork routines have been greatly exacerbated.

Recommendations

FCPS and FEA FCPS’ ability to modify its teacher evaluation system is constrained by the VA Department of Education’s Guidelines

document. However, FCPS can improve the system by reconvening the Evaluation Task Force and charging it with

streamlining the system and improving the validity and reliability. The Task Force should consider:

Studying the reliability and validity of TPEP, especially inter-rater reliability

Revising the Standards and the supporting Data Sources to eliminate substantial overlap. For example, a teacher

who needs to augment formative assessments of students can currently be down-rated in Standards 1, 2, 3, and

4 because all four Standards include “Analysis of data”, “Needs assessments and results”, etc. This overlap can

lead to a teacher’s receipt of an overall rating of “Ineffective” when modest effort will make the teacher

“Effective”. FEA staff have seen examples of this overlap problem in some members’ evaluations.

Providing better guidance on appropriate SMARTR goals, especially for educators whose primary duties involve

little, if any, instruction of students (such as counselors, librarians, and secondary SBTS). This should include

direction that successful completion of the goals must be a direct result of the employee’s own efforts, not

depending on the performance of other educators.

Ensuring that evaluators are familiar with the subject area being observed and evaluated

Requiring that at least one formal observation conducted in a summative year last an entire lesson

For teachers’ evaluations, providing evaluators with guidance that formal observations should be of instructional

activities

Requiring a minimum of two observations

Reducing the number/length of required paperwork, possibly including

o Reducing the total number of structured interview questions

o Streamlining the document log by reducing the number of artifacts and/or encouraging the use of

existing documents

Allowing educators to provide additional supporting evidence in response to evaluator concerns

Requiring evaluators to provide specific suggestions for improvement to accompany any “Ineffective” or

“Developing/ Needs Improvement” ratings

Providing an appeal process in cases of

o Inappropriate SMARTR goals being required by administrators

o Evaluations not conforming to standards set forth in the TPEP manual

Virginia Department of Education and VEA The onerous burden of the TPEP evaluation system is the direct result of requirements set forth in the VA Department of

Education’s Guidelines document. The Guidelines should be revised to:

Streamline the evaluation process, reducing the amount of paperwork and the time required to complete it

Reduce the overlap between Standards

Eliminate entirely, or greatly reduce the weight of the Value Added Measure (Standard 7). Peer reviewed

academic research has discredited the use of VAMs in teacher evaluation.

Page 19: FEA Survey about FCPS’ Teacher Performance Evaluation ... · FCPS purchased the TPEP system from James H. Stronge, the author of the Guidelines. It has not been significantly revised

19

VEA should lead this effort by:

Collecting data on the effects of the current Guidelines on evaluation systems statewide, as this survey has done

for Fairfax County Public Schools;

Lobbying for revisions to the Guidelines as outlined above.

Lobbying for a change in § 22.1-253.13:5 B. of the Code of Virginia to eliminate the requirement that all licensed

educators be evaluated by student progress – it just doesn’t make sense for some educators (e.g., counselors),

and academic research has shown it to be unreliable generally.

Page 20: FEA Survey about FCPS’ Teacher Performance Evaluation ... · FCPS purchased the TPEP system from James H. Stronge, the author of the Guidelines. It has not been significantly revised

20

Appendix: Respondents’ comments: selected responses

Suggestion for improvement/comment about TPEP Commenter Years evaluated

The current system is time consuming for teaches and overwhelming for administators who in some cases are dealing with 20+ staff on cycle and those are just summative numbers. The back and forth required with the online tool is not efficient. There is talk of all employees moving to the online system. We can barely keep up with teachers - I can't imagine going this method with support staff who do not utilize email or computers on a regular basis. The mid-year standard 7 rating is highly subjective since often the window for data collection and timeline for mid-years are not aligned.

HS Administrator

TPEP is one of the reasons I am seeking employment outside of FCPS. It is overly burdensome to both teachers and administrators. It is too time-consuming and complicated to be a constructive or meaningful form of feedback for teachers. The student performance metric is weighted too heavily in the overall calculation.

HS Teacher 2014 – 2015

Multiple administrators should provide feedback for each teacher.

ES Teacher 2012 – 2013 2013 – 2014 2014 – 2015

Admin accountability. If a source of information is used, it should be complete (e.g. memo in local employee file should be accompanied automatically by rebuttal where applicable). The teacher should be allowed to provide a statement even after he/she signed the evaluation if the admin made changes after the teacher signed.

HS Teacher 2014 – 2015

Teachers who aren't meeting standards according to the evaluator should also be observed by a neutral party from outside the school. This neutral party should have knowledge in that subject area.

ES Teacher 2013 – 2014

The doc log should be shortened to 1 piece of evidence per category, or the system revised to remove it entirely. It took FOREVER. It was stressful and time consuming.

ES Teacher 2014 – 2015

Allowing teachers to create their own SMARTR goals. More planned observations with a discussion ahead of time of what areas to focus on.

Other Other school 2014 – 2015

Stringent standards for determining the line between 2 and 3, and the line between 3 and 4, in evaluator ratings

ES Teacher 2012 – 2013 2015 – 2016

SMARTR goals should be teacher-created not mandated because they are school improvement goals. Teachers should not be rated ineffective based on unreasonable goals that were mandated.

ES Teacher 2013 – 2014

All teachers should receive feedback on positives as well as areas for improvement, with specific suggestions for improvement and a team approach to making this improvement happen.

MS Teacher 2015 – 2016

New teachers need to be required to sit down with an administrator in the first three weeks of school to discuss SMARTR goals and give them direction. It is too

ES Teacher 2015 – 2016

Page 21: FEA Survey about FCPS’ Teacher Performance Evaluation ... · FCPS purchased the TPEP system from James H. Stronge, the author of the Guidelines. It has not been significantly revised

21

overwhelming for first year teachers to figure this out on their own and mentors differ widely in their skill level. New teachers should have frequent meetings with their evaluator in the first months of school and time spent with mentors should be documented so it happens. We are losing young teachers who are overwhelmed.

By administrators giving written feedback on a more regular basis.

ES Teacher 2014 – 2015

This system is humiliating, non-collaborative, way too time-consuming for all involved, useless as far as student achievement is concerned, and resources could be far better allocated for max benefit. Eval'ing almost half the staff every single year is a colossal waste of time...where is the ROI in all this? How many central staff are invovled, and at what cost?? And going back to the beginning of this program 4-5 years ago, why was this humiliating, time-consuming, top-down program put in place in the first place?? What was wrong with the previous system, and/or what changed so that it was necessary or advisable to put this terrible system in place? Admins should be able to walk in with a clipboard, observe a few things, check the boxes, do it a handful of times, and move on to more important things like working on curriculum improvements...we need relief from this oppressive program, and fixing this is the single most impactful thing to do to improve working conditions.

ES Teacher 2013 – 2014 2014 – 2015

More clarity from admin on deadlines, and I don't think that our goals should be based solely on SOL performance (which they are for all teachers at my school this year)

ES Teacher 2013 – 2014 2014 – 2015 2015 – 2016

This is just one more evaluation system. It's the 5th or 6th one since I started teaching. I've gone up to 6 years without being observed. I have a written evaluation every 3 years, but it's rare for anyone to come into my classroom and actually watch instruction.

ES Teacher 2013 – 2014

consistency between the administrators evaluating you; restricting the evaluation to ONLY administrators and not coaches; rotating the evaluator each evaluation cycle

ES Teacher 2014 – 2015

The document log is a waste of time. I did not gain anything or learn anything about myself or my profession from completing it.

MS Teacher 2012 – 2013 2013 – 2014 2014 – 2015

Making all adminstration follow the same standards

ES Teacher 2014 – 2015

I was observed less than 2 weeks before the end of the year,and it was not very good.No write up but I was put on an improvement plan.I have talked to other teachers that were pets of the principal who received mostly exceeds and were never observed.The whole system is awful

ES Other 2014 – 2015

Waste of time; my evaluator has taught only one year about 20 years ago....and doesn't give me appropriate or useful feedback.

ES Teacher 2015 – 2016

Observers often come in for a part of the lesson and don't see the "good" part of the lesson. I had an adminstrator come in and observie me while I was doing a read aloud of a test. How is that useful?

MS Teacher 2015 – 2016

involve content area experts in the process

MS Teacher 2014 – 2015

Page 22: FEA Survey about FCPS’ Teacher Performance Evaluation ... · FCPS purchased the TPEP system from James H. Stronge, the author of the Guidelines. It has not been significantly revised

22

The evaluator has never been a classroom teacher, so her opinions and evaluations did not hold validity for me. They were textbook comments without real world application.

ES Teacher 2014 – 2015

SMARTR goal should not be tied to your TPEP points. Administrator should have training to know what to do to help teachers improve in their areas which need growth. It is too much too cover in one cycle. Maybe every year should be summative for a different evaluation focus area. That way, you are always working on it.

ES Teacher 2014 – 2015

Where is the accountability for evaluators? I know of a principal that has abused the process to get rid of employees he doesn't like and he still has his job! I am talking at least a dozen people! His school went for a semester with two classroom positions unfilled because so many teachers have left and nothing is being done!@

ES Teacher 2012 – 2013 2013 – 2014 2014 – 2015

Cross raters- have staff from different schools or central offices observe

Other Administrator

I believe strongly that evaluations should include observation by others who teach the same subject, across the District, who have a better understanding of best practices, and what the actual challenges are in the classroom.

HS Teacher 2013 – 2014

I have yet to be observed this year even though I am on the summative evaluation piece of the TPEP cycle. Based on my colleagues' experiences who have different administrators, there is a GREAT difference between evaluators at the high school level.

HS Teacher 2015 – 2016

I believe there needs to be more than one person evaluating with specific feedback and suggestions given for improvement.

HS Teacher 2015 – 2016

I have been observed twice unofficially by admins who aren't mine, but not once by my assigned admin. My mid year evaluation was all "meets" but there was no observation to even come to that conclusion. Other admins put much more effort in and I feel like I'm jumping through hoops for an admin won't even look at the material I'm gathering and that will give me the same summative evaluation regardless of anything I do.

HS Teacher 2015 – 2016

My Principal came into my room for 10 minutes this year to observe me. She made comments on the observation that would have been answered if she'd stayed longer and watched more of the lesson. Also, she would not approve my SMARTR goal this year even though see approved last years. It was the same goal with different stats. Why would one year's goal be acceptable and not the following year? She loaded my class with 26 ESOL students even though we're Title 1. I'm being held accountable for the learning off all these students without the proper resources. (Title 1 3rd grade should be exceed 23 students.)

ES Teacher 2012 – 2013 2015 – 2016

I wish specialist such as reading teachers did not have to follow the teacher standards. I wish we follwed the same standards or procedures as other specialist such as AART. Our role is very different with coaching and other school responsibities that is not a direct match to a classroom teacher.

ES Teacher 2013 – 2014

Arbitary, for Sped the evaluators were not knowledgeable about the curruciulum to offer constructive criticsm. Some evaluatators made suggestions that budget or time wise were impossible or close to impossible

HS Teacher 2012 – 2013

Page 23: FEA Survey about FCPS’ Teacher Performance Evaluation ... · FCPS purchased the TPEP system from James H. Stronge, the author of the Guidelines. It has not been significantly revised

23

There needs to be a system in place for inter-rater reliability to make sure that an evaluation does not depend on who evaluates you.

HS Teacher 2014 – 2015 2015 – 2016

Someone could develop good instructions on how to use the system. A training class on tpep would be helpful.

HS Teacher 2015 – 2016

I would rather have periodic, documented observations by other department members as a large part of how I am evaluated as an effective teacher.

HS Teacher 2013 – 2014

My principal observed me once this year and has never observed me before in the three years I have been here. How can she really know what I do? I know she never read my previous SMARTR goals because she told me I made an error on the percentage of improvement I wanted to make and I figured my last two goals the exact same way. What a colossal waste of time!!!

ES Other 2015 – 2016

Make it discussion based. Each administrator has a 1 hour discussion with each teacher during which time they talk about successes and areas of concern. Part of that would include constructive comments for improvement.

HS Teacher 2013 – 2014

The online interface is not at all user friendly. The old forms had prompts that helped you understand what was being asked for in the self assessment, for example. Administrators without content knowledge cannot adequately advise specialists. There should still be an option to have someone from the central office in your specialty area give you an evaluation for actual useful feedback. I value the reflection in the process, but without outside feedback, it's hard to really improve.

ES Teacher 2013 – 2014 2014 – 2015 2015 – 2016

It is slightly easier to do it on-line, but still confusing and hard to find things. It needs to be simplified and not take nearly as much time to complete as it does.There should be much less emphasis on the SMARTr goal portion, especially since principals often dictate what they want those goals to look like. I spend so much time in CLTs, IEPs, Data Days and working on things related to evaluation that I sometimes wonder when I'm actually supposed to spend time with students.

ES Teacher 2013 – 2014 2014 – 2015 2015 – 2016

As special ed deaprtment chair I have no classroom teaching duties. Eventhough the county has a different TPEP for us, my Principal requires us to have SMARTR goals tied directly to studednt achievement. I think my goals should be tied to something I have direct control over. I tried to write a goal for having my teacher complete effective transition plans, or standards based goals on IEPs, but that was not allowed. I resent having to create a teaching lessons and additonal work if I want control over my goal. I recieved the highest rating, but I am wishing I could retire before I have to go through that process again!

MS Other 2014 – 2015

It is ridiculous for librarians to have to provide numerical data that impacts student learning. I was fortunate that my administrator let me use circulation data in a variety of ways.

HS Other 2013 – 2014

Have the option for professional evaluators from Central office to observe and provide unbiased feedback when local admin is unable to provide valuable and/or unbiased feedback.

HS Teacher 2013 – 2014

Most useful for teacher and evaluator, when evaluator has actually been in a classroom as a teacher. The lack of experience in a teaching evironment can skew the results.

HS Teacher 2015 – 2016

Page 24: FEA Survey about FCPS’ Teacher Performance Evaluation ... · FCPS purchased the TPEP system from James H. Stronge, the author of the Guidelines. It has not been significantly revised

24

Some administrators use the TPEP process to bully teachers. If an administrator does not like a teacher on a personal front, they use the evaluation process to look for any deficit that other administrators wouldn't think was important.

ES Teacher 2013 – 2014

The system is too subjective and leaves to much discretion to the evaluator. If the standards are more specific or lent themselves to more quantitative evaluation, the results would be more consistent across evaluators. There should also be more explicit requirements for evaluator to consider artifacts from all three years of the evaluation cycle, rather than focusing on only the most recent year. The current system leaves to much room for evaluator preferences and personality considerations. This can lead to a high performing teacher receiving a poor evaluation. My evaluator this year has been critical of my artifacts as she feels they should show the impact of my work; however, the performance standards focus on actions and not outcomes. If it is desired for the focus of performance standards to be on outcomes, then the standards should explicitly require the teacher’s artifacts to address outcomes.

ES SBTS 2015 – 2016

My evaluator had only visted me once this year, last monday, and gave me feedback that I felt was irrelevant and socred me based on items that she had no previous observations to compare. I felt that system was very unfair and subjective based on the evaluators "thoughts" about what the advanced classroom should look like.

ES Teacher 2014 – 2015 2015 – 2016

I would very much like for my Music class teaching to be judged by A MUSICIAN. I would like for the most important objectives in my class be the ones for which I am held accountable. (Exposing children to Music, engaging them constantly. (Not just focusing on the SMARTR goal of the most paper-focused activities.) Classes should be judged differently. The Fine Arts Office could create our Music Teacher Evaluation, for example. Many administrators need to admit that they know little or nothing about the best practices for Music. Many other staff members have different job descriptions and responsibilities than just a "regular" classroom teacher who focuses on SOL's. It is not fair to be judged by a completely different standard than what one's job actually is. if possible, teachers should not be evaluated by someone new to the building, when being judged on the previous three years of work. Paper work should be much less. The old system WAS FINE!!!!!!

ES Teacher 2014 – 2015

Peer evaluations need to be a part of the process. Other professions are evaluated by peers who understand the demands of the profession. Administrators in education often do not have the necessary content knowledge to make effective observations.

HS Teacher 2014 – 2015

Administrators need to spend more time in the classroom observing.

HS Teacher 2014 – 2015

My principal could actually observe me!! She has had ample opportunity but yet she doesn't do it! I don't think it should be up to me to track her down and make sure that she is actually observed me. The last time I went through this (different administrator), it was mostly a lie because he hadn't observed me more than ONE TIME!! Or at least that was all that he wrote up... how can that be?! AND Just because they don't see you do something, they think that YOU can't do it!!?? WHAT?! Doesn't make sense!! Of course EVEN if i did recieve all EXEMPLARY or whatever the highest rank is... it doesn't mean a thing! I don't get more pay, I don't get any kind of compensation, not even an 'atta girl' from administration... so WHY WOULD ANYONE TRY?? It doesn't make much sense. The current TPEP is a waste of time. In fact, my principal didn't even watch the video I gave her of a lesson that I had done, wouldn' that have counted as an observation. Right now she is scrambling to observe me

ES Other 2015 – 2016

Page 25: FEA Survey about FCPS’ Teacher Performance Evaluation ... · FCPS purchased the TPEP system from James H. Stronge, the author of the Guidelines. It has not been significantly revised

25

Have someone observe me who understands my content/subject area and actually understands my classroom set up and if I have any idea what I'm doing.

HS Teacher 2014 – 2015

I continue to be concerned about my performance review being tied to student performance on summative assessments. I'm a special ed teacher and my kids don't necessarily demonstrate consistent academic progress (ex: 1year's progress every year). It irks me that AAP teachers don't have to worry about this aspect of their evaluation, but my evaluation is resting on whether the 15 kids I teach in a self-contained setting - some in and out of the prison system, some psychologically unstable, some enrolled in ID classes the rest of the day - can demonstrate a year's worth of academic progress. I do believe my admins are trying to be as understanding and fair as they can, but their hands are tied by the "one size fits all" state criteria.

MS Teacher 2012 – 2013 2015 – 2016

Evaluators don't always understand the curriculum of the different areas and specialties so an evaluation of those areas isn't realistic. Specialties can't be lumped in with classroom teachers. They are so very different. They see the students for a fraction of the time that classroom teachers see the same students. Environments are so very different.

ES Teacher 2013 – 2014

I have never been formally observed in all 3 years I have worked in FCPS

ES Teacher 2012 – 2013 2013 – 2014 2014 – 2015 2015 – 2016

We are told to encourage our students but as teachers we are discouraged from expecting an exceeds rating. After 26 years of dedicating myself to my career, I have only received 1 exceeds...I have presented at conferences, written curriculum, been Dept. Chair, attended Nat'l Conferences, worked in classroom, as an itinerant, and with adults...I feel disheartened not to be recognized for my hard work. I fear I may cry this year at what will be my final evaluation. I will retire in 2 years. I can't wait.

HS Teacher 2015 – 2016