session 6 addressing data quality.ppt · water utilities performance review report 2013/2014....

24
2/19/2015 1 Session 6 Addressing Data Quality and Alternative Methodologies for Measuring Performance Sanford Berg and Michelle Phillips With material from Rui Marques Addressing Data Quality Key Performance Indicators Productivity Trends Statistical Analysis Data Envelopment Analysis (DEA) 2

Upload: others

Post on 18-Aug-2020

1 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Session 6 Addressing Data Quality.ppt · Water Utilities Performance Review Report 2013/2014. Regional and National Project Water Utilities and Dawasco. December 2014. 2/19/2015 3

2/19/2015

1

Session 6Addressing Data Quality and Alternative Methodologies for

Measuring Performance

Sanford Berg and Michelle Phillips

With material from Rui Marques 

Addressing Data Quality

Key Performance Indicators 

Productivity Trends 

Statistical Analysis 

Data Envelopment Analysis (DEA)

2

Page 2: Session 6 Addressing Data Quality.ppt · Water Utilities Performance Review Report 2013/2014. Regional and National Project Water Utilities and Dawasco. December 2014. 2/19/2015 3

2/19/2015

2

Reliability Providing reliable and accurate information is vital for 

making meaningful comparative assessmentsReliability: Confidence regarding how the data was 

gathered

3

AccuracyAccuracy is a number indicating its likely range of error. Example: Tanzania Water Utilities Performance Review.

4

Accuracy Band Associated Uncertainty

1 (0-5%): Better than or equal to +/- 5%

2 (5-20%): Worse than +/- 5%, but better than or equal to +/- 20%

3 (20-50%): Worse than +/- 20%, but better than or equal to +/- 50%

4 (>50%): Worse than +/- 50%

  Source: The United Republic of Tanzania. EWURA. Water Utilities Performance Review Report 2013/2014. Regional and National Project Water Utilities and Dawasco. December 2014.

Page 3: Session 6 Addressing Data Quality.ppt · Water Utilities Performance Review Report 2013/2014. Regional and National Project Water Utilities and Dawasco. December 2014. 2/19/2015 3

2/19/2015

3

Standardized Confidence Indicator

Built upon the reliability and accuracy factors by joining the letter with the number

5

Source: The United Republic of Tanzania. EWURA. Water Utilities Performance Review Report 2013/2014. Regional and National Project Water Utilities and Dawasco. December 2014.

Exercise: Uncertainty about Indicator

Using the Handout from Session 2

You assigned weights to the indicators

What if you discovered that your “most important indicator” had a huge measurement error (+ or –30%) instead of the + or – 5% that you originally thought.

Does this new information change the weights you assign to the different KPIs when creating the OPI?

6

Page 4: Session 6 Addressing Data Quality.ppt · Water Utilities Performance Review Report 2013/2014. Regional and National Project Water Utilities and Dawasco. December 2014. 2/19/2015 3

2/19/2015

4

Importance of Making Comparisons

Performance comparisons are necessary but not sufficient for sound policy

Benchmarking represents an important tool for:Documenting past performance Establishing baselines for gauging improvements Making comparisons across service providers Designing incentives

7

Can an Index Capture Complexity? A single index of utility performance will be neither 

comprehensive nor fully diagnostic.

Physician can have information on a patient’s temperature, pulse, height and weight. 

Patient is in trouble: dangerous fever and/or is significantly overweight. 

Blood tests provide more detailed information

Diagnosing and treating mental health issues would require other diagnostics and treatments . . . Still, temperature and weight provide useful information.

8

Page 5: Session 6 Addressing Data Quality.ppt · Water Utilities Performance Review Report 2013/2014. Regional and National Project Water Utilities and Dawasco. December 2014. 2/19/2015 3

2/19/2015

5

Five Benchmarking Methodologies

Core Indicators and a Summary or Overall Performance Indicator (partial metric method)

Performance Scores based on Production or Cost Estimates (“total” methods)

Performance Relative to a Model Company (engineering approach) 

Process Benchmarking (involving detailed analysis of operating characteristics)

Customer Survey Benchmarking (identifying customer perceptions) 

9

KPIs and Overall Performance IndicatorsSpecific Key Indices, such as water delivered per worker, quality of service (continuity, water quality, complaints), unaccounted for water, coverage, and key financial data (operating expenses relative to total revenues, collections). 

partial measures provide the simplest way to perform comparisons: trends direct attention to potential problem areas

Overall Performance Indicator (OPI) combines the specific core indices into a summary index 

Peru: 9 KPIs: summed by SUNASS (3 quality, 2 coverage, 3 management, 1 finance (Cost/Revenues)

Canada: 61 KPIs

10

Page 6: Session 6 Addressing Data Quality.ppt · Water Utilities Performance Review Report 2013/2014. Regional and National Project Water Utilities and Dawasco. December 2014. 2/19/2015 3

2/19/2015

6

Performance Scores Based on Production or Cost Estimates

Rankings can be based on the analysis of production patterns and/or cost structures.  

Production function studies (requiring data on inputs and outputs) show how inputs affect utility outputs (such as volume of water delivered, number of customers, and service quality). Utilities that produce far less output than other utilities (who are using the same input levels) are deemed to be relatively inefficient. 

Cost functions show how outputs, inputs and input prices affect costs; such models have heavy data requirements. Excessively high costs would trigger more in‐depth studies to determine the source of poor performance. 

11

Use of Performance ScoresIdentify the best performers and the weakest performers in a group of utilities:

Technical/engineering efficiency, 

Cost efficiency, 

Production efficiency, 

Scale efficiency,

Economies of scope (water and wastewater) or transmission and distribution)

Efficiency change 

12

Page 7: Session 6 Addressing Data Quality.ppt · Water Utilities Performance Review Report 2013/2014. Regional and National Project Water Utilities and Dawasco. December 2014. 2/19/2015 3

2/19/2015

7

Engineering/Model Company Requires the development of an optimized economic and 

engineering model Idealized benchmark specific to each utility‐ incorporating the 

topology, customer demand patterns, and density of the service territory

“Artificial” firm has optimized its network design and minimized its operating costs 

Production relationships can be obscured through a set of assumed coefficients used in the optimization process. 

Chile and Argentina used this approach to establish infrastructure performance targets

US telecom interconnection pricing and battles of “models”

13

Process BenchmarkingFocuses on individual production processes 

Detailed examination of facilities and their operations 

Identifies stages of the production process needing attention: pumping up, intake, transport, clarification and filtration, purification and treatment. 

Studies of distribution processes (network design, pipeline construction and maintenance), sales processes (meter reading, data processing, billing, collections, and customer relations), and general processes (planning, staff recruitment and retention, and public relations). 

Provides a mechanism for identifying potential benchmarking partners, undertaking benchmarking visits, and implementing best practices

Mats Larsson, et al (2002).  Process Benchmarking in the Water Industry: Towards a Worldwide Approach, IWA

14

Page 8: Session 6 Addressing Data Quality.ppt · Water Utilities Performance Review Report 2013/2014. Regional and National Project Water Utilities and Dawasco. December 2014. 2/19/2015 3

2/19/2015

8

Customer Survey BenchmarkingCustomer Complaints: one indicator SERVQUAL identifies five dimensions of service quality

as perceived by customers: External characteristics (tidy workplace, employee appearances)

Reliability (meeting deadlines, consistency in interactions)

Responsiveness (providing service promptly), Consideration (personnel who are courteous, friendly, and helpful) 

Empathy (giving individual care and attention)  Parasuraman et. al. (1985) Journal of Marketing R. Parena (1999). The IWSA Benchmarking Initiatives:

15

Steps for Conducting Production or Cost Benchmarking Studies1. Identify objectives, select methodology, and gather data

2. Screen and analyze data 

3. Utilize specific analytic techniques

4. Conduct consistency/sensitivity tests

5. Develop policy implications 

16

Page 9: Session 6 Addressing Data Quality.ppt · Water Utilities Performance Review Report 2013/2014. Regional and National Project Water Utilities and Dawasco. December 2014. 2/19/2015 3

2/19/2015

9

Step 1. Identify Objectives, Select Methodology and Gather Data

Decide issues to be addressed, time period to be analyzed, and types of comparisons

Choices will reflect capabilities, initial understanding of data availability, and preliminary methodological choices. 

The objectives of any benchmarking study will depend on most important policy issues under consideration.  

Staff requirements can be substantial

17

Step 2. Screen and Analyze DataScreen data 

timeframe,

sample size, and 

statistical techniques. 

Check data quality 

inconsistent definitions, 

missing data or 

extreme data values

Analysis is an iterative process

18

Page 10: Session 6 Addressing Data Quality.ppt · Water Utilities Performance Review Report 2013/2014. Regional and National Project Water Utilities and Dawasco. December 2014. 2/19/2015 3

2/19/2015

10

Step 3. Utilize Specific Analytic Techniques

Based on data and objectives, choose technique

Core Overall Performance Indicators

Total Factor Productivity Indices 

Relative Performance using Data Envelopment Analysis (DEA) 

Relative Performance using Statistical Techniques 

Utilize specialists for sophisticated models

“A Primer on Efficiency Measurement for Utilities and Transport Regulators”, by Coelli, Estache, Perelman, & Trujillo, World Bank Institute

19

Metric Methodologies: Four Categories Core Overall Performance Indicators (OPIs combine partial 

indicators of operating or financial performance; these are summary indices), 

Total Factor Productivity Indices (an index number approach using output per unit input—multiple inputs are taken into consideration to gauge efficiency levels and changes), 

*Relative Performance using Data Envelopment Analysis (a non‐parametric technique that makes no assumptions about the functional form of production or cost functions), 

*Relative Performance using Statistical Techniques (parametric approaches that involve assumptions about functional relationships)

20

Page 11: Session 6 Addressing Data Quality.ppt · Water Utilities Performance Review Report 2013/2014. Regional and National Project Water Utilities and Dawasco. December 2014. 2/19/2015 3

2/19/2015

11

Data Envelopment Analysis (DEA)Linear programming technique applied to a selected set of variables to calculate an efficiency coefficient for each water utility.  

Adapted from the multi‐input, multi‐output production function:Production Function: how much output can be produced with a given basket of inputs. 

DEA benchmarks firms only against the best producers (so it is a non‐parametric frontier analysis). Differs from OLS that bases comparisons with respect to an average producer.

*Applied to cost functions as well   

21

Production Observations

Output

(Q)

Input

x

x

x

xx

x

x x

x

x

Each X represents datafor a water utility

FrontierFrom DEA

One observation far from the frontier!

22

Page 12: Session 6 Addressing Data Quality.ppt · Water Utilities Performance Review Report 2013/2014. Regional and National Project Water Utilities and Dawasco. December 2014. 2/19/2015 3

2/19/2015

12

Unit Cost Observations

Unit Cost

(TC/Q)

Output (Q)

x

x

xx

x

x

x

x

x

x

Each X represents an observationon a utility

23

Frontier Analysis

Frontier Analysis

Unit Cost

Output

x

x

xx

x

x

x

x

x

x

24

Page 13: Session 6 Addressing Data Quality.ppt · Water Utilities Performance Review Report 2013/2014. Regional and National Project Water Utilities and Dawasco. December 2014. 2/19/2015 3

2/19/2015

13

Statistical Analysis Identify relationship between firm performance and 

market conditions and characteristics of the production processes.  

Coefficients: the roles of multiple variables 

Calculated expected cost, given actual output. 

Predicted versus actual cost can provide a measure of relative performance. 

Widely used econometric techniques include  Ordinary Least Squares (OLS) 

Corrected OLS (COLS)

Stochastic Frontier Analysis (SFA). 

25

Unit Cost Averaging

Unit Cost Averaging

Unit Cost

Output

x

x

xx

x

x

x

x

x

x

Compare to “average”, but hereScale (output level) seems to matter

26

Page 14: Session 6 Addressing Data Quality.ppt · Water Utilities Performance Review Report 2013/2014. Regional and National Project Water Utilities and Dawasco. December 2014. 2/19/2015 3

2/19/2015

14

Linear RegressionLinear Regression: AC = a – b Q

Unit Cost

Output

x

x

xx

x

x

x

x

x

x

Functional form Variables Error terms

27

interceptSlope (-)

Relative Inefficiency (OLS?)

Linear Regression Unit Cost

Output

x

x

xx

x

x

x

x

x

x

“Error” difference between Actual and Predicted Cost: define as inefficiency oras noise?120

111

100028

Page 15: Session 6 Addressing Data Quality.ppt · Water Utilities Performance Review Report 2013/2014. Regional and National Project Water Utilities and Dawasco. December 2014. 2/19/2015 3

2/19/2015

15

Frontier Regression

AC = a – b QUnit

Cost

Output

x

x

xx

x

x

x

x

x

xError terms (random noise?)

Corrected OLS Stochastic Frontier

Analysis (SFA, with multiple components of error term)

OLS

29

www.purc.ufl.edu“Leadership in Infrastructure Policy”

SFA: Stochastic Frontier Analysis

TCTotalcost

Output

Differences due to both errors in variables ANDInefficiency (requires Assumptions aboutDistribution of errors).

COLS

SFA

Much more complicated estimation process.

30

Page 16: Session 6 Addressing Data Quality.ppt · Water Utilities Performance Review Report 2013/2014. Regional and National Project Water Utilities and Dawasco. December 2014. 2/19/2015 3

2/19/2015

16

Model SpecificationFunctional forms:  Linear or non‐linear Interpretation of error terms

Cross Section, Time Series, Panel?Excluded Variables?Service QualityOther Determinants:• Customer Density? Age of network?• Topography, Hydrology, Ownership• Input prices (for Cost function)

31

Relative Inefficiency (frontier/non-linear)

Frontier Analysis

Unit Cost

Output

x

x

xx

Z

Y

x

x

x

x

Relative to Predicted Cost

Develop  Scores  Rankings

32

Page 17: Session 6 Addressing Data Quality.ppt · Water Utilities Performance Review Report 2013/2014. Regional and National Project Water Utilities and Dawasco. December 2014. 2/19/2015 3

2/19/2015

17

Step 4. Conduct Consistency/Sensitivity Tests

Accuracy and robustness of inefficiency estimates are important due to financial or social impacts. 

Cost Function vs. Production Function Functional form (linear, nonlinear) Outputs and inputs (e.g., network length vs. fixed assets)Alternative methodologies (e.g., DEA vs. SFA).   Need to check whether estimated inefficiency scores or rankings are sensitive to the benchmarking method

33

Step 5. Develop Policy Implications

Analyze Scores and Rankings  Explore in greater detail the potential determinants of 

inefficiencies across firms and over time.  Firms should not be ranked as poor performers if they 

operate under conditions that differ from those of the other firms. 

Identify the impact of factors like region, population density, regulatory environment, ownership structure, and network vintage.

Seek public comments. Design incentives to improve performance. 

34

Page 18: Session 6 Addressing Data Quality.ppt · Water Utilities Performance Review Report 2013/2014. Regional and National Project Water Utilities and Dawasco. December 2014. 2/19/2015 3

2/19/2015

18

Creating Appropriate Yardsticks

Regulators want to induce outcomes comparable to those achieved under competition.

Pass‐Fail Standards?

Reward outstanding performance

Penalize weak performance

Benchmarking provides Yardsticks

35

Rui Cunha Marques

Many benefits:

Very useful. They are effective in calling attention of stakeholders to issues!... 

Sunshine (name and shame) regulation;

Consequences for tariff‐setting (carrot and stick approach).

Several shortcomings:

Selection of performance measures (often partial ones);Weighting the indicators; Dealing with bad and missing data; Reliability;  Stability (over time) and Rankings are NOT additive

Presentation Techniques: Rankings

36

Page 19: Session 6 Addressing Data Quality.ppt · Water Utilities Performance Review Report 2013/2014. Regional and National Project Water Utilities and Dawasco. December 2014. 2/19/2015 3

2/19/2015

19

Rui Cunha Marques

Empirical patterns: Minimum and maximum scores; The same trend;  Comparable means, standard deviations, and 

other distributional properties; Stability over time; Correlation with intuitive partial measures.

Statistical tests: Spearman and Pearson tests (Kendall Tau when 

there are ties)

Robustness of Scores

37

Improving Health: “Do No Harm”

Benchmarking specialists produce and critique studies that utilize various methodologies.

Rankings can be manipulated by choice of variables, model specification, sample size, time frame, and treatment of outliers.

Results can be misinterpreted and misused. 

The stakes are high, since affected parties have an interest in the relative and absolute performance comparisons prepared by analysts.

38

Page 20: Session 6 Addressing Data Quality.ppt · Water Utilities Performance Review Report 2013/2014. Regional and National Project Water Utilities and Dawasco. December 2014. 2/19/2015 3

2/19/2015

20

Benchmarking is Part of Regulatory and Managerial Tool-kits

The application of the techniques summarized here can improve service quality, expand networks, and optimize operations.  

Any benchmarking study will have limitations, but sound studies can be used to place the burden of proof on other parties who might argue that the analysis is incomplete or incorrect.  

Over time, data availability will improve and studies will be strengthened as professionals gain experience with these quantitative techniques.

39

Summing Up

Performance Scores & Rankings can serve as catalysts for better stewardship of water and other resources. 

If regulators cannot identify historical trends, determine today’s baseline performance, and quantify relative performance across utilities, then as an Indian regulator said, they may as well be writing “pretty poetry”. 

40

Page 21: Session 6 Addressing Data Quality.ppt · Water Utilities Performance Review Report 2013/2014. Regional and National Project Water Utilities and Dawasco. December 2014. 2/19/2015 3

2/19/2015

21

Total Factor Productivity (Outputs/Inputs)Efficiency vs. ProductivityClassification of Methods (parametric and nonparametric)

Efficiency ClassificationsFeatures of Metric Techniques

Technical AppendixFrom Rui Cunha Marques

41

Global Productivity Measures: Outputs/Inputs

M

i ii 1N

j jj 1

a yWeighted Sum of Outputs

TFPWeighted Sumof Inputs

b x

Total factor productivity (TFP):

Linear relationship between inputs and outputs;

Constant weights for all the elements beingcompared;

Different results can be reached according to thecomposition of weights adopted.

42

Page 22: Session 6 Addressing Data Quality.ppt · Water Utilities Performance Review Report 2013/2014. Regional and National Project Water Utilities and Dawasco. December 2014. 2/19/2015 3

2/19/2015

22

Global Productivity Measures: Outputs/Inputs

Productivity change measures:

43

t 1 tt 1t, t 1

t t 1 t

f y , yTFPTFP

TFP g x , x

The inputs and outputs choice and the way they can beaggregated characterize the different indexes than can bebuilt

They are non‐parametric and non‐frontier measures ofperformance evaluation, called Index Numbers.

The concept of efficiency is different from that of productivity

Productivity is the ratio between the products (outputs) and thefactors (inputs) used. The productivity of an organization (activity)only coincides with its efficiency in particular situations. For example,the operation at different scales or distinct operational environmentsleads to different productivities.

Definitions

Efficiency

Static nature

Productivity

Dynamic nature

Y     (Output)Fp2

c Fp1

d

b a

X (Input)44

Page 23: Session 6 Addressing Data Quality.ppt · Water Utilities Performance Review Report 2013/2014. Regional and National Project Water Utilities and Dawasco. December 2014. 2/19/2015 3

2/19/2015

23

Stochastic frontiers (SFA)Characterization:

It is the main technique that competes with DEA;

The determination of the frontier is realised based on themethod of maximum likelihood;

Differs from other methodologies by separating thestatistical error term from the inefficiency term.

SFA

Inefficiency

Error

iinii vu,yfc

Input

45

Output

Methodsy COLS SFA

x DEA

x xx x

x xx

x

x

OLS

Benchmarking technique

Parametric Non-parametric

Frontier Non-frontier Frontier Non-frontier

SFA, COLS OLS DEA PI, TFP

46

Page 24: Session 6 Addressing Data Quality.ppt · Water Utilities Performance Review Report 2013/2014. Regional and National Project Water Utilities and Dawasco. December 2014. 2/19/2015 3

2/19/2015

24

Efficiency ClassificationsEfficiency

Productive Allocative

Static Dynamic

Technical Price

Scale/ScopePure 

technical

w/o cong. CongestionRui Cunha Marques

47

Features DEA SFA COLS OLS

Specification of the functional form

No Yes Yes Yes

Integration of multiple inputs and outputs

Yes Difficult Difficult Difficult

Identification of best practices Yes No Yes No

Detail of efficiency measures High Low Low Low

Statistical inference Difficult Yes Yes Yes

Adjusting for operational environment

More difficult

Multico-linearity

Multico-linearity

Multico--linearity

Accounting for noise No Yes No No

Sensitivity to outliers Very Sensitive Very Little

Rui Cunha Marques

Features of Major Metric Techniques

48