1 a forecast evaluation tool (fet) for cpc operational forecasts edward olenic chief, operations...

Download 1 A Forecast Evaluation Tool (FET) for CPC Operational Forecasts Edward OLenic Chief, Operations Branch, NOAA-NWS-CPC

If you can't read please download the document

Upload: barnard-chapman

Post on 18-Jan-2018

215 views

Category:

Documents


0 download

DESCRIPTION

3 Background People and organizations need planning tools. For decades, climate forecasts have been sought for use in planning. The obvious signs of climate change have made this desire more urgent. The uncertain nature of climate forecasts can lead the uninformed to make poor choices. Users need an “honest broker” they can trust.

TRANSCRIPT

1 A Forecast Evaluation Tool (FET) for CPC Operational Forecasts Edward OLenic Chief, Operations Branch, NOAA-NWS-CPC 2 Outline CPC operational products We have discovered users Users are diverse We need to know how users use products We need a new service paradigm FET addresses these needs Tour of FET Current status Implications 3 Background People and organizations need planning tools. For decades, climate forecasts have been sought for use in planning. The obvious signs of climate change have made this desire more urgent. The uncertain nature of climate forecasts can lead the uninformed to make poor choices. Users need an honest broker they can trust. 4 Current Operational Capabilities CPC products, include extreme events for days 3-14, extended-range forecasts of T, P for week 2, 1-month and 3-month outlooks, and ENSO forecasts. These products have a scientific basis, have skill, and therefore ought to be useful, but - We have a less-than adequate understanding of user requirements, or actual practical use. CPC has neither the staff, nor the expertise to effectively pursue understanding these things. 5 Challenges from Users Western Governors Association (Jones, 2007): - More accurate, finer-resolution long range forecasts - Continued and expanded funding for data collection, monitoring and prediction - Partnerships with federal and state climatologists, RCCs, agricultural extension services, resource management agencies, federal, state and local governments. USDA ARS Grazinglands Research Laboratory (Schneider, 2002) - Fewer EC forecasts - Better correspondence between F probability and O frequency - Forecast more useful than climatology - Forecasts of impacts, not meteorological variables 6 Bridging Research, Operations, Users Assumed User Needs ? GOVT. PROVIDERS Use it if you can Basic data and Forecasts Basic Research Know WHO stakeholders are and HOW they USE climate products, (relationships), LEARN from stakeholders WHAT we need to provide (iterative refinement of requirements, relationships), Implement USER-, and SCIENCE- VETTED products, (iterative refinement of requirements, relationships), Operationally support evolving user and producer REQUIREMENTS (extension function, relationships, operations, research) 7 Bridging Research, Operations, Users Intermediary Applications Products CTB-RISA/PRIVATE/RCC/SC-SBIR, NIDIS Decision-Support Development ? GOV PROVIDERS (CPC/EMC) NCEP/NCDC/USGS CTB SUPPOPRT Iterative Use refinement Transfer User-Vetted products PROVIDERS Iterative Technical Refinement Use it if you can Basic data, Forecasts RESEARCH, MODELING (dyn, stat), OBSERVATIONS GOV OPS PRIVATE OPS Transfer User-Vetted products O2R: Model Test Facility R2O : CTB users Research, Modeling, Obs O2R R2O REQUIREMENTS Assumed User Needs O2R R2O 8 , 9 Trust, but verify 10 , , Verify, and Trust 11 Trust is at the core of a successful product suite. Transparent verification is one way to secure trust. , , Verify, and Trust 12 Trust is at the core of a successful product suite. Transparent verification is one way to secure trust. VERIFICATION: A MEASURE OF FORECAST QUALITY, SKILL AND VALUE , , Verify, and Trust 13 Types of Verification Accuracy e.g., AC, error, #correct, rmse, mae, etc Skill e.g., HSS, RPSS, Brier SS Bias forecast too high/low? Resolution how well are different events discriminated? Sharpness able to predict extreme events? 14 How to Proceed? Over the last decade, Dr. Holly C. Hartmann, and programmers Ellen Lay and Damian Hammond, of CLIMAS, have developed an interactive, on-line Forecast Evaluation Tool (FET) which allows users to evaluate the meaning and skill of CPC 3-Month Outlooks of temperature and precipitation. CPC proposes to: 1)Make the FET CPCs outlet to users for forecast skill information, 2)Become a partner with CLIMAS and others to make the FET a community resource. 3)Expand the capabilities of FET 15 Forecast Evaluation Tool: Example of a Means to Address Gaps What FET provides: User-centric forecast evaluation and data access and display capability. Leveraging of community software development capabilities. Opportunity to DISCOVER and collect user requirements. 16 CPC Temperature and Precipitation Outlooks Probability of three categories Bottom, middle, top 10 years, Maps show probability of likeliest category Each category has a value at every point Probabilities sum to 100% If middle favored, borrow from extremes EC means pr(b,n,a)=33.33, 33.33, 33.33 17 CPC Temperature and Precipitation Outlooks Probability of three categories Bottom, middle, top 10 years, Maps show probability of likeliest category Each category has a value at every point Probabilities sum to 100% If middle favored, borrow from extremes EC means pr(b,n,a)=33.33, 33.33, Pr(b,n,a)=40, 33.33, Pr(b,n,a)= 16.67, 33.33, 50 18 Downscaling: Probabilities to Temperatures 19 Downscaling: Probabilities to Amounts 20 Recent CPC Skill Improvements HSS=% improvement the forecast makes over random forecasts. More non-EC forecasts, and a large HSS are GOOD. Non-EC forecasts OLenic et al (2008) compared the HSS and % non-EC of official (OFF) forecasts made in real-time from 1995 through 2004 with forecasts made using an objective consolidation (CON) of the identical four main forecast tools which were used to prepare the real-time forecasts. Period ( ) mean T, P forecast skills for OFF/CON were 22/26 for T, and 8.8/12.1 for P. CON % non-EC is also higher. CPC began using CON as a first guess in 2006. 21 +20% +8% +18% +16% Top 2 rows: HSS (lines) of 3-month P Outlooks, Official (OFF) and Consolidation (CON). Colors are the fraction of the time non-EC is predicted (%). Bottom row: Difference, CON-OFF (lines and colors). (See OLenic et al, 2008) DIF CON HSS DIF CON HSS OFF DIF SPRING SUMMER FALL WINTER FMA, MAM, AMJ MJJ, JJA, JAS ASO, SON, OND NDJ, DJF, JFM DIF 22 SPRING SUMMER FALL WINTER FMA, MAM, AMJ MJJ, JJA, JAS ASO, SON, OND NDJ, DJF, JFM +11% +31% +40% DIF CON HSS DIF +55% CON HSS OFF DIF Top 2 rows: HSS (lines) of 3-month T Outlooks, Official (OFF) and Consolidation (CON). Colors are the fraction of the time non-EC is predicted (%). Bottom row: Difference, CON-OFF (lines and colors). DIF 23 GPRA Score Official Skill Metric: 48-Mo. Running Mean U.S. Average T HSS 24 What is the Forecast Evaluation Tool (FET)? A web tool that allows users to interact with a database of CPC 3-Month Mean Temperature, Total Precipitation Outlooks, and verifying observations from1995-present, A tracker of user preferences. A self-teaching tool to enable a wide range of users to learn what CPC forecasts are, what they mean, and what their implications are for user applications. A collector of user requirements. A community tool for CPC, IRI, IPCC, ??? A laboratory for growing services to users 25 26 Introduction to the FET 27 Tutorial 28 Look at the latest forecasts 29 Look at the latest forecasts 30 Interact with observations 31 32 Las Vegas Historical La Nina Observed P LA NINA NEUTRALEL NINO 33 Central Florida Historical El Nino Observed P LA NINA NEUTRAL EL NINO 34 Oregon (east) Historical El Nino Observed P PDO - PDO neut PDO + 35 Verification(regime, score, years) 36 Verification(regime, score, years) 37 Frequency of non-EC Forecasts 38 Probability of Detection 39 False Alarm Rate 40 Brier Skill Score 41 Ranked Probability Skill Score 42 FUTURE: Expand FETs Capabilities 43 CTB User-Centric Forecast Tools Progress Simple Object Access Protocol (SOAP) Tested Secured Go-Ahead to Place FET on NWS Web Operations Center (WOC) Trained CPC Staff in JAVA language Scheduled Ellen Lay Training session in Nov. 44 FUTURE of the FET Next 1-4 months: Finalize and implement FET project plan at CPC. Ellen Lay (CLIMAS) to train CPC personnel on FET version control and bug tracking at CPC, November 18-21, Necessary software (APACHE TOMCAT, JAVA, Desktop View) acquired and installed at CPC. Forecast, observations datasets in-place at CPC. FET code ported to CPC, installed, tested. FET installed to NWS Web Operations Center (WOC) servers 45 FUTURE of the FET + In partnership with CLIMAS and community we will add: Other forecasts and organizations Time and space aggregation options Significance tests/cautions to users Requirements requests option Questions option The stakes are high.. 46 Source: The Washington Post Outlook Section, July 13, 2008 47 Summary Users want partnership, accuracy, specificity, flexibility Relationship is synonymous with partnership TRUST (honest brokerage) is central to these requirements. Producers must learn WHO users are, HOW they use products and WHAT their evolving requirements are Need to involve users and producers in iteratively optimizing products A continuous flow of requirements from users toward research may avert VOD. Means to fund an ever-expanding, perpetual product suite needed Stakes are high: 6 of top 20 news/media June 2008 sites were weather-related. 10s-100s of B$ at stake. Climate will only add to this.