119
Model Output Statistics (MOS) - Objective Interpretation of NWP Model Output University of Maryland April 4, 2012 Mark S. Antolik Meteorological Development Laboratory Statistical Modeling Branch NOAA/National Weather Service Silver Spring, MD (301) 713-0023 ext. 110 email: [email protected]

Model Output Statistics (MOS) - Objective Interpretation

  • Upload
    others

  • View
    9

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Model Output Statistics (MOS) - Objective Interpretation

Model Output Statistics (MOS) -

Objective Interpretation of NWP

Model Output

University of Maryland – April 4, 2012

Mark S. Antolik

Meteorological Development Laboratory

Statistical Modeling Branch

NOAA/National Weather Service

Silver Spring, MD

(301) 713-0023 ext. 110

email: [email protected]

Page 2: Model Output Statistics (MOS) - Objective Interpretation

MOS Operational System “Fun Facts”

With apologies to David Letterman, of course!

● 9 million regression equations ● 75 million forecasts per day ● 1200 products sent daily

● 400,000 lines of code – mostly FORTRAN ● 180 min. supercomputer time daily

● All developed and maintained by ~ 12 MDL / SMB meteorologists!

X 8

Page 3: Model Output Statistics (MOS) - Objective Interpretation

OUTLINE

1. Why objective statistical guidance?

2. What is MOS? Definition and characteristics The “traditional” MOS product suite (GFS, NAM) Other additions to the lineup

3. Simple regression examples / REEP

4. Development strategy -

MOS in the “real world”

5. Verification

6. Dealing with NWP model changes

7. Where we’re going – GMOS and the future

Page 4: Model Output Statistics (MOS) - Objective Interpretation

WHY STATISTICAL GUIDANCE?

● Add value to direct NWP model output Objectively interpret model

- remove systematic biases

- quantify uncertainty

Predict what the model does not

Produce site-specific forecasts

(i.e. a “downscaling” technique)

● Assist forecasters “First Guess” for expected local conditions

“Built-in” model/climo memory for new staff

Page 5: Model Output Statistics (MOS) - Objective Interpretation

A SIMPLE STATISTICAL MODEL

Relative Frequency of Precipitation as a Function of

12-24 Hour NGM Model-Forecast Mean RH

NGM MEAN RELATIVE HUMIDITY (%)

OB

SE

RV

ED

RE

L. F

RE

QU

EN

CY

0 10 20 30 40 50 60 70 80 90 100 0.0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1.0

3-YR SAMPLE; 200 STATIONS

1987-1990 COOL SEASON

47%

Page 6: Model Output Statistics (MOS) - Objective Interpretation

MOS Max Temp vs. Direct Model Output

2

3

4

5

6

7

8

MA

E (

de

g F

)

24 48 72 96 120 144 168 192

Projection (hours)

NEW MOS CLIMO DMO

MRF MOS Max Temp1999 Warm Season - CONUS/AK

MOS

DMO

CLIMO

Page 7: Model Output Statistics (MOS) - Objective Interpretation

What is MOS?

Page 8: Model Output Statistics (MOS) - Objective Interpretation

MODEL OUTPUT STATISTICS (MOS)

1. Numerical Weather Prediction (NWP) Model

Forecasts

2. Prior Surface Weather Observations

3. Geoclimatic Information

Current Statistical Method:

MULTIPLE LINEAR REGRESSION

(Forward Selection)

Relates observed weather elements (PREDICTANDS)

to appropriate variables (PREDICTORS) via a

statistical approach.

Predictors are obtained from:

Page 9: Model Output Statistics (MOS) - Objective Interpretation

MODEL OUTPUT STATISTICS (MOS)

Properties

● Mathematically simple, yet powerful

● Need historical record of observations

at forecast points (Hopefully a long, stable one!)

● Equations are applied to future run of

similar forecast model

Page 10: Model Output Statistics (MOS) - Objective Interpretation

MODEL OUTPUT STATISTICS (MOS)

Properties (cont.)

● Non-linearity can be modeled by using

NWP variables and transformations

● Probability forecasts possible from a

single run of NWP model

● Other statistical methods can be used e.g. Polynomial or logistic regression;

Neural networks

Page 11: Model Output Statistics (MOS) - Objective Interpretation

MODEL OUTPUT STATISTICS (MOS)

● ADVANTAGES Recognition of model predictability

Removal of some systematic model bias

Optimal predictor selection

Reliable probabilities

Specific element and site forecasts

● DISADVANTAGES Short samples

Changing NWP models

Availability & quality of observations

Page 12: Model Output Statistics (MOS) - Objective Interpretation

MAJOR CHALLENGE TO MOS DEVELOPMENT:

RAPIDLY EVOLVING NWP MODELS AND

OBSERVATION PLATFORMS

RAPIDLY EVOLVING NWP MODELS AND

OBSERVATION PLATFORMS

2. DIFFICULT COLLECTION OF APPROPRIATE

PREDICTAND DATA

1. SHORT, UNREPRESENTATIVE

DATA SAMPLES

MODELS

NCEP

OBS

Mesonets

New observing systems: (ASOS, WSR-88D, Satellite)

(Co-Op, Mesonets)

“Old” predictands: The elements don’t change!

Can make for:

Page 13: Model Output Statistics (MOS) - Objective Interpretation

“Traditional” MOS

text products

Page 14: Model Output Statistics (MOS) - Objective Interpretation

GFS MOS GUIDANCE MESSAGE

FOUS21-26 (MAV) KLNS GFS MOS GUIDANCE 11/29/2004 1200 UTC

DT /NOV 29/NOV 30 /DEC 1 /DEC 2

HR 18 21 00 03 06 09 12 15 18 21 00 03 06 09 12 15 18 21 00 06 12

N/X 28 48 35 49 33

TMP 43 44 39 36 33 32 31 39 46 45 41 38 37 39 41 44 45 44 40 40 35

DPT 27 27 28 29 29 29 29 33 35 35 36 35 36 39 41 42 37 34 30 30 28

CLD CL BK BK BK OV OV OV OV OV OV OV OV OV OV OV OV OV BK CL CL CL

WDR 34 36 00 00 00 00 00 00 00 14 12 12 10 11 12 19 28 29 29 29 28

WSP 06 02 00 00 00 00 00 00 00 01 02 04 04 06 07 08 15 17 18 09 05

P06 0 0 4 3 11 65 94 96 7 0 0

P12 6 19 94 96 0

Q06 0 0 0 0 0 3 4 4 0 0 0

Q12 0 0 4 2 0

T06 0/ 0 0/18 0/ 3 0/ 0 0/ 0 0/18 2/ 1 10/ 4 0/ 3 1/ 0

T12 0/26 0/17 0/27 10/25 1/38

POZ 2 0 0 1 2 4 4 0 1 1 2 3 3 1 1 0 2 1 2 3 1

POS 13 2 1 2 1 0 0 0 0 0 0 0 0 2 0 0 0 3 0 9 28

TYP R R R R R R R R R R R R R R R R R R R R R

SNW 0 0 0

CIG 8 8 8 8 7 7 7 8 8 7 7 7 4 2 3 3 6 7 8 8 8

VIS 7 7 7 7 7 7 7 7 7 7 7 7 5 5 4 2 6 7 7 7 7

OBV N N N N N N N N N N N N BR BR BR BR N N N N N

Page 15: Model Output Statistics (MOS) - Objective Interpretation

NAM MOS GUIDANCE MESSAGE

FOUS44-49 (MET)

KBWI NAM MOS GUIDANCE 2/27/2009 1200 UTC

DT /FEB 27/FEB 28 /MAR 1 /MAR 2

HR 18 21 00 03 06 09 12 15 18 21 00 03 06 09 12 15 18 21 00 06 12

N/X 38 46 32 41 24

TMP 59 58 55 54 49 43 38 38 43 45 40 38 37 35 33 34 37 38 33 29 25

DPT 46 47 48 46 37 30 24 22 22 22 24 27 28 26 25 24 24 21 17 12 10

CLD OV OV OV OV OV SC SC SC CL BK OV OV OV OV OV OV OV OV OV OV BK

WDR 21 20 22 25 31 32 34 36 01 03 05 04 01 36 35 35 35 34 35 33 34

WSP 15 09 08 06 10 11 10 12 10 09 08 10 12 13 14 16 11 13 15 16 17

P06 89 10 3 2 2 76 73 13 17 27 19

P12 10 3 81 17 30

Q06 1 0 0 0 0 4 1 0 0 0 0

Q12 0 0 4 0 0

T06 2/ 9 0/ 5 0 /0 0/ 5 3/ 1 5/ 3 0/ 0 0/ 2 2/ 5 0/ 0

T12 2/ 9 0/ 5 5/ 3 1/ 2 7/ 5

SNW 0 0 0

CIG 6 6 4 5 7 8 8 8 8 8 7 6 4 3 4 3 4 4 7 6 7

VIS 7 7 6 7 7 7 7 7 7 7 7 7 3 6 5 7 7 7 7 7 7

OBV N N N N N N N N N N N N BR N BR N N N N N N

Page 16: Model Output Statistics (MOS) - Objective Interpretation

Short-range (GFS / NAM) MOS

● STATIONS: Now at approx. 1990 Forecast Sites

(CONUS, AK, HI, PR, Canada)

● FORECASTS: Available at projections of 6-84 hours GFS available for 0600 and 1800 UTC cycles

● RESOLUTION: GFS predictors on 95.25 km grid; NAM on 32 km Predictor fields available at 3-h timesteps

● DEPENDENT SAMPLE NOT “IDEAL”: Fewer seasons than older MOS systems Non-static underlying NWP model

Page 17: Model Output Statistics (MOS) - Objective Interpretation

Approx.

1990 sites

Page 18: Model Output Statistics (MOS) - Objective Interpretation

Short-range (GFS / NAM) MOS

● STATIONS: Now at approx. 1990 Forecast Sites

(CONUS, AK, HI, PR)

● FORECASTS: Available at projections of 6-84 hours GFS available for 0600 and 1800 UTC cycles

● RESOLUTION: GFS predictors on 95.25 km grid; NAM on 32 km Predictor fields available at 3-h timesteps

● DEPENDENT SAMPLE NOT “IDEAL”: Fewer seasons than older MOS systems Non-static underlying NWP model

Page 19: Model Output Statistics (MOS) - Objective Interpretation

GFSX MOS GUIDANCE MESSAGE

FEUS21-26 (MEX)

KCXY GFSX MOS GUIDANCE 11/26/2004 0000 UTC

FHR 24| 36 48| 60 72| 84 96|108 120|132 144|156 168|180 192

FRI 26| SAT 27| SUN 28| MON 29| TUE 30| WED 01| THU 02| FRI 03 CLIMO

X/N 43| 29 47| 40 55| 35 51| 29 45| 32 40| 36 42| 30 45 31 46

TMP 37| 32 43| 43 46| 37 41| 32 39| 35 36| 38 37| 33 37

DPT 24| 27 37| 40 32| 28 28| 26 31| 32 30| 32 27| 24 25

CLD PC| OV OV| OV PC| CL PC| PC OV| OV OV| PC CL| CL CL

WND 10| 5 11| 11 16| 10 10| 5 9| 6 10| 12 14| 12 12

P12 0| 5 13| 91 13| 3 9| 14 24| 52 54| 48 21| 12 25 20 18

P24 | 16| 100| 9| 26| 62| 72| 25 29

Q12 0| 0 0| 3 0| 0 0| 0 0| 2 2| 2 |

Q24 | 0| 3| 0| 0| 4| |

T12 0| 0 0| 3 0| 0 0| 0 4| 6 4| 3 1| 1 1

T24 | 0 | 3 | 0 | 0 | 6 | 4 | 1

PZP 12| 9 12| 4 3| 5 6| 10 8| 8 3| 16 10| 12 8

PSN 62| 15 3| 0 0| 10 9| 15 24| 1 0| 9 32| 27 18

PRS 26| 24 7| 0 17| 18 20| 13 15| 1 2| 18 9| 11 11

TYP S| RS R| R R| R R| R RS| R R| R RS| RS R

SNW | 0| 0| 0| 0| | |

Page 20: Model Output Statistics (MOS) - Objective Interpretation

MOS station-oriented products:

Other additions

Page 21: Model Output Statistics (MOS) - Objective Interpretation

44004 GFS MOS GUIDANCE 11/22/2005 1200 UTC

DT /NOV 22/NOV 23 /NOV 24 /NOV 25

HR 18 21 00 03 06 09 12 15 18 21 00 03 06 09 12 15 18 21 00 03 06

TMP 58 53 49 49 50 48 46 44 44 45 47 48 51 54 56 60 62 61 59 51 47

WD 23 25 27 28 28 29 29 28 28 27 27 25 22 22 22 23 23 23 24 27 28

WS 33 31 29 25 23 22 24 25 23 18 14 12 14 19 26 29 30 29 29 28 24

WS10 36 34 31 26 25 24 26 27 25 19 15 13 15 21 28 31 32 31 31 30 26

DT /NOV 25 /

HR 09 12 15 18 21 00

TMP 45 45 45 47 47 47

WD 29 29 28 30 29 34

WS 18 15 10 10 13 12

WS10 20 16 11 11 14 13

Marine MOS

Marine MOS sites

Standard MOS sites

Page 22: Model Output Statistics (MOS) - Objective Interpretation

GFS-BASED MOS COOP MAX/MIN GUIDANCE 3/01/05 1800 UTC

WED 02| THU 03| FRI 04

ANNM2 26 46| 24 45| 25 46

BERM2 28 41| 25 39| 25 43

BTVM2 23 39| 21 38| 20 43

CBLM2 20 40| 18 39| 20 46

CHEM2 25 42| 21 39| 21 44

CNWM2 21 42| 21 40| 20 45

DMAM2 20 37| 18 37| 20 42

ELCM2 25 41| 21 41| 18 45

EMMM2 23 42| 20 41| 20 43

FREM2 23 46| 21 42| 23 44

FRSM2 17 27| 13 27| 13 36

GLDM2 21 37| 18 39| 18 43

HAGM2 23 43| 18 43| 19 45

KAPG 27 41| 23 37| 22 43

LRLM2 23 44| 21 42| 22 46

MECM2 24 47| 20 42| 20 45

MILM2 25 48| 22 41| 20 39

MLLM2 22 39| 18 37| 18 41

OLDM2 18 31| 13 28| 12 35

OXNM2 23 42| 22 40| 23 48

PRAM2 22 49| 22 45| 18 45

Max/Min Guidance for Co-op Sites

Glenn Dale, MD

Laurel 3 W

Beltsville, MD

Page 23: Model Output Statistics (MOS) - Objective Interpretation

Wake, US

Midway, US

Saipan, ROM

Western Pacific MOS Guidance

NSTU GFS MOS GUIDANCE 11/07/2008 1200 UTC

DT /NOV 7/NOV 8 /NOV 9 /NOV 10

HR 18 21 00 03 06 09 12 15 18 21 00 03 06 09 12 15 18 21 00 06 12

TMP 84 85 85 85 82 82 81 79 80 83 84 83 81 81 80 79 81 84 86 82 80

DPT 77 77 78 77 76 77 76 75 77 78 77 77 76 77 76 75 77 78 77 77 76

WDR 08 08 08 09 08 07 05 04 06 07 08 07 05 02 35 01 02 07 07 08 10

WSP 17 17 15 13 11 08 07 07 07 08 09 08 07 05 04 04 04 08 09 06 06

P06 36 37 47 46 50 43 25 35 43 30 31

P12 60 66 60 59 47

Page 24: Model Output Statistics (MOS) - Objective Interpretation

Application of Linear Regression

to MOS Development

Page 25: Model Output Statistics (MOS) - Objective Interpretation

MOS LINEAR REGRESSION JANUARY 1 - JANUARY 30, 1994 0000 UTC

KCMH

18-H NGM 850-1000 MB THICKNESS (M)

TO

DA

Y'S

MA

X (°

F)

1150 1200 1250 1300 1350

10

20

30

40

50

60

0

-10

Page 26: Model Output Statistics (MOS) - Objective Interpretation

MOS LINEAR REGRESSION JANUARY 1 - JANUARY 30, 1994 0000 UTC

KCMH

18-H NGM 850-1000 MB THICKNESS (M)

TO

DA

Y'S

MA

X (°

F)

1150 1200 1250 1300 1350

10

20

30

40

50

60

0

-10

RV=93.1%

MAX T = -352 + (0.3 x 850-1000 mb THK)

Page 27: Model Output Statistics (MOS) - Objective Interpretation

REDUCTION OF VARIANCE

A measure of the “goodness” of fit and

Predictor / Predictand correlation

PREDICTOR

PR

ED

ICTA

ND

MEAN

RV

RV

Variance - Standard Error =

Variance

{ } UNEXPLAINED VARIANCE

*

Page 28: Model Output Statistics (MOS) - Objective Interpretation

MOS LINEAR REGRESSION JANUARY 1 - JANUARY 30, 1994 0000 UTC

KUIL

18-H NGM 850-1000 MB THICKNESS (M)

TO

DA

Y'S

MA

X (°

F)

1250 1300 1350 1400 30

40

50

60

RV=26.8%

Different site,

Different relationship!

Same predictor,

Page 29: Model Output Statistics (MOS) - Objective Interpretation

MOS LINEAR REGRESSION

DECEMBER 1 1993 - MARCH 5 1994 0000 UTC

KCMH

AVG. 12-24 H NGM ~1000 - 500 MB RH

12

-24

H P

RE

CIP

ITA

TIO

N ≥

.0

1"

10 20 30 40 50 60 70 80 90 100

0

1

Page 30: Model Output Statistics (MOS) - Objective Interpretation

MOS LINEAR REGRESSION

DECEMBER 1 1993 - MARCH 5 1994 0000 UTC

KCMH

AVG. 12-24 H NGM ~1000 - 500 MB RH

12

-24

H P

RE

CIP

ITA

TIO

N ≥

.0

1"

10 20 30 40 50 60 70 80 90 100

0

1

RV=36.5%

Page 31: Model Output Statistics (MOS) - Objective Interpretation

MOS LINEAR REGRESSION

DECEMBER 1 1993 - MARCH 5 1994 0000 UTC

KCMH

AVG. 12-24 H NGM ~1000 - 500 MB RH

12

-24

H P

RE

CIP

ITA

TIO

N ≥

.0

1"

10 20 30 40 50 60 70 80 90 100

0

1

RV=36.5%

RV=42.4%

Page 32: Model Output Statistics (MOS) - Objective Interpretation

MOS LINEAR REGRESSION

DECEMBER 1 1993 - MARCH 5 1994 0000 UTC

KCMH

AVG. 12-24 H NGM ~1000 - 500 MB RH

12

-24

H P

RE

CIP

ITA

TIO

N ≥

.0

1"

10 20 30 40 50 60 70 80 90 100

0

1

RV=36.5%

RV=42.4%

RV=44.9%

POP = -0.234 + (0.007 X MRH) +

(0.478 X BINARY MRH (70%))

Page 33: Model Output Statistics (MOS) - Objective Interpretation

EXAMPLE REGRESSION EQUATIONS

Y = a + bX

CMH MAX TEMPERATURE EQUATION

CMH PROBABILITY OF PRECIPITATION EQUATION

MAX T = -352 + (0.3 x 850 -1000 mb THICKNESS)

POP = -0.234 + (0.007 x MEAN RH)

+ (0.478 x BINARY MEAN RH CUTOFF AT 70%)*

*(IF MRH ≥ 70% BINARY MRH = 1; else BINARY MRH = 0)

Page 34: Model Output Statistics (MOS) - Objective Interpretation

If the predictand is BINARY,

MOS regression equations produce

estimates of event PROBABILITIES... KCMH

AVG. 12-24 H NGM ~1000 - 500 MB RH

12

-24

H P

RE

CIP

ITA

TIO

N ≥

.0

1"

10 20 30 40 50 60 70 80 90 100

0

1

3 Events

7 Events

RF= 30% P = 30%

Page 35: Model Output Statistics (MOS) - Objective Interpretation

Making a PROBABILISTIC

statement...

Quantifies the uncertainty !

Page 36: Model Output Statistics (MOS) - Objective Interpretation

DEFINITION of PROBABILITY

(Wilks, 2006)

● The degree of belief, or quantified judgment,

about the occurrence of an uncertain event.

OR

● The long-term relative frequency of an event.

Page 37: Model Output Statistics (MOS) - Objective Interpretation

PROBABILITY FORECASTS

● POINT PROBABILITIES

● AREAL PROBABILITIES

● CONDITIONAL PROBABILITIES

Assessment of probability is EXTREMELY dependent

upon how predictand “event” is defined:

Some things to keep in mind

-Time period of consideration

-Area of occurrence

-Dependent upon another event?

MOS forecasts can be:

Page 38: Model Output Statistics (MOS) - Objective Interpretation

3H Eta MOS thunderstorm probability forecasts

valid 0000 UTC 8/27/2002 (21-24h proj)

40-km gridbox

10% contour interval 20-km gridbox

10% contour interval

AREAL PROBABILITIES

What if these were 6-h

forecasts?

Page 39: Model Output Statistics (MOS) - Objective Interpretation

PROPERTIES OF

MOS PROBABILITY FORECASTS

● Unbiased Average forecast probability equals

long-term relative frequency of event

● Reliable Conditionally or “Piecewise” unbiased

over entire range of forecast probabilities

● Reflect predictability of event Range narrows and approaches event RF

as NWP model skill declines

- extreme forecast projection

- rare events

Page 40: Model Output Statistics (MOS) - Objective Interpretation

Reliable Probabilities…

Reliabilty of 12-h PQPF > 0.25", 48h Forecasts

Cool Seasons 05-06 and 06-07, 335 sites

0%

10%

20%

30%

40%

50%

60%

70%

80%

90%

100%

0% 10% 20% 30% 40% 50% 60% 70% 80% 90% 100%

Forecast

Re

lati

ve

Fre

qu

en

cy

65

182

400

638

800

1159

1437

2108

3467

8611

6192933811

0

10000

20000

30000

40000

50000

60000

700000.0

%2.5

%10.0

%20.0

%30.0

%40.0

%50.0

%60.0

%70.0

%80.0

%90.0

%97.5

%

Mean: 4.7%

Even for rare events

12-h Precip > 0.25”

Page 41: Model Output Statistics (MOS) - Objective Interpretation

Designing an

Operational MOS System:

Putting theory into practice…

Page 42: Model Output Statistics (MOS) - Objective Interpretation

DEVELOPMENTAL CONSIDERATIONS

● Selection (and QC!) of Suitable

Observational Datasets

ASOS? Remote sensor? Which mesonet?

MOS in the real world

Page 43: Model Output Statistics (MOS) - Objective Interpretation

Suitable observations?

Appropriate Sensor?

Good siting? Real or Memorex?

Photo Courtesy W. Shaffer

Page 44: Model Output Statistics (MOS) - Objective Interpretation

MOS Snowfall Guidance

36-hr forecast

12Z 12/05/03 – 12Z 12/06/03

6 - <8

4 - < 6

2 - < 4

>Trace - 2

Verification

Uses Observations from Cooperative Observer Network

Page 45: Model Output Statistics (MOS) - Objective Interpretation

DEVELOPMENTAL CONSIDERATIONS

● Selection (and QC!) of Suitable

Observational Datasets

ASOS? Remote sensor? Which mesonet?

● Predictand Definition Must be precise !!

MOS in the real world

Page 46: Model Output Statistics (MOS) - Objective Interpretation

PREDICTAND DEFINITION

Daytime Maximum Temperature “Daytime” is 0700 AM - 0700 PM LST *

Nighttime Minimum Temperature “Nighttime” is 0700 PM - 0800 AM LST *

* CONUS – differs in AK

Probability of Precipitation Precipitation occurrence is accumulation

of ≥ 0.01 inches of liquid-equivalent at a

gauge location within a specified period

Max/Min and PoP

Page 47: Model Output Statistics (MOS) - Objective Interpretation

PREDICTAND DEFINITION

● Determined from 13 consecutive hourly ASOS observations, satellite augmented

● Assign value to each METAR report: CLR; FEW; SCT; BKN; OVC

0 ; 0.15; 0.38; 0.69; 1

● Take weighted average of above

● Categorize: CL < .3125 ≤ PC ≤ .6875 < OV

GFSX 12-h Average Cloud Amount

Page 48: Model Output Statistics (MOS) - Objective Interpretation

Creating a Gridded Predictand

××××× ×

××× ×××

××× ×

××××××

××× ×

××× ×

××××× ×

××× ×××

××× ×

××××××

××× ×

××× ×

= thunderstorm = no thunderstorm

Lightning strikes are summed over the “appropriate” time

period and assigned to the center of “appropriate” grid boxes

A thunderstorm is deemed to have occurred when one or

more lightning strikes are observed within a given gridbox:

Page 49: Model Output Statistics (MOS) - Objective Interpretation

DEVELOPMENTAL CONSIDERATIONS

● Selection (and QC!) of Suitable

Observational Datasets

ASOS? Remote sensor? Which mesonet?

● Predictand Definition Must be precise !!

● Choice of Predictors “Appropriate” formulation

Binary or other transform?

MOS in the real world

Page 50: Model Output Statistics (MOS) - Objective Interpretation

“APPROPRIATE” PREDICTORS

● DESCRIBE PHYSICAL PROCESSES ASSOCIATED

WITH OCCURRENCE OF PREDICTAND

● “MIMIC” FORECASTER THOUGHT PROCESS

(VERTICAL VELOCITY) X (MEAN RH)

PRECIPITABLE WATER

VERTICAL VELOCITY

MOISTURE DIVERGENCE

MODEL PRECIPITATION

1000-500 MB THK

TROPOPAUSE HGT X i.e. for POP:

Page 51: Model Output Statistics (MOS) - Objective Interpretation

POINT BINARY PREDICTOR

24-H MEAN RH CUTOFF = 70% INTERPOLATE ; STATION RH ≥ 70% , BINARY = 1

BINARY = 0 OTHERWISE

96 86 89 94

87 73 76 90

76 60 69 92

64 54 68 93

RH ≥70% ; BINARY AT KCMH = 1

• KCMH

(71%)

Page 52: Model Output Statistics (MOS) - Objective Interpretation

GRID BINARY PREDICTOR

24 H MEAN RH CUTOFF = 70% WHERE RH ≥ 70% ; GRIDPOINT = 1 ; INTERPOLATE

1 1 1 1

1 1 1 1

1 0 0 1

0 0 0 1

0 ≤ VALUE AT KCMH ≤ 1

• KCMH

(.21 )

Page 53: Model Output Statistics (MOS) - Objective Interpretation

-0.2

0

0.2

0.4

0.6

0.8

1

1.2

250 255 260 265 270 275 280 285 290

850 MB TEMP

PR

OB

. o

f F

RO

ZE

N (P

OZ

)Logit Transform Example

POZ = - 0.01 + 0.9444 (LOGIT TRAN (850 T))

RV = 0.7209

1 + e- (a + bx)

________1

POZ = 12.5 - 0.0446 (850 T)

RV = 0.6136

KPIA (Peoria, IL) 0000 UTC ; 18-h projection

Page 54: Model Output Statistics (MOS) - Objective Interpretation

DEVELOPMENTAL CONSIDERATIONS

● Terms in Equations; Selection Criteria

(cont.)

Page 55: Model Output Statistics (MOS) - Objective Interpretation

“REAL” REGRESSION EQUATIONS

MULTIVARIATE , of form:

Y = a + a X + a X + ... + a X

large:

FORWARD SELECTION procedure determines the

2 N N 1 0 1 2

MOS regression equations are

Where,

the "a's" represent COEFFICIENTS

the "X's" represent PREDICTOR variables

The maximum number of terms, N, can be

For GFS QPF, N = 15 For GFS VIS, N = 20

The

predictors and the order in which they appear.

QUITE

Page 56: Model Output Statistics (MOS) - Objective Interpretation

FORWARD SELECTION

● METHOD OF PREDICTOR SELECTION

ACCORDING TO CORRELATION WITH

PREDICTAND

● “BEST” OR STATISTICALLY MOST IMPORTANT

PREDICTORS CHOSEN FIRST

FIRST predictor selected accounts for greatest reduction

of variance (RV)

Subsequent predictors chosen that give greatest RV in conjunction with predictors already selected

selection when desired maximum number of terms is reached or new predictors provide less than a user-specified minimum RV

● STOP

Page 57: Model Output Statistics (MOS) - Objective Interpretation

DEVELOPMENTAL CONSIDERATIONS

● Terms in Equations; Selection Criteria

● Dependent Data Sample Size, Stability, Representativeness

AVOID OVERFIT !!

Stratification - Seasons

Pooling – Regions

(cont.)

Page 58: Model Output Statistics (MOS) - Objective Interpretation

MOS LINEAR REGRESSION

OCTOBER 1 1993 - MARCH 31 1994 0000 UTC

KUIL

12-24 H NGM PRECIPITATION AMOUNT (IN.)

12

-24 H

PR

EC

IPIT

AT

ION

≥ 1

.0"

0.00 0.25 0.50 0.75 1.00

0

1

0

RV=14.2%

Few observed cases,

Limited skill!

Short sample,

Page 59: Model Output Statistics (MOS) - Objective Interpretation

GFS MOS Cool Season PoP/QPF Regions

With GFS MOS forecast sites (1720) + PRISM

Page 60: Model Output Statistics (MOS) - Objective Interpretation

DEVELOPMENTAL CONSIDERATIONS

● Terms in Equations; Selection Criteria

● Dependent Data Sample Size, Stability, Representativeness

AVOID OVERFIT !!

Stratification - Seasons

Pooling – Regions

● Categorical Forecasts?

(cont.)

Page 61: Model Output Statistics (MOS) - Objective Interpretation

0

20

40

60

80

0.01" 0.10" 0.25" 0.50" 1.00" 2.00"

PRECIPITATION AMOUNT EQUAL TO OR EXCEEDING

FORECAST

THRESHOLD

MOS BEST CATEGORY SELECTION

KDCA 12-Hour QPF Probabilities 48-Hour Projection valid 1200 UTC 10/31/93

PR

OB

AB

ILIT

Y (

%)

THRESHOLD

EXCEEDED?

TO MOS GUIDANCE MESSAGES

Page 62: Model Output Statistics (MOS) - Objective Interpretation

How well do we do?

MOS Verification

Page 63: Model Output Statistics (MOS) - Objective Interpretation

2-M TEMPERATURE MAE at 1591 STATIONS

1

2

3

4

5

6

7

8

6 18 30 42 54 66 78 90 102 114 126 138 150 162 174 190

PROJECTION (HOURS)

MA

E (

DE

GR

EE

S F

)

GFS DMO

GFS MOS

Temperature Verification - 0000 UTC GFS MOS vs. GFS DMO (4/2004 - 5/2006)

Page 64: Model Output Statistics (MOS) - Objective Interpretation

Temperature Verification - 0000 UTC

Warm Season: April – September, 2010

Mean Absolute Error - 00Z Temperatures

CONUS (300 stations)

April 1 - September 30, 2010

1.5

2

2.5

3

3.5

4

6 12 18 24 30 36 42 48 54 60 66 72 78 84

Projection (hours)

MA

E (

de

gre

es

F)

NAM GFS

Page 65: Model Output Statistics (MOS) - Objective Interpretation

Temperature Bias - 0000 UTC Warm Season: April – September, 2010

Bias - 00Z Temperature

CONUS (300 stations)

April 1 - September 30, 2010

-2

-1

0

1

2

6 12 18 24 30 36 42 48 54 60 66 72 78 84Projection (hours)

Me

an

Alg

. E

rro

r (

F)

NAM GFS

Page 66: Model Output Statistics (MOS) - Objective Interpretation

Temperature Bias - 0000 UTC 10/06; 01/07; 03/08

Bias - 00z Temperature

CONUS - 300 Stations

Oct. 1 - 31, 2006, Jan. 1 - 31 2007, Mar. 1 - 31 2008

-2

-1

0

1

2

6 12 18 24 30 36 42 48 54 60 66 72 78 84

Projection (hrs)

Me

an

Alg

. E

rro

r (

F)

NAM GFS

Projection (Hours)

Having a representative verification sample is

important, too!

Page 67: Model Output Statistics (MOS) - Objective Interpretation

PoP Verification - 0000 UTC 10/06; 01/07; 03/08

6-h PoP Verif; Oct 2006, Jan 2007, Mar 2008

0.04

0.05

0.06

0.07

0.08

0.09

0.1

12 18 24 30 36 42 48 54 60 66 72 78 84

Forecast Projection (Hrs)

Bri

er

Sc

ore

NAM GFS

Page 68: Model Output Statistics (MOS) - Objective Interpretation

0%

10%

20%

30%

40%

50%

97-98 98-99 99-00 00-01 01-02 02-03 03-04

F48 F96

F144 F192

GFSX 12-h Forecast Skill - 0000 UTC Max Temperatures and PoP

0%

10%

20%

30%

40%

50%

60%

70%

97-98 98-99 99-00 00-01 01-02 02-03 03-04

Day 2 Day 4

Day 6 Day 8

% Improvement over Climate

Cool Season 1997 - 2003

Max T PoP

Page 69: Model Output Statistics (MOS) - Objective Interpretation

45-yr Max Temperature Verification

Guidance / WFO; Cool Season 1966 - 2010

2.0

2.5

3.0

3.5

4.0

4.5

5.0

5.5

6.0

6.5

Mean

Ab

so

lute

Err

or

(F)

NAM

LFM NGM

EDAS AVN

Day/Nite

Perf. Pg. /

PE MOS

24-h

48-h

2010 1970 1980 1990 2000

GFS

48-h Guidance

48-h WFO

24-h Guidance

24-h WFO

Page 70: Model Output Statistics (MOS) - Objective Interpretation

Dealing with

NWP model changes

Page 71: Model Output Statistics (MOS) - Objective Interpretation

1. Improved model realism better model = better statistical system

2. Coarse, consistent archive grid smoothing of fine-scale detail

constant mesh length for grid-sensitive calculations

3. Enlarged geographic regions larger data pools help to stabilize equations

4. Use of “robust” predictor variables fewer boundary layer variables

variables likely immune to known model changes;

(e.g. combinations of state variables only)

Mitigating the effects on development

To help reduce the impact of model changes and small sample size, we rely upon...

MODELS

NCEP

OBS

Mesonets

Page 72: Model Output Statistics (MOS) - Objective Interpretation

• Parallel evaluation Run MOS…new vs. old NWP model

Assess impacts on MOS skill

Responding to NWP Model Changes

Page 73: Model Output Statistics (MOS) - Objective Interpretation

6h PoP Verification

335 Stations, 00Z runs, March-April 2006

0.06

0.07

0.08

0.09

0.10

12 18 24 30 36 42 48 54 60 66 72 78 84

PROJECTION (h)

BR

IER

SC

OR

E

ETAMOS

ETA-NMM

GFSMOS

Responding to NWP Model Changes

Eta MOS PoP: Eta vs. NMM output

Page 74: Model Output Statistics (MOS) - Objective Interpretation

• Parallel evaluation Run MOS…new vs. old NWP model

Assess impacts on MOS skill

• Do nothing? OK if impacts are minimal

But, often they aren’t! (GFS wind / temps)

Responding to NWP Model Changes

Page 75: Model Output Statistics (MOS) - Objective Interpretation

2009 - 2011 GFS MOS Wind Bias

2009

2010 May-Jul

2011

Jan-Apr 2011

Page 76: Model Output Statistics (MOS) - Objective Interpretation

-1

0

1

2

3

4

5

6

7

8

6 9 12 15 18 21 24 27 30 33 36 39 42 45 48 51 54 57 60 63 66 69 72 75 78 81 84

Wind Speed Bias for KABQ July - Sept. 2010 (00Z Cycle)

OPER PARA

GFSMOS”fix”

GFSMOSold

Bia

s (

kn

ots

)

Forecast Projection

0

Page 77: Model Output Statistics (MOS) - Objective Interpretation

• Parallel evaluation Run MOS…new vs. old NWP model

Assess impacts on MOS skill

• Do nothing? OK if impacts are minimal

But, often they aren’t! (GFS wind / temps)

• OK, now what? • Model changes may be recent

i.e. limited sample available from new model version

• Error characteristics significantly different

• Undesirable effects on MOS performance

Responding to NWP Model Changes

Page 78: Model Output Statistics (MOS) - Objective Interpretation

• Bias Correction for MOS?

Responding to NWP Model Changes

Page 79: Model Output Statistics (MOS) - Objective Interpretation

Daily Bias Correction based on past N (7, 10, 20 or 30)- day forecast errors

Bias correction:

F' = F (t) – Bias

N

1t

O(t)][F(t)N

1Bias

Daily biases can be treated equally or

weighted to favor most recent days, etc.

Today

P1

Future

P2

Past

Past N days

t = N 1

F = Forecasts ; O = Observations

N = Days in training sample

(typically, N = 7, 10, 20, or 30)

Page 80: Model Output Statistics (MOS) - Objective Interpretation

Raw / Corrected GFS MOS Wind MAE

KABQ – 00UTC, 96-h Projection

Page 81: Model Output Statistics (MOS) - Objective Interpretation

Raw / Corrected GFS MOS Temp MAE

Southwest U.S. – 00UTC, 48-h Projection

Page 82: Model Output Statistics (MOS) - Objective Interpretation

• Bias Correction for MOS? Apply to Temps? Winds?

Run continuously in background?

Satisfactory in rapidly-varying conditions?

• Redevelop? Short sample from new model or “mixed”?

Full System, selected elements?

Biggest impacts on single-station equations (Temp, Wind)

Responding to NWP Model Changes

Page 83: Model Output Statistics (MOS) - Objective Interpretation

MOS DEWPOINT BIAS

OVERALL

-0.50

0.00

0.50

1.00

1.50

2.00

2.50

3.00

3.50

6 12 18 24 30 36 42 48 54 60 66 72 78 84

PROJECTION (h)

BIA

S (

deg

F)

NMM

ETAonNMM

ETA

MOS DEWPOINT MAE

OVERALL

2.00

2.50

3.00

3.50

4.00

4.50

5.00

5.50

6 12 18 24 30 36 42 48 54 60 66 72 78 84

PROJECTION (h)

MA

E (

deg

F)

NMM

ETAonNMM

ETA

0.00 0.00

350 Stations; CONUS + AK

NAM

NAM

NAM / Eta MOS Dewpoint Comparison Jul 15-31, 2006 and May 1-15, 2007

Using even just a little data

from the new NWP model

version can be helpful!

Page 84: Model Output Statistics (MOS) - Objective Interpretation

KORD ETA MOS GUIDANCE 9/27/2007 1200 UTC DT /SEPT 27/SEPT 28 /SEPT 29 /SEPT 30 HR 18 21 00 03 06 09 12 15 18 21 00 03 06 09 12 15 18 21 00 06 12 N/X 50 72 50 78 57 TMP 67 69 64 60 56 53 52 64 70 70 65 58 55 53 53 65 74 77 71 61 58 DPT 54 52 51 50 48 46 46 48 45 44 44 47 47 47 48 50 49 49 51 55 53 CLD OV BK SC SC CL CL CL CL CL CL CL CL CL CL CL CL CL CL CL CL FW WDR 23 27 30 30 29 28 30 32 33 34 06 11 16 17 17 18 18 19 17 19 20 WSP 09 10 08 06 05 05 04 07 08 07 06 02 02 02 04 08 09 10 08 08 08 P06 19 3 6 1 0 0 1 3 3 8 10 P12 6 1 1 6 12 Q06 0 0 0 0 0 0 0 0 0 0 0 Q12 0 0 0 0 0 T06 1/ 0 9/ 7 0/ 0 0/ 7 0/ 0 0/ 1 0/ 0 0/ 8 2/ 0999/99 T12 9/ 7 0/ 7 0/ 1 0/ 8 999/99 SNW 0 0 0 CIG 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 VIS 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 OBV N N N N N N N N N N N N N N N N N N N N N

“Classic” Eta MOS Hybrid “NAM MOS”

KORD NMM MOS GUIDANCE 9/27/2007 1200 UTC DT /SEPT 27/SEPT 28 /SEPT 29 /SEPT 30 HR 18 21 00 03 06 09 12 15 18 21 00 03 06 09 12 15 18 21 00 06 12 N/X 51 74 49 76 58 TMP 68 65 62 60 57 52 52 63 71 73 69 59 53 51 50 63 72 76 73 60 58 DPT 54 53 53 51 49 48 48 48 47 45 45 47 47 47 48 50 50 50 51 52 55 CLD OV OV SC FW CL CL CL CL CL CL CL CL CL CL CL CL CL CL CL FW FW WDR 23 27 30 30 29 28 30 32 33 34 04 10 16 16 17 18 18 19 17 19 20 WSP 08 10 08 06 05 05 05 07 08 07 05 03 02 03 04 08 09 10 08 08 08 P06 21 3 2 1 1 0 0 3 5 10 12 P12 3 1 0 7 16 Q06 0 0 0 0 0 0 0 0 0 0 0 Q12 0 0 0 0 0 T06 1/ 0 6/ 1 0/ 0 0/ 3 0/ 0 0/ 0 0/ 0 0/ 1 4/ 0999/99 T12 6/ 1 0/ 3 0/ 0 0/ 1 999/99 SNW 0 0 0 CIG 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 VIS 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 OBV N N N N N N N N N N N N N N N N N N N N N

• All Eta model input

• All elements used eta-based

equations

• All NMM model input

• Redeveloped elements use new

NMM-based equations

• Other elements use older Eta-

based equations applied to NMM

Eta MOS Replacement

December 9, 2008

Page 85: Model Output Statistics (MOS) - Objective Interpretation

• Bias Correction for MOS? Apply to Temps? Winds?

Run continuously in background?

Satisfactory in rapidly-varying conditions?

• Redevelop? Short sample from new model or “mixed”?

Full System, selected elements?

Biggest impacts on single-station equations (Temp, Wind)

• Reforecasts? 1-2 year sample probably sufficient for T, Wind

Rare elements need longer or “mixed” sample?

Requires additional supercomputer resources

Responding to NWP Model Changes

Page 86: Model Output Statistics (MOS) - Objective Interpretation

MOS: Today and Beyond

Page 87: Model Output Statistics (MOS) - Objective Interpretation

● GFS / GFSX MOS: Update GFSX Sky Cover equations

(Completes 1200 UTC text message)

Expand GFSX to Day 10 for some elements Update climate normals (1981-2010 NCDC) Bias-corrected T, Td, Max/Min, windspeed

● NAM MOS (Eta MOS replacement): Add precipitation type suite (TYP, POZ, POS) Add 0600 and 1800 UTC cycles? Update remaining eta-based elements Update temperature suite with NMM-b data

The Future of MOS “Traditional” Station-oriented Products

Page 88: Model Output Statistics (MOS) - Objective Interpretation

The Future of MOS

● Western Pacific MOS: Add new elements (Sky Cover, CIG)

● General: Evaluate impacts of NWP model changes Periodic addition of new CONUS sites Gradual phaseout of station-oriented graphics

“Traditional” Station-oriented Products (contd.)

Page 89: Model Output Statistics (MOS) - Objective Interpretation

GFS MOS 24-hr Conditional Probability

of Precipitation ≥ 0.25”

http://www.nws.noaa.gov/mdl/synop

Page 90: Model Output Statistics (MOS) - Objective Interpretation

End of an era? KDCA GFS MOS GUIDANCE 11/27/2006 1200 UTC

DT /NOV 27/NOV 28 /NOV 29 /NOV 30

HR 18 21 00 03 06 09 12 15 18 21 00 03 06 09 12 15 18 21 00 06 12

N/X 44 66 47 63 51

TMP 61 60 54 49 47 45 45 54 62 63 57 51 49 48 49 54 61 61 57 55 53

DPT 39 39 41 44 43 42 42 44 43 44 45 45 46 46 47 49 50 51 52 52 52

CLD OV BK BK FW CL CL SC SC SC SC BK BK BK OV OV OV OV OV OV OV OV

WDR 22 22 23 23 22 24 18 17 17 15 15 16 17 17 15 16 19 19 20 20 21

WSP 03 04 03 02 01 01 01 02 03 03 03 03 01 01 02 02 04 04 04 05 04

P06 0 0 1 1 0 1 4 3 5 8 12

P12 2 4 7 8 16

Q06 0 0 0 0 0 0 0 0 0 0 0

Q12 0 0 0 0 0

T06 0/24 0/ 1 0/ 0 0/ 0 0/16 0/ 0 0/ 0 0/ 0 0/20 0/ 0

T12 0/24 0/ 0 0/16 0/ 0 0/20

POZ 0 0 1 1 1 1 0 0 0 1 2 1 2 1 0 0 1 1 2 2 0

POS 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0

TYP R R R R R R R R R R R R R R R R R R R R R

SNW 0 0 0

CIG 8 8 8 8 8 8 8 8 8 8 8 8 8 4 4 3 4 4 1 2 1

VIS 7 7 7 7 7 7 7 7 7 7 7 7 6 3 1 5 7 7 6 5 1

OBV N N N N N N N N N N N N BR FG FG BR N N HZ BR FG

WANTED! High-resolution, gridded guidance for NDFD

Page 91: Model Output Statistics (MOS) - Objective Interpretation

Gridded MOS

GFS-based CONUS-wide @ 5km

http://www.weather.gov/mdl/synop/

gridded/sectors/index.php

Max / Min

Temp / Td

Winds

RH

Tstm

PoP

Gusts

QPF

Snowfall

Sky Cover

Page 92: Model Output Statistics (MOS) - Objective Interpretation

Wait…Stop the Presses!! 2.5-km CONUS GMOS - “live” on Feb. 27, 2012

Max Temperature

15 UTC 02/13/12

Max Temperature

15 UTC 02/13/12

Page 93: Model Output Statistics (MOS) - Objective Interpretation

Alaska Gridded MOS AK GMOS: GFS-based, 3-km grid

All elements complete

January, 2010

3-km grid

Page 94: Model Output Statistics (MOS) - Objective Interpretation

Hawaii Gridded MOS Hawaii GMOS: GFS-based, 2.5-km grid

Implemented

November, 2010

Max / Min

Temp / Td

Winds

RH

PoP

Gusts

2.5-km grid

Page 95: Model Output Statistics (MOS) - Objective Interpretation

The Future of MOS

“Enhanced-Resolution” Gridded MOS Systems

● “MOS at any point” (e.g. GMOS) Support NWS digital forecast database

2.5 km - 5 km resolution

Equations valid away from observing sites

Emphasis on high-density surface networks

Use high-resolution geophysical data

Page 96: Model Output Statistics (MOS) - Objective Interpretation

• METAR

• Buoys/C-MAN

• Mesonet (RAWS/SNOTEL/Other)

• NOAA cooperative observer network

• RFC-supplied sites

Surface observation systems used in GMOS

Page 97: Model Output Statistics (MOS) - Objective Interpretation

Approx.

1990 sites

Page 98: Model Output Statistics (MOS) - Objective Interpretation
Page 99: Model Output Statistics (MOS) - Objective Interpretation

Approx.

11,000 sites!

Page 100: Model Output Statistics (MOS) - Objective Interpretation

Gridded MOS – Central CA

Page 101: Model Output Statistics (MOS) - Objective Interpretation

Geophysical Datasets

5-km Terrain 5-km Land Cover

Page 102: Model Output Statistics (MOS) - Objective Interpretation

Gridded MOS Concept - Step 1

“Blending” first guess and high-density station forecasts

First guess field from

Generalized Operator Equation

or other source

First guess + guidance

at all available sites

Day 1 Max Temp

00 UTC 03/03/05

Day 1 Max Temp

00 UTC 03/03/05

Page 103: Model Output Statistics (MOS) - Objective Interpretation

Developing the “First Guess” Field

• Generalized operator equation (GOE)

Pool observations regionally

Develop equations for all elements, projections

Apply equations at all grid points within region

• Use average field value at all stations

• Use other user-specified constant

• Use NWP model forecast

Some options

Page 104: Model Output Statistics (MOS) - Objective Interpretation

Gridded MOS Concept - Step 2

First guess + station forecasts +

terrain First guess + guidance

at all available sites

Add further detail to analysis with high-resolution

geophysical data and “smart” interpolation

Day 1 Max Temp

00 UTC 03/03/05

Day 1 Max Temp

00 UTC 03/03/05

Page 105: Model Output Statistics (MOS) - Objective Interpretation

GMOS Analysis

• Method of successive corrections (“BCDG”)

Bergthorssen and Doos (1955); Cressman (1959);

Glahn (1985, LAMP vertical adjustment)

• Elevation (“lapse rate”) adjustment

Inferred from forecasts at different elevations Calculations done “on the fly” from station data Can vary by specific element, synoptic situation

• Land/water gridpoints treated differently

Basic Methodology (Glahn, et al. 2009, WaF)

Page 106: Model Output Statistics (MOS) - Objective Interpretation

GMOS Analysis

• Special, terrain-following smoother

• ROI can be adjusted to account for variations

in density of observed data

• Nudging can be performed to help preserve

nearby station data

• Parameters can be adjusted individually for

each weather element

Other Features

Page 107: Model Output Statistics (MOS) - Objective Interpretation

GMOS Analysis

• Not optimized for all weather elements and synoptic situations

Need situation specific, dynamic models?

• May not capture localized variations in vertical structure

Vertical adjustment uses several station “neighbors”

• May have problems in data-sparse regions over flat terrain Defaults to pure Cressman analysis with small ROI Can result in some “bulls-eye” features

Some Issues

Page 108: Model Output Statistics (MOS) - Objective Interpretation

NDGD vs. NDFD

NDFD Max T NDGD Max T

Which is “better”?

Max Temperature

00 UTC 03/11/06

Max Temperature

00 UTC 03/11/06

Page 109: Model Output Statistics (MOS) - Objective Interpretation

NDGD vs. NDFD

NDFD RH NDGD R H

Which is “better”?

Relative Humidity

21 UTC 03/10/06

Relative Humidity

21 UTC 03/10/06

Forecasters adding detail: Which is “better”? More accurate?

Fewer obs available to analysis = less detail in GMOS

Page 110: Model Output Statistics (MOS) - Objective Interpretation

AK GMOS Temps & Observing Sites

3-km grid

Even fewer obs available – Yikes!

Page 111: Model Output Statistics (MOS) - Objective Interpretation

The Future of MOS

“Enhanced-Resolution”, Gridded MOS Systems

● “MOS at any point” (e.g. GMOS) Support NWS digital forecast database

2.5 km - 5 km resolution

Equations valid away from observing sites

Emphasis on high-density surface networks

Use high-resolution geophysical data

● “True” gridded MOS Observations and forecasts valid on fine grid

Use remotely-sensed predictand data

e.g. WSR-88D QPE, Satellite clouds, NLDN

Page 112: Model Output Statistics (MOS) - Objective Interpretation

Remotely-sensed precipitation data

Page 113: Model Output Statistics (MOS) - Objective Interpretation

Gridded MOS: Where do we go from here?

• Additions to current CONUS GMOS system Present weather grid NAM-based companion system (short-range) Probabilistic and/or ensemble-based products

The Future of MOS

Page 114: Model Output Statistics (MOS) - Objective Interpretation

Sample Forecast as Quantile Function (CDF)

25

30

35

40

45

50

0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1

Probability

Te

mp

era

ture

(72-h Temp KBWI 12/14/2004)

Page 115: Model Output Statistics (MOS) - Objective Interpretation

Ensemble MOS [5%, 95%] Quantile Temps.

Page 116: Model Output Statistics (MOS) - Objective Interpretation

Gridded MOS: Where do we go from here?

• Additions to current CONUS GMOS system “Predominant” weather grid NAM-based companion system (short-range) Probabilistic and/or ensemble-based products

• Increase CONUS resolution from 5-km to 2.5-km NCEP jobstream Feb. 2012; awaits comms upgrade

• Update land / water mask based on WFO input

• Improve GMOS interpolation procedures

The Future of MOS

Page 117: Model Output Statistics (MOS) - Objective Interpretation

• Increase utilization of mesonet data Investigate MADIS archive (NCO/TOC/ESRL) ~20,000 additional sites?

• Incorporate remotely-sensed data where possible SCP augmented clouds (already in use) WSR-88D QPF: March 13, 2012 NSSL MRMS (Multi-radar, Multi-sensor) dataset?

• Expand GMOS for AK; add other OCONUS areas AK: Increase grid extent; improve marine winds Hawaii: add QPF, Sky Cover Puerto Rico

Gridded MOS: Where do we go from here?

The Future of MOS

Page 118: Model Output Statistics (MOS) - Objective Interpretation

REFERENCES…the “classics”

Wilks, D.: Statistical Methods in the

Atmospheric Sciences, 2nd Ed., Chap. 6,

p. 179 - 254.

Draper, N.R., and H. Smith: Applied

Regression Analysis, Chap. 6, p. 307 - 308.

Glahn, H.R., and D. Lowry, 1972: The use of

model output statistics in objective weather

forecasting, JAM, 11, 1203 - 1211.

Carter, G.M., et al., 1989: Statistical forecasts

based on the NMC’s NWP System, Wea. &

Forecasting, 4, 401 - 412.

Page 119: Model Output Statistics (MOS) - Objective Interpretation

REFERENCES (GMOS)

Glahn, H.R., et al., 2009: The Gridding of MOS.,

Wea. & Forecasting, 24, 520 – 529.