Towards an NWP-testbed

Similar documents
Developing Continuous SCM/CRM Forcing Using NWP Products Constrained by ARM Observations

Sub-grid cloud parametrization issues in Met Office Unified Model

Evaluating the Impact of Cloud-Aerosol- Precipitation Interaction (CAPI) Schemes on Rainfall Forecast in the NGGPS

A SURVEY OF CLOUD COVER OVER MĂGURELE, ROMANIA, USING CEILOMETER AND SATELLITE DATA

Partnership to Improve Solar Power Forecasting

Solar Energy Forecasting Using Numerical Weather Prediction (NWP) Models. Patrick Mathiesen, Sanyo Fellow, UCSD Jan Kleissl, UCSD

Investigations on COSMO 2.8Km precipitation forecast

MSG-SEVIRI cloud physical properties for model evaluations

How To Find Out How Much Cloud Fraction Is Underestimated

Application of Numerical Weather Prediction Models for Drought Monitoring. Gregor Gregorič Jožef Roškar Environmental Agency of Slovenia

Guy Carpenter Asia-Pacific Climate Impact Centre, School of energy and Environment, City University of Hong Kong

Mixed-phase layer clouds

Towards assimilating IASI satellite observations over cold surfaces - the cloud detection aspect

Cloud verification: a review of methodologies and recent developments

Diurnal Cycle: Cloud Base Height clear sky

Parameterization of Cumulus Convective Cloud Systems in Mesoscale Forecast Models

How To Find Out How Much Cloud Fraction Is Underestimated

ASSESSMENT OF THE CAPABILITY OF WRF MODEL TO ESTIMATE CLOUDS AT DIFFERENT TEMPORAL AND SPATIAL SCALES

Long-term Observations of the Convective Boundary Layer (CBL) and Shallow cumulus Clouds using Cloud Radar at the SGP ARM Climate Research Facility

Physical properties of mesoscale high-level cloud systems in relation to their atmospheric environment deduced from Sounders

High-resolution Regional Reanalyses for Europe and Germany

Validation n 2 of the Wind Data Generator (WDG) software performance. Comparison with measured mast data - Flat site in Northern France

SAFNWC/MSG Cloud type/height. Application for fog/low cloud situations

Fire Weather Index: from high resolution climatology to Climate Change impact study

MOGREPS status and activities

A system of direct radiation forecasting based on numerical weather predictions, satellite image and machine learning.

Very High Resolution Arctic System Reanalysis for

Continental and Marine Low-level Cloud Processes and Properties (ARM SGP and AZORES) Xiquan Dong University of North Dakota

Cloud Profiling at the Lindenberg Observatory

A Microwave Retrieval Algorithm of Above-Cloud Electric Fields

REGIONAL CLIMATE AND DOWNSCALING

Assimilation of cloudy infrared satellite observations: The Met Office perspective

Clouds and Convection

High Resolution Modeling, Clouds, Precipitation and Climate

Evalua&ng Downdra/ Parameteriza&ons with High Resolu&on CRM Data

Improving Low-Cloud Simulation with an Upgraded Multiscale Modeling Framework

Why aren t climate models getting better? Bjorn Stevens, UCLA

RPG Radiometer Deployment Experience

Solarstromprognosen für Übertragungsnetzbetreiber

Validation of SEVIRI cloud-top height retrievals from A-Train data

II. Related Activities

Improving Accuracy of Solar Forecasting February 14, 2013

Chapter 2 False Alarm in Satellite Precipitation Data

Combining Satellite High Frequency Microwave Radiometer & Surface Cloud Radar Data for Determination of Large Scale 3 D Cloud IWC 서은경

Marko Markovic Department of Earth and Atmospheric Sciences. University of Quebec at Montreal

GCMs with Implicit and Explicit cloudrain processes for simulation of extreme precipitation frequency

Comparison of the Vertical Velocity used to Calculate the Cloud Droplet Number Concentration in a Cloud-Resolving and a Global Climate Model

Cloud/Hydrometeor Initialization in the 20-km RUC Using GOES Data

Cloud-Resolving Simulations of Convection during DYNAMO

Large Eddy Simulation (LES) & Cloud Resolving Model (CRM) Françoise Guichard and Fleur Couvreux

Project Title: Quantifying Uncertainties of High-Resolution WRF Modeling on Downslope Wind Forecasts in the Las Vegas Valley

Forecaster comments to the ORTECH Report

Frank and Charles Cohen Department of Meteorology The Pennsylvania State University University Park, PA, U.S.A.

Comparative Evaluation of High Resolution Numerical Weather Prediction Models COSMO-WRF

Near Real Time Blended Surface Winds

Simulations of Clouds and Sensitivity Study by Wearther Research and Forecast Model for Atmospheric Radiation Measurement Case 4

Fog and low cloud ceilings in the northeastern US: climatology and dedicated field study

A model to observation approach to evaluating cloud microphysical parameterisations using polarimetric radar

Forecasting of Solar Radiation

Improved diagnosis of low-level cloud from MSG SEVIRI data for assimilation into Met Office limited area models

Nowcasting: analysis and up to 6 hours forecast

Assessing the performance of a prognostic and a diagnostic cloud scheme using single column model simulations of TWP ICE

Impact of microphysics on cloud-system resolving model simulations of deep convection and SpCAM

Catastrophe Bond Risk Modelling

NOWCASTING OF PRECIPITATION Isztar Zawadzki* McGill University, Montreal, Canada

Data Integration and long-term planning of the Observing Systems as a cross-cutting process in a NMS

Assessing Cloud Spatial and Vertical Distribution with Infrared Cloud Analyzer

Development of a. Solar Generation Forecast System

Meteorological Forecasting of DNI, clouds and aerosols


Diurnal Cycle of Convection at the ARM SGP Site: Role of Large-Scale Forcing, Surface Fluxes, and Convective Inhibition

Transcription:

Towards an NWP-testbed Ewan O Connor and Robin Hogan University of Reading, UK

Overview Cloud schemes in NWP models are basically the same as in climate models, but easier to evaluate using ARM because: NWP models are trying to simulate the actual weather observed They are run every day In Europe at least, NWP modelers are more interested in comparisons with ARM-like data than climate modelers (not true in US?) But can we use these comparisons to improve the physics? Can compare different models which have different parameterizations But each model uses different data assimilation system Cleaner test if the setup is identical except one aspect of physics SCM-testbed is the crucial addition to the NWP-testbed How do we set such a system up? Start by interfacing Cloudnet processing with ARM products Metrics: test both bias and skill (can only test bias of climate model) Diurnal compositing to evaluate boundary-layer physics

Level 1b Minimum instrument requirements at each site Cloud radar, lidar, microwave radiometer, rain gauge, model or sondes Radar Lidar

Level 1c Instrument Synergy product Example of target classification and data quality fields: Ice Aerosol Liquid Rain

Level 2a/2b Cloud products on (L2a) observational and (L2b) model grid Water content and cloud fraction L2a IWC on radar/lidar grid L2b Cloud fraction on model grid

Cloud fraction Chilbolton Observations Met Office Mesoscale Model ECMWF Global Model Meteo-France ARPEGE Model KNMI RACMO Model Swedish RCA model

Cloud fraction ARM SGP Observations NCEP GFS model ECMWF model ERA Interim Analyses Met Office Global Model

Cloud fraction in 7 models Mean & PDF for 2004 for Chilbolton, Paris and Cabauw 0-7 km All models except DWD underestimate mid-level cloud Some have separate radiatively inactive snow (ECMWF, DWD); Met Office has combined ice and snow but still underestimates cloud fraction Wide range of low cloud amounts in models Not enough overcast boxes, particularly in Met Office model Illingworth et al. (BAMS 2007)

Cloud fraction in 5 models Mean for ARM SGP 2005 2006 All models again underestimate mid-level cloud Météo france shows improvement from 2005 to 2006

Cloud fraction components ECMWF model at ARM SGP for 2005 Mean cloud fraction underestimated. Improves slightly with the inclusion of snow. Clouds are forecast often enough, when snow is included, except in BL. Underestimate of the mid-level cloud fraction amounts, even when snow is included.

Seasonal variation ECMWF NCEP UK Met Office

Diurnal variation Model cloud fraction always lower than observed. Not enough boundary layer cloud during the day. Can we simply scale the model cloud fraction?

Winter 2004

Summer 2004

Omega at 500 mb Model cloud fraction always lower than observed. Not enough cloud in anticyclonic conditions, especially boundary layer cloud. Can we scale cloud fraction? Only in large-scale ascent.

Skill Scores Met Office Global model has much lower skill for high cloud fraction amounts. Most models show more skill in the mid-levels than in the BL. NB. Not all models are shown with the same forecast leadtime!

Skill Scores Met Office Global model has much lower skill for high cloud fraction amounts. Most models show more skill in the mid-levels than in the BL.

Skill score Six years of cloud fraction evaluation over SGP Clearly less skill in summer, often no better than persistence ERA Interim Reanalysis no different to forecast Any improvement in the cloud fraction forecast over time? For Météo France, yes..

Summary and future work Six years of evaluation over SGP (extending to nine) All models underestimate mid- and low-level cloud Skill may be robustly quantified: less skill in summer Infrastructure to interface ARM and Cloudnet data has been tested with cloud fraction, IWC/LWC ongoing So far Met Office, NCEP, ECMWF, Météo-France and ERA Interim processed. Analyses do not show much improvement over NWP forecasts. Next implement code at BNL, with other ARM sites and models. Question: have cloud forecasts improved in 10 years? Next compare with results from SCM-testbed We have the tools to quantify objectively improvements in both bias and skill with changed parameterizations in NWP models and SCMs. Other metrics of performance or compositing methods required?

Joint PDFs of cloud fraction b d a c Raw (1 hr) resolution 1 year from Murgtal DWD COSMO model 6-hr averaging or use a simple contingency table

Contingency tables Model cloud Observed cloud Observed clear-sky a: Cloud hit b: False alarm a = 7194 b = 4098 Model clear-sky c: Miss d: Clear-sky hit c = 4502 d = 41062 DWD model, Murgtal For given set of observed events, only 2 degrees of freedom in all possible forecasts (e.g. a & b), because 2 quantities fixed: - Number of events that occurred n =a +b +c +d - Base rate (observed frequency of occurrence) p =(a +c)/n

Skill versus lead time 2004 2007 Only possible for UK Met Office 12-km model and German DWD 7-km model Steady decrease of skill with lead time Both models appear to improve between 2004 and 2007 Generally, UK model best over UK, German best over Germany An exception is Murgtal in 2007 (Met Office model wins)

Met Office DWD Forecast half life 2004 2007 S( t) = S t / τ 0 2 Fit an inverse-exponential: S 0 is the initial score and τ 1/2 is the half-life Noticeably longer half-life fitted after 36 hours Same thing found for Met Office rainfall forecast (Roberts 2008) First timescale due to data assimilation and convective events Second due to more predictable large-scale weather systems 1/ 2

Why is half-life less for clouds than pressure? Different spatial scales? Convection? Average temporally before calculating skill scores: Absolute score and half-life increase with number of hours averaged