Agenda and Summary for WG3 and WG5 workshops

Wednesday 9 March 2005, Langen

Contents

1. Agenda

Wednesday 9 May 2005 (ppt, 5.5Mb)
09:00-12:45 Joint WG3-WG5 session
WP 5.6.2 Verification of new schemes from WGs 1-3
09:00-09:15 Welcome address Patrizio Emilianiand Marco Arpagaus
09:15-09:30 Introduction to WP 5.6.2 Marco Arpagaus and Matthias Raschendorfer
09:30-10:15 A first attempt to WP 5.6.2 (ppt, 6Mb) Francis Schubiger, Pirmin Kaufmann, Dominique Ruffieux, and Matteo Buzzi
10:15-10:45 Discussion, next steps, i.e.,
  1. Who should run the test cases, WGs 1-3 or WG5? (all)
  2. Test cases for WP 5.6.2: defined by WGs 1-3 or WG5 (or both)? (all)
  3. What can be improved?-First experience from WG5 members … (all)
  4. What could be the next (WG3) candidates for WP 5.6.2? (Arpagaus)
10:45-11:15 coffee break
WP 5.5.1 Validation of near-surface boundary layer processes
11:15-11:45 Impact of soil moisture on near surface atmospheric layers (ppt, 8Mb) Gerd Vogel & Felix Ament
Other results from WG5
11:45-12:15 Analysis of LM behaviour in different configurations on some test cases (ppt, 3.2Mb) Chiara Marsigli
12:15-12:45 Some systematic LM-errors (ppt, 11.7Mb) Ulrich Damrath
12:45-13:45 lunch break
13:45-16:00 WG5 session
13:45-14:00 Verification and Case Studies Overview Patrizio Emiliani
14:00-15:00 Discussion on possible Common Verification Suite implementation using software packages running at other Centres and guidelines definition in relation to further developments
15:00-1600 Discussion of WPs 5.1.7 and 5.2.6 to define common LM, how to have results from these two packages.

2. Summary

The workshop discussion was mainly based on the definition of verification procedures for the needs of WG3 members (generally called "Scientific verification") in order to localize specific problems of the operational model or to verify new LM version. The discussion starts within the new WP 5.6.2.

In relation to this argument, during the meeting, specific topics were discussed: in particular has been clarify that the better verification techniques to use for the needs of the WG3 are the so called "Conditional verifications". This kind of verifications allows a simplification of the relations between different meteorological fields and a reduction of the problem to the analysis of some selected parameters; in this way it's more "easy" to find and solve problems or to optimize the variables parameterization.

At the present time are not available, at different Centres, conditional verification tools able to satisfy the WG3 request. So, it will be necessary to develop this kind of instruments and the best way to do it will be probably with the implementation of the Common Verification Suite.

This work is behind the initial tasks of the CVS, so to do it is necessary the Steering Committee clearance and to find new resources (as discussed during the workshop in the afternoon WG5 session); after that will be created the new Work Package "implementation of the Common Verification Suite for conditional verification".

It's important to remember that the realization of the conditional verification package needs specific request, since the general criteria of the verification and the parameters that must be analyzed should be still defined (standard parameters, special parameters..). The list of criteria for conditional verification will be defined by WG3.

The heart of the conditional verifications probably will be to provide a tool ("data finding tool") that allows to find days / grid-points which satisfy specific criteria, in fact localized the couple "fcs, obs", which satisfy the requested criteria, it will be quite easy to apply the already available verification routine.

The development of the "data finding tool" routine will allow also to extend and/or adapt the already available list of test cases.

At the moment WG5 can produce only partial conditional verification, as showed in the joint workshop with the preliminary results of Schubiger et al.; for WG3 is not enough but we're working to do more and more.

There's a last point to clarify concerning the run of the new LM versions on selected WG3 test cases; this argument will be discussed at the next WG Coord. Meeting.

During WG5 workshop session, held in the afternoon, other two topics was discussed:

06/04/2005, Patrizio Emiliani

3. Wishes and decisions

The main discussion at the joint COSMO WG3-WG5 workshop was devoted to the way forward concerning conditional verification of test cases, test periods, and operational model suites (for the full agenda of the meeting, check the COSMO web-site).

The following decisions have been taken (point 1) and wishes have been formulated (points 2 to 5):

  1. WG3 provides a list of criteria to use for conditional verification

    Deadline for delivery of list: end of April 2005.

  2. WG5 is invited to provide a tool ("data finding tool") that allows to easily and efficiently find days/grid-points which satisfy specific criteria (e.g., the ones specified in the list mentioned under 1; since most current verification packages use in-situ observations, only grid-points for which an observation exists need to be considered at this stage). The selected days/grid-points can then be used for conditional verification and/or for the selection of test cases.

  3. To save manpower (development of the tool!), there should be one universal data finding tool for all the COSMO members, rather than different tools at every service or even different tools connected to each verification package.

    Data finding tool should be delivered by: asap, priority 1.

  4. WG5 is kindly asked to supply a conditional verification of the operational model suites on a regular basis (e.g., every season, as is already done for "normal" verification of the operational model suites).

    Whether every verification should be extended to allow for conditional verifications or whether one should try to implement theses features into the common verification package is left to WG5 (or the COSMO Steering Committee) to decide. However, considering the necessary manpower to extend all the verification packages, an extended common verification package is certainly preferable.

    Start conditional verification of operational model suites: asap, priority 1.

  5. Testing of the reference version should also include conditional verifications. Therefore, the common verification package should be extended accordingly, and WG5 is kindly requested to perform conditional verifications of the reference version once the process of verifying the reference version is established.

    Concerning the selection of test cases and/or test periods for testing the reference version, the data finding tool mentioned under point 2 can be used to extend and/or adapt the already available list of test cases (maintained by Chiara Marsigli)

    Start conditional verification of reference version: asap, priority 2.

  6. Finally, WG3 would very much appreciate if a common and easy-to-use verification tool that allows non-WG5 members to do conditional verifications themselves could be provided, maintained (and used) by WG5. The obvious candidate for this tool is the common verification package.

    Common and easy-to-use conditional verification package should be delivered: after completion of points 2 and 3, priority 3.

In view of the points above, the common verification package gets more and more important and enough manpower should be devoted to its development to ensure it will be able to live up to expectations.

For the minutes: Marco Arpagaus, 24.3.2005.

4. Conditional verification of NWP forecast products

1. General remarks

The classical verification of forecast products is generally based on the evaluation of single elements (e.g. T2m, RR) over specific domains in space and time. The resulting numbers, tables and plots present measures of the overall performance of the model with regard to the product considered. Potential interdependencies between various products are a priori ignored (e.g. cloud cover & near surface temperature). The interpretation of classical verification results with regard to systematic deficiencies in the model simulation of specific processes is also far from trivial. Even though obvious failures of the model may be established, finding the reasons behind these failures is hampered by the integral properties of the classical approach, i.e. the averaging over spatial domains (and time) without consideration of specific details of the atmospheric and/or surface situation. However, if tools existed, which would allow a process or situation dependent verification, e.g. the evaluation of the model forecast quality with regard to the diurnal cycle of near surface temperatures in the absence of clouds, the task of finding the cause of model shortcomings might be made much easier and model improvement would be facilitated. Formulating the verification of forecast products in conjunction with the existence of additional criteria, which are to be met, will be considered as, "conditional verification" (CV).

2. A possible approach to conditional verification of forecast products

The prime purpose of CV is the systematic evaluation of model performance in order to reveal typical shortcomings and the reasons behind them. Applied in a routine mode, it should also have the potential to provide information to the forecasters with regard to the situation and product dependent model reliability (e.g. typical clear sky forecast errors for T2M in contrast to cloudy sky condition errors). The typical approach to CV could consist of the selection of one or several forecast products and one or several mask variables, which would be used to define thresholds for the product, verification (e.g. verification of T2M for grid points with zero cloud cover in model and observations only). However, the masking requirements may occasionally be rather complex as in cases, where the forecast product depends strongly on the, history of the evolution rather than the current state of a certain variable. For this reason, one could define at least four classes of, masks:

  1. time-independent masks
  2. time dependent masks concurrent to the product
  3. time dependent masks with possible time-lag to the verified product
  4. masks depending on an atmospheric, "situation classification"

The first class is certainly the most simple one to handle, but carries nevertheless a certain potential with regard to the purpose of the exercise. Apart from obvious masks like the application of a land-sea filter or the stratification of verification results with regard to orographic height, one could envisage masks like the model soil type, plant cover, roughness length etc. The application of such masks might well reveal that certain deficiencies are more pronounced within certain ranges of a particular masks than outside of that range and thereby lead to the detection of a more specific model deficiency.

The second class concerns more the direct interaction between various model products. Typical examples would be the verification of T2M in the presence or absence of snow, but also the verification of products like surface radiative fluxes stratified by cloud cover.

The third class allows in addition to take into account the "history" of the model (or observation) evolution. E.g. the near surface temperature as an instantaneous model product depends among others on the cloud cover which was present in the period before the verification time. In order to obtain a meaningful CV for such a product/mask combination, one would evaluate the verified variable with respect to the previous evolution and the current state of the mask variable. With this type of masks one might also try to exploit the potential of mask variables like the time-integrated or time-averaged Bowen ratio or other indicators of process relevant model quantities. For the example mentioned, the stratification of results might provide indications with regard to unrealistic soil moisture conditions or shortcomings in the soil and surface scheme.

Finally, the fourth class may need a sophisticated algorithm or even manual intervention in the stratification of all cases with regard to the mask criteria. E.g. if we are interested in the performance of the model in "stable" wintertime conditions, it may not be sufficient to check for a temperature inversion near the earth’s surface or relatively high pressure over the domain considered..

Masking conditions may be formulated in model or observation space or both, depending on the application. E.g. in order to obtain a detailed insight in the models ability to simulate the diurnal cycle of near surface temperatures in cloud-free conditions, neither simulated nor observed cloud cover must exceed the specified threshold value.

Generally several masks may be applied at once, e.g. simultaneous application of cloud cover, soil moisture and plant cover thresholds. Obviously the size of the ensemble will depend on the number of criteria applied. Even though this could be detrimental for the evaluation of a few case studies, in an operational environment this problem would be alleviated by the large number of forecasts generated.

One drawback of CV may be the multitude of verification products arising, if various mask criteria applied to a set of forecasts. In order to avoid excessive demands with regard to the evaluation and interpretation of the results, methods should be developed which enable the automatic distinction between "insignificant" verification results and those which, if closely inspected, bear the potential for the further insight into the model behaviour. E.g. the correlation between a verification measure (e.g. mean error) and corresponding threshold values for the mask criteria could serve for an automatic detection.

3. Scope of the CV-system for future enhancements/improvements

One should attempt to develop tools for CV in a rather generic fashion, so that modifications or extensions with regard to the criteria applied can be handled within an existing framework rather than requiring extensive updates or redesign and -development of the software. As a simple example, one could imagine that a tool which allows the CV of T2M with an upper threshold in soil moisture, should require only a modification of a configuration file to be suitable for CV of the near surface wind within a given range of the surface roughness length.

4. Proposal for a first set of conditional verification set up

A generalised CV tool, which can handle all the potential applications mentioned above requires development resources which exceed those available for the time being. Consequently, in a first step one should try to create a tool which is able to handle at least the more basic applications mentioned above. However, this should be done in a flexible, if possible user configurable way and allow future extensions towards a fully flexible tool that may be applied both to operational and experimental forecast products. Depending on the situation restrictions in data archives will have to be taken into account, when certain CV products are considered for realisation. In the following a first attempt is made to define a small set of products to be verified in conjunction with corresponding criteria. This set should only be considered as a starting point, as it is to be expected that, additional and/or modified configurations will be considered in the future. No attempt is made to order the following table with regard to the expected benefit.

Verified variable Mask variable(s) Criteria Remarks
T2m CLC(t); local time lower & upper thresholds in CLC; local time slots cloud cover thresholds should be applied over the time period preceding the verification time and both to model and observations
T2m Wsoil lower & upper thresholds in (relative to field capacity) soil moisture soil moisture is a multi-layer variable and it may be useful to compute an "effective" soil moisture as average over several layers
T2m SHF/LHF thresholds in Bowen ratio Bowen ratio as an indirect measure of soil wetness needs to be considered as an average over time
CLC(L) vertical stability index "stable" versus unstable situations differences in temperatures at various pressure levels may be used as a stability index; the distinction with regard to stability may be considered as an example for situation dependent masking, e.g. to focus on low level stratus or convective regimes
RR as above as above regime dependent precipitation verification
U10m z0 low, medium, large z0 correlation between wind speed errors and roughness length may point to problems in external parameters
T2m U10m upper threshold in wind speed exclude advection dominated situations in temperature verification
Td2m Wsoil lower & upper thresholds in (relative to field capacity) soil moisture determine the error of dew point temperatures in the case of dry soils versus wet soils
T2m Wsnow No snow/broken snow/snow The temperature error is likely to depend on snow cover yes/no, a broken snow deck might be an indicator for melting snow