-
Notifications
You must be signed in to change notification settings - Fork 24
/
point_stat.cc
2225 lines (1767 loc) · 81.1 KB
/
point_stat.cc
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
// *=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*
// ** Copyright UCAR (c) 1992 - 2022
// ** University Corporation for Atmospheric Research led(UCAR)
// ** National Center for Atmospheric Research (NCAR)
// ** Research Applications Lab (RAL)
// ** P.O.Box 3000, Boulder, Colorado, 80307-3000, USA
// *=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*
////////////////////////////////////////////////////////////////////////
//
// Filename: point_stat.cc
//
// Description:
// Based on user-specified parameters, this tool compares a
// a gridded forecast field to the point observation output of
// the point pre-processing tools. It computes many verification
// scores and statistics, including confidence intervals, to
// summarize the comparison.
//
// Mod# Date Name Description
// ---- ---- ---- -----------
// 000 04/18/07 Halley Gotway New
// 001 12/20/07 Halley Gotway Allow verification for level 0
// for verifying PRMSL
// 002 01/24/08 Halley Gotway In compute_cnt, print a warning
// message when bad data values are encountered.
// 003 01/24/08 Halley Gotway Add support for writing the matched
// pair data to the MPR file.
// 004 02/04/08 Halley Gotway Modify to read new format of the
// intermediate NetCDF point-observation format.
// 005 02/11/08 Halley Gotway Remove BIAS from the CNT line
// since it's the same as the ME.
// 006 02/12/08 Halley Gotway Fix bug in writing COP line to
// write out OY and ON rather than FY and FN.
// 007 02/12/08 Halley Gotway Enable Point-Stat to read the
// NetCDF output of PCP-Combine.
// 008 02/25/08 Halley Gotway Write the output lines using the
// routines in the vx_met_util library.
// 009 07/01/08 Halley Gotway Add the rank_corr_flag to the
// config file to disable computing rank correlations.
// 010 07/18/08 Halley Gotway Fix bug in the write_mpr routine.
// Replace i_fho with i_mpr.
// 011 09/23/08 Halley Gotway Change argument sequence for the
// GRIB record access routines.
// 012 11/03/08 Halley Gotway Use the get_file_type routine to
// determine the input file types.
// 013 03/13/09 Halley Gotway Add support for verifying
// probabilistic forecasts.
// 014 04/21/09 Halley Gotway Fix bug for resetting obs_var.
// 015 05/07/10 Halley Gotway Rename process_grid() to
// setup_first_pass() and modify its logic.
// 016 05/24/10 Halley Gotway Rename command line options:
// From -ncfile to -point_obs
// From -valid_beg to -obs_valid_beg
// From -valid_end to -obs_valid_end
// 017 05/27/10 Halley Gotway Add -fcst_valid and -fcst_lead
// command line options.
// 018 06/08/10 Halley Gotway Add support for multi-category
// contingency tables.
// 019 06/15/10 Halley Gotway Dump reason codes for why
// point observations were rejected.
// 020 06/30/10 Halley Gotway Enhance grid equality checks.
// 021 10/20/11 Holmes Added use of command line class to
// parse the command line arguments.
// 022 11/14/11 Holmes Added code to enable reading of
// multiple config files.
// 023 02/02/12 Bullock Set default output directory to "."
// 024 03/05/12 Halley Gotway Fix bug in processing command line
// for setting valid end times.
// 025 04/16/12 Halley Gotway Switch to using vx_config library.
// 026 04/27/12 Halley Gotway Move -fcst_valid and -fcst_lead
// command line options to config file.
// 027 02/25/13 Halley Gotway Add duplicates rejection counts.
// 028 08/21/13 Halley Gotway Fix sizing of output tables for 12
// or more probabilstic thresholds.
// 029 07/09/14 Halley Gotway Add station id exclusion option.
// 030 03/02/15 Halley Gotway Add automated regridding.
// 031 07/30/15 Halley Gotway Add conditional continuous verification.
// 032 09/11/15 Halley Gotway Add climatology to config file.
// 033 02/27/17 Halley Gotway Add HiRA verification.
// 034 05/15/17 Prestopnik P Add shape for HiRA, interp and regrid.
// 035 06/16/17 Halley Gotway Add ECLV line type.
// 036 11/01/17 Halley Gotway Add binned climatologies.
// 037 02/06/18 Halley Gotway Restructure config logic to make
// all options settable for each verification task.
// 038 08/15/18 Halley Gotway Add mask.llpnt type.
// 039 08/24/18 Halley Gotway Add ECNT output for HiRA.
// 040 04/01/10 Fillmore Add FCST and OBS units.
// 041 04/08/19 Halley Gotway Add percentile thresholds.
// 042 10/14/19 Halley Gotway Add support for climo distribution
// percentile thresholds.
// 043 11/15/19 Halley Gotway Apply climatology bins to
// continuous and probabilistic statistics.
// 044 01/24/20 Halley Gotway Add HiRA RPS output.
// 045 03/28/21 Halley Gotway Add mpr_column and mpr_thresh
// filtering options.
// 046 05/28/21 Halley Gotway Add MCTS HSS_EC output.
// 047 08/23/21 Seth Linden Add ORANK line type for HiRA.
// 048 09/13/21 Seth Linden Changed obs_qty to obs_qty_inc.
// Added code for obs_qty_exc.
// 049 12/11/21 Halley Gotway MET #1991 Fix VCNT output.
// 050 02/11/22 Halley Gotway MET #2045 Fix HiRA output.
// 051 07/06/22 Howard Soh METplus-Internal #19 Rename main to met_main
//
////////////////////////////////////////////////////////////////////////
using namespace std;
#include <cstdio>
#include <cstdlib>
#include <ctype.h>
#include <dirent.h>
#include <fstream>
#include <math.h>
#include <sys/stat.h>
#include <sys/types.h>
#include <unistd.h>
#include "main.h"
#include "point_stat.h"
#include "vx_statistics.h"
#include "vx_nc_util.h"
#include "vx_regrid.h"
#include "vx_log.h"
#include "seeps.h"
#include "nc_obs_util.h"
#include "nc_point_obs_in.h"
#ifdef WITH_PYTHON
#include "data2d_nc_met.h"
#include "pointdata_python.h"
#endif
////////////////////////////////////////////////////////////////////////
////////////////////////////////////////////////////////////////////////
#define BUFFER_SIZE (DEF_NC_BUFFER_SIZE/2)
static void process_command_line (int, char **);
static void setup_first_pass (const DataPlane &, const Grid &);
static void setup_txt_files();
static void setup_table (AsciiTable &);
static void build_outfile_name(unixtime, int, const char *,
ConcatString &);
static void process_fcst_climo_files();
static void process_obs_file(int);
static void process_scores();
static void do_cts (CTSInfo *&, int, const PairDataPoint *);
static void do_mcts (MCTSInfo &, int, const PairDataPoint *);
static void do_cnt_sl1l2 (const PointStatVxOpt &, const PairDataPoint *);
static void do_vl1l2 (VL1L2Info *&, int, const PairDataPoint *, const PairDataPoint *);
static void do_pct (const PointStatVxOpt &, const PairDataPoint *);
static void do_hira_ens ( int, const PairDataPoint *);
static void do_hira_prob ( int, const PairDataPoint *);
static void do_seeps_agg (const PairDataPoint *);
static void finish_txt_files();
static void clean_up();
static void usage();
static void set_point_obs(const StringArray &);
static void set_ncfile(const StringArray &);
static void set_obs_valid_beg_time(const StringArray &);
static void set_obs_valid_end_time(const StringArray &);
static void set_outdir(const StringArray &);
////////////////////////////////////////////////////////////////////////
int met_main(int argc, char *argv[]) {
int i;
// Process the command line arguments
process_command_line(argc, argv);
// Process the forecast and climo files
process_fcst_climo_files();
// Process each observation netCDF file
for(i=0; i<obs_file.n(); i++) {
process_obs_file(i);
}
// Calculate and print observation summaries
for(i=0; i<conf_info.get_n_vx(); i++) {
conf_info.vx_opt[i].vx_pd.calc_obs_summary();
conf_info.vx_opt[i].vx_pd.print_obs_summary();
}
// Compute the scores and write them out
process_scores();
// Close the text files and deallocate memory
clean_up();
return(0);
}
////////////////////////////////////////////////////////////////////////
const string get_tool_name() {
return "point_stat";
}
////////////////////////////////////////////////////////////////////////
void process_command_line(int argc, char **argv) {
CommandLine cline;
int i;
GrdFileType ftype;
ConcatString default_config_file;
out_dir = ".";
// Check for zero arguments
if(argc == 1) usage();
// Parse the command line into tokens
cline.set(argc, argv);
// Set the usage function
cline.set_usage(usage);
// Add the options function calls
cline.add(set_point_obs, "-point_obs", 1);
cline.add(set_ncfile, "-ncfile", 1);
cline.add(set_obs_valid_beg_time, "-obs_valid_beg", 1);
cline.add(set_obs_valid_end_time, "-obs_valid_end", 1);
cline.add(set_outdir, "-outdir", 1);
// Parse the command line
cline.parse();
// Check for error. There should be three arguments left:
// forecast, observation, and config filenames
if(cline.n() != 3) usage();
// Check that the end_ut >= beg_ut
if(obs_valid_beg_ut != (unixtime) 0 &&
obs_valid_end_ut != (unixtime) 0 &&
obs_valid_beg_ut > obs_valid_end_ut) {
mlog << Error << "\nprocess_command_line() -> "
<< "the ending time ("
<< unix_to_yyyymmdd_hhmmss(obs_valid_end_ut)
<< ") must be greater than the beginning time ("
<< unix_to_yyyymmdd_hhmmss(obs_valid_beg_ut)
<< ").\n\n";
exit(1);
}
// Store the input forecast and observation file names
fcst_file = cline[0];
obs_file.insert(0, cline[1].c_str());
config_file = cline[2];
// Create the default config file names
default_config_file = replace_path(default_config_filename);
// List the config files
mlog << Debug(1)
<< "Default Config File: " << default_config_file << "\n"
<< "User Config File: " << config_file << "\n";
// Read the config files
conf_info.read_config(default_config_file.c_str(), config_file.c_str());
// Get the forecast file type from config, if present
ftype = parse_conf_file_type(conf_info.conf.lookup_dictionary(conf_key_fcst));
// Read forecast file
if(!(fcst_mtddf = mtddf_factory.new_met_2d_data_file(fcst_file.c_str(), ftype))) {
mlog << Error << "\nTrouble reading forecast file \""
<< fcst_file << "\"\n\n";
exit(1);
}
// Use a variable index from var_name instead of GRIB code
bool use_var_id = is_using_var_id(obs_file[0].c_str());
// Store the forecast file type
ftype = fcst_mtddf->file_type();
// Process the configuration
conf_info.process_config(ftype, use_var_id);
// Set the model name
shc.set_model(conf_info.model.c_str());
// Use the first verification task to set the random number generator
// and seed value for bootstrap confidence intervals
rng_set(rng_ptr,
conf_info.vx_opt[0].boot_info.rng.c_str(),
conf_info.vx_opt[0].boot_info.seed.c_str());
// List the input files
mlog << Debug(1)
<< "Forecast File: " << fcst_file << "\n";
for(i=0; i<obs_file.n(); i++) {
mlog << Debug(1)
<< "Observation File: " << obs_file[i] << "\n";
}
return;
}
////////////////////////////////////////////////////////////////////////
void setup_first_pass(const DataPlane &dp, const Grid &data_grid) {
// Unset the flag
is_first_pass = false;
// Determine the verification grid
grid = parse_vx_grid(conf_info.vx_opt[0].vx_pd.fcst_info->regrid(),
&(data_grid), &(data_grid));
// Process the masks
conf_info.process_masks(grid);
// Process the geography data
conf_info.process_geog(grid, fcst_file.c_str());
// Setup the VxPairDataPoint objects
conf_info.set_vx_pd();
// Store the lead and valid times
if(fcst_valid_ut == (unixtime) 0) fcst_valid_ut = dp.valid();
if(is_bad_data(fcst_lead_sec)) fcst_lead_sec = dp.lead();
return;
}
////////////////////////////////////////////////////////////////////////
void setup_txt_files() {
int i, j;
int max_col, max_prob_col, max_mctc_col, max_orank_col;
int n_prob, n_cat, n_eclv, n_ens;
ConcatString base_name;
// Create output file names for the stat file and optional text files
build_outfile_name(fcst_valid_ut, fcst_lead_sec, "", base_name);
/////////////////////////////////////////////////////////////////////
//
// Setup the output STAT file
//
/////////////////////////////////////////////////////////////////////
// Get the maximum number of data columns
n_prob = max(conf_info.get_max_n_fprob_thresh(),
conf_info.get_max_n_hira_prob());
n_cat = conf_info.get_max_n_cat_thresh() + 1;
n_eclv = conf_info.get_max_n_eclv_points();
n_ens = conf_info.get_max_n_hira_ens();
max_prob_col = get_n_pjc_columns(n_prob);
max_mctc_col = get_n_mctc_columns(n_cat);
max_orank_col = get_n_orank_columns(n_ens);
// Determine the maximum number of data columns
max_col = (max_prob_col > max_stat_col ? max_prob_col : max_stat_col);
if (max_mctc_col > max_col) max_col = max_mctc_col;
if (max_orank_col > max_col) max_col = max_orank_col;
// Add the header columns
max_col += n_header_columns + 1;
// Initialize file stream
stat_out = (ofstream *) 0;
// Build the file name
stat_file << base_name << stat_file_ext;
// Create the output STAT file
open_txt_file(stat_out, stat_file.c_str());
// Setup the STAT AsciiTable
stat_at.set_size(conf_info.n_stat_row() + 1, max_col);
setup_table(stat_at);
// Write the text header row
write_header_row((const char **) 0, 0, 1, stat_at, 0, 0);
// Initialize the row index to 1 to account for the header
i_stat_row = 1;
/////////////////////////////////////////////////////////////////////
//
// Setup each of the optional output text files
//
/////////////////////////////////////////////////////////////////////
// Loop through output file type
for(i=0; i<n_txt; i++) {
// Only set it up if requested in the config file
if(conf_info.output_flag[i] == STATOutputType_Both) {
// Initialize file stream
txt_out[i] = (ofstream *) 0;
// Build the file name
txt_file[i] << base_name << "_" << txt_file_abbr[i]
<< txt_file_ext;
// Create the output text file
open_txt_file(txt_out[i], txt_file[i].c_str());
// Get the maximum number of columns for this line type
switch(i) {
case(i_mctc):
max_col = get_n_mctc_columns(n_cat) + n_header_columns + 1;
break;
case(i_pct):
max_col = get_n_pct_columns(n_prob) + n_header_columns + 1;
break;
case(i_pstd):
max_col = get_n_pstd_columns(n_prob) + n_header_columns + 1;
break;
case(i_pjc):
max_col = get_n_pjc_columns(n_prob) + n_header_columns + 1;
break;
case(i_prc):
max_col = get_n_prc_columns(n_prob) + n_header_columns + 1;
break;
case(i_eclv):
max_col = get_n_eclv_columns(n_eclv) + n_header_columns + 1;
break;
case(i_orank):
max_col = get_n_orank_columns(n_ens) + n_header_columns + 1;
break;
default:
max_col = n_txt_columns[i] + n_header_columns + 1;
break;
} // end switch
// Setup the text AsciiTable
txt_at[i].set_size(conf_info.n_txt_row(i) + 1, max_col);
setup_table(txt_at[i]);
// Write the text header row
switch(i) {
case(i_mctc):
write_mctc_header_row(1, n_cat, txt_at[i], 0, 0);
break;
case(i_pct):
write_pct_header_row(1, n_prob, txt_at[i], 0, 0);
break;
case(i_pstd):
write_pstd_header_row(1, n_prob, txt_at[i], 0, 0);
break;
case(i_pjc):
write_pjc_header_row(1, n_prob, txt_at[i], 0, 0);
break;
case(i_prc):
write_prc_header_row(1, n_prob, txt_at[i], 0, 0);
break;
case(i_eclv):
write_eclv_header_row(1, n_eclv, txt_at[i], 0, 0);
break;
case(i_orank):
write_orank_header_row(1, n_ens, txt_at[i], 0, 0);
break;
default:
write_header_row(txt_columns[i], n_txt_columns[i], 1,
txt_at[i], 0, 0);
break;
} // end switch
// Initialize the row index to 1 to account for the header
i_txt_row[i] = 1;
}
}
return;
}
////////////////////////////////////////////////////////////////////////
void setup_table(AsciiTable &at) {
// Justify the STAT AsciiTable objects
justify_stat_cols(at);
// Set the precision
at.set_precision(conf_info.conf.output_precision());
// Set the bad data value
at.set_bad_data_value(bad_data_double);
// Set the bad data string
at.set_bad_data_str(na_str);
// Don't write out trailing blank rows
at.set_delete_trailing_blank_rows(1);
return;
}
////////////////////////////////////////////////////////////////////////
void build_outfile_name(unixtime valid_ut, int lead_sec,
const char *suffix, ConcatString &str) {
//
// Create output file name
//
// Append the output directory and program name
str << cs_erase << out_dir << "/" << program_name;
// Append the output prefix, if defined
if(conf_info.output_prefix.nonempty())
str << "_" << conf_info.output_prefix;
// Append the timing information
str << "_"
<< sec_to_hhmmss(lead_sec) << "L_"
<< unix_to_yyyymmdd_hhmmss(valid_ut) << "V";
// Append the suffix
str << suffix;
return;
}
////////////////////////////////////////////////////////////////////////
void process_fcst_climo_files() {
int i, j;
int n_fcst;
DataPlaneArray fcst_dpa, cmn_dpa, csd_dpa;
unixtime file_ut, beg_ut, end_ut;
// Loop through each of the fields to be verified and extract
// the forecast and climatological fields for verification
for(i=0; i<conf_info.get_n_vx(); i++) {
// Read the gridded data from the input forecast file
n_fcst = fcst_mtddf->data_plane_array(
*conf_info.vx_opt[i].vx_pd.fcst_info, fcst_dpa);
mlog << Debug(2)
<< "\n" << sep_str << "\n\n"
<< "Reading data for "
<< conf_info.vx_opt[i].vx_pd.fcst_info->magic_str()
<< ".\n";
// Check for zero fields
if(n_fcst == 0) {
mlog << Warning << "\nprocess_fcst_climo_files() -> "
<< "no fields matching "
<< conf_info.vx_opt[i].vx_pd.fcst_info->magic_str()
<< " found in file: "
<< fcst_file << "\n\n";
continue;
}
// Setup the first pass through the data
if(is_first_pass) setup_first_pass(fcst_dpa[0], fcst_mtddf->grid());
// Regrid, if necessary
if(!(fcst_mtddf->grid() == grid)) {
mlog << Debug(1)
<< "Regridding " << fcst_dpa.n_planes()
<< " forecast field(s) for "
<< conf_info.vx_opt[i].vx_pd.fcst_info->magic_str()
<< " to the verification grid.\n";
// Loop through the forecast fields
for(j=0; j<fcst_dpa.n_planes(); j++) {
fcst_dpa[j] = met_regrid(fcst_dpa[j], fcst_mtddf->grid(), grid,
conf_info.vx_opt[i].vx_pd.fcst_info->regrid());
}
}
// Rescale probabilities from [0, 100] to [0, 1]
if(conf_info.vx_opt[i].vx_pd.fcst_info->p_flag()) {
for(j=0; j<fcst_dpa.n_planes(); j++) {
rescale_probability(fcst_dpa[j]);
}
} // end for j
// Read climatology data
cmn_dpa = read_climo_data_plane_array(
conf_info.conf.lookup_array(conf_key_climo_mean_field, false),
i, fcst_dpa[0].valid(), grid);
csd_dpa = read_climo_data_plane_array(
conf_info.conf.lookup_array(conf_key_climo_stdev_field, false),
i, fcst_dpa[0].valid(), grid);
// Store data for the current verification task
conf_info.vx_opt[i].vx_pd.set_fcst_dpa(fcst_dpa);
conf_info.vx_opt[i].vx_pd.set_climo_mn_dpa(cmn_dpa);
conf_info.vx_opt[i].vx_pd.set_climo_sd_dpa(csd_dpa);
// Get the valid time for the first field
file_ut = fcst_dpa[0].valid();
// If obs_valid_beg_ut and obs_valid_end_ut were set on the command
// line, use them. If not, use beg_ds and end_ds.
if(obs_valid_beg_ut != (unixtime) 0 ||
obs_valid_end_ut != (unixtime) 0) {
beg_ut = obs_valid_beg_ut;
end_ut = obs_valid_end_ut;
}
else {
beg_ut = file_ut + conf_info.vx_opt[i].beg_ds;
end_ut = file_ut + conf_info.vx_opt[i].end_ds;
}
// Store the valid times for this VxPairDataPoint object
conf_info.vx_opt[i].vx_pd.set_fcst_ut(fcst_valid_ut);
conf_info.vx_opt[i].vx_pd.set_beg_ut(beg_ut);
conf_info.vx_opt[i].vx_pd.set_end_ut(end_ut);
// Dump out the number of levels found
mlog << Debug(2)
<< "For " << conf_info.vx_opt[i].vx_pd.fcst_info->magic_str()
<< " found " << n_fcst << " forecast levels, "
<< cmn_dpa.n_planes() << " climatology mean levels, and "
<< csd_dpa.n_planes() << " climatology standard deviation levels.\n";
} // end for i
// Check for no data
if(is_first_pass) {
mlog << Error << "\nprocess_fcst_climo_files() -> "
<< "no requested forecast data found! Exiting...\n\n";
exit(1);
}
mlog << Debug(2)
<< "\n" << sep_str << "\n\n";
return;
}
////////////////////////////////////////////////////////////////////////
void process_obs_file(int i_nc) {
int j, i_obs;
float obs_arr[OBS_ARRAY_LEN], hdr_arr[HDR_ARRAY_LEN];
float prev_obs_arr[OBS_ARRAY_LEN];
ConcatString hdr_typ_str;
ConcatString hdr_sid_str;
ConcatString hdr_vld_str;
ConcatString obs_qty_str;
unixtime hdr_ut;
NcFile *obs_in = (NcFile *) 0;
const char *method_name = "process_obs_file() -> ";
// Set flags for vectors
bool vflag = conf_info.get_vflag();
bool is_ugrd, is_vgrd;
// Open the observation file as a NetCDF file.
// The observation file must be in NetCDF format as the
// output of the PB2NC or ASCII2NC tool.
bool status;
bool use_var_id = true;
bool use_arr_vars = false;
bool use_python = false;
MetNcPointObsIn nc_point_obs;
MetPointData *met_point_obs = 0;
#ifdef WITH_PYTHON
MetPythonPointDataFile met_point_file;
string python_command = obs_file[i_nc];
bool use_xarray = (0 == python_command.find(conf_val_python_xarray));
use_python = use_xarray || (0 == python_command.find(conf_val_python_numpy));
if (use_python) {
int offset = python_command.find("=");
if (offset == std::string::npos) {
mlog << Error << "\n" << method_name
<< "trouble parsing the python command " << python_command << ".\n\n";
exit(1);
}
if(!met_point_file.open(python_command.substr(offset+1).c_str(), use_xarray)) {
met_point_file.close();
mlog << Error << "\n" << method_name
<< "trouble getting point observation file from python command "
<< python_command << ".\n\n";
exit(1);
}
met_point_obs = met_point_file.get_met_point_data();
}
else {
#endif
if( !nc_point_obs.open(obs_file[i_nc].c_str()) ) {
nc_point_obs.close();
mlog << Warning << "\n" << method_name
<< "can't open observation netCDF file: "
<< obs_file[i_nc] << "\n\n";
return;
}
nc_point_obs.read_dim_headers();
nc_point_obs.check_nc(obs_file[i_nc].c_str(), method_name);
nc_point_obs.read_obs_data_table_lookups();
met_point_obs = (MetPointData *)&nc_point_obs;
use_var_id = nc_point_obs.is_using_var_id();
use_arr_vars = nc_point_obs.is_using_obs_arr();
#ifdef WITH_PYTHON
}
#endif
int hdr_count = met_point_obs->get_hdr_cnt();
int obs_count = met_point_obs->get_obs_cnt();
mlog << Debug(2)
<< "Searching " << obs_count
<< " observations from " << hdr_count
<< " messages.\n";
ConcatString var_name("");
StringArray var_names;
StringArray obs_qty_array = met_point_obs->get_qty_data();
if( use_var_id ) var_names = met_point_obs->get_var_names();
const int buf_size = ((obs_count > BUFFER_SIZE) ? BUFFER_SIZE : (obs_count));
int obs_qty_idx_block[buf_size];
float obs_arr_block[buf_size][OBS_ARRAY_LEN];
// Process each observation in the file
int str_length, block_size;
int prev_grib_code = bad_data_int;
for(int i_block_start_idx=0; i_block_start_idx<obs_count; i_block_start_idx+=buf_size) {
block_size = (obs_count - i_block_start_idx);
if (block_size > buf_size) block_size = buf_size;
#ifdef WITH_PYTHON
if (use_python)
status = met_point_obs->get_point_obs_data()->fill_obs_buf(
block_size, i_block_start_idx, (float *)obs_arr_block, obs_qty_idx_block);
else
#endif
status = nc_point_obs.read_obs_data(block_size, i_block_start_idx,
(float *)obs_arr_block,
obs_qty_idx_block, (char *)0);
if (!status) exit(1);
int hdr_idx;
for(int i_block_idx=0; i_block_idx<block_size; i_block_idx++) {
i_obs = i_block_start_idx + i_block_idx;
for (j=0; j<OBS_ARRAY_LEN; j++) {
obs_arr[j] = obs_arr_block[i_block_idx][j];
}
int qty_offset = use_arr_vars ? i_obs : obs_qty_idx_block[i_block_idx];
obs_qty_str = obs_qty_array[qty_offset];
int headerOffset = met_point_obs->get_header_offset(obs_arr);
// Range check the header offset
if(headerOffset < 0 || headerOffset >= hdr_count) {
mlog << Warning << "\n" << method_name
<< "range check error for header index " << headerOffset
<< " from observation number " << i_obs
<< " of point observation file: " << obs_file[i_nc]
<< "\n\n";
continue;
}
// Read the corresponding header array for this observation
// - the corresponding header type, header Station ID, and valid time
#ifdef WITH_PYTHON
if (use_python)
met_point_obs->get_header(headerOffset, hdr_arr, hdr_typ_str, hdr_sid_str, hdr_vld_str);
else
#endif
nc_point_obs.get_header(headerOffset, hdr_arr, hdr_typ_str,
hdr_sid_str, hdr_vld_str);
// Store the variable name
int org_grib_code = met_point_obs->get_grib_code_or_var_index(obs_arr);
int grib_code = org_grib_code;
if (prev_grib_code != org_grib_code) {
if (use_var_id && grib_code < var_names.n()) {
var_name = var_names[grib_code];
grib_code = bad_data_int;
}
else {
var_name = "";
}
// Check for wind components
is_ugrd = ( use_var_id && var_name == ugrd_abbr_str ) ||
(!use_var_id && nint(grib_code) == ugrd_grib_code);
is_vgrd = ( use_var_id && var_name == vgrd_abbr_str ) ||
(!use_var_id && nint(grib_code) == vgrd_grib_code);
prev_grib_code = org_grib_code;
}
// If the current observation is UGRD, save it as the
// previous. If vector winds are to be computed, UGRD
// must be followed by VGRD
if(vflag && is_ugrd) {
for(j=0; j<4; j++) prev_obs_arr[j] = obs_arr[j];
}
// If the current observation is VGRD and vector
// winds are to be computed. Make sure that the
// previous observation was UGRD with the same header
// and at the same vertical level.
if(vflag && is_vgrd) {
if(!met_point_obs->is_same_obs_values(obs_arr, prev_obs_arr)) {
mlog << Error << "\n" << method_name
<< "for observation index " << i_obs
<< ", when computing VL1L2 and/or VAL1L2 vector winds "
<< "each UGRD observation must be followed by a VGRD "
<< "observation with the same header and at the same "
<< "level.\n\n";
exit(1);
}
}
// Convert string to a unixtime
hdr_ut = timestring_to_unix(hdr_vld_str.c_str());
// Check each conf_info.vx_pd object to see if this observation
// should be added
for(j=0; j<conf_info.get_n_vx(); j++) {
// Check for no forecast fields
if(conf_info.vx_opt[j].vx_pd.fcst_dpa.n_planes() == 0) continue;
// Attempt to add the observation to the conf_info.vx_pd object
conf_info.vx_opt[j].vx_pd.add_point_obs(
hdr_arr, hdr_typ_str.c_str(), hdr_sid_str.c_str(),
hdr_ut, obs_qty_str.c_str(), obs_arr,
grid, var_name.c_str());
}
met_point_obs->set_grib_code_or_var_index(obs_arr, org_grib_code);
}
} // end for i_block_start_idx
// Deallocate and clean up
#ifdef WITH_PYTHON
if (use_python) met_point_file.close();
else
#endif
nc_point_obs.close();
return;
}
////////////////////////////////////////////////////////////////////////
void process_scores() {
int i, j, k, l, m;
int n_cat, n_wind;
ConcatString cs;
// Initialize pointers
PairDataPoint *pd_ptr = (PairDataPoint *) 0;
CTSInfo *cts_info = (CTSInfo *) 0;
MCTSInfo mcts_info;
VL1L2Info *vl1l2_info = (VL1L2Info *) 0;
mlog << Debug(2)
<< "\n" << sep_str << "\n\n";
// Setup the output text files as requested in the config file
setup_txt_files();
// Store the maximum number of each threshold type
n_cat = conf_info.get_max_n_cat_thresh();
n_wind = conf_info.get_max_n_wind_thresh();
// Allocate space for output statistics types
cts_info = new CTSInfo [n_cat];
vl1l2_info = new VL1L2Info [n_wind];
// Compute scores for each PairData object and write output
for(i=0; i<conf_info.get_n_vx(); i++) {
// Check for no forecast fields
if(conf_info.vx_opt[i].vx_pd.fcst_dpa.n_planes() == 0) continue;
// Store the description
shc.set_desc(conf_info.vx_opt[i].vx_pd.desc.c_str());
// Store the forecast variable name
shc.set_fcst_var(conf_info.vx_opt[i].vx_pd.fcst_info->name_attr());
// Store the forecast variable units
shc.set_fcst_units(conf_info.vx_opt[i].vx_pd.fcst_info->units_attr());
// Set the forecast level name
shc.set_fcst_lev(conf_info.vx_opt[i].vx_pd.fcst_info->level_attr().c_str());
// Store the observation variable name
shc.set_obs_var(conf_info.vx_opt[i].vx_pd.obs_info->name_attr());
// Store the observation variable units
cs = conf_info.vx_opt[i].vx_pd.obs_info->units_attr();
if(cs.empty()) cs = na_string;
shc.set_obs_units(cs);
// Set the observation level name
shc.set_obs_lev(conf_info.vx_opt[i].vx_pd.obs_info->level_attr().c_str());
// Set the forecast lead time
shc.set_fcst_lead_sec(conf_info.vx_opt[i].vx_pd.fcst_dpa[0].lead());
// Set the forecast valid time
shc.set_fcst_valid_beg(conf_info.vx_opt[i].vx_pd.fcst_dpa[0].valid());
shc.set_fcst_valid_end(conf_info.vx_opt[i].vx_pd.fcst_dpa[0].valid());
// Set the observation lead time
shc.set_obs_lead_sec(0);
// Set the observation valid time
shc.set_obs_valid_beg(conf_info.vx_opt[i].vx_pd.beg_ut);
shc.set_obs_valid_end(conf_info.vx_opt[i].vx_pd.end_ut);
// Loop through the message types
for(j=0; j<conf_info.vx_opt[i].get_n_msg_typ(); j++) {
// Store the message type in the obtype column
shc.set_obtype(conf_info.vx_opt[i].msg_typ[j].c_str());
// Loop through the verification masking regions
for(k=0; k<conf_info.vx_opt[i].get_n_mask(); k++) {
// Store the verification masking region
shc.set_mask(conf_info.vx_opt[i].mask_name[k].c_str());
// Loop through the interpolation methods
for(l=0; l<conf_info.vx_opt[i].get_n_interp(); l++) {
// Store the interpolation method and width being applied
shc.set_interp_mthd(conf_info.vx_opt[i].interp_info.method[l],
conf_info.vx_opt[i].interp_info.shape);
shc.set_interp_wdth(conf_info.vx_opt[i].interp_info.width[l]);
pd_ptr = &conf_info.vx_opt[i].vx_pd.pd[j][k][l];
mlog << Debug(2)
<< "Processing "
<< conf_info.vx_opt[i].vx_pd.fcst_info->magic_str()
<< " versus "
<< conf_info.vx_opt[i].vx_pd.obs_info->magic_str()
<< ", for observation type " << pd_ptr->msg_typ
<< ", over region " << pd_ptr->mask_name
<< ", for interpolation method "
<< shc.get_interp_mthd() << "("
<< shc.get_interp_pnts_str()
<< "), using " << pd_ptr->n_obs << " matched pairs.\n";
// List counts for reasons why observations were rejected
cs << cs_erase
<< "Number of matched pairs = " << pd_ptr->n_obs << "\n"
<< "Observations processed = " << conf_info.vx_opt[i].vx_pd.n_try << "\n"
<< "Rejected: station id = " << conf_info.vx_opt[i].vx_pd.rej_sid << "\n"
<< "Rejected: obs var name = " << conf_info.vx_opt[i].vx_pd.rej_var << "\n"
<< "Rejected: valid time = " << conf_info.vx_opt[i].vx_pd.rej_vld << "\n"
<< "Rejected: bad obs value = " << conf_info.vx_opt[i].vx_pd.rej_obs << "\n"
<< "Rejected: off the grid = " << conf_info.vx_opt[i].vx_pd.rej_grd << "\n"
<< "Rejected: topography = " << conf_info.vx_opt[i].vx_pd.rej_topo << "\n"
<< "Rejected: level mismatch = " << conf_info.vx_opt[i].vx_pd.rej_lvl << "\n"
<< "Rejected: quality marker = " << conf_info.vx_opt[i].vx_pd.rej_qty << "\n"
<< "Rejected: message type = " << conf_info.vx_opt[i].vx_pd.rej_typ[j][k][l] << "\n"
<< "Rejected: masking region = " << conf_info.vx_opt[i].vx_pd.rej_mask[j][k][l] << "\n"
<< "Rejected: bad fcst value = " << conf_info.vx_opt[i].vx_pd.rej_fcst[j][k][l] << "\n"
<< "Rejected: bad climo mean = " << conf_info.vx_opt[i].vx_pd.rej_cmn[j][k][l] << "\n"