#PBS -N hwrf%STORMNUM%_relocate_gdas_3_%CYC% #PBS -j oe #PBS -S /bin/bash #PBS -q %QUEUE% #PBS -A %PROJ%-%PROJENVIR% #PBS -l walltime=00:25:00 #PBS -l select=1:mpiprocs=1:ompthreads=24:ncpus=24 export NODES=1 export TOTAL_TASKS=24 %include module purge module load ecflow %include export cyc="%CYC%" export storm_num="%STORMNUM%" # versions file for hwrf sets $model_ver and $code_ver model=hwrf %include source ${HOMEhwrf}/versions/run.ver module use ${HOMEhwrf}/modulefiles/ module load HWRF/run module list export INIT_MODEL=GDAS1 INIT_FHR="3" ${HOMEhwrf}/jobs/JHWRF_RELOCATE %include %manual TASK relocate_gdas_3 PURPOSE: Reads the output of the init_gfs and init_gdas_3 jobs. Relocates the GDAS vortex, and merges it with the prior HWRF six hour forecast (if any) and possibly a bogus vortex. Adjusts intensity and structure to match values reported by forecasters in the message files and tcvitals database. Those data are only available every six hours, so they are interpolated to 3 and 9 hour times. The output of this job is used as the first guess for the gsi_d02 and gsi_d03 jobs. TROUBLESHOOTING: All errors seen in the 2016 HWRF pre-implementation testing were due to node crashes, filesystem problems or quota violations which caused data corruption. Check for node and filesystem problems first. If all hardware is running well, then the problem may be a genuine code or script issue. This job is made up of many tiny fortran programs with limited error checking. Usually if one fails, the problem was actually in an earlier one, or possibly in an earlier job. Detecting which one is the source of the trouble requires expert help. Please call on-call EMC staff for assistance. %end