#! /bin/ksh # @ step_name=ungrib # @ output = ungrib.log # @ error = ungrib.log # @ notification = never # @ wall_clock_limit = 00:12:00 # @ arguments = YSTARTMSTARTDSTART HSTART # @ job_type = parallel # @ resources=ConsumableCPUS(1)ConsumableMemory(1 GB) # @ class=dev # @ group=devonprod # @ account_no=HRW-T2O # @ executable = /meso/save/wx20py/NMMB_init_to_share/NPS/run_ungrib.ll # @ queue # @ step_name=geo_grid # @ output =geo.log # @ error = geo.log # @ notification = never # @ wall_clock_limit = 00:08:00 # @ job_type = parallel # @ total_tasks = 1 # @ arguments = NMM # @ class=dev # @ group=devonprod # @ resources=consumablecpus(1)consumablememory(4000 MB) # @ node_usage=shared # @ account_no=HRW-T2O # @ network.MPI = csss,shared,us # @ executable = /meso/save/wx20py/NMMB_init_to_share/NPS/run_geo.ll # @ queue # @ dependency = (ungrib == 0 && geo_grid == 0 ) # @ step_name = run_wps_met # @ output = met.log # @ error = met.log # @ notification = never # @ wall_clock_limit = 00:12:00 # @ job_type = parallel # @ total_tasks = 1 # @ node=1 # @ arguments = NMM # @ class=dev # @ group=devonprod # @ resources=ConsumableCPUS(1)ConsumableMemory(4000 MB) # @ node_usage=shared # @ account_no=HRW-T2O # @ network.MPI = csss,shared,us # @ executable = /meso/save/wx20py/NMMB_init_to_share/NPS/run_met.ll # @ queue # @ dependency = (run_wps_met == 0 ) # @ step_name = run_wps_nemsinterp # @ output = nemsint.log # @ error = nemsint.log # @ notification = never # @ wall_clock_limit = 00:12:00 # @ job_type = parallel # @ total_tasks = 4 # @ node=1 # @ arguments = NMM # @ class=dev # @ group=devonprod # @ resources=ConsumableCPUS(1)ConsumableMemory(1000 MB) # @ node_usage=shared # @ account_no=HRW-T2O # @ network.MPI = csss,shared,us # @ executable = /meso/save/wx20py/NMMB_init_to_share/NPS/run_nemsinterp.ll_fromload # @ queue ## @ dependency = (real_nmm == 0) ## @ step_name = run_nmmb ## @ output = real.log ## @ error = real.log ## @ wall_clock_limit = 00:12:00 ## @ job_type = parallel ## @ total_tasks = 31 ## @ node=1 ## @ class=dev ## @ group=devonprod ## @ resources=ConsumableCPUS(1) ## @ node_usage=not_shared ## @ account_no=HRW-T2O ## @ network.MPI = csss,shared,us ## @ executable = /meso/save/wx20py/NMMB_init_to_share/test_umo/run_umo.ll ## @ queue