Configure a run
As noted earlier, the many configuration files in GCHP can be
overwhelming but you should be able to accomplish most if not all of
what you wish to configure from one place in
setCommonRunSettings.sh
. Use this section to learn what to
change in the run directory based on what you would like to do.
Note
If there is topic not covered on this page that you would like to see added please create an issue on the GCHP issues page with your request.
Compute resources
Set number of nodes and cores
To change the number of nodes and cores for your run you must update
settings in two places: (1) setCommonRunSettings.sh
, and (2)
your run script.
The setCommonRunSettings.sh
file contains detailed
instructions on how to set resource parameter options and what they
mean. Look for the Compute Resources section in the script.
Update your resource request in your run script to match the resources
set in setCommonRunSettings.sh
.
It is important to be smart about your resource allocation. To do this it is useful to understand how GCHP works with respect to distribution of nodes and cores across the grid. At least one unique core is assigned to each face on the cubed sphere, resulting in a constraint of at least six cores to run GCHP. The same number of cores must be assigned to each face, resulting in another constraint of total number of cores being a multiple of six. There is also an upper limit on total number of cores for any given grid resolution. Advection also requires that each core cover a domain size of at least 4x4 grid cells. For C24 this means there needs to be at least 36 cores per face because a 24x24 face can be broken into 36 4x4 regions. This translates into a maximum of 216 cores for C24, 864 cores for C48, 3174 cores for C90, 12150 cores for C180, and so on.
Communication between the cores occurs only during transport processes and you will typically start to see negative effects due to excessive communication if a core is handling less than around one hundred grid cells. You can determine approximately how many grid cells are handled per core by analyzing your grid resolution and resource allocation. Total number of grid cells is equal to N*N*6 for grid resolution CN, e.g. 24*24*6 for C24. Number of grid cells handled per core is then N*N*6 divided by total number of cores. This number is approximate since number of grid cells per core can vary across each face, such as if you run C48 with 360 cores.
Excessive communication may also occur if the grid cell region of a
core is not approximately square. Run-time parameters NX
and NY
quantify how each face will be decomposed and can be
set to maximize squareness. The product of these parameters is equal
to the total number of cores and each face will be divided into
NX
x NY/6
regions. Setting NX
and
NY/6
as close as possible in value will therefore ensure
optimal communication. Maximizing squareness of grid cells per core
is done automatically for you within
setCommonRunSettings.sh
. You may disable that feature by
changing variable AutoUpdate_NXNY
to OFF
in the
“DOMAIN DECOMPOSITON” section of the file. Beware that disabling it
will require you to set NX
and NY
yourself in
the file.
If using mass flux input data then there is an additional constraint that the model grid divides evenly across cores. In other words NX and NY/6 must divide evenly into N where the run grid resolution is CN. For example, a C30 run can have each face split into 6x6 regions but not 4x4 regions. The former gives a 5x5 grid cell region per core while the latter gives 7.5x7.5 which is non-integer. This has implications for total number of cores you can run with when using mass flux inputs since a given total number of cores can evenly divide every possible grid resolution.
Change domain stack size
For runs at very high resolution or small number of processors you may
run into a domains stack size error. This is caused by exceeding the
domains stack size memory limit set at run-time. The error will be
apparent from the message in your log file. If this occurs you can
increase the domains stack size in file input.nml
.
Considerations for very high core counts
GCHP has the capability of running on tens of thousands of cores. If running at very high grid resolution and you have access to this level of compute power then there are a few things to consider prior to submitting your run.
First, very high resolution runs make model I/O more challenging. Restart files will be very large and reading and writing them will be intensive. Second, MPI configuration, file system, and compute hardware can make or break your run at this scale. This often comes back to challenges with I/O. For example, we have found that using OpenMPI on certain compute clusters requires turning on the MAPL O-server for checkpoint write to prevent the model from hanging.
There are a few settings in configuration file GCHP.rc
that
you can play with to try to improve performance in this area.
NUM_READERS
parallelizes restart read and setting it to
values between 6 and 24 may yield
improvement. WRITE_RESTART_BY_OSERVER
assigns restart write
to extra nodes by running what is called the O-server. The O-server is
off by default but can be toggled on manually. NUM_WRITERS
affects how many processes to write with. You can try anywhere from 6
to 24 to see if it makes a differences. Note that none of these
settings are dependable ways to speed up or fix a run because they are
dependent on file system and MPI stack.
Lastly, your specific grid resolution will impact how performance scales with more cores. FV3 advection has a point of diminishing returns due to excessive communication if core count gets too high for a given run resolution. See earlier section on setting number of nodes and cores.
Basic run settings
Set cubed-sphere grid resolution
GCHP uses a cubed sphere grid rather than the traditional lat-lon grid used in GEOS-Chem Classic. While regular lat-lon grids are typically designated as ΔLat ⨉ ΔLon (e.g. 4⨉5), cubed sphere grids are designated by the side-length of the cube. In GCHP we specify this as CX (e.g. C24 or C180). The simple rule of thumb for determining the roughly equivalent lat-lon resolution for a given cubed sphere resolution is to divide the side length by 90. Using this rule you can quickly match C24 with about 4x5, C90 with 1 degree, C360 with quarter degree, and so on.
To change your grid resolution in the run directory edit
CS_RES
in the “GRID RESOLUTION” section of
setCommonRunSettings.sh
. The paramter should be an
integer value of the cube side length you wish to use.
To use a uniform global grid resolution make sure
STRETCH_GRID
is set to OFF
in the “STRETCHED
GRID” section of the file. To use a stretched grid rather
han a globally uniform grid see the section on this page for
setting stretched grid parameters.
Set stretched grid parameters
GCHP has the capability to run with a stretched grid, meaning one
portion of the globe is stretched to fine resolution. Set stretched
grid parameter in setCommonRunSettings.sh
section “STRETCHED
GRID”. See instructions in that section of the file. For more
detailed information see the stretched grid section of the
Supplemental Guides section of the GCHP ReadTheDocs.
Turn on/off model components
You can toggle most primary GEOS-Chem components that are set in
geoschem_config.yml
from the “GEOS-CHEM COMPONENTS” section of
setCommonRunSettings.sh
. The settings in that file will
update geoschem_config.yml
automatically so be
sure to check that the settings there are as you
intend. For emissions you should directly edit
HEMCO_Config.rc
.
Change model timesteps
Model timesteps, including chemistry, dynamic, and RRTMG, are
configured within the “TIMESTEPS” section of
setCommonRunSettings.sh
. By default, the RRTMG timestep is set
to 3 hours. All other GCHP timesteps are automatically set based
on grid resolution. Chemistry and dynamic timesteps are 20 and
10 minutes respectively for grid resolutions coarser than C180,
and 10 and 5 minutes for C180 and higher. Meteorology read
frequency for PS2, SPHU2, and TMPU2 are automatically updated in
ExtData.rc
accordingly. To change the default timesteps
settings edit the “TIMESTEPS” section of
setCommonRunSettings.sh
.
Set simulation start date and duration
Unlike GEOS-Chem Classic, GCHP uses a start date and run duration
rather than start and end dates. Set simulation start date in
cap_restart
using string format YYYYMMDD
HHmmSS
. Set simulation duration in section “SIMULATION
DURATION” in setCommonRunSettings.sh
using the same
format as start date. For example, a 1-year run starting 15
January 2019 would have 20190115 000000
in
cap_restart
and 00010000 000000
in
setCommonRunSettings.sh
.
Under the hood cap_restart
is used directly by the MAPL
software in GCHP, and setCommonRunSettings.sh
auto-updates the
run duration in GCHP config file CAP.rc
. Please be aware that
MAPL overwrites cap_restart
at the end of the simulation to
contain the new start date (end of last run) so be sure to check it
every time you run GCHP.
If you poke around the GCHP configuration files you may notice that
file CAP.rc
contains entries for BEG_DATE
and
END_DATE
. You can ignore these fields for most
cases. BEG_DATE
is not used for start date if
cap_restart
is present. However, it must be prior to your
start date for use in GEOS-Chem’s “ELAPSED_TIME” variable. We set it
to year 1960 to be safe. BEG_DATE
can also be ignored as long
as it is the same as or later than your start date plus run
duration. For safety we set it to year 2200. The only time you would
need to adjust these settings is for simulations way in the past or
way into the future.
Inputs
Change restart file
All GCHP run directories come with symbolic links to initial restart
files for commonly used cubed sphere resolutions. These are located in
the Restarts
directory in the run directory. All initial
restart files contain start date and grid resolution in the filename
using the start date in cap_restart
. Prior to running GCHP,
either you or your run script will execute setRestartLink.sh
to create a symbolic link gchp_restart.nc4
to point to the
appropriate restart file given configured start date and grid
resolution. gchp_restart.nc4
will always be used as the
restart file for all runs since it is specified as the restart file in
GCHP.rc
.
If you want to change the restart file then you should put the restart
file you want to use in the Restarts
directory using the
expected filename format with the start date you configure in
cap_restart
and the grid resolution you configure in
setCommonRunSettings.sh
. The expected format is
GEOSChem.Restarts.YYYYMMDD_HHmmz.cN.nc4
. Running
setRestartLink.sh
will update
gchp_restart.nc4
to use it.
If you do not want to rename your restart file then you can create a
symbolic link in the Restarts
folder that points to it.
Please note that unlike GC-Classic, GCHP does not use a separate HEMCO restart file. All HEMCO restart variables are included in the main GCHP restart.
Enable restart file to have missing species
Most simulations by default do not allow missing species in the
restart file. The model will exit with an error if species are not
found. However, there is a switch in setCommonRunSetting.sh
to
disable this behavior. This toggle is located in the section on
infrequently changed settings under the header REQUIRE ALL
SPECIES IN INITIAL RESTART FILE
. Setting the switch to NO
will use background values set in species_database.yml
as
initial values for species that are missing.
Turn on/off emissions inventories
Because file I/O impacts GCHP performance it is a good idea to turn
off file read of emissions that you do not need. You can turn
individual emissions inventories on or off the same way you would in
GEOS-Chem Classic, by setting the inventories to true or false at the
top of configuration file HEMCO_Config.rc
. All emissions that
are turned off in this way will be ignored when GCHP uses
ExtData.rc
to read files, thereby speeding up the model.
For emissions that do not have an on/off toggle at the top of the
file, you can prevent GCHP from reading them by commenting them out in
HEMCO_Config.rc
. No updates to ExtData.rc
would be
necessary. If you alternatively comment out the emissions in
ExtData.rc
but not HEMCO_Config.rc
then GCHP will fail
with an error when looking for the file information.
Another option to skip file read for certain files is to replace the
file path in ExtData.rc
with /dev/null
. However,
if you want to turn these inputs back on at a later time you should
preserve the original path by commenting out the original line.
Change input meteorology
Input meteorology source and grid resolution are set in config file
ExtData.rc
during run directory creation. You will be prompted
to choose between MERRA2 and GEOS-FP, and grid resolution is
automatically set to the native grid lat-lon resolution. If you
would like to change the meteorology inputs, for example using a
different grid resolution, then you would need to change the
met-field entries in run directory file ExtData.rc
after
creating a run directory. Simply open the file, search for the
meteorology section, and edit file paths as needed. Please note
that while MAPL will automatically regrid met-fields to the run
resolution you specify in setCommonRunSettings.sh
, you
will achieve best performance using native resolution inputs.
Add new emissions files
There are two steps for adding new emissions inventories to GCHP. They
are (1) add the inventory information to HEMCO_Config.rc
,
and (2) add the inventory information to ExtData.rc
.
To add inventory information to HEMCO_Config.rc
, follow the
same rules as you would for adding a new emission inventory to
GEOS-Chem Classic. Note that not all information in
HEMCO_Config.rc
is used by GCHP. This is because HEMCO is
only used by GCHP to handle emissions after they are read,
e.g. scaling and applying hierarchy. All functions related to HEMCO
file read are skipped. This means that you could put garbage for the
file path and units in HEMCO_Config.rc
without running into
problems with GCHP, as long as the syntax is what HEMCO expects.
However, we recommend that you fill in HEMCO_Config.rc
in the
same way you would for GEOS-Chem Classic for consistency and also to
avoid potential format check errors.
To add inventory information to ExtData.rc
follow the
guidelines listed at the top of the file and use existing inventories
as examples. Make sure that you stay consistent with the information
you put into HEMCO_Config.rc
. You can ignore all entries in
HEMCO_Config.rc
that are copies of another entry (i.e. mostly
filled with dashes). Putting these in ExtData.rc
would
result in reading the same variable in the same file twice.
A few common errors encountered when adding new input emissions files to GCHP are:
Your input file contains integer values. Beware that the MAPL I/O component in GCHP does not read or write integers. If your data contains integers then you should reprocess the file to contain floating point values instead.
Your data latitude and longitude dimensions are in the wrong order. Lat must always come before lon in your inputs arrays, a requirement true for both GCHP and GEOS-Chem Classic.
Your 3D input data are mapped to the wrong levels in GEOS-Chem (silent error). If you read in 3D data and assign the resulting import to a GEOS-Chem state variable such as
State_Chm
orState_Met
, then you must flip the vertical axis during the assignment. See filesIncludes_Before_Run.H
and settingState_Chm%Species
inChem_GridCompMod.F90
for examples.
You have a typo in either
HEMCO_Config.rc
orExtData.rc
. Errors inHEMCO_Config.rc
typically result in the model crashing right away. Errors inExtData.rc
typically result in a problem later on during ExtData read. Always try a short run with all debug prints enabled when first implementing new emissions. See the debugging section of the user manual for more information. Another useful strategy is to find config file entries for similar input files and compare them against the entry for your new file. Directly comparing the file metadata may also lead to insights into the problem.
Outputs
Output diagnostics data on a lat-lon grid
See documentation in the HISTORY.rc
config file for
instructions on how to output diagnostic collection on lat-lon grids,
as well as the configuration files section at the top of this page for
more information on that file. If outputting on a lat-lon grid you may
also output regional data instead of global. Make sure that whatever
grid you choose is listed under GRID_LABELS
and is not
commented out in HISTORY.rc
.
Output restart files at regular frequency
The MAPL component in GCHP has the option to output restart files
(also called checkpoint files) prior to run end. These periodic
restart files are output to the main level of the run directory with
filename gcchem_internal_checkpoint.YYYYMMDD_HHssz.nc4
.
Outputting restart files beyond the end of the run is a good idea if you plan on doing a long simulation and you are not splitting your run into multiple jobs.
If the run crashes unexpectedly then you can restart mid-run rather
than start over from the beginning. Update settings for checkpoint
restart outputs in setCommonRunSettings.sh
section “MID-RUN
CHECKPOINT FILES”. Instructions for configuring restart frequency are
included in the file.
Turn on/off diagnostics
To turn diagnostic collections on or off, comment (“#”) collection
names in the “COLLECTIONS” list at the top of file
HISTORY.rc
. Collections cannot be turned on/off from
setCommonRunSettings.sh
.
Set diagnostic frequency, duration, and mode
All diagnostic collections that come with the run directory have
frequency and duration auto-set within
setCommonRunSettings.sh
. The file contains a list of
time-averaged collections and instantaneous collections, and
allows setting a frequency and duration to apply to all
collections listed for each. Time-avraged collections also have
a monthly mean option (see separate section on this page about
monthly mean). To avoid auto-update of a certain collection, remove
it from the list in setCommonRunSettings.sh
, or set
“AutUpdate_Diagnostics” to OFF
. See section “DIAGNOSTICS”
within setCommonRunSettings.sh
for examples.
Add a new diagnostics collection
Adding a new diagnostics collection in GCHP is the same as for
GEOS-Chem Classic netcdf diagnostics. You must add your collection to
the collection list in HISTORY.rc
and then define it further
down in the file. Any 2D or 3D arrays that are stored within
GEOS-Chem objects State_Met
, State_Chm
, or
State_Diag
, may be included as fields in a collection.
State_Met
variables must be preceded by “Met_”,
State_Chm
variables must be preceded by “Chem_”, and
State_Diag
variables should not have a prefix. Collections
may have a combination of 2D and 3D variables, but all 3D variables
must have the same number of levels. See the HISTORY.rc
file
for examples.
Generate monthly mean diagnostics
You can toggle monthly mean diagnostics on/off from within
setCommonRunSettings.sh
in the “DIAGNOSTICS” section if you
also set auto-update of diagnostics it that file to on. All
time-averaged diagnostic collections will then automatically be
configured to compute monthly mean. Alternatively, you can edit
HISTORY.rc
directly and set the “monthly” field to value
1 for each collection you wish to output monthly diagnostics for.
Prevent overwriting diagnostic files
By default all GCHP run directories are configured to allow
overwriting diagnostics files present in OutputDir
over the
course a simulation. You may disable this feature by setting
Allow_Overwrite=.false.
at the top of configuration file
HISTORY.rc
.