diff --git a/config/acme/machines/config_machines.xml b/config/acme/machines/config_machines.xml
index 6bf273f03ded..b82eb059f7f8 100644
--- a/config/acme/machines/config_machines.xml
+++ b/config/acme/machines/config_machines.xml
@@ -60,7 +60,7 @@
NERSC XC30, os is CNL, 24 pes/node, batch system is SLURM
edison
acme_developer
- intel,intel17,gnu,cray
+ intel,gnu
mpt
$ENV{CSCRATCH}/acme_scratch/edison
$CIME_OUTPUT_ROOT/$CASE/run
@@ -85,7 +85,8 @@
--label
-n $TOTALPES
- -c $ENV{OMP_NUM_THREADS}
+ -c $SHELL{echo 48/`./xmlquery --value MAX_TASKS_PER_NODE`*$ENV{OMP_NUM_THREADS} | bc }
+ $SHELL{if [ 24 -ge `./xmlquery --value MAX_TASKS_PER_NODE` ]; then echo "--cpu_bind=cores"; else echo "--cpu_bind=threads"; fi}
@@ -117,7 +118,6 @@
craype-ivybridge
craype
papi
- cmake
cray-petsc
esmf
@@ -125,68 +125,54 @@
PrgEnv-intel
intel
- intel/15.0.1.133
+ intel/17.0.2.174
cray-libsci
- cray-mpich/7.2.5
-
-
- PrgEnv-intel
- intel
- intel/17.0.1.132
- cray-libsci
- cray-mpich/7.5.1
-
-
- PrgEnv-cray
- cce cce/8.5.1
- cray-libsci/16.07.1
- cray-mpich/7.5.1
PrgEnv-gnu
- gcc gcc/6.2.0
- cray-libsci/16.07.1
- cray-mpich/7.5.1
+ gcc
+ gcc/6.3.0
+ cray-libsci
+ cray-libsci/17.06.1
+ cray-mpich/7.6.0
+
craype
- craype/2.5.5
+ craype/2.5.12.3
craype-ivybridge
pmi
- pmi/5.0.10-1.0000.11050.0.0.ari
+ pmi/5.0.12
+ cray-netcdf-hdf5parallel
+ cray-hdf5-parallel
+ cray-parallel-netcdf
cray-hdf5/1.8.16
cray-netcdf/4.4.0
+ cray-netcdf-hdf5parallel
cray-netcdf-hdf5parallel/4.4.0
cray-hdf5-parallel/1.8.16
cray-parallel-netcdf/1.6.1
-
- git/2.4.6
- papi/5.4.3.2
- perl/5.20.0
- cmake/3.3.2
-
-
1
1
1
+
64M
+ spread
+ threads
yes
-
- yes
-
@@ -221,7 +207,7 @@
-n $TOTALPES
-c 2
-
+ --cpu_bind=cores
@@ -289,10 +275,14 @@
+ cray-netcdf-hdf5parallel
+ cray-hdf5-parallel
+ cray-parallel-netcdf
cray-hdf5/1.8.16
cray-netcdf/4.4.0
+ cray-netcdf-hdf5parallel
cray-netcdf-hdf5parallel/4.4.0
cray-hdf5-parallel/1.8.16
cray-parallel-netcdf/1.7.0
@@ -302,7 +292,6 @@
git/2.9.1
cmake/3.3.2
pmi/5.0.10-1.0000.11069.183.8.ari
- papi/5.4.3.2
zlib/1.2.8
@@ -423,7 +412,9 @@
- cray-netcdf-hdf5parallel
+ cray-netcdf-hdf5parallel
+ cray-hdf5-parallel
+ cray-parallel-netcdf
cray-hdf5/1.8.16
cray-netcdf/4.4.0
@@ -437,9 +428,7 @@
git/2.9.1
cmake/3.3.2
pmi/5.0.10-1.0000.11069.183.8.ari
- papi/5.4.3.2
zlib/1.2.8
-
@@ -529,7 +518,6 @@
$CIME_OUTPUT_ROOT/archive/$CASE
csm/$CASE
/sems-data-store/ACME/baselines
- /sems-data-store/ACME/timings
/sems-data-store/ACME/cprnc/build.new/cprnc
jgfouca at sandia dot gov
@@ -594,7 +582,6 @@
$CIME_OUTPUT_ROOT/archive/$CASE
csm/$CASE
/sems-data-store/ACME/baselines
- /sems-data-store/ACME/timings
/sems-data-store/ACME/cprnc/build/cprnc
jgfouca at sandia dot gov
@@ -633,6 +620,7 @@
sems-openmpi/1.8.7
sems-cmake/2.8.12
sems-netcdf/4.4.1/exo_parallel
+ sems-boost/1.58.0/base
@@ -659,7 +647,6 @@
csm/$CASE
/home/climate1/acme/baselines
/home/climate1/acme/cprnc/build/cprnc
- $CIME_OUTPUT_ROOT/timings
jgfouca at sandia dot gov
@@ -722,8 +709,7 @@
$CIME_OUTPUT_ROOT/archive/$CASE
USERDEFINED_optional_run
/projects/ccsm/ccsm_baselines
- /projects/ccsm/cprnc/build.new/cprnc_wrap
- /projects/ccsm/timings
+ /projects/ccsm/cprnc/build.toss3/cprnc_wrap
slurm
jgfouca at sandia dot gov
8
@@ -736,21 +722,20 @@
mpiexec
- --bind-to-core
--n $TOTALPES
- --npernode $PES_PER_NODE
+ --map-by ppr:{{ tasks_per_numa }}:socket:PE=$ENV{OMP_NUM_THREADS} --bind-to core
- /usr/share/Modules/init/python.py
- /usr/share/Modules/init/perl.pm
- /usr/share/Modules/init/sh
- /usr/share/Modules/init/csh
- /usr/bin/modulecmd python
- /usr/bin/modulecmd perl
+ /usr/share/lmod/lmod/init/python.py
+ /usr/share/lmod/lmod/init/perl.pm
+ /usr/share/lmod/lmod/init/sh
+ /usr/share/lmod/lmod/init/csh
+ /usr/share/lmod/lmod/libexec/lmod python
+ /usr/share/lmod/lmod/libexec/lmod perl
module
module
@@ -760,27 +745,18 @@
sems-python/2.7.9
sems-cmake
gnu/4.9.2
- intel/intel-15.0.3.187
- libraries/intel-mkl-15.0.2.164
+ sems-intel/16.0.2
+ mkl/16.0
- openmpi-intel/1.6
- sems-hdf5/1.8.12/parallel
+ sems-openmpi/1.10.5
sems-netcdf/4.4.1/exo_parallel
- sems-hdf5/1.8.12/base
sems-netcdf/4.4.1/exo
-
-
-
-
-
-
-
$ENV{SEMS_NETCDF_ROOT}
$ENV{SEMS_NETCDF_ROOT}/include
$ENV{SEMS_NETCDF_ROOT}/lib
@@ -808,7 +784,6 @@
USERDEFINED_optional_run
/projects/ccsm/ccsm_baselines
/projects/ccsm/cprnc/build.new/cprnc_wrap
- /projects/ccsm/timings
slurm
jgfouca at sandia dot gov
8
@@ -816,14 +791,13 @@
16
1
TRUE
- fy150001
+ fy150001P
mpiexec
- --bind-to-core
--n $TOTALPES
- --npernode $PES_PER_NODE
+ --map-by ppr:{{ tasks_per_numa }}:socket:PE=$ENV{OMP_NUM_THREADS} --bind-to core
@@ -859,13 +833,6 @@
-
-
-
-
-
-
-
$ENV{SEMS_NETCDF_ROOT}
$ENV{SEMS_NETCDF_ROOT}/include
$ENV{SEMS_NETCDF_ROOT}/lib
@@ -876,14 +843,80 @@
+
+ SNL clust
+ ghost-login
+ wwwproxy.sandia.gov:80
+ acme_integration
+ intel
+ openmpi,mpi-serial
+ LINUX
+ /gscratch/$USER/acme_scratch/ghost
+ $CIME_OUTPUT_ROOT/$CASE/run
+ $CIME_OUTPUT_ROOT/$CASE/bld
+ /projects/ccsm/inputdata
+ /projects/ccsm/inputdata/atm/datm7
+ $CIME_OUTPUT_ROOT/archive/$CASE
+ USERDEFINED_optional_run
+ /projects/ccsm/ccsm_baselines
+ /projects/ccsm/cprnc/build.toss3/cprnc_wrap
+ slurm
+ jgfouca at sandia dot gov
+ 8
+ 36
+ 36
+ 1
+ TRUE
+ fy150001
+
+
+ mpiexec
+
+ --n $TOTALPES
+ --map-by ppr:{{ tasks_per_numa }}:socket:PE=$ENV{OMP_NUM_THREADS} --bind-to core
+
+
+
+
+
+
+ /usr/share/lmod/lmod/init/python.py
+ /usr/share/lmod/lmod/init/perl.pm
+ /usr/share/lmod/lmod/init/sh
+ /usr/share/lmod/lmod/init/csh
+ /usr/share/lmod/lmod/libexec/lmod python
+ /usr/share/lmod/lmod/libexec/lmod perl
+ module
+ module
+
+
+ sems-env
+ sems-git
+ sems-python/2.7.9
+ sems-cmake
+ gnu/4.9.2
+ sems-intel/16.0.2
+ sems-openmpi/1.10.5
+ mkl/16.0
+ sems-netcdf/4.4.1/exo_parallel
+
+
+
+ $ENV{SEMS_NETCDF_ROOT}
+ $ENV{SEMS_NETCDF_ROOT}
+ $ENV{SEMS_NETCDF_ROOT}/include
+ $ENV{SEMS_NETCDF_ROOT}/lib
+ 64M
+
+
+
ANL/LCRC Linux Cluster
- b.*.lcrc.anl.gov
+ blogin.*.lcrc.anl.gov
acme_integration
gnu,pgi,intel,nag
mvapich,mpich,openmpi
/lcrc/project/$PROJECT/$USER/acme_scratch
- /lcrc/project/$PROJECT
$CIME_OUTPUT_ROOT/$CASE/run
$CIME_OUTPUT_ROOT/$CASE/bld
/home/ccsm-data/inputdata
@@ -906,6 +939,12 @@
-n $TOTALPES
+
+ mpiexec
+
+ -n $TOTALPES
+
+
@@ -914,21 +953,39 @@
/etc/profile.d/a_softenv.sh
soft
soft
+
+ +cmake-2.8.12
+ +python-2.7
+
+gcc-5.2
+netcdf-4.3.3.1-gnu5.2-serial
- +cmake-2.8.12
- +python-2.7
+mvapich2-2.2b-gcc-5.2
- +cmake-2.8.12
- +python-2.7
+intel-15.0
+pnetcdf-1.6.1-mvapich2-2.2a-intel-15.0
+mvapich2-2.2b-intel-15.0
+mkl-11.2.1
+
+ +pgi-15.7
+ +binutils-2.27
+ +netcdf-c-4.4.1-f77-4.4.4-pgi-15.7-serial
+
+
+ +mvapich2-2.2-pgi-15.7
+ +pnetcdf-1.7.0-pgi-15.7-mvapich2-2.2
+
+
+ +nag-6.0
+ +hdf5-1.8.12-serial-nag
+ +netcdf-4.3.1-serial-nag
+
+
+ +mpich3-3.1.4-nag-6.0
+ +pnetcdf-1.6.1-mpich-3.1.4-nag-6.0
+
/soft/netcdf_serial/4.3.3.1/gnu-5.2
@@ -947,6 +1004,18 @@
/soft/climate/pnetcdf/1.6.1/intel-15.0.1/mvapich2-2.2a-intel-15.0
+
+ /soft/spack-0.10.0/opt/spack/linux-centos6-x86_64/pgi-15.7-0/netcdf-4.4.1-4el54ak5ic4ukbfxknv64ejvlum2kmy7
+
+
+ /soft/spack-0.10.0/opt/spack/linux-centos6-x86_64/pgi-15.7-0/parallel-netcdf-1.7.0-htrulk5hgp3jjci2f72srr5ujkg6q4i6
+
+
+ /soft/netcdf/4.3.1-serial/nag-6.0
+
+
+ /home/robl/soft/pnetcf-trunk-mpich-3.1.4-nag-6.0
+
64M
@@ -954,7 +1023,7 @@
ANL/LCRC Linux Cluster
- b.*.lcrc.anl.gov
+ blogin.*.lcrc.anl.gov
acme_integration
intel,gnu,pgi
mvapich,openmpi
@@ -1044,7 +1113,7 @@
$SHELL{which pnetcdf_version | xargs dirname | xargs dirname}
-
+
256M
spread
threads
@@ -1052,10 +1121,89 @@
1
1
+
+ 1
+ 1
+ 1
+
+
+
+
+ ANL/LCRC Cluster, Cray CS400, 352-nodes Xeon Phi 7230 KNLs 64C/1.3GHz + 672-nodes Xeon E5-2695v4 Broadwells 36C/2.10GHz, Intel Omni-Path network, SLURM batch system, Lmod module environment.
+ beboplogin.*
+ acme_developer
+ intel,gnu
+ impi,mpich,mvapich,openmpi,mpi-serial
+ /lcrc/group/acme/$USER/acme_scratch/bebop
+ $CIME_OUTPUT_ROOT/$CASE/run
+ $CIME_OUTPUT_ROOT/$CASE/bld
+ /home/ccsm-data/inputdata
+ /home/ccsm-data/inputdata/atm/datm7
+ $CIME_OUTPUT_ROOT/archive/$CASE
+ csm/$CASE
+ /lcrc/group/acme/acme_baselines/bebop
+ /lcrc/group/acme/tools/cprnc/cprnc
+ LINUX
+ slurm
+ acme
+ 8
+ 36
+ 36
+ acme
+ TRUE
+ -D PIO_BUILD_TIMING:BOOL=ON
+
+ mpirun
+
+ -n $TOTALPES
+
+
+
+ /home/software/spack-0.10.1/opt/spack/linux-centos7-x86_64/gcc-4.8.5/lmod-7.4.9-ic63herzfgw5u3na5mdtvp3nwxy6oj2z/lmod/lmod/init/sh
+ /home/software/spack-0.10.1/opt/spack/linux-centos7-x86_64/gcc-4.8.5/lmod-7.4.9-ic63herzfgw5u3na5mdtvp3nwxy6oj2z/lmod/lmod/init/csh
+ /home/software/spack-0.10.1/opt/spack/linux-centos7-x86_64/gcc-4.8.5/lmod-7.4.9-ic63herzfgw5u3na5mdtvp3nwxy6oj2z/lmod/lmod/init/perl
+ /home/software/spack-0.10.1/opt/spack/linux-centos7-x86_64/gcc-4.8.5/lmod-7.4.9-ic63herzfgw5u3na5mdtvp3nwxy6oj2z/lmod/lmod/init/env_modules_python.py
+ /home/software/spack-0.10.1/opt/spack/linux-centos7-x86_64/gcc-4.8.5/lmod-7.4.9-ic63herzfgw5u3na5mdtvp3nwxy6oj2z/lmod/lmod/libexec/lmod perl
+ /home/software/spack-0.10.1/opt/spack/linux-centos7-x86_64/gcc-4.8.5/lmod-7.4.9-ic63herzfgw5u3na5mdtvp3nwxy6oj2z/lmod/lmod/libexec/lmod python
+ module
+ module
+
+
+
+
+ intel/17.0.4-74uvhji
+ intel-mkl/2017.3.196-jyjmyut
+
+
+ gcc/7.1.0-4bgguyp
+
+
+ intel-mpi/2017.3-dfphq6k
+
+
+ cmake
+ netcdf/4.4.1.1-prsuusl
+ netcdf-fortran/4.4.4-ojwazvy
+ parallel-netcdf/1.8.1
+
+
+
+ $SHELL{which nf-config | xargs dirname | xargs dirname}
+ $SHELL{which pnetcdf_version | xargs dirname | xargs dirname}
+ /lcrc/group/acme/soft/perl/5.26.0/bin:$ENV{PATH}
+
+
+ 128M
+ spread
+ threads
+
+
+ shm:tmi
+
- ANL IBM BG/Q, os is BGP, 16 cores/node, batch system is cobalt
+ ANL IBM BG/Q, os is BGQ, 16 cores/node, batch system is cobalt
cetus
acme_developer
ibm
@@ -1069,7 +1217,6 @@
/home/$USER/csm/$CASE/
/projects/ccsm/ccsm_baselines/
/projects/ccsm/tools/cprnc/cprnc
- /projects/$PROJECT
BGQ
cobalt
jayesh -at- mcs.anl.gov
@@ -1077,21 +1224,16 @@
4
64
TRUE
- HiRes_EarthSys_2
+ ClimateEnergy_2
-D PIO_BUILD_TIMING:BOOL=ON
/usr/bin/runjob
- --label short
-
- --ranks-per-node $PES_PER_NODE
-
- --np $TOTALPES
- --block $COBALT_PARTNAME $LOCARGS
- --envs BG_THREADLAYOUT=1
- --envs XL_BG_SPREADLAYOUT=YES
- --envs OMP_STACKSIZE=64M
- --envs OMP_NUM_THREADS=$ENV{OMP_NUM_THREADS}
+ --label short
+ --ranks-per-node $PES_PER_NODE
+ --np $TOTALPES
+ --block $COBALT_PARTNAME $LOCARGS
+ $ENV{BGQ_SMP_VARS}
@@ -1108,8 +1250,10 @@
10000
- FALSE
- 64M
+
+
+
+ --envs BG_THREADLAYOUT=1 XL_BG_SPREADLAYOUT=YES OMP_DYNAMIC=FALSE OMP_STACKSIZE=64M OMP_NUM_THREADS=$ENV{OMP_NUM_THREADS}
@@ -1133,7 +1277,6 @@
16
16
lc_slurm
- /usr/gdata/climdat/timing_performance
@@ -1190,7 +1333,6 @@
16
16
lc_slurm
- /usr/gdata/climdat/timing_performance
@@ -1227,8 +1369,65 @@
+
+
+ LLNL Linux Cluster, Linux (pgi), 36 pes/node, batch system is Slurm
+ intel
+ mpich,mpi-serial
+ /p/lscratchh/$CCSMUSER/ACME/$CASE/run
+ /p/lscratchh/$CCSMUSER/$CASE/bld
+ /p/lscratchh/$USER
+ /usr/gdata/climdat/ccsm3data/inputdata
+ /usr/gdata/climdat/ccsm3data/inputdata/atm/datm7
+ /p/lscratchh/$CCSMUSER/archive/$CASE
+ FALSE
+ UNSET
+ /p/lscratchh/$CCSMUSER/ccsm_baselines
+ /p/lscratchd/ma21/ccsm3data/tools/cprnc/cprnc
+ LINUX
+ donahue5 -at- llnl.gov
+ 8
+ 36
+ 36
+ lc_slurm
+
+
+
+
+ srun
+
+
+ /usr/share/lmod/lmod/init/env_modules_python.py
+ /usr/share/lmod/lmod/init/perl
+ /usr/share/lmod/lmod/init/sh
+ /usr/share/lmod/lmod/init/csh
+ module
+ module
+ /usr/share/lmod/lmod/libexec/lmod python
+ /usr/share/lmod/lmod/libexec/lmod perl
+
+ mvapich2
+ intel
+ python/3.5.1
+ git/2.8.3
+ intel/17.0.0
+ mvapich2/2.2
+ netcdf-fortran-4.4.4-intel-17.0.0-4naprkr
+ parallel-netcdf-1.8.0-intel-17.0.0-gy2hic6
+
+
+
+ /usr/workspace/wsa/climdat/spack/opt/spack/linux-rhel7-x86_64/intel-17.0.0/netcdf-fortran-4.4.4-4naprkre2m7kriadyxwboauil7nc3jtc/
+ /usr/workspace/wsa/climdat/spack/opt/spack/linux-rhel7-x86_64/intel-17.0.0/parallel-netcdf-1.8.0-gy2hic6n3hkygcczpumkcqmzw5sam7hn/
+
+
+
- ANL IBM BG/Q, os is BGP, 16 cores/node, batch system is cobalt
+ ANL IBM BG/Q, os is BGQ, 16 cores/node, batch system is cobalt
mira.*
acme_developer
ibm
@@ -1250,21 +1449,16 @@
4
64
TRUE
- HiRes_EarthSys_2
+ ClimateEnergy_2
-D PIO_BUILD_TIMING:BOOL=ON
/usr/bin/runjob
- --label short
-
- --ranks-per-node $PES_PER_NODE
-
- --np $TOTALPES
- --block $COBALT_PARTNAME $LOCARGS
- --envs BG_THREADLAYOUT=1
- --envs XL_BG_SPREADLAYOUT=YES
- --envs OMP_STACKSIZE=64M
- --envs OMP_NUM_THREADS=$ENV{OMP_NUM_THREADS}
+ --label short
+ --ranks-per-node $PES_PER_NODE
+ --np $TOTALPES
+ --block $COBALT_PARTNAME $LOCARGS
+ $ENV{BGQ_SMP_VARS}
@@ -1281,8 +1475,10 @@
10000
- FALSE
- 64M
+
+
+
+ --envs BG_THREADLAYOUT=1 XL_BG_SPREADLAYOUT=YES OMP_DYNAMIC=FALSE OMP_STACKSIZE=64M OMP_NUM_THREADS=$ENV{OMP_NUM_THREADS}
@@ -1292,21 +1488,20 @@
acme_developer
intel,gnu,cray
mpt
- /projects/EarlyPerf_theta/acme/scratch/$USER
- /projects/EarlyPerf_theta/acme/inputdata
- /projects/EarlyPerf_theta/acme/inputdata/atm/datm7
+ /projects/$PROJECT/$USER
+ /projects/$PROJECT/acme/inputdata
+ /projects/$PROJECT/acme/inputdata/atm/datm7
$CIME_OUTPUT_ROOT/archive/$CASE
- /projects/EarlyPerf_theta/acme/csm/$CASE
- /projects/EarlyPerf_theta/acme/baselines
- /projects/EarlyPerf_theta/acme/tools/cprnc
- /projects/EarlyPerf_theta/acme
+ $CIME_OUTPUT_ROOT/csm/$CASE
+ /projects/$PROJECT/acme/baselines
+ /projects/$PROJECT/acme/tools/cprnc/cprnc
CNL
cobalt_theta
acme
8
64
64
- Theta_ESP
+ OceanClimate
TRUE
-D PIO_BUILD_TIMING:BOOL=ON
@@ -1329,47 +1524,47 @@
module
module
+ craype-mic-knl
PrgEnv-intel
PrgEnv-cray
PrgEnv-gnu
intel
cce
+ cray-mpich
cray-parallel-netcdf
cray-hdf5-parallel
- pmi
- cray-libsci
- cray-mpich
- cray-netcdf
cray-hdf5
+ cray-netcdf
cray-netcdf-hdf5parallel
+ cray-libsci
craype
+
+ craype/2.5.11
+
- PrgEnv-intel/6.0.3
- intel intel/17.0.2.174
- cray-libsci
+ PrgEnv-intel/6.0.4
+ intel/17.0.4.196
- PrgEnv-cray/6.0.3
- cce cce/8.5.4
+ PrgEnv-cray/6.0.4
+ cce/8.6.0
- PrgEnv-gnu/6.0.3
- gcc gcc/6.2.0
-
-
- craype craype/2.5.9
+ PrgEnv-gnu/6.0.4
+ gcc/6.3.0
- cray-libsci/16.09.1
+ cray-libsci/17.06.1
- cray-mpich/7.5.3
+ craype-mic-knl
+ cray-mpich/7.6.0
- cray-netcdf-hdf5parallel/4.4.1.1
- cray-hdf5-parallel/1.10.0.1
- cray-parallel-netcdf/1.8.0
+ cray-netcdf-hdf5parallel/4.4.1.1.3
+ cray-hdf5-parallel/1.10.0.3
+ cray-parallel-netcdf/1.8.1.3
@@ -1397,7 +1592,6 @@
UNSET
/lustre/climate/acme_baselines
/lustre/climate/acme_baselines/cprnc/cprnc
- /lustre/climate/timing_acme
8
slurm
balwinder.singh -at- pnnl.gov
@@ -1471,7 +1665,6 @@
UNSET
/dtemp/sing201/acme/acme_baselines
/dtemp/sing201/acme/acme_baselines/cprnc/cprnc
- /dtemp/sing201/acme/timing_acme
slurm
balwinder.singh -at- pnnl.gov
8
@@ -1523,7 +1716,7 @@
intel
-
+
PNL Haswell cluster, OS is Linux, batch system is SLURM
LINUX
@@ -1539,7 +1732,6 @@
UNSET
/pic/projects/climate/acme_baselines
/pic/projects/climate/acme_baselines/cprnc/cprnc
- /pic/projects/climate/csmdata/
slurm
balwinder.singh -at- pnnl.gov
8
@@ -1795,7 +1987,7 @@
cli115
-D PIO_BUILD_TIMING:BOOL=ON
- aprun
+ aprun
+
+ $COMPILER
+ $MPILIB
+ 1
+ 1
+ 1
+ 128M
+ 128M
+
-
-
- $COMPILER
- $MPILIB
- 1
- 1
- 1
- 128M
- 128M
-
-
-
-
-
- istanbul
- 1
- /opt/cray/netcdf-hdf5parallel/4.4.1.1/PGI/15.3/
-
-
-
- /opt/cray/parallel-netcdf/1.7.0/PGI/15.3
-
+
+
+
+ istanbul
+ 1
+/opt/cray/netcdf-hdf5parallel/4.4.1.1/PGI/15.3/
+
-
-
- istanbul
- /opt/cray/netcdf-hdf5parallel/4.4.1.1/PGI/15.3/
-
+
+/opt/cray/parallel-netcdf/1.7.0/PGI/15.3
+
-
- /opt/cray/parallel-netcdf/1.7.0/PGI/15.3
-
+
+
+ istanbul
+/opt/cray/netcdf-hdf5parallel/4.4.1.1/PGI/15.3/
+
-
- dynamic
- /opt/cray/netcdf-hdf5parallel/4.4.1.1/INTEL/15.0/
-
+
+/opt/cray/parallel-netcdf/1.7.0/PGI/15.3
+
-
- /opt/cray/parallel-netcdf/1.7.0/INTEL/15.0
-
+
+ dynamic
+/opt/cray/netcdf-hdf5parallel/4.4.1.1/INTEL/15.0/
+
-
- /opt/cray/netcdf-hdf5parallel/4.4.1.1/CRAY/8.3/
-
+
+/opt/cray/parallel-netcdf/1.7.0/INTEL/15.0
+
-
- /opt/cray/parallel-netcdf/1.7.0/CRAY/8.3
-
+
+/opt/cray/netcdf-hdf5parallel/4.4.1.1/CRAY/8.3/
+
-
+
+/opt/cray/parallel-netcdf/1.7.0/CRAY/8.3
+
@@ -2047,7 +2238,7 @@
- LANL Linux Cluster, 36 pes/node, batch system Moab
+ LANL Linux Cluster, 36 pes/node, batch system slurm
gr-fe.*.lanl.gov
acme_developer
gnu,intel
@@ -2061,7 +2252,6 @@
UNSET
/lustre/scratch3/turquoise/$ENV{USER}/ACME/input_data/ccsm_baselines
/lustre/scratch3/turquoise/$ENV{USER}/ACME/scratch
- $CASEROOT/timings
/turquoise/usr/projects/climate/SHARED_CLIMATE/software/wolf/cprnc/v0.40/cprnc
/usr/share/Modules/init/perl.pm
@@ -2101,7 +2291,7 @@
romio_ds_write=disable;romio_ds_read=disable;romio_cb_write=enable;romio_cb_read=enable
- moab
+ slurm
mpirun
@@ -2132,7 +2322,7 @@
- LANL Linux Cluster, 16 pes/node, batch system Moab
+ LANL Linux Cluster, 16 pes/node, batch system slurm
wf-fe.*.lanl.gov
acme_developer
intel,gnu
@@ -2143,7 +2333,7 @@
/lustre/scratch3/turquoise/$ENV{USER}/ACME/archive/$CASE
UNSET
LINUX
- moab
+ slurm
jonbob -at- lanl.gov
4
16
@@ -2154,7 +2344,6 @@
/lustre/scratch3/turquoise/$ENV{USER}/ACME/input_data/atm/datm7
/lustre/scratch3/turquoise/$ENV{USER}/ACME/input_data/ccsm_baselines
/turquoise/usr/projects/climate/SHARED_CLIMATE/software/wolf/cprnc/v0.40/cprnc
- $CASEROOT/timings
mpirun
@@ -2414,7 +2603,6 @@
16
16
TRUE
- /global/scratch/$ENV{USER}
mpirun
@@ -2478,7 +2666,6 @@
12
12
TRUE
- /global/scratch/$ENV{USER}
mpirun
@@ -2553,6 +2740,111 @@
+
+ ORNL pre-Summit testbed. Node: 2x POWER8 + 4x Tesla P100, 20 cores/node, 8 HW threads/core.
+ summitdev-*
+ Received node event ec_node
+ acme_developer
+ ibm,pgi
+ openmpi,mpi_serial
+ $ENV{HOME}/acme_scratch/$PROJECT
+ /lustre/atlas/scratch/$ENV{USER}/$PROJECT/$CASE/run
+ $CIME_OUTPUT_ROOT/$CASE/bld
+ /lustre/atlas1/cli900/world-shared/cesm/inputdata
+ /lustre/atlas1/cli900/world-shared/cesm/inputdata/atm/datm7
+ /lustre/atlas/scratch/$ENV{USER}/$PROJECT/archive/$CASE
+ csm/$CASE
+ /lustre/atlas1/cli900/world-shared/cesm/baselines
+ /lustre/atlas1/cli900/world-shared/cesm/tools/cprnc/cprnc
+ /lustre/atlas/proj-shared/$PROJECT
+ LINUX
+ lsf
+ acme
+ 32
+ 20
+ 160
+ TRUE
+ csc249
+ -D PIO_BUILD_TIMING:BOOL=ON
+
+
+ mpirun
+
+ -np $TOTALPES
+ --map-by core:PE=$ENV{OMP_NUM_THREADS} --bind-to core
+ -x OMP_NUM_THREADS=$ENV{OMP_NUM_THREADS}
+ --report-bindings
+ --display-map
+
+
+
+
+
+
+ /sw/summitdev/lmod/7.4.0/rhel7.2_gnu4.8.5/lmod/7.4/init/sh
+ /sw/summitdev/lmod/7.4.0/rhel7.2_gnu4.8.5/lmod/7.4/init/csh
+ /sw/summitdev/lmod/7.4.0/rhel7.2_gnu4.8.5/lmod/7.4/init/env_modules_python.py
+ /sw/summitdev/lmod/7.4.0/rhel7.2_gnu4.8.5/lmod/7.4/init/perl
+
+ module
+ module
+ module
+ module
+
+
+
+
+
+
+ DefApps
+ python/3.5.2
+ subversion/1.9.3
+ git/2.13.0
+ cmake/3.6.1
+ essl/5.5.0-20161110
+ netlib-lapack/3.6.1
+
+
+
+ xl
+ pgi/17.4
+
+
+
+ pgi
+ xl/20161123
+
+
+
+
+
+ netcdf/4.4.1
+ netcdf-fortran/4.4.4
+
+
+ spectrum_mpi/10.1.0.2-20161221
+ netcdf/4.4.1
+ netcdf-fortran/4.4.4
+ parallel-netcdf/1.7.0
+ hdf5/1.10.0-patch1-parallel
+
+
+
+ $COMPILER
+ $MPILIB
+ 128M
+ $ENV{OLCF_NETCDF_ROOT}
+ $ENV{OLCF_NETCDF_FORTRAN_ROOT}
+ $ENV{OLCF_HDF5_ROOT}
+ $ENV{OLCF_ESSL_ROOT}
+
+
+ $ENV{OLCF_PARALLEL_NETCDF_ROOT}
+
+
+
${EXEROOT}/acme.exe
diff --git a/config/xml_schemas/env_mach_specific.xsd b/config/xml_schemas/env_mach_specific.xsd
index 45a4ddfc6564..a8a35b581813 100644
--- a/config/xml_schemas/env_mach_specific.xsd
+++ b/config/xml_schemas/env_mach_specific.xsd
@@ -7,6 +7,7 @@
+
@@ -80,6 +81,7 @@
+