diff --git a/config/acme/machines/config_machines.xml b/config/acme/machines/config_machines.xml index 6bf273f03ded..b82eb059f7f8 100644 --- a/config/acme/machines/config_machines.xml +++ b/config/acme/machines/config_machines.xml @@ -60,7 +60,7 @@ NERSC XC30, os is CNL, 24 pes/node, batch system is SLURM edison acme_developer - intel,intel17,gnu,cray + intel,gnu mpt $ENV{CSCRATCH}/acme_scratch/edison $CIME_OUTPUT_ROOT/$CASE/run @@ -85,7 +85,8 @@ --label -n $TOTALPES - -c $ENV{OMP_NUM_THREADS} + -c $SHELL{echo 48/`./xmlquery --value MAX_TASKS_PER_NODE`*$ENV{OMP_NUM_THREADS} | bc } + $SHELL{if [ 24 -ge `./xmlquery --value MAX_TASKS_PER_NODE` ]; then echo "--cpu_bind=cores"; else echo "--cpu_bind=threads"; fi} @@ -117,7 +118,6 @@ craype-ivybridge craype papi - cmake cray-petsc esmf @@ -125,68 +125,54 @@ PrgEnv-intel intel - intel/15.0.1.133 + intel/17.0.2.174 cray-libsci - cray-mpich/7.2.5 - - - PrgEnv-intel - intel - intel/17.0.1.132 - cray-libsci - cray-mpich/7.5.1 - - - PrgEnv-cray - cce cce/8.5.1 - cray-libsci/16.07.1 - cray-mpich/7.5.1 PrgEnv-gnu - gcc gcc/6.2.0 - cray-libsci/16.07.1 - cray-mpich/7.5.1 + gcc + gcc/6.3.0 + cray-libsci + cray-libsci/17.06.1 + cray-mpich/7.6.0 + craype - craype/2.5.5 + craype/2.5.12.3 craype-ivybridge pmi - pmi/5.0.10-1.0000.11050.0.0.ari + pmi/5.0.12 + cray-netcdf-hdf5parallel + cray-hdf5-parallel + cray-parallel-netcdf cray-hdf5/1.8.16 cray-netcdf/4.4.0 + cray-netcdf-hdf5parallel cray-netcdf-hdf5parallel/4.4.0 cray-hdf5-parallel/1.8.16 cray-parallel-netcdf/1.6.1 - - git/2.4.6 - papi/5.4.3.2 - perl/5.20.0 - cmake/3.3.2 - - 1 1 1 + 64M + spread + threads yes - - yes - @@ -221,7 +207,7 @@ -n $TOTALPES -c 2 - + --cpu_bind=cores @@ -289,10 +275,14 @@ + cray-netcdf-hdf5parallel + cray-hdf5-parallel + cray-parallel-netcdf cray-hdf5/1.8.16 cray-netcdf/4.4.0 + cray-netcdf-hdf5parallel cray-netcdf-hdf5parallel/4.4.0 cray-hdf5-parallel/1.8.16 cray-parallel-netcdf/1.7.0 @@ -302,7 +292,6 @@ git/2.9.1 cmake/3.3.2 pmi/5.0.10-1.0000.11069.183.8.ari - papi/5.4.3.2 zlib/1.2.8 @@ -423,7 +412,9 @@ - cray-netcdf-hdf5parallel + cray-netcdf-hdf5parallel + cray-hdf5-parallel + cray-parallel-netcdf cray-hdf5/1.8.16 cray-netcdf/4.4.0 @@ -437,9 +428,7 @@ git/2.9.1 cmake/3.3.2 pmi/5.0.10-1.0000.11069.183.8.ari - papi/5.4.3.2 zlib/1.2.8 - @@ -529,7 +518,6 @@ $CIME_OUTPUT_ROOT/archive/$CASE csm/$CASE /sems-data-store/ACME/baselines - /sems-data-store/ACME/timings /sems-data-store/ACME/cprnc/build.new/cprnc jgfouca at sandia dot gov @@ -594,7 +582,6 @@ $CIME_OUTPUT_ROOT/archive/$CASE csm/$CASE /sems-data-store/ACME/baselines - /sems-data-store/ACME/timings /sems-data-store/ACME/cprnc/build/cprnc jgfouca at sandia dot gov @@ -633,6 +620,7 @@ sems-openmpi/1.8.7 sems-cmake/2.8.12 sems-netcdf/4.4.1/exo_parallel + sems-boost/1.58.0/base @@ -659,7 +647,6 @@ csm/$CASE /home/climate1/acme/baselines /home/climate1/acme/cprnc/build/cprnc - $CIME_OUTPUT_ROOT/timings jgfouca at sandia dot gov @@ -722,8 +709,7 @@ $CIME_OUTPUT_ROOT/archive/$CASE USERDEFINED_optional_run /projects/ccsm/ccsm_baselines - /projects/ccsm/cprnc/build.new/cprnc_wrap - /projects/ccsm/timings + /projects/ccsm/cprnc/build.toss3/cprnc_wrap slurm jgfouca at sandia dot gov 8 @@ -736,21 +722,20 @@ mpiexec - --bind-to-core --n $TOTALPES - --npernode $PES_PER_NODE + --map-by ppr:{{ tasks_per_numa }}:socket:PE=$ENV{OMP_NUM_THREADS} --bind-to core - /usr/share/Modules/init/python.py - /usr/share/Modules/init/perl.pm - /usr/share/Modules/init/sh - /usr/share/Modules/init/csh - /usr/bin/modulecmd python - /usr/bin/modulecmd perl + /usr/share/lmod/lmod/init/python.py + /usr/share/lmod/lmod/init/perl.pm + /usr/share/lmod/lmod/init/sh + /usr/share/lmod/lmod/init/csh + /usr/share/lmod/lmod/libexec/lmod python + /usr/share/lmod/lmod/libexec/lmod perl module module @@ -760,27 +745,18 @@ sems-python/2.7.9 sems-cmake gnu/4.9.2 - intel/intel-15.0.3.187 - libraries/intel-mkl-15.0.2.164 + sems-intel/16.0.2 + mkl/16.0 - openmpi-intel/1.6 - sems-hdf5/1.8.12/parallel + sems-openmpi/1.10.5 sems-netcdf/4.4.1/exo_parallel - sems-hdf5/1.8.12/base sems-netcdf/4.4.1/exo - - - - - - - $ENV{SEMS_NETCDF_ROOT} $ENV{SEMS_NETCDF_ROOT}/include $ENV{SEMS_NETCDF_ROOT}/lib @@ -808,7 +784,6 @@ USERDEFINED_optional_run /projects/ccsm/ccsm_baselines /projects/ccsm/cprnc/build.new/cprnc_wrap - /projects/ccsm/timings slurm jgfouca at sandia dot gov 8 @@ -816,14 +791,13 @@ 16 1 TRUE - fy150001 + fy150001P mpiexec - --bind-to-core --n $TOTALPES - --npernode $PES_PER_NODE + --map-by ppr:{{ tasks_per_numa }}:socket:PE=$ENV{OMP_NUM_THREADS} --bind-to core @@ -859,13 +833,6 @@ - - - - - - - $ENV{SEMS_NETCDF_ROOT} $ENV{SEMS_NETCDF_ROOT}/include $ENV{SEMS_NETCDF_ROOT}/lib @@ -876,14 +843,80 @@ + + SNL clust + ghost-login + wwwproxy.sandia.gov:80 + acme_integration + intel + openmpi,mpi-serial + LINUX + /gscratch/$USER/acme_scratch/ghost + $CIME_OUTPUT_ROOT/$CASE/run + $CIME_OUTPUT_ROOT/$CASE/bld + /projects/ccsm/inputdata + /projects/ccsm/inputdata/atm/datm7 + $CIME_OUTPUT_ROOT/archive/$CASE + USERDEFINED_optional_run + /projects/ccsm/ccsm_baselines + /projects/ccsm/cprnc/build.toss3/cprnc_wrap + slurm + jgfouca at sandia dot gov + 8 + 36 + 36 + 1 + TRUE + fy150001 + + + mpiexec + + --n $TOTALPES + --map-by ppr:{{ tasks_per_numa }}:socket:PE=$ENV{OMP_NUM_THREADS} --bind-to core + + + + + + + /usr/share/lmod/lmod/init/python.py + /usr/share/lmod/lmod/init/perl.pm + /usr/share/lmod/lmod/init/sh + /usr/share/lmod/lmod/init/csh + /usr/share/lmod/lmod/libexec/lmod python + /usr/share/lmod/lmod/libexec/lmod perl + module + module + + + sems-env + sems-git + sems-python/2.7.9 + sems-cmake + gnu/4.9.2 + sems-intel/16.0.2 + sems-openmpi/1.10.5 + mkl/16.0 + sems-netcdf/4.4.1/exo_parallel + + + + $ENV{SEMS_NETCDF_ROOT} + $ENV{SEMS_NETCDF_ROOT} + $ENV{SEMS_NETCDF_ROOT}/include + $ENV{SEMS_NETCDF_ROOT}/lib + 64M + + + ANL/LCRC Linux Cluster - b.*.lcrc.anl.gov + blogin.*.lcrc.anl.gov acme_integration gnu,pgi,intel,nag mvapich,mpich,openmpi /lcrc/project/$PROJECT/$USER/acme_scratch - /lcrc/project/$PROJECT $CIME_OUTPUT_ROOT/$CASE/run $CIME_OUTPUT_ROOT/$CASE/bld /home/ccsm-data/inputdata @@ -906,6 +939,12 @@ -n $TOTALPES + + mpiexec + + -n $TOTALPES + + @@ -914,21 +953,39 @@ /etc/profile.d/a_softenv.sh soft soft + + +cmake-2.8.12 + +python-2.7 + +gcc-5.2 +netcdf-4.3.3.1-gnu5.2-serial - +cmake-2.8.12 - +python-2.7 +mvapich2-2.2b-gcc-5.2 - +cmake-2.8.12 - +python-2.7 +intel-15.0 +pnetcdf-1.6.1-mvapich2-2.2a-intel-15.0 +mvapich2-2.2b-intel-15.0 +mkl-11.2.1 + + +pgi-15.7 + +binutils-2.27 + +netcdf-c-4.4.1-f77-4.4.4-pgi-15.7-serial + + + +mvapich2-2.2-pgi-15.7 + +pnetcdf-1.7.0-pgi-15.7-mvapich2-2.2 + + + +nag-6.0 + +hdf5-1.8.12-serial-nag + +netcdf-4.3.1-serial-nag + + + +mpich3-3.1.4-nag-6.0 + +pnetcdf-1.6.1-mpich-3.1.4-nag-6.0 + /soft/netcdf_serial/4.3.3.1/gnu-5.2 @@ -947,6 +1004,18 @@ /soft/climate/pnetcdf/1.6.1/intel-15.0.1/mvapich2-2.2a-intel-15.0 + + /soft/spack-0.10.0/opt/spack/linux-centos6-x86_64/pgi-15.7-0/netcdf-4.4.1-4el54ak5ic4ukbfxknv64ejvlum2kmy7 + + + /soft/spack-0.10.0/opt/spack/linux-centos6-x86_64/pgi-15.7-0/parallel-netcdf-1.7.0-htrulk5hgp3jjci2f72srr5ujkg6q4i6 + + + /soft/netcdf/4.3.1-serial/nag-6.0 + + + /home/robl/soft/pnetcf-trunk-mpich-3.1.4-nag-6.0 + 64M @@ -954,7 +1023,7 @@ ANL/LCRC Linux Cluster - b.*.lcrc.anl.gov + blogin.*.lcrc.anl.gov acme_integration intel,gnu,pgi mvapich,openmpi @@ -1044,7 +1113,7 @@ $SHELL{which pnetcdf_version | xargs dirname | xargs dirname} - + 256M spread threads @@ -1052,10 +1121,89 @@ 1 1 + + 1 + 1 + 1 + + + + + ANL/LCRC Cluster, Cray CS400, 352-nodes Xeon Phi 7230 KNLs 64C/1.3GHz + 672-nodes Xeon E5-2695v4 Broadwells 36C/2.10GHz, Intel Omni-Path network, SLURM batch system, Lmod module environment. + beboplogin.* + acme_developer + intel,gnu + impi,mpich,mvapich,openmpi,mpi-serial + /lcrc/group/acme/$USER/acme_scratch/bebop + $CIME_OUTPUT_ROOT/$CASE/run + $CIME_OUTPUT_ROOT/$CASE/bld + /home/ccsm-data/inputdata + /home/ccsm-data/inputdata/atm/datm7 + $CIME_OUTPUT_ROOT/archive/$CASE + csm/$CASE + /lcrc/group/acme/acme_baselines/bebop + /lcrc/group/acme/tools/cprnc/cprnc + LINUX + slurm + acme + 8 + 36 + 36 + acme + TRUE + -D PIO_BUILD_TIMING:BOOL=ON + + mpirun + + -n $TOTALPES + + + + /home/software/spack-0.10.1/opt/spack/linux-centos7-x86_64/gcc-4.8.5/lmod-7.4.9-ic63herzfgw5u3na5mdtvp3nwxy6oj2z/lmod/lmod/init/sh + /home/software/spack-0.10.1/opt/spack/linux-centos7-x86_64/gcc-4.8.5/lmod-7.4.9-ic63herzfgw5u3na5mdtvp3nwxy6oj2z/lmod/lmod/init/csh + /home/software/spack-0.10.1/opt/spack/linux-centos7-x86_64/gcc-4.8.5/lmod-7.4.9-ic63herzfgw5u3na5mdtvp3nwxy6oj2z/lmod/lmod/init/perl + /home/software/spack-0.10.1/opt/spack/linux-centos7-x86_64/gcc-4.8.5/lmod-7.4.9-ic63herzfgw5u3na5mdtvp3nwxy6oj2z/lmod/lmod/init/env_modules_python.py + /home/software/spack-0.10.1/opt/spack/linux-centos7-x86_64/gcc-4.8.5/lmod-7.4.9-ic63herzfgw5u3na5mdtvp3nwxy6oj2z/lmod/lmod/libexec/lmod perl + /home/software/spack-0.10.1/opt/spack/linux-centos7-x86_64/gcc-4.8.5/lmod-7.4.9-ic63herzfgw5u3na5mdtvp3nwxy6oj2z/lmod/lmod/libexec/lmod python + module + module + + + + + intel/17.0.4-74uvhji + intel-mkl/2017.3.196-jyjmyut + + + gcc/7.1.0-4bgguyp + + + intel-mpi/2017.3-dfphq6k + + + cmake + netcdf/4.4.1.1-prsuusl + netcdf-fortran/4.4.4-ojwazvy + parallel-netcdf/1.8.1 + + + + $SHELL{which nf-config | xargs dirname | xargs dirname} + $SHELL{which pnetcdf_version | xargs dirname | xargs dirname} + /lcrc/group/acme/soft/perl/5.26.0/bin:$ENV{PATH} + + + 128M + spread + threads + + + shm:tmi + - ANL IBM BG/Q, os is BGP, 16 cores/node, batch system is cobalt + ANL IBM BG/Q, os is BGQ, 16 cores/node, batch system is cobalt cetus acme_developer ibm @@ -1069,7 +1217,6 @@ /home/$USER/csm/$CASE/ /projects/ccsm/ccsm_baselines/ /projects/ccsm/tools/cprnc/cprnc - /projects/$PROJECT BGQ cobalt jayesh -at- mcs.anl.gov @@ -1077,21 +1224,16 @@ 4 64 TRUE - HiRes_EarthSys_2 + ClimateEnergy_2 -D PIO_BUILD_TIMING:BOOL=ON /usr/bin/runjob - --label short - - --ranks-per-node $PES_PER_NODE - - --np $TOTALPES - --block $COBALT_PARTNAME $LOCARGS - --envs BG_THREADLAYOUT=1 - --envs XL_BG_SPREADLAYOUT=YES - --envs OMP_STACKSIZE=64M - --envs OMP_NUM_THREADS=$ENV{OMP_NUM_THREADS} + --label short + --ranks-per-node $PES_PER_NODE + --np $TOTALPES + --block $COBALT_PARTNAME $LOCARGS + $ENV{BGQ_SMP_VARS} @@ -1108,8 +1250,10 @@ 10000 - FALSE - 64M + + + + --envs BG_THREADLAYOUT=1 XL_BG_SPREADLAYOUT=YES OMP_DYNAMIC=FALSE OMP_STACKSIZE=64M OMP_NUM_THREADS=$ENV{OMP_NUM_THREADS} @@ -1133,7 +1277,6 @@ 16 16 lc_slurm - /usr/gdata/climdat/timing_performance @@ -1190,7 +1333,6 @@ 16 16 lc_slurm - /usr/gdata/climdat/timing_performance @@ -1227,8 +1369,65 @@ + + + LLNL Linux Cluster, Linux (pgi), 36 pes/node, batch system is Slurm + intel + mpich,mpi-serial + /p/lscratchh/$CCSMUSER/ACME/$CASE/run + /p/lscratchh/$CCSMUSER/$CASE/bld + /p/lscratchh/$USER + /usr/gdata/climdat/ccsm3data/inputdata + /usr/gdata/climdat/ccsm3data/inputdata/atm/datm7 + /p/lscratchh/$CCSMUSER/archive/$CASE + FALSE + UNSET + /p/lscratchh/$CCSMUSER/ccsm_baselines + /p/lscratchd/ma21/ccsm3data/tools/cprnc/cprnc + LINUX + donahue5 -at- llnl.gov + 8 + 36 + 36 + lc_slurm + + + + + srun + + + /usr/share/lmod/lmod/init/env_modules_python.py + /usr/share/lmod/lmod/init/perl + /usr/share/lmod/lmod/init/sh + /usr/share/lmod/lmod/init/csh + module + module + /usr/share/lmod/lmod/libexec/lmod python + /usr/share/lmod/lmod/libexec/lmod perl + + mvapich2 + intel + python/3.5.1 + git/2.8.3 + intel/17.0.0 + mvapich2/2.2 + netcdf-fortran-4.4.4-intel-17.0.0-4naprkr + parallel-netcdf-1.8.0-intel-17.0.0-gy2hic6 + + + + /usr/workspace/wsa/climdat/spack/opt/spack/linux-rhel7-x86_64/intel-17.0.0/netcdf-fortran-4.4.4-4naprkre2m7kriadyxwboauil7nc3jtc/ + /usr/workspace/wsa/climdat/spack/opt/spack/linux-rhel7-x86_64/intel-17.0.0/parallel-netcdf-1.8.0-gy2hic6n3hkygcczpumkcqmzw5sam7hn/ + + + - ANL IBM BG/Q, os is BGP, 16 cores/node, batch system is cobalt + ANL IBM BG/Q, os is BGQ, 16 cores/node, batch system is cobalt mira.* acme_developer ibm @@ -1250,21 +1449,16 @@ 4 64 TRUE - HiRes_EarthSys_2 + ClimateEnergy_2 -D PIO_BUILD_TIMING:BOOL=ON /usr/bin/runjob - --label short - - --ranks-per-node $PES_PER_NODE - - --np $TOTALPES - --block $COBALT_PARTNAME $LOCARGS - --envs BG_THREADLAYOUT=1 - --envs XL_BG_SPREADLAYOUT=YES - --envs OMP_STACKSIZE=64M - --envs OMP_NUM_THREADS=$ENV{OMP_NUM_THREADS} + --label short + --ranks-per-node $PES_PER_NODE + --np $TOTALPES + --block $COBALT_PARTNAME $LOCARGS + $ENV{BGQ_SMP_VARS} @@ -1281,8 +1475,10 @@ 10000 - FALSE - 64M + + + + --envs BG_THREADLAYOUT=1 XL_BG_SPREADLAYOUT=YES OMP_DYNAMIC=FALSE OMP_STACKSIZE=64M OMP_NUM_THREADS=$ENV{OMP_NUM_THREADS} @@ -1292,21 +1488,20 @@ acme_developer intel,gnu,cray mpt - /projects/EarlyPerf_theta/acme/scratch/$USER - /projects/EarlyPerf_theta/acme/inputdata - /projects/EarlyPerf_theta/acme/inputdata/atm/datm7 + /projects/$PROJECT/$USER + /projects/$PROJECT/acme/inputdata + /projects/$PROJECT/acme/inputdata/atm/datm7 $CIME_OUTPUT_ROOT/archive/$CASE - /projects/EarlyPerf_theta/acme/csm/$CASE - /projects/EarlyPerf_theta/acme/baselines - /projects/EarlyPerf_theta/acme/tools/cprnc - /projects/EarlyPerf_theta/acme + $CIME_OUTPUT_ROOT/csm/$CASE + /projects/$PROJECT/acme/baselines + /projects/$PROJECT/acme/tools/cprnc/cprnc CNL cobalt_theta acme 8 64 64 - Theta_ESP + OceanClimate TRUE -D PIO_BUILD_TIMING:BOOL=ON @@ -1329,47 +1524,47 @@ module module + craype-mic-knl PrgEnv-intel PrgEnv-cray PrgEnv-gnu intel cce + cray-mpich cray-parallel-netcdf cray-hdf5-parallel - pmi - cray-libsci - cray-mpich - cray-netcdf cray-hdf5 + cray-netcdf cray-netcdf-hdf5parallel + cray-libsci craype + + craype/2.5.11 + - PrgEnv-intel/6.0.3 - intel intel/17.0.2.174 - cray-libsci + PrgEnv-intel/6.0.4 + intel/17.0.4.196 - PrgEnv-cray/6.0.3 - cce cce/8.5.4 + PrgEnv-cray/6.0.4 + cce/8.6.0 - PrgEnv-gnu/6.0.3 - gcc gcc/6.2.0 - - - craype craype/2.5.9 + PrgEnv-gnu/6.0.4 + gcc/6.3.0 - cray-libsci/16.09.1 + cray-libsci/17.06.1 - cray-mpich/7.5.3 + craype-mic-knl + cray-mpich/7.6.0 - cray-netcdf-hdf5parallel/4.4.1.1 - cray-hdf5-parallel/1.10.0.1 - cray-parallel-netcdf/1.8.0 + cray-netcdf-hdf5parallel/4.4.1.1.3 + cray-hdf5-parallel/1.10.0.3 + cray-parallel-netcdf/1.8.1.3 @@ -1397,7 +1592,6 @@ UNSET /lustre/climate/acme_baselines /lustre/climate/acme_baselines/cprnc/cprnc - /lustre/climate/timing_acme 8 slurm balwinder.singh -at- pnnl.gov @@ -1471,7 +1665,6 @@ UNSET /dtemp/sing201/acme/acme_baselines /dtemp/sing201/acme/acme_baselines/cprnc/cprnc - /dtemp/sing201/acme/timing_acme slurm balwinder.singh -at- pnnl.gov 8 @@ -1523,7 +1716,7 @@ intel - + PNL Haswell cluster, OS is Linux, batch system is SLURM LINUX @@ -1539,7 +1732,6 @@ UNSET /pic/projects/climate/acme_baselines /pic/projects/climate/acme_baselines/cprnc/cprnc - /pic/projects/climate/csmdata/ slurm balwinder.singh -at- pnnl.gov 8 @@ -1795,7 +1987,7 @@ cli115 -D PIO_BUILD_TIMING:BOOL=ON - aprun + aprun + + $COMPILER + $MPILIB + 1 + 1 + 1 + 128M + 128M + - - - $COMPILER - $MPILIB - 1 - 1 - 1 - 128M - 128M - - - - - - istanbul - 1 - /opt/cray/netcdf-hdf5parallel/4.4.1.1/PGI/15.3/ - - - - /opt/cray/parallel-netcdf/1.7.0/PGI/15.3 - + + + + istanbul + 1 +/opt/cray/netcdf-hdf5parallel/4.4.1.1/PGI/15.3/ + - - - istanbul - /opt/cray/netcdf-hdf5parallel/4.4.1.1/PGI/15.3/ - + +/opt/cray/parallel-netcdf/1.7.0/PGI/15.3 + - - /opt/cray/parallel-netcdf/1.7.0/PGI/15.3 - + + + istanbul +/opt/cray/netcdf-hdf5parallel/4.4.1.1/PGI/15.3/ + - - dynamic - /opt/cray/netcdf-hdf5parallel/4.4.1.1/INTEL/15.0/ - + +/opt/cray/parallel-netcdf/1.7.0/PGI/15.3 + - - /opt/cray/parallel-netcdf/1.7.0/INTEL/15.0 - + + dynamic +/opt/cray/netcdf-hdf5parallel/4.4.1.1/INTEL/15.0/ + - - /opt/cray/netcdf-hdf5parallel/4.4.1.1/CRAY/8.3/ - + +/opt/cray/parallel-netcdf/1.7.0/INTEL/15.0 + - - /opt/cray/parallel-netcdf/1.7.0/CRAY/8.3 - + +/opt/cray/netcdf-hdf5parallel/4.4.1.1/CRAY/8.3/ + - + +/opt/cray/parallel-netcdf/1.7.0/CRAY/8.3 + @@ -2047,7 +2238,7 @@ - LANL Linux Cluster, 36 pes/node, batch system Moab + LANL Linux Cluster, 36 pes/node, batch system slurm gr-fe.*.lanl.gov acme_developer gnu,intel @@ -2061,7 +2252,6 @@ UNSET /lustre/scratch3/turquoise/$ENV{USER}/ACME/input_data/ccsm_baselines /lustre/scratch3/turquoise/$ENV{USER}/ACME/scratch - $CASEROOT/timings /turquoise/usr/projects/climate/SHARED_CLIMATE/software/wolf/cprnc/v0.40/cprnc /usr/share/Modules/init/perl.pm @@ -2101,7 +2291,7 @@ romio_ds_write=disable;romio_ds_read=disable;romio_cb_write=enable;romio_cb_read=enable - moab + slurm mpirun @@ -2132,7 +2322,7 @@ - LANL Linux Cluster, 16 pes/node, batch system Moab + LANL Linux Cluster, 16 pes/node, batch system slurm wf-fe.*.lanl.gov acme_developer intel,gnu @@ -2143,7 +2333,7 @@ /lustre/scratch3/turquoise/$ENV{USER}/ACME/archive/$CASE UNSET LINUX - moab + slurm jonbob -at- lanl.gov 4 16 @@ -2154,7 +2344,6 @@ /lustre/scratch3/turquoise/$ENV{USER}/ACME/input_data/atm/datm7 /lustre/scratch3/turquoise/$ENV{USER}/ACME/input_data/ccsm_baselines /turquoise/usr/projects/climate/SHARED_CLIMATE/software/wolf/cprnc/v0.40/cprnc - $CASEROOT/timings mpirun @@ -2414,7 +2603,6 @@ 16 16 TRUE - /global/scratch/$ENV{USER} mpirun @@ -2478,7 +2666,6 @@ 12 12 TRUE - /global/scratch/$ENV{USER} mpirun @@ -2553,6 +2740,111 @@ + + ORNL pre-Summit testbed. Node: 2x POWER8 + 4x Tesla P100, 20 cores/node, 8 HW threads/core. + summitdev-* + Received node event ec_node + acme_developer + ibm,pgi + openmpi,mpi_serial + $ENV{HOME}/acme_scratch/$PROJECT + /lustre/atlas/scratch/$ENV{USER}/$PROJECT/$CASE/run + $CIME_OUTPUT_ROOT/$CASE/bld + /lustre/atlas1/cli900/world-shared/cesm/inputdata + /lustre/atlas1/cli900/world-shared/cesm/inputdata/atm/datm7 + /lustre/atlas/scratch/$ENV{USER}/$PROJECT/archive/$CASE + csm/$CASE + /lustre/atlas1/cli900/world-shared/cesm/baselines + /lustre/atlas1/cli900/world-shared/cesm/tools/cprnc/cprnc + /lustre/atlas/proj-shared/$PROJECT + LINUX + lsf + acme + 32 + 20 + 160 + TRUE + csc249 + -D PIO_BUILD_TIMING:BOOL=ON + + + mpirun + + -np $TOTALPES + --map-by core:PE=$ENV{OMP_NUM_THREADS} --bind-to core + -x OMP_NUM_THREADS=$ENV{OMP_NUM_THREADS} + --report-bindings + --display-map + + + + + + + /sw/summitdev/lmod/7.4.0/rhel7.2_gnu4.8.5/lmod/7.4/init/sh + /sw/summitdev/lmod/7.4.0/rhel7.2_gnu4.8.5/lmod/7.4/init/csh + /sw/summitdev/lmod/7.4.0/rhel7.2_gnu4.8.5/lmod/7.4/init/env_modules_python.py + /sw/summitdev/lmod/7.4.0/rhel7.2_gnu4.8.5/lmod/7.4/init/perl + + module + module + module + module + + + + + + + DefApps + python/3.5.2 + subversion/1.9.3 + git/2.13.0 + cmake/3.6.1 + essl/5.5.0-20161110 + netlib-lapack/3.6.1 + + + + xl + pgi/17.4 + + + + pgi + xl/20161123 + + + + + + netcdf/4.4.1 + netcdf-fortran/4.4.4 + + + spectrum_mpi/10.1.0.2-20161221 + netcdf/4.4.1 + netcdf-fortran/4.4.4 + parallel-netcdf/1.7.0 + hdf5/1.10.0-patch1-parallel + + + + $COMPILER + $MPILIB + 128M + $ENV{OLCF_NETCDF_ROOT} + $ENV{OLCF_NETCDF_FORTRAN_ROOT} + $ENV{OLCF_HDF5_ROOT} + $ENV{OLCF_ESSL_ROOT} + + + $ENV{OLCF_PARALLEL_NETCDF_ROOT} + + + ${EXEROOT}/acme.exe diff --git a/config/xml_schemas/env_mach_specific.xsd b/config/xml_schemas/env_mach_specific.xsd index 45a4ddfc6564..a8a35b581813 100644 --- a/config/xml_schemas/env_mach_specific.xsd +++ b/config/xml_schemas/env_mach_specific.xsd @@ -7,6 +7,7 @@ + @@ -80,6 +81,7 @@ +