#!/bin/bash ######################################################## # This file loads the required modules and sets the paths for simulations # To modify the paths: # 1) In the function, find the section corresponding to your machine # 2) Modify the variables, which are documented in the *) section of # See the end of for the explanation of each ######################################################## function get_hostname { if grep -q "Adastra" /etc/motd; then hostname="adastra" elif which idrquota &> /dev/null; then hostname="jean-zay" else hostname=$(hostname) fi } function set_env { # Platform-specific case ${hostname:0:5} in jean-) module purge compilo=19.0.4 # available 2013.0, 2017.2 module load intel-compilers/$compilo #module load intel-mpi/$compilo module load intel-mkl/$compilo module load hdf5/1.10.5-mpi module load netcdf/4.7.2-mpi module load netcdf-fortran/4.5.2-mpi module load subversion/1.9.7 #Pour module gcc, voir : https://trac.lmd.jussieu.fr/LMDZ/wiki/PortageJeanZay #module load gcc/6.5.0 module load nco module load cdo # Imputation de la consommation sur le groupe (projet) actif par defaut, # idrproj indique le groupe (projet) actif par defaut # idrproj -d newproj redefinit "newproj" en tant que projet actif, # alors $STORE, $WORK etc vont designer les espaces de "newproj") account="lmd" # $(idrproj | grep active | awk '{ print $1}') doesn't work on compute nodes root_dir="$WORK/LMDZ_Setup"; ARCH="X64_JEANZAY" SIMRUNBASEDIR="$SCRATCH/" LMDZD="$root_dir/LMDZD" LMDZ_INIT="$root_dir/LMDZ_Init" NB_MPI_MAX=2000 NB_OMP_MAX=20 MPICMD="srun -n" RUNBASHCMD="srun -A $account@cpu --label -n 1 -c" SUBMITCMD="sbatch -A $account@cpu" ;; spiri) module purge module load subversion/1.13.0 module load gcc/11.2.0 module load openmpi/4.0.7 module load cdo/2.3.0 root_dir="/data/abarral/LMDZ_SETUP_ROOT/LMDZ_Setup"; ARCH="X64_MESOIPSL-GNU" SIMRUNBASEDIR="$SCRATCH/" LMDZD="$root_dir/LMDZD" LMDZ_INIT="$root_dir/LMDZ_Init" NB_MPI_MAX=5 NB_OMP_MAX=1 NB_CORE_PER_NODE_MAX=0 N_HYPERTHREADING=1 MPICMD="mpirun -np" # on spirit, we can't run MPI using srun from within sbatch RUNBASHCMD="bash" SUBMITCMD="sbatch" ;; adast) module purge module load PrgEnv-gnu module load gcc/13.2.0 # required, see https://dci.dci-gitlab.cines.fr/webextranet/user_support/index.html#prgenv-and-compilers function cdo { # cdo is available as a spack cmd which requires a specific, incompatible env unset cdo module purge module load GCC-CPU-3.1.0 /opt/software/gaia-external/CPU/cdo-2.4.0/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spa/cdo-2.4.0-gcc-12.1-b3qr/bin/cdo "$@" set_env } function ferret { unset ferret module purge module load GCC-CPU-3.1.0 module load ferret ferret "$@" set_env } account=$(/usr/sbin/my_project.py -l 2>&1 | head -1 | cut -d " " -f 3- | cut -c 5-) root_dir="$WORKDIR/LMDZ_Setup"; ARCH="X64_ADASTRA-GNU" SIMRUNBASEDIR="$SCRATCHDIR/" LMDZD="$root_dir/LMDZD" LMDZ_INIT="$root_dir/LMDZ_Init" NB_MPI_MAX=2000 NB_OMP_MAX=200 NB_CORE_PER_NODE_MAX=96 N_HYPERTHREADING=2 MPICMD="srun -n" RUNBASHCMD="srun --label --account=$account --constraint=GENOA --ntasks-per-node=1 -n 1 --time=00:10:00 -c" SUBMITCMD="sbatch --constraint=GENOA --account=$account" ;; *) echo "WARNING: RUNNING THIS SCRIPT ON A LOCAL COMPUTER IS DISCOURAGED (lackluster cpus and memory)" root_dir="/home/abarral/PycharmProjects/installLMDZ/LMDZ_Setup"; # Where you have extracted LMDZ_Setup. Can't use $(pwd) since this script gets copied and ran from several locations ARCH="local-gfortran-parallel" # The arch file to use SIMRUNBASEDIR="/tmp/SCRATCH/" # Where the simulations will be executed ($SIMRUNBASEDIR/LMDZ_Setup/...) LMDZD="$root_dir/LMDZD" # Where the sources will be downloaded and compiled LMDZ_INIT="$root_dir/LMDZ_Init" # Where to store shared files used for initialisation NB_MPI_MAX=33 # Max number of MPI cores (only for running simulations) NB_OMP_MAX=100 # Max number of OMP threads (only for running simulations) NB_CORE_PER_NODE_MAX=0 # Max number of cores per node (real cores, not hyperthreading - only for running simulations, cluster-specific) N_HYPERTHREADING=1 # How many hyperthreading threads per physical core MPICMD="mpirun -np" # command to run an mpi executable, as $MPICMD