cn:csn4:calcolo:suma:galileo_howto
Differences
This shows you the differences between two versions of the page.
| Both sides previous revisionPrevious revisionNext revision | Previous revision | ||
| cn:csn4:calcolo:suma:galileo_howto [2015/02/20 15:38] – [Batch scheduler] roberto.alfieri@infn.it | cn:csn4:calcolo:suma:galileo_howto [2015/02/24 16:58] (current) – [GALILEO login] roberto.alfieri@infn.it | ||
|---|---|---|---|
| Line 1: | Line 1: | ||
| + | ====== GALILEO ====== | ||
| + | |||
| + | === NEWS === | ||
| + | |||
| + | 2015/ | ||
| + | |||
| + | 2015/01/28 Galileo is open to the INFN users for the " | ||
| + | The MIC accelerators are not available yet. | ||
| + | |||
| + | [[http:// | ||
| + | |||
| + | ===== GALILEO login ===== | ||
| + | |||
| + | * In order to become a CINECA user you have to register yourself on the CINECA UserDB ( https:// | ||
| + | |||
| + | * Each user must be associated to the Account raleted to the " | ||
| + | |||
| + | * At the end of the previous step you can access the Galileo front-end | ||
| + | |||
| + | The following command displays | ||
| + | |||
| + | login> saldo -b | ||
| + | | ||
| + | ===== GALILEO usage ===== | ||
| + | |||
| + | |||
| + | [[http:// | ||
| + | - | ||
| + | [[http:// | ||
| + | |||
| + | |||
| + | |||
| + | ==== Architecture ==== | ||
| + | |||
| + | [[ http:// | ||
| + | |||
| + | < | ||
| + | Model: IBM NeXtScale | ||
| + | Nodes: 516 | ||
| + | Processors: 8-cores Intel Haswell 2.40 GHz (2 per node) | ||
| + | Cores: 16 cores/node, 8256 cores in total | ||
| + | Accelerators: | ||
| + | RAM: 128 GB/node, 8 GB/core | ||
| + | Internal Network: Infiniband with 4x QDR switches | ||
| + | Disk Space:2,500 TB of local storage | ||
| + | Peak Performance: | ||
| + | </ | ||
| + | |||
| + | To get on-line details: | ||
| + | |||
| + | login> pbsnodes -a | egrep ' | ||
| + | | ||
| + | ==== Batch scheduler ==== | ||
| + | |||
| + | THe job management facility adopted by CINECA is PBS: | ||
| + | [[http:// | ||
| + | |||
| + | Routing Queue " | ||
| + | You have only to declare how many resources you need and your job will be directed into the right queue with a right priority. | ||
| + | Normal parallel jobs will be routed to the " | ||
| + | |||
| + | |||
| + | Script example (script.pbs) | ||
| + | < | ||
| + | #!/bin/bash | ||
| + | #PBS -N prova | ||
| + | #PBS -l walltime=02: | ||
| + | #PBS -l select=16: | ||
| + | #PBS -A INFNG_test | ||
| + | # | ||
| + | module load intel/ | ||
| + | module load intelmpi/ | ||
| + | cd working_dir | ||
| + | mpirun executable | ||
| + | </ | ||
| + | |||
| + | Submit your job | ||
| + | qsub script.pbs | ||
| + | Monitor your job | ||
| + | qstat [-u username] | ||
| + | Cancel your job | ||
| + | qdel JOB.id | ||
| + | |||
| + | |||
| + | Interactive example (option -I): | ||
| + | |||
| + | qsub -l select=1: | ||
| + | > cat $PBS_NODEFILE | ||
| + | > exit | ||
| + | |||
| + | == Default values assigned by the queue manager == | ||
| + | |||
| + | * 1 CPU | ||
| + | * 8GB of memory (each node has 128 GB ram) | ||
| + | * Max Walltime: 30 minutes | ||
| + | * MICs : 0 | ||
| + | * MPI processes : 1 per node | ||
| + | * cores allocation: Pack (try to pack requested CPU on smallest number of nodes) | ||
| + | |||
| + | The default walltime is 30 minutes. | ||
| + | |||
| + | == More complex requests == | ||
| + | |||
| + | qsub -A INFNG_test -I -l ncpus=16, | ||
| + | qsub -A INFNG_test -I -l select=2: | ||
| + | qsub -A INFNG_test -I -l select=16: | ||
| + | qsub -A INFNG_test -I -l select=16: | ||
| + | qsub -A INFNG_test -I -l select=2: | ||
| + | qsub -A INFNG_test -I -l select=2: | ||
| + | qsub -A INFNG_test -I -l select=2: | ||
| + | |||
| + | ==== Storage ==== | ||
| + | |||
| + | CINECA documentation: | ||
| + | |||
| + | |||
| + | $HOME (/ | ||
| + | $CINECA_SCRATCH | ||
| + | $WORK ( / | ||
| + | Use the local command " | ||
| + | |||
| + | cindata | ||
| + | |||
| + | ==== Software Environment ==== | ||
| + | |||
| + | * OS: RedHat CentOS release 7, 64 bit | ||
| + | * Compilers, scientific libraries and tools are installed using the **software modules** mechanism. | ||
| + | |||
| + | CINECA Documentation: | ||
| + | - | ||
| + | [[ http:// | ||
| + | - | ||
| + | [[http:// | ||
| + | |||
| + | |||
| + | ===== MIC job submission (Work in progress) | ||
| + | |||
| + | [[http:// | ||
| + | - | ||
| + | [[http:// | ||
| + | |||
| + | ==Compilation== | ||
| + | |||
| + | * login on one mic-node using command | ||
| + | |||
| + | qsub -A INFNG_test -I -l select=1: | ||
| + | |||
| + | * load needed modules and set variables | ||
| + | |||
| + | | ||
| + | | ||
| + | | ||
| + | |||
| + | * compile | ||
| + | * exit | ||
| + | |||
| + | ==Execution on mic-node == | ||
| + | |||
| + | qsub -A INFNG_test -I -l select=1: | ||
| + | | ||
| + | | ||
| + | | ||
| + | | ||
| + | |||
| + | == Execution using PBS from front-end == | ||
| + | |||
| + | Example of PBS file | ||
| + | |||
| + | < | ||
| + | #!/bin/bash | ||
| + | #PBS -l select=1: | ||
| + | #PBS -l walltime=00: | ||
| + | #PBS -A INFNG_test | ||
| + | |||
| + | # load required modules | ||
| + | module load intel intelmpi mkl | ||
| + | source $INTEL_HOME/ | ||
| + | export I_MPI_MIC=enable | ||
| + | export MIC0=$(head -n 1 $PBS_NODEFILE | sed " | ||
| + | export MIC1=$(head -n 1 $PBS_NODEFILE | sed " | ||
| + | cd < | ||
| + | |||
| + | export MIC_PATH= | ||
| + | export MIC_PATH=$MIC_PATH:/ | ||
| + | export MIC_PATH=$MIC_PATH:/ | ||
| + | |||
| + | mpirun -genv LD_LIBRARY_PATH $MIC_PATH -host ${MIC0}, | ||
| + | </ | ||
| + | |||
| + | |||
| + | ---- | ||
| + | |||
| + | // 2015/ | ||
