Getting started with Niagara Cluster Access for ISC22 SCC

To get access you will need a username and Initial password that will be sent to you. 

Getting Ready

Quickstart for Niagara and Mist documentation page can be found here:

 

$ ssh -Y lclmaoroph@niagara.scinet.utoronto.ca Warning: Permanently added 'niagara.scinet.utoronto.ca' (RSA) to the list of known hosts. Password: =============================================================================== SciNet welcomes you to the NIAGARA supercomputer. This is a Niagara login node. Use this node to develop and compile code, to run short tests, and to submit computations to the scheduler. Remember that /scratch is never backed-up. Documentation: https://docs.scinet.utoronto.ca/index.php/Niagara_Quickstart Support: support@scinet.utoronto.ca or niagara@computecanada.ca =============================================================================== lclmaoroph@nia-login06:~$

 

Once you get access, you should change your password using this command on Niagara login node:

ldappasswd  -H ldaps://ldapk9.scinet.utoronto.ca -D "uid=lclmaoroph,ou=People,dc=computecanada,dc=ca" -W -S

 

After that you MUST set up your SSH keys for login to Niagara, and two factor verification setup. The proecure is very simple, just follow the following docs:

Compute resource

Use slurm to allocate and run jobs on the compute partition.

The cluster has Intel Skylake CPUs and InfiniBand EDR networking:

$ salloc -N 1 -p compute -t 1:00:00 SALLOC WARNING: Number of tasks not specified to salloc, so the number of tasks per node is set to the default of 80 tasks per 40-core node, giving a total of 80 tasks. If this is not what you intended, specify --ntasks or --ntasks-per-node. SALLOC: 1 warning was found. SALLOC: Job submission will be attempted despite warnings. SALLOC: Pending job allocation 4753951 SALLOC: job 4753951 queued and waiting for resources SALLOC: job 4753951 has been allocated resources SALLOC: Granted job allocation 4753951 SALLOC: Waiting for resource configuration SALLOC: Nodes nia0040 are ready for job [lclmaoroph@nia0040 ~]$ lscpu Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Byte Order: Little Endian CPU(s): 80 On-line CPU(s) list: 0-79 Thread(s) per core: 2 Core(s) per socket: 20 Socket(s): 2 NUMA node(s): 2 Vendor ID: GenuineIntel CPU family: 6 Model: 85 Model name: Intel(R) Xeon(R) Gold 6148 CPU @ 2.40GHz Stepping: 4 CPU MHz: 3099.169 CPU max MHz: 3700.0000 CPU min MHz: 1000.0000 BogoMIPS: 4800.00 Virtualization: VT-x L1d cache: 32K L1i cache: 32K L2 cache: 1024K L3 cache: 28160K NUMA node0 CPU(s): 0-19,40-59 NUMA node1 CPU(s): 20-39,60-79 Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc art arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc aperfmperf eagerfpu pni pclmulqdq dtes64 ds_cpl vmx smx est tm2 ssse3 sdbg fma cx16 xtpr pdcm pcid dca sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm 3dnowprefetch epb cat_l3 cdp_l3 intel_ppin intel_pt ssbd mba ibrs ibpb stibp tpr_shadow vnmi flexpriority ept vpid fsgsbase tsc_adjust bmi1 hle avx2 smep bmi2 erms invpcid rtm cqm mpx rdt_a avx512f avx512dq rdseed adx smap clflushopt clwb avx512cd avx512bw avx512vl xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local dtherm ida arat pln pts hwp hwp_act_window hwp_epp hwp_pkg_req pku ospke md_clear spec_ctrl intel_stibp flush_l1d [lclmaoroph@nia0040 ~]$ lspci | grep Mel 06:00.0 Infiniband controller: Mellanox Technologies MT27800 Family [ConnectX-5] [lclmaoroph@nia0040 ~]$ ibstat CA 'mlx5_0' CA type: MT4119 Number of ports: 1 Firmware version: 16.25.6000 Hardware version: 0 Node GUID: 0xec0d9a0300849332 System image GUID: 0xec0d9a0300849332 Port 1: State: Active Physical state: LinkUp Rate: 100 Base lid: 510 LMC: 0 SM lid: 1 Capability mask: 0x2651e848 Port GUID: 0xec0d9a0300849332 Link layer: InfiniBand [lclmaoroph@nia0040 ~]$

Scratch Area

Scratch area is available under $SCRATCH