Niagara Cluster Access for ISC21 SCC
To get access you will need a username and Initial password that will be sent to you.Â
Getting Ready
Quickstart for Niagara and Mist documentation page can be found here:
Â
$ ssh -Y lclmaoroph@niagara.scinet.utoronto.ca
Warning: Permanently added 'niagara.scinet.utoronto.ca' (RSA) to the list of known hosts.
Password:
===============================================================================
SciNet welcomes you to the NIAGARA supercomputer.
This is a Niagara login node. Use this node to develop and compile code,
to run short tests, and to submit computations to the scheduler.
Remember that /scratch is never backed-up.
Documentation: https://docs.scinet.utoronto.ca/index.php/Niagara_Quickstart
Support: support@scinet.utoronto.ca or niagara@computecanada.ca
===============================================================================
lclmaoroph@nia-login06:~$
Â
Once you get access, you should change your password using this command on Niagara login node:
ldappasswd -H ldaps://ldapk9.scinet.utoronto.ca -D "uid=lclmaoroph,ou=People,dc=computecanada,dc=ca" -W -S
Â
After that you MUST set up your SSH keys for login to Niagara, and two factor verification setup. The proecure is very simple, just follow the following docs:
Compute resource
Use slurm to allocate and run jobs on the compute partition.
The cluster has Intel Skylake CPUs and InfiniBand EDR networking:
$ salloc -N 1 -p compute -t 1:00:00
SALLOC WARNING:
Number of tasks not specified to salloc, so the number of tasks per node is
set to the default of 80 tasks per 40-core node, giving a total of 80 tasks.
If this is not what you intended, specify --ntasks or --ntasks-per-node.
SALLOC: 1 warning was found.
SALLOC: Job submission will be attempted despite warnings.
SALLOC: Pending job allocation 4753951
SALLOC: job 4753951 queued and waiting for resources
SALLOC: job 4753951 has been allocated resources
SALLOC: Granted job allocation 4753951
SALLOC: Waiting for resource configuration
SALLOC: Nodes nia0040 are ready for job
[lclmaoroph@nia0040 ~]$ lscpu
Architecture: x86_64
CPU op-mode(s): 32-bit, 64-bit
Byte Order: Little Endian
CPU(s): 80
On-line CPU(s) list: 0-79
Thread(s) per core: 2
Core(s) per socket: 20
Socket(s): 2
NUMA node(s): 2
Vendor ID: GenuineIntel
CPU family: 6
Model: 85
Model name: Intel(R) Xeon(R) Gold 6148 CPU @ 2.40GHz
Stepping: 4
CPU MHz: 3099.169
CPU max MHz: 3700.0000
CPU min MHz: 1000.0000
BogoMIPS: 4800.00
Virtualization: VT-x
L1d cache: 32K
L1i cache: 32K
L2 cache: 1024K
L3 cache: 28160K
NUMA node0 CPU(s): 0-19,40-59
NUMA node1 CPU(s): 20-39,60-79
Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc art arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc aperfmperf eagerfpu pni pclmulqdq dtes64 ds_cpl vmx smx est tm2 ssse3 sdbg fma cx16 xtpr pdcm pcid dca sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm 3dnowprefetch epb cat_l3 cdp_l3 intel_ppin intel_pt ssbd mba ibrs ibpb stibp tpr_shadow vnmi flexpriority ept vpid fsgsbase tsc_adjust bmi1 hle avx2 smep bmi2 erms invpcid rtm cqm mpx rdt_a avx512f avx512dq rdseed adx smap clflushopt clwb avx512cd avx512bw avx512vl xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local dtherm ida arat pln pts hwp hwp_act_window hwp_epp hwp_pkg_req pku ospke md_clear spec_ctrl intel_stibp flush_l1d
[lclmaoroph@nia0040 ~]$ lspci | grep Mel
06:00.0 Infiniband controller: Mellanox Technologies MT27800 Family [ConnectX-5]
[lclmaoroph@nia0040 ~]$ ibstat
CA 'mlx5_0'
CA type: MT4119
Number of ports: 1
Firmware version: 16.25.6000
Hardware version: 0
Node GUID: 0xec0d9a0300849332
System image GUID: 0xec0d9a0300849332
Port 1:
State: Active
Physical state: LinkUp
Rate: 100
Base lid: 510
LMC: 0
SM lid: 1
Capability mask: 0x2651e848
Port GUID: 0xec0d9a0300849332
Link layer: InfiniBand
[lclmaoroph@nia0040 ~]$
Scratch Area
Scratch area is available under $SCRATCH
Â
Â