ISC23 SCC Getting Started with Bridges-2 Cluster

 

Follow the following procedures to setup an account for PSC Bridges-2

 

Setting up the account

 

Start by creating one account per team on the Access portal:

https://identity.access-ci.org/new-user - Two way authentication is needed via DUO mobile app

Login here: https://cilogon.org/

 

Later on, we will need your team username to be able to apply the allocation to the team.

 

Check your username here: https://allocations.access-ci.org/profile

You will see the “Pittsburgh Supercomputing Center” username.

 

To reset the password, see here: https://apr.psc.edu/enter_info

And follow the steps.

 

Login to Bridges2

 

Login using your psc username and password:

bash-4.1$ ssh maor@bridges2.psc.edu maor@bridges2.psc.edu's password: Warning: No xauth data; using fake authentication data for X11 forwarding. ********************************* W A R N I N G ******************************** You have connected to br014.ib.bridges2.psc.edu, a login node of Bridges 2. This computing resource is the property of the Pittsburgh Supercomputing Center. It is for authorized use only. By using this system, all users acknowledge notice of, and agree to comply with, PSC polices including the Resource Use Policy, available at http://www.psc.edu/index.php/policies. Unauthorized or improper use of this system may result in administrative disciplinary action, civil charges/criminal penalties, and/or other sanctions as set forth in PSC policies. By continuing to use this system you indicate your awareness of and consent to these terms and conditions of use. LOG OFF IMMEDIATELY if you do not agree to the conditions stated in this warning ********************************* W A R N I N G ******************************** For documentation on Bridges 2, please see www.psc.edu/resources/bridges-2/user-guide/ Please contact help@psc.edu with any comments/concerns. Last login: Mon Nov 15 19:58:05 2021 from 149.165.168.51 Projects --------------------------------------------------------------------------------------------------------------------------------------- Project: bio220064p [Default charging account] PI: Ophir Maor Regular Memory 999,979 SU remain of 1,000,000 SU Active: Yes Ocean /ocean/projects/bio220064p 16k used of 4.883T

 

 

Use slurm to allocate a node

[maor@bridges2-login014 ~]$ sinfo PARTITION AVAIL TIMELIMIT NODES STATE NODELIST RM* up infinite 1 down* r059 RM* up infinite 1 drng r384 RM* up infinite 1 drain r191 RM* up infinite 2 resv r[099,174] RM* up infinite 34 mix r[053-055,057-058,063,068,071,073-075,077,080-081,084-085,090,092,094,109,126-127,129,135,137,143,151,157,161,172-173,175,190,282] RM* up infinite 393 alloc r[050-052,056,060-062,064-067,069-070,072,076,078-079,082-083,086-089,091,093,095-098,100-108,110-125,128,130-134,136,138-142,144-150,152-156,158-160,162-171,176-189,192-246,248-251,253-272,274-281,283-360,362-376,378-383,386-446,448-488] RM* up infinite 2 idle r[247,377] RM* up infinite 5 down r[252,273,361,385,447] RM-512 up infinite 16 alloc l[001-016] RM-shared up infinite 1 down* r059 RM-shared up infinite 2 resv r[099,174] RM-shared up infinite 70 mix r[005-008,010-013,018,020-025,027-029,031-037,039-049,053-055,057-058,063,068,071,073-075,077,080-081,084-085,090,092,094,109,126-127,129,135,137,143,151,157,161,172-173,175,190,282] RM-shared up infinite 158 alloc r[009,014-017,019,026,030,038,050-052,056,060-062,064-067,069-070,072,076,078-079,082-083,086-089,091,093,095-098,100-108,110-125,128,130-134,136,138-142,144-150,152-156,158-160,162-171,176,180-181,192-196,199-204,208,223,227-229,245-246,250,257-258,260,263-264,285,289-291,293,295,301,310,319,323-325,328-330,332-333,336-337,342,344-345,347,353-354,356-360,367] RM-small up infinite 2 mix r[001-002] RM-small up infinite 2 idle r[003-004] GPU up infinite 18 mix v[003-006,008-012,015-016,018,025,027-030,033] GPU up infinite 13 alloc v[007,013-014,017,019-024,026,031-032] GPU up infinite 1 idle v034 GPU-shared up infinite 18 mix v[003-006,008-012,015-016,018,025,027-030,033] GPU-shared up infinite 13 alloc v[007,013-014,017,019-024,026,031-032] GPU-small up infinite 1 alloc v002 GPU-small up infinite 1 idle v001 EM up infinite 1 mix e001 EM up infinite 3 alloc e[002-004] BatComputer up infinite 1 mix dv001 BatComputer up infinite 3 alloc dv[002-004]

 

7. Allocate a node.

 

Note: For ISC23 use only RM partition and 4 node allocation.

Here is an example to allocate single node.

[maor@bridges2-login014 ~]$ salloc -N 1 -p RM salloc: Pending job allocation 5062923 salloc: job 5062923 queued and waiting for resources salloc: job 5062923 has been allocated resources salloc: Granted job allocation 5062923 salloc: Waiting for resource configuration salloc: Nodes r352 are ready for job [maor@r352 ~]$ ibstat CA 'mlx5_0' CA type: MT4123 Number of ports: 1 Firmware version: 20.30.1004 Hardware version: 0 Node GUID: 0x9440c9ffffac407c System image GUID: 0x9440c9ffffac407c Port 1: State: Active Physical state: LinkUp Rate: 200 Base lid: 636 LMC: 0 SM lid: 35 Capability mask: 0x2651e848 Port GUID: 0x9440c9ffffac407c Link layer: InfiniBand [maor@r352 ~]$ lscpu Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Byte Order: Little Endian CPU(s): 128 On-line CPU(s) list: 0-127 Thread(s) per core: 1 Core(s) per socket: 64 Socket(s): 2 NUMA node(s): 2 Vendor ID: AuthenticAMD CPU family: 23 Model: 49 Model name: AMD EPYC 7742 64-Core Processor Stepping: 0 CPU MHz: 3335.192 BogoMIPS: 4491.35 Virtualization: AMD-V L1d cache: 32K L1i cache: 32K L2 cache: 512K L3 cache: 16384K NUMA node0 CPU(s): 0-63 NUMA node1 CPU(s): 64-127 Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc rep_good nopl nonstop_tsc cpuid extd_apicid aperfmperf pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_llc mwaitx cpb cat_l3 cdp_l3 hw_pstate ssbd mba ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 xsaves cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr wbnoinvd arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip rdpid overflow_recov succor smca [maor@r352 ~]$

Installing and using HPC-X

 

  1. Download the above file.

  2. tar xjfp hpcx-2.13.1.tbz

  3. module use <path to hpcx-2.13.1>/modulefiles

  4. module load hpcx

  5. which mpirun