GPAW for ISC21
- 1 References
- 2 GPAW Overview
- 3 Description
- 4 Requirements
- 5 Downloading and installing requirements
- 6 Downloading and installing GPAW
- 7 Testing the installation
- 8 Running GPAW
- 9 Sample output
- 10 GPAW Example on Niagara cluster
- 10.1 Download and Install
- 10.2 How to use HPC-X module
- 10.3 Install dataset
- 10.4 Running GPAW
- 10.5 Sample output
- 11 Tasks and Submissions
- 11.1 Input Files
- 11.2 Building and running GPAW
- 11.3 Visualization
- 11.4 Profiling
- 11.5 Performance tuning
- 11.6 Bonus task: bug fix
References
GPAW Overview
Slides:
Description
GPAW is a density-functional theory (DFT) code based on the projector-augmented wave (PAW) method and various basis sets. The wave functions can be described with:
Uniform real-space grids
Plane waves
Localized atomic orbitals
GPAW is implemented in Python and C programming languages and it relies on high performance libraries for linear algebra operations, FFTs, etc.
Parallelization is primarily with MPI, complementary OpenMP parallelization can improve the performance in some cases.
Requirements
Python 3.6 or later
NumPy 1.9 or later (base N-dimensional array package)
SciPy 0.14 or later (library for scientific computing)
ASE 3.18.0 or later (atomic simulation environment)
a C-compiler
LibXC 3.x or 4.x
BLAS library
an MPI library (required for parallel calculations)
Optional, but highly recommended:
Downloading and installing requirements
The required Python packages (NumPy, SciPy, ASE) will be downloaded and installed (if not present in the system) automatically by the GPAW’s installation process, and do not normally need to be separately installed. Typically, LibXC is the only dependency that needs to be separately installed (for BLAS, MPI, etc. one should rely on high performance system libraries). Installing LibXC from source (also package management tools yum
, apt
etc. can be used when appropriate):
libxc_version=4.3.4
wget --content-disposition http://www.tddft.org/programs/libxc/down.php?file=$libxc_version/libxc-$libxc_version.tar.gz
tar xzf libxc-$libxc_version.tar.gz
cd libxc-$libxc_version
export CFLAGS="-O3 -fPIC" # -fPIC is needed as libxc will be used from shared library
./configure --prefix=$HOME/libxc/$libxc_version
make
make install
export CPATH=$CPATH:$HOME/libxc/$libxc_version/include
export LIBRARY_PATH=$LIBRARY_PATH:$HOME/libxc/$libxc_version/lib
Downloading and installing GPAW
There are several ways to install GPAW, refer to https://wiki.fysik.dtu.dk/gpaw/install.html for all options. Here, we recommend installing GPAW from source, version 21.1.0 will be used for the competition.
//download from gitlab
git clone -b 21.1.0 https://gitlab.com/gpaw/gpaw.git
cd gpaw
By default GPAW looks for BLAS library libblas
in the default locations, and if it is available installation can be done at simplest with:
export PYTHONUSERBASE=some_installation_root
pip3 install --user --verbose .
For some testing one should install also the optional pytest
package
By default, the same compiler and compiler options as when building the Python interpreter are used. In HPC systems it is, however, often recommended to customize the installation in order to use other libraries, compilers or compiler options. This can accomplished via siteconfig.py
file, sample file is provided in siteconfig_example.py
. As an example, in order to use icc compiler with extra optimization flags and MKL BLAS library one should add into siteconfig.py
:
One should also set the GPAW_CONFIG corresponding to this siteconfig.py
file:
and proceed then with pip3 install …
. (Note: siteconfig.py
can be also named differently when playing with different installation options, i.e. siteconfig_intel.py
, siteconfig_gcc.py
… )
More information about customizing the installation can be found in GPAW wiki
Once the installation is complete, the PATH variable needs to be set for GPAW bins:
Finally, one needs to install the PAW datasets
Testing the installation
Once installation is complete and PYTHONUSERBASE and PATH are set, one can check the installation information as:
Next, one should make short serial test calculation
followed by short parallel calculation
GPAW contains also an extensive test set which should be run when developing the code, more details in GPAW wiki
Running GPAW
Sample output
In short benchmarks we will be looking the time for SCF-cycle
(neglecting the initialization of the calculation).
GPAW Example on Niagara cluster
Download and Install
There are several ways to install GPAW, refer to https://wiki.fysik.dtu.dk/gpaw/install.html for all options.
Here is an basic example with CentOS 7/8:
- 1 References
- 2 GPAW Overview
- 3 Description
- 4 Requirements
- 5 Downloading and installing requirements
- 6 Downloading and installing GPAW
- 7 Testing the installation
- 8 Running GPAW
- 9 Sample output
- 10 GPAW Example on Niagara cluster
- 10.1 Download and Install
- 10.2 How to use HPC-X module
- 10.3 Install dataset
- 10.4 Running GPAW
- 10.5 Sample output
- 11 Tasks and Submissions
- 11.1 Input Files
- 11.2 Building and running GPAW
- 11.3 Visualization
- 11.4 Profiling
- 11.5 Performance tuning
- 11.6 Bonus task: bug fix
Note: In case you get the error during installation, export the below variables and rebuild gpaw.
Building GPAW with Intel compilers:
Create siteconfig.py with blow lines.
Install GPAW
How to use HPC-X module
Check the bin files:
Install dataset
Running GPAW
Sample output
Tasks and Submissions
Note: You will need to supply runs on both clusters Niagara and Aspire-1 (4 CPU nodes only).
Input Files
Can be download from this location: GitHub - jussienko/gpaw-isc-2021: GPAW related information for ISC 2021 Studen Cluster Comptetition
Building and running GPAW
Build the code both on NCSS and Niagara. Run the provided input copper.py on one, two, and four nodes using pure MPI parallelization. The text output is in the file output_M_xxx.txt (xxx = number of MPI tasks). Input file has check for the correctness, so there is an error message if the results are wrong. Runtime should be less than 45 minutes with a single node.
Discuss the differences in performance and scalability in the two clusters. No modifications to the input file and source code are allowed.
Submit the output files to your shared folder in OneDrive.
Visualization
Run the provided input nanoribbon.py in any of the clusters (it is recommended to use one node, running time should be max. few minutes). As a result you will obtain a file named elf_ribbon.cube which contains electron localization function of the simulated system. Make a visualization (picture or animation) of the electron localization function. See e.g.
https://wiki.fysik.dtu.dk/gpaw/tutorials/plotting/plot_wave_functions.html
for hints for visualizing .cube files.
Submit the figure or animation to your team's shared folder.
Profiling
Use IPM profiler to profile the application copper.py over 4 node run.
Submit the profiler output as text or PDF file.
Performance tuning
Try to optimize the performance of the copper.py input in one of the clusters. Use the result in task 1 as baseline. You can try different compilers, compiler options, and high performance libraries. You can try also hybrid OpenMP / MPI parallelization with different number of OpenMP threads. You are allowed to modify parallelization options in copper.py (see https://wiki.fysik.dtu.dk/gpaw/documentation/parallel_runs/parallel_runs.html#manual-parallelization-types),
and modify also the source code as long as the correctness check in the input passes. Any modifications to the source code have to be made available e.g. in github. Submit report about the steps you did for performance tuning and about the performance improvements you achieved.
Bonus task: bug fix
The scalapack parallelization in GPAW fails with certain input:
Using Scalapack does not work (#269) · Issues · gpaw / gpaw · GitLab
Try to fix the bug and if successful make a merge request to GPAW's master branch.
Note that you need to build GPAW with Scalapack support for this task.