Distributed-Data Massively-Parallel Self-Consistent Field

Adrian T. Wong and Robert J. Harrison

Recent significant optimizations to our distributed data SCF algorithm are described. These result in superior scaling for smaller molecules and an up to 7 fold increase in performance in very large sparse systems.

Background

The simplest N-electron wavefunction in common use is a single antisymmetric product (or ``Slater-determinant'') of one-electron functions (or molecular orbitals, MOs) which are orthonormal linear combinations of the atomic orbital (AO) basis functions. The expansion coefficients are determined by minimization of the energy in the Self Consistent Field (SCF) approach. This simplest of wavefunctions determines many properties with surprising accuracy and recovers over 99% of the total energy.

The computational kernel of the SCF algorithm is construction of the Fock matrix F by contracting the integrals (ij|kl) against the density matrix D

        F   = h   + (1/2) SUM [ 2 (ij|kl) - (ik|jl) ] D
         ij    ij         kl                           kl

Symmetry and sparsity in the matrix of integrals (ij|kl) must be exploited to avoid redundant computation. The major computation is evaluation of the non-zero integrals and the largest data requirements are due to the Fock and density matrices, both O(N*N) with N = O(1000). There are two major problems:

Improved distributed data algorithm

Our distributed data algorithm is modeled after that of a blocked matrix-multiply algorithm for a high-performance workstation with a small fast memory cache. This is because non-uniform memory access (NUMA) is the fundamental performance issue for both workstations and MPPs. The cost of accessing an element of the distributed density or Fock matrices must be offset by using this element multiple times. To achieve this re-use of data we simply stripmined the fourfold loop over the basis function indices. Since the computation is a quartic function of the block size while the communication is only a quadratic function, the block size may be adjusted to ensure the computation time dominates the communication time.

We previously stripmined by grouping basis functions according to their (usually atomic) centers. This had the advantage that the geometrical sparsity may be exploited in the outer stripmining loops. However, the granularity was fixed and was inadequate for calculations on large molecules in small basis sets, or when using machines with very high latencies. The revised algorithm groups atoms into blocks, looping over quartets of blocks of atoms, and then evaluating all integrals within a quartet of blocks of atoms. Sparsity may still be exploited in the outer loops and the performance gains include

All calculations will benefit from the vectorization of integral evaluation, but most calculations benefit by only 10% from the reduction in communication. However, the time to solution for our ZSM-5 zeolite fragment test-case (389 atoms, 2053 function, STO-3G basis, no symmetry) on 64 nodes of the IBM SP2 was reduced from over 40 hours to less than six. This is because the minimal basis set and high latency of the IBM violated all the criteria for good performance with the old algorithm. With the new code the Fock-build is completely dominated by integral computation.

The global array tools make it very easy to express blocked algorithms since the library provides routines that enable each process to access patches of arrays independently of all other processes. This asynchrony makes the program easier to write and also more efficient since there are no unnecessary synchronizations between processes. In addition, since the data is readily accessed by any process we can use full dynamic load balancing, implemented by using an element of a global array as a shared counter.

Conventional SCF with memory caching

Previous versions of NWChem have been fully-direct. The latest version includes the ability to perform conventional SCF calculations, caching integrals in memory where possible and storing additional integrals on disk. The blocking described in the above section is constrained to keep blocks less than 256 functions, and we can thus pack integral labels one per byte independent of the actual basis dimension. Data compression is also used to compress the integral values to 32 bits, maintaining a fixed overall precision which defaults to 10^-9. Thus, an integral and its labels are packed into 64 bits, a 33-100 percent saving over most other codes.

The frugal memory usage of the distributed data algorithm also permits much more memory to be used to cache integrals in memory. For instance, in the 324 function in-core calculation described in the benchmarking section, a replicated-data approach would have wasted over 200 Mbytes of memory that NWChem uses to cache integrals.

This functionality will be extended to include semi-direct algorithms.

Reducing linear algebra overhead

Linear algebra components in a conventional SCF algorithm include

To this list our quadratically convergent algorithm adds With PeIGS and ScaLAPACK we now have quite efficient matrix diagonalization and orthogonalization routines, but these operations still take much longer than a matrix multiplication. This is because the latter both parallelizes more readily and also drives a single processor at peak speed.

We have restructured the SCF code to rephrase as much linear algebra as possible as matrix multiplication operations. Most novel is the new matrix exponentiation routine that provides an accurate unitary approximation, which eliminates the necessity of orthogonalization and stabilizes the initial search step when large rotations are applied. The new matrix exponentiation algorithm (a modification of one due to Moler) is

  1. scale the matrix by a power of two so the largest element is less than 0.01,
  2. use a 3rd order taylor series approximation to the exponential,
  3. repeatedly square the result to undo the scaling, and
  4. if the largest element of the original was greater than 0.001 apply one iteration of a quadratically convergent symmetric orthogonalization (two matrix multiplications).
This is tailored to the antisymmetric matrices typical of our applications and may not be stable for more general matrices.

The net result is that in a typical calculation

As a result, small and in-core calculations are up to 30% improved on large numbers of processors.

Acknowledgments

This work was performed under the auspices of the High Performance Computing and Communications Program of the Office of Scientific Computing, U.S. Department of Energy under contract DE-AC6-76RLO 1830 with Battelle Memorial Institute which operates the Pacific Northwest National Laboratory, a multiprogram national laboratory. This work has made extensive use of software developed by the Molecular Science Software Group of the Environmental Molecular Sciences Laboratory project managed by the Office Health and Energy Research.


Prepared by RJ Harrison: Email: nwchem-support@emsl.pnl.gov.