The initial Hessian matrix needs of a special consideration. The matrix shape employed in section 3.1 displayed in page with a small squared Hessian for a core and a diagonal vector for the environment was suitable for a very localized reaction. But when many atoms have to rearrange during the optimization a bigger square Hessian would be needed.
The sparsity of the Hessian matrix:
It is interesting to see how a Hessian looks like.
In figure 3.7 there is a representation of the Hessian elements bigger than a threshold.
The Hessian matrix corresponds to 148 atoms in the active site of Mandelate Racemase enzyme. The Hessian is calculated
numerically in Cartesian coordinates representation.
We can see that the biggest elements are in the diagonal, but there are also elements bigger than
10 kcal/(molÅ) even quite far away from the diagonal. This fact is probably due to the Cartesian coordinates
coupling and the long-range interaction in the enzymatic system.
In the next section we will have to recall this fact, in other words, this kind of Hessian is not like a MC-SCF Hamiltonian
which is rather sparse, concentrating the significative elements in the diagonal or near-diagonal region[243].
|
When there are many significative elements out of the diagonal the initial Hessian matrix represented as a small matrix plus a vector is not the most suitable. A sparse Hessian when only some elements are stored has been suggested for the Newton-like optimizations[157].
Optimization with a sparse Hessian:
We have tested how important is the exclusion of some elements in the Hessian matrix.
The antamanide polypeptide studied in section 3.1 is used here to carry out several TS optimizations using RFO method.
The tests consist in
taking as initial Hessian matrix a sparse Hessian where only the elements bigger than a threshold are stored.
Procedure: In order to test only the influence of a sparse Hessian we work with the same LAPACK diagonalization subroutine [246] that needs the whole matrix in memory. In the next section we implement a real save memory strategy with the iterative Lanczos diagonalization.4.7
We calculate the Hessian numerically. When the calculated element matrix is smaller than the given threshold this element is set to zero. This matrix element will be zero all along the optimization process and the Powell Hessian update will not be permitted to change it. RFO equations are solved and the optimization process is performed as explained in section 3.1. The convergence criteria is 10 kcal/(molÅ) on the gradient norm.
Results:
In table 3.17 we show the results of two different structure optimizations.
We represent the steps required to converge depending on the initial sparse Hessian matrix.
From the table 3.17 we can extract some conclusions. When the initial structure is too far from the stationary point () the initial Hessian has few influence on the number of iterations. After one hundred of steps the Hessian has been updated one hundred of times and there is no trace of the initial matrix shape. On the contrary, when the initial gradient is closer to the stationary point the optimization obviously takes less steps to converge. It can be seen that we obtain the same results storing the elements bigger than 10.0 kcal/(molÅ), that is, storing only the 41.1 % of a Hessian, than storing all the Hessian, In addition, we obtain reasonable results storing only the 22.1 % and even the 10.5 %. The optimization of both structures fail when the threshold is too big (10 kcal/(molÅ)), this means that the essential matrix elements are not stored.
From the above results a rapid conclusion is that a considerable number of Hessian elements can be discarded in the storage and consequently a method to save memory is a worthy task.
In the next section we will test these kind of Hessians with the general shape displayed in figure 3.8.
|