Conjugate gradient optimization: Difference between revisions

From VASP Wiki
No edit summary
No edit summary
 
Line 13: Line 13:
<references/>
<references/>
----
----
[[Category:Electronic Minimization]][[Category:Electronic Minimization Methods]][[Category:Theory]]
[[Category:Electronic minimization]][[Category:Theory]]

Latest revision as of 10:48, 6 April 2022

Instead of the previous iteration scheme, which is just some kind of Quasi-Newton scheme, it also possible to optimize the expectation value of the Hamiltonian using a successive number of conjugate gradient steps. The first step is equal to the steepest descent step in section Single band steepest descent scheme. In all following steps the preconditioned gradient [math]\displaystyle{ g^N_{n} }[/math] is conjugated to the previous search direction. The resulting conjugate gradient algorithm is almost as efficient as the algorithm given in Efficient single band eigenvalue-minimization. For further reading see [1][2][3].

References