Hello,
I am running vibrational frequencies with HSE06 in VASP 6.5.1 (Intel/IMPI). The job fails immediately/early with job.*err, while the same system and workflow runs fine with VASP 5.4.4 on the same machine.
In VASP 6.5.1, Slurm reports multiple oom_kill events and srun: Out Of Memory, and the step gets cancelled.
I can reproduce the same OOM-kill behavior in VASP 6.5.1 using 112, 224, 336, and 448 MPI ranks (tested on the same system and input). The job is terminated by Slurm with oom_kill / Out Of Memory messages.
Key environment/modules:
module load mkl impi intel hdf5/1.10.11 ucx vasp/6.5.1 and srun .../vasp_std.
Job layout: 2 nodes, 224 MPI tasks.
System: 42 atoms (Mg/O/C), k-mesh 1×2×1 (Gamma-centered).
Main INCAR:
ISTART = 0
ICHARG = 2
ALGO = All
NELMDL = -15
EDIFF = 1E-8
ENCUT = 600
EDIFFG = -0.01
IBRION = 5 ! Quasi Newton, faster-efficient close local minima
ISIF = 2
NFREE = 2
NELMIN = 8 ! Use together with IBRION = 1
NSW = 1
POTIM = 0.015
ISPIN = 1
#MAGMOM = 60*0.6
LORBIT = 11
NEDOS = 10001
PREC = Accurate
ISMEAR = 0
SIGMA = 0.05
LREAL= .TRUE.
#Hybrid Functional calculations:
LHFCALC = .TRUE. ! A hybrid XC potential should be used
TIME = 0.40 ! Trial time step for IALGO=5X
LMAXFOCK = 4 ! It might be required to increase, if the system contains f-electrons.
HFSCREEN= 0.207 ! Switch from the PBE0 to HSE03
NPAR = 224 ! IF
VASP OUTPUTs
NWRITE = 2
LCHARG = .TRUE.
LWAVE = .TRUE.
Parallelization: I tested NPAR=224 (current), and also tried NPAR=1 / other values: VASP 6.5.1 still gets OOM-killed.
job.*err excerpt (summary):
slurmstepd: error: Detected ... oom_kill events ...
srun: error: ... tasks ... Out Of Memory
I will attach a minimal reproducible zip
Specific questions
1. Is the OOM behavior in VASP 6.5.1 expected for hybrid + frequencies (IBRION=5)?
I am surprised because the same setup runs with VASP 5.4.4, and the system is small (42 atoms, 1×2×1 k-points).
2. Is my parallelization setting inappropriate for hybrid calculations in 6.5.1?
In particular: should NPAR be removed (or forced to 1) and instead use NCORE?
Could large MPI task counts (224 ranks) trigger excessive replicated-memory overhead in 6.5.1 for exact exchange?
3. Is there any recommended low-memory setting for HSE06 in VASP 6.5.1 for force-constant/frequency runs?
e.g., recommended combinations of MPI ranks vs. OMP threads, or specific tags that reduce Fock-memory footprint.

