Page 1 of 1

Memory issue when using more than 1 core

Posted: Mon Sep 23, 2024 10:23 am
by Ponnappa
Dear Yambo Developers and Users,
I recently installed yambo 5.2.3. The configration I used was (./configure FC=ifort CC=icc MPIFC=mpiifort --enable-memory-profile). The installtion is complete but when I try to use more than 1 core I face memory issue and do not start any calculation (The calculation can be started with only 1 core :( ). I have attached the error file. Kindly help to resolve the issue. Thank you in advance
H1.txt

Re: Memory issue when using more than 1 core

Posted: Tue Sep 24, 2024 3:04 pm
by Daniele Varsano
Dear Ponnappa,

it could there are some problems with your MPI compilation, can you please provide some more information, e.g. report and log files if any?
If not, can you try to run the code interactively?
> mpirun -np 2 $path/yambo

Also, having a look at the config.log file can be useful,

Best,

Daniele

Re: Memory issue when using more than 1 core

Posted: Sat Sep 28, 2024 2:54 pm
by Ponnappa
Dear Daniele,
THe cluster which I am has 24 cores per per node. Looking at your suggestion I used used cores which was serving the purpose and later gradually increased. I could find that I shows the segmentation fault only if I use complete 24 cores. I can even run the calcuation with n*22 cores but not with n*24 cores.

Currently I used 8*22 cores the calcualtion started but ended with some other memory or Parallelization issue. I am attaching the input
gw.in.txt
and log file
l-jdir_HF_and_locXC_gw0_dyson_rim_cut_em1d_ppa_el_el_corr_CPU_1.txt
l-jdir_HF_and_locXC_gw0_dyson_rim_cut_em1d_ppa_el_el_corr_CPU_1_02.txt
of the same. Can I get more insights how to we solve the memory or paralleliztion issue in general. Thank you for the reply.

Regards,
Ponnappa

Re: Memory issue when using more than 1 core

Posted: Mon Sep 30, 2024 8:22 am
by Daniele Varsano
Dear Ponnappa,

most probably using all the core in the node, you are filling all the available memory in the node.
In order to distribute memory try to parallelize on bands as much as possible as you did in X_and_IO_CPU.
You can fine-tuning the distribution moving cpus from "v" to "c", but as I can't have a look at the report file I do not know how many bands are occupied, in any
case try to balance the cpus according to the number of occupied and empty states.

Regarding the Self energy, avoid assigning cpu on "q" and assing them on "b" as much as possible e.g.

Code: Select all

SE_CPU= "1 2 88"   
If it fails, the strategy is to use less cpus per node.

Best,
Daniele