Page 3 of 4
Re: Issues compiling yambo 5.0 on HPC
Posted: Thu Mar 18, 2021 4:11 pm
by pyadav
Dear Andrea,
Thank you for your reply. Yes it seems that it's not able to unzip the tar file as I got the following message
Code: Select all
tar xvfz libxc-2.2.3.tar.gz
gzip: stdin: unexpected end of file
tar: Child returned status 1
tar: Error is not recoverable: exiting now
-------------------------------------
tar -zxvf libxc-2.2.3.tar.gz
gzip: stdin: unexpected end of file
tar: Child returned status 1
tar: Error is not recoverable: exiting now
How can I get rid of it?, Should I replace the tar files with freshly downloaded ones? and where can I get these files?
best,
pushpendra
Re: Issues compiling yambo 5.0 on HPC
Posted: Thu Mar 18, 2021 5:05 pm
by andrea.ferretti
Dear Marzieh,
unless I am mistaken, it looks all good to me.
In terms of possible causes to be further investigated
* the MPI issue does not look promising. Can you check with your sys-admin ?
* otherwise, I would try to either use external hdf5 and recompile internal netcdf and netcdff,
* or simply to compile the whole IO stack (netcdf, netcdff, hdf5) internally (just drop the related lines from configure invocation)
take care
Andrea
Re: Issues compiling yambo 5.0 on HPC
Posted: Thu Mar 18, 2021 6:37 pm
by pyadav
Dear Andrea,
When I'm downloading a fresh libxc-2.2.3.tar.gz file, it successfully unzips but when I put it in the archive folder and trying to build the code it shows the error "libxc build failed". Later on when I'm trying to unzip the same "libxc-2.2.3.tar.gz" in the archive, it shows the error
Code: Select all
gzip: stdin: unexpected end of file
tar: Child returned status 1
tar: Error is not recoverable: exiting now
How can I fix this?
Thanks,
Pushpendra
Re: Issues compiling yambo 5.0 on HPC
Posted: Fri Mar 19, 2021 1:29 pm
by Marzieh
Dear Andrea
I recompiled yambo with these modules an configuration:
Code: Select all
module --force purge
module load releases/2019b
module load intel/2019b
module load libxc/4.3.4-iccifort-2019.5.281
module load HDF5/1.10.5-iimpi-2019b
module load netCDF/4.7.1-iimpi-2019b
module load netCDF-Fortran/4.5.2-iimpi-2019b
Code: Select all
./configure FC="ifort" PFC="mpiifort" F77="ifort" CC="mpiicc" CXX="mpiicpc" \
--enable-hdf5-par-io \
--enable-mpi \
--enable-open-mp \
--enable-dp \
--with-blas-libs="-lmkl_intel_lp64 -lmkl_intel_thread -lmkl_core -liomp5 -lpthread -lm" \
--with-lapack-libs="-lmkl_intel_lp64 -lmkl_intel_thread -lmkl_core -liomp5 -lpthread -lm" \
--with-fft-includedir="-I${MKLROOT}/include/fftw" \
--with-fft-libs="-lmkl_intel_lp64 -lmkl_intel_thread -lmkl_core -liomp5 -lpthread -lm" \
--with-blacs-libs="-lmkl_scalapack_lp64 -lmkl_blacs_intelmpi_lp64 -lmkl_core -liomp5 -lpthread -lm" \
--with-scalapack-libs="-lmkl_scalapack_lp64 -lmkl_blacs_intelmpi_lp64 -lmkl_core -liomp5 -lpthread -lm" \
--with-netcdf-libs="-L$EBROOTNETCDF/lib64/" \
--with-netcdf-path="$EBROOTNETCDF/" \
--with-netcdf-libdir="$EBROOTNETCDF/lib64/" \
--with-netcdf-includedir="$EBROOTNETCDF/include/" \
--with-netcdff-libs="-L$EBROOTNETCDFMINFORTRAN/lib/" \
--with-netcdff-path="$EBROOTNETCDFMINFORTRAN/" \
--with-netcdff-libdir="$EBROOTNETCDFMINFORTRAN/lib/" \
--with-netcdff-includedir="$EBROOTNETCDFMINFORTRAN/include/" \
--with-hdf5-libs="-L$EBROOTHDF5/lib/" \
--with-hdf5-path="$EBROOTHDF5/" \
--with-hdf5-libdir="$EBROOTHDF5/lib/" \
--with-hdf5-includedir="$EBROOTHDF5/include/"
The output of the configuaration was as:
Code: Select all
# [VER] 5.0.0 r.19466
#
# - GENERAL CONFIGURATIONS -
#
# [SYS] linux@x86_64
# [SRC] /home/ucl/modl/afekete/src/yambo-5.0.0
# [BRN]
# [CMP] /home/ucl/modl/afekete/src/yambo-5.0.0
# [TGT] /home/ucl/modl/afekete/src/yambo-5.0.0
# [BIN] /home/ucl/modl/afekete/src/yambo-5.0.0/bin
# [LIB] /home/ucl/modl/afekete/src/yambo-5.0.0/lib/external
#
# [EDITOR] vim
# [ MAKE ] make
#
# [X] Double precision
# [X] Keep object files
# [X] Run-Time timing profile
# [-] Run-Time memory profile
#
# - SCRIPTS -
#
# [-] YDB: Yambo DataBase
# [-] YAMBOpy: Yambo Python scripts
#
# - PARALLEL SUPPORT -
#
# [X] MPI
# [X] OpenMP
#
# - LIBRARIES [E=external library; I?=internal library (c=to be compiled / f=found already compiled); X=system default; -=not used;] -
#
# > I/O: (NETCDF with large files support)
#
# [ - ] FUTILE :
#
# [ - ] YAML :
#
# [ If] IOTK : /home/ucl/modl/afekete/src/yambo-5.0.0/lib/external/intel/mpiifort/lib/libiotk.a (QE hdf5-support)
# -I/home/ucl/modl/afekete/src/yambo-5.0.0/lib/external/intel/mpiifort/include/
# [ - ] ETSF_IO :
#
# [ E ] NETCDF : -L/opt/sw/arch/easybuild/2019b/software/netCDF-Fortran/4.5.2-iimpi-2019b/lib/ -lnetcdff -L/opt/sw/arch/easybuild/2019b/software/netCDF/4.7.1-iimpi-2019b/lib64/ -lnetcdf
# -I/opt/sw/arch/easybuild/2019b/software/netCDF-Fortran/4.5.2-iimpi-2019b/include/ -I/opt/sw/arch/easybuild/2019b/software/netCDF/4.7.1-iimpi-2019b/include/
# [ E ] HDF5 : -L/opt/sw/arch/easybuild/2019b/software/HDF5/1.10.5-iimpi-2019b/lib/ -lhdf5hl_fortran -lhdf5_fortran -lhdf5_hl -lhdf5 -lz -lm -ldl -lcurl
# -I/opt/sw/arch/easybuild/2019b/software/HDF5/1.10.5-iimpi-2019b/include/
#
# > MATH: (FFTW MKL)
#
# [ E ] FFT : -lmkl_intel_lp64 -lmkl_intel_thread -lmkl_core -liomp5 -lpthread -lm
#
# [ E ] BLAS : -lmkl_intel_lp64 -lmkl_intel_thread -lmkl_core -liomp5 -lpthread -lm
# [ E ] LAPACK : -lmkl_intel_lp64 -lmkl_intel_thread -lmkl_core -liomp5 -lpthread -lm
# [ E ] SCALAPACK : -lmkl_scalapack_lp64 -lmkl_blacs_intelmpi_lp64 -lmkl_core -liomp5 -lpthread -lm
# [ E ] BLACS : -lmkl_scalapack_lp64 -lmkl_blacs_intelmpi_lp64 -lmkl_core -liomp5 -lpthread -lm
# [ - ] PETSC :
#
# [ - ] SLEPC :
#
#
# > OTHERs
#
# [ If] LibXC : /home/ucl/modl/afekete/src/yambo-5.0.0/lib/external/intel/mpiifort/lib/libxcf90.a /home/ucl/modl/afekete/src/yambo-5.0.0/lib/external/intel/mpiifort/lib/libxc.a
# -I/home/ucl/modl/afekete/src/yambo-5.0.0/lib/external/intel/mpiifort/include
# [ E ] MPI : -L/opt/sw/arch/easybuild/2019b/software/impi/2018.5.288-iccifort-2019.5.281/intel64/lib -Xlinker --enable-new-dtags -Xlinker -rpath -Xlinker /opt/sw/arch/easybuild/2019b/software/impi/2018.5.288-iccifort-2019.5.281/intel64/lib/release_mt -Xlinker -rpath -Xlinker /opt/sw/arch/easybuild/2019b/software/impi/2018.5.288-iccifort-2019.5.281/intel64/lib -Xlinker -rpath -Xlinker /opt/intel/mpi-rt/2017.0.0/intel64/lib/release_mt -Xlinker -rpath -Xlinker /opt/intel/mpi-rt/2017.0.0/intel64/lib -lmpifort -lmpi -lmpigi -ldl -lrt -lpthread
# -I/opt/sw/arch/easybuild/2019b/software/impi/2018.5.288-iccifort-2019.5.281/intel64/include
# [ Ic] Ydriver : 1.0.0
#
# - COMPILERS -
#
# FC kind = intel ifort version 19.0.5.281
# MPI kind= Intel(R) MPI Library 2018 Update 5 for Linux* OS
#
# [ CPP ] mpiicc -E -ansi -D_HDF5_LIB -D_HDF5_IO -D_PAR_IO -D_MPI -D_FFTW -D_FFTW_OMP -D_SCALAPACK -D_DOUBLE -D_OPENMP -D_TIMING -D_P2Y_QEXSD_HDF5
# [ FPP ] fpp -free -P -D_HDF5_LIB -D_HDF5_IO -D_PAR_IO -D_MPI -D_FFTW -D_FFTW_OMP -D_SCALAPACK -D_DOUBLE -D_OPENMP -D_TIMING
# [ CC ] mpiicc -O2 -D_C_US -D_FORTRAN_US
# [ FC ] mpiifort -assume bscc -O3 -g -ip -qopenmp
# [ FCUF] -assume bscc -O0 -g
# [ F77 ] mpiifort -assume bscc -O3 -g -ip
# [ F77U] -assume bscc -O0 -g
# [Cmain] -nofor_main
The previous error was solved but I do not have any values for E-Eo [eV] and Sc|Eo [eV] in o.qp
Code: Select all
# K-point Band Eo [eV] E-Eo [eV] Sc|Eo [eV]
#
1 1 -39.73587 NaN NaN
1 2 -19.57856 NaN NaN
1 3 -1.099598 NaN NaN
1 4 -0.251880 NaN NaN
1 5 -0.205098 NaN NaN
....
Could you please tell me where I made mistake?
Best,
Marzieh
Re: Issues compiling yambo 5.0 on HPC
Posted: Fri Mar 19, 2021 2:53 pm
by Daniele Varsano
Dear Marzieh,
can you check in the report file if you have NaN also in the HF part of the calculation, or in the Sc only?
Otherwise, If you attach your report file we can have a look.
Best,
Daniele
Re: Issues compiling yambo 5.0 on HPC
Posted: Fri Mar 19, 2021 3:09 pm
by Marzieh
Dear Daniele,
I have attached the generated files.
Best,
Marzieh
Re: Issues compiling yambo 5.0 on HPC
Posted: Fri Mar 19, 2021 3:23 pm
by Daniele Varsano
Dear Marzieh,
the problem is in the correlation part of the self energy. Not easy to spot, but note the parallelism is incomplete.
Can you remove the ./SAVE/ndb.QP and ./SAVE/ndb.pp* files,
and run again the code after setting:
SE_CPU= "1 1 2" # [PARALLEL] CPUs for each role
SE_ROLEs= "q qp b" # [PARALLEL] CPUs roles (q,qp,b)
X_and_IO_CPU= "1 1 1 2 1" # [PARALLEL] CPUs for each role
X_and_IO_ROLEs= "q g k c v" # [PARALLEL] CPUs roles (q,g,k,c,v)
Note that this is just a test, in your input file you have "NGsBlkXp= 1 RL" , and this make your screening unreliable.
Best,
Daniele
Re: Issues compiling yambo 5.0 on HPC
Posted: Fri Mar 19, 2021 7:14 pm
by Marzieh
Dear Daniele,
I did what you suggested me. But I still have that problem.
Best,
Marzieh
Re: Issues compiling yambo 5.0 on HPC
Posted: Mon Mar 22, 2021 12:44 pm
by Marzieh
Dear Daniele,
Since I did not see these parts during run by yambo 5.0.0 (in both parallel and without parallel):
Code: Select all
[WF-HF/Rho] Performing Wave-Functions I/O from ./SAVE
[WF-GW] Performing Wave-Functions I/O from ./SAVE
It seems yambo 5.0.0 can not read files from ./SAVE.
( As I mentioned before I use abinit (of course a2y) to generate files.
So I had to compile yambo 4.5.3 on the cluster. It works.
I would like to ask you if there are main differences between yambo-5.0.0 and yambo4.5.03 especially in nonlinear optics? because I am going to calculate nlo.
Best,
Marzieh
Re: Issues compiling yambo 5.0 on HPC
Posted: Sun Mar 28, 2021 5:46 pm
by claudio
Dear Marzieh
there are not particular differences in the non-linear optics part from the 4.5.3 and the 5.0, as far as I remember,
you can continue with the 4.5.3.
In the while we are trying to solve all compilation problems of the 5.0
best
Claudio
NB: as you see the databases of the 4.5.3 are not compatible with the 5.0