[molpro-user] Problems with a large eom-ccsd calculation

Manhui Wang wangm9 at cardiff.ac.uk
Thu Apr 7 11:53:02 BST 2011


Dear Evgeniy,

Have you built a working  Molpro2010.1 with GA 5-0-2 now?  Does your job
work with the GA-Molpro?

Best wishes,
Manhui

Manhui Wang wrote:
> Dear Evgeniy,
> 
> Just remind you that when configure GA
> 
>       ./configure  --with-tcgmsg --with-mpi="-ldl
> -I/software/intel/mpi/4.0.0.025/intel64/include
> -L/software/intel/mpi/4.0.0.025/intel64/lib -lmpi -lmpigf -lmpigi -lpthread
> -lrt" --with-openib="/usr/lib64 -libverbs -I/usr/include/infiniband"
> F77=ifort CC=icc CXX=icpc  --prefix=${GA_path}-install
> 
> the content of --with-mpi=" " comes from MPI link options (eg, mpiifort
> -show, mpiifort -link_info, mpif90 -show), and you might have to replace
> this with your own's .
> 
> Best wishes,
> Manhui
> 
> 
> 
> Evgeniy Gromov wrote:
>> Manhui,
>>
>> Thanks for your help.
>> Ok, molpro configure finally passed. The problem was
>> that I followed the manual where it is said that ga_libs
>> should be in lib/LINUX64 (new ga put libs in /ga-path/lib.)
>> Well, I did't recompile the whole code, only the libraries.
>> And the new binary still does not work. I will try to recompile
>> the code completely. Maybe it helps.
>>
>> Best wishes,
>> Evgeniy
>>
>> Manhui Wang wrote:
>>> Dear Evgeniy,
>>>
>>> Evgeniy Gromov wrote:
>>>> Dear Manhui,
>>>>
>>>> GA runs OK (I did test.x for 16 CPUs).
>>>> To run molpro configure in the fresh molpro2010.1
>>>> I use your example (with the corrected paths):
>>>>
>>>> ./configure -batch -icc -ifort -mpp -mppbase
>>>> /bwfs/scratch/temphomes/hd/f81/prog/ga-5-0-2 -var LIBS="-libverbs"
>>>> -blaspath /opt/bwgrid/compiler/intel/ct_4.0/mkl/10.2.5.035/lib/em64t
>>>>
>>>> molpro configure crashes then after the following:
>>>>
>>>> ga_GA_MP_LIBS     = -lmpi -lmpigf -lmpigi -lpthread
>>>> ga_GA_MP_LDFLAGS  =
>>>> -L/opt/bwgrid/compiler/intel/ct_4.0/impi/4.0.0.028/intel64/lib
>>>> ga_GA_MP_CPPFLAGS =
>>>> -I/opt/bwgrid/compiler/intel/ct_4.0/impi/4.0.0.028/intel64/include
>>>> ga_TARGET         = LINUX64
>>>> ga_MSG_COMMS      = TCGMSGMPI
>>>> ga_prefix         = /bwfs/scratch/temphomes/hd/f81/ga-5-0-2-install
>>>>
>>>> MPPLIB=
>>>> parallel build specified but unable to find Global Arrays
>>> It is trying to find the generated GA lib files in directory
>>> /path/ga-5-0-2-install/lib/
>>>
>>> A typical examples:
>>> ls  /mypatch/ga-5-0-2-install/lib/
>>> libarmci.a  libarmci.la  libga.a  libga.la
>>>
>>> Best wishes,
>>> Manhui
>>>
>>>> Result so far of configure written to CONFIG.errout
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> Apperently molpro configure finds the config.log in the ga-5-0-2
>>>> directory. But what it needs then I don't understand. What is
>>>> this MPPLIB ?
>>>>
>>>> Best,
>>>> Evgeniy
>>>>
>>>>
>>>>
>>>> Manhui Wang wrote:
>>>>> Dear Evgeniy,
>>>>>
>>>>> Could you please check your GA is working properly?
>>>>> Could you please provide more details about how do you run configure in
>>>>> a fresh Molpro2010.1 PL20?
>>>>>
>>>>> Best wishes,
>>>>> Manhui
>>>>>
>>>>> Evgeniy Gromov wrote:
>>>>>> Dear Manhui,
>>>>>>
>>>>>> If I specify the original GA directory configure cannot find some
>>>>>> libraries whatever and crashes:
>>>>>>
>>>>>> ga_GA_MP_LIBS     = -lmpi -lmpigf -lmpigi -lpthread
>>>>>> ga_GA_MP_LDFLAGS  =
>>>>>> -L/opt/bwgrid/compiler/intel/ct_4.0/impi/4.0.0.028/intel64/lib
>>>>>> ga_GA_MP_CPPFLAGS =
>>>>>> -I/opt/bwgrid/compiler/intel/ct_4.0/impi/4.0.0.028/intel64/include
>>>>>> ga_TARGET         = LINUX64
>>>>>> ga_MSG_COMMS      = TCGMSGMPI
>>>>>> ga_prefix         = /bwfs/scratch/temphomes/hd/f81/ga-5-0-2-install
>>>>>>
>>>>>> MPPLIB=
>>>>>> parallel build specified but unable to find Global Arrays
>>>>>>
>>>>>>
>>>>>> It looks like the structure of the GA directory has been changed
>>>>>> in the latest release. At least there is neither include no lib
>>>>>> directories in the GA directory. Should I take some older GA realese
>>>>>> then? Which one?
>>>>>>
>>>>>> Best,
>>>>>> Evgeniy
>>>>>>
>>>>>> Manhui Wang wrote:
>>>>>>> Dear Evgeniy,
>>>>>>>
>>>>>>>
>>>>>>> It appears your CONFIG is not right. As I mentioned in previous
>>>>>>> email:
>>>>>>>
>>>>>>> ./configure -batch  -icc -ifort -blaspath
>>>>>>> /software/intel/mkl/10.2.5.035/lib/em64t -mpp -mppbase ${GA_path}
>>>>>>> -var
>>>>>>> LIBS="-libverbs"
>>>>>>>
>>>>>>> Please note ${GA_path} is not the GA installation directory, but the
>>>>>>> original GA directory where you type"configure...make". Molpro
>>>>>>> configure
>>>>>>> takes some info from ${GA_path} to determine the GA mode and the
>>>>>>> underlying MPI library etc.
>>>>>>> -var LIBS="-libverbs" will enable IntelMPI to be linked with
>>>>>>> Infiniband
>>>>>>> library if your IntelMPI on the machine can't do this automatically.
>>>>>>> Once the CONFIG is generated, we don't recommend to modify it
>>>>>>> manually
>>>>>>> since it may introduce unexpected problems.
>>>>>>>
>>>>>>> Best wishes,
>>>>>>> Manhui
>>>>>>>
>>>>>>> Evgeniy Gromov wrote:
>>>>>>>> Dear Manhui,
>>>>>>>>
>>>>>>>> Thanks for your response.
>>>>>>>> There is unfortunately no progress with GA. I compiled molpro
>>>>>>>> but the binary doesn't work at all, although the compilation
>>>>>>>> went OK. There is somehow again no ABDATA in lib, but it doesn't
>>>>>>>> complain to that. What it complains looks strange:
>>>>>>>>
>>>>>>>> Attempting to use an MPI routine before initializing MPICH
>>>>>>>>
>>>>>>>>
>>>>>>>> Well I used intel MPI and I don't understand why it refers to MPICH.
>>>>>>>>
>>>>>>>> Just in case my CONFIG looks the following:
>>>>>>>>
>>>>>>>> # MOLPRO CONFIG generated at Di 5. Apr 10:21:18 CEST 2011, for host
>>>>>>>> frbw3
>>>>>>>>
>>>>>>>> CONFIGURE_OPTIONS="-i8" "-openmp" "-icc" "-ifort" "-nohdf5"
>>>>>>>> "-nocuda"
>>>>>>>> "-mpp" "-mppbase" "/bwfs/scratch/temphomes/hd/f81/ga-5-0-2-install"
>>>>>>>> "-blaspath"
>>>>>>>> "/opt/bwgrid/compiler/intel/ct_4.0/mkl/10.2.5.035/lib/em64t"
>>>>>>>>
>>>>>>>> AR=ar
>>>>>>>> ARCHNAME=Linux/x86_64
>>>>>>>> ARFLAGS=-rS
>>>>>>>> AWK=awk
>>>>>>>> BIBTEX=/usr/bin/bibtex
>>>>>>>> BLASLIB=-L/opt/bwgrid/compiler/intel/ct_4.0/mkl/10.2.5.035/lib/em64t
>>>>>>>> -lmkl_intel_ilp64 -lmkl_sequential -lmkl_core
>>>>>>>> BUILD=p
>>>>>>>> CAT=cat
>>>>>>>> CC=/opt/bwgrid/compiler/intel/ct_4.0/Compiler/11.1/072/bin/intel64/icc
>>>>>>>>
>>>>>>>> CCVERSION=11.1
>>>>>>>> CC_FRONT=
>>>>>>>> CDEBUG=-g $(addprefix $(CDEFINE),_DEBUG)
>>>>>>>> CDEFINE=-D
>>>>>>>> CFLAGS=-ftz -vec-report0 -DINT64 -DZLIB -openmp
>>>>>>>> CLEARSPEEDLIB=
>>>>>>>> CMPPINCLUDE=/bwfs/scratch/temphomes/hd/f81/ga-5-0-2-install/include
>>>>>>>> COPT=-O2
>>>>>>>> COPT0=-O0
>>>>>>>> COPT1=-O1
>>>>>>>> COPT2=-O2
>>>>>>>> COPT3=-O3
>>>>>>>> CP=cp -p
>>>>>>>> CPROFILE=-p
>>>>>>>> CSCN=cscn
>>>>>>>> CSFLAGS=-O3 -I. --dynamic
>>>>>>>> CUDACC=
>>>>>>>> CUDACCVERSION=
>>>>>>>> CUDACDEBUG=-g $(addprefix $(CUDACDEFINE),_DEBUG)
>>>>>>>> CUDACDEFINE=-D
>>>>>>>> CUDACFLAGS=
>>>>>>>> CUDACOPT=
>>>>>>>> CUDACOPT0=-O0
>>>>>>>> CUDACOPT1=-O1
>>>>>>>> CUDACOPT2=-O2
>>>>>>>> CUDACOPT3=-O3
>>>>>>>> CUDACPROFILE=-p
>>>>>>>> CXX=/opt/bwgrid/compiler/intel/ct_4.0/Compiler/11.1/072/bin/intel64/icpc
>>>>>>>>
>>>>>>>>
>>>>>>>> CXXFLAGS=$(CFLAGS)
>>>>>>>> DOXYGEN=/usr/bin/doxygen
>>>>>>>> ECHO=echo
>>>>>>>> EXPORT=export
>>>>>>>> F90FLAGS=
>>>>>>>> FC=/opt/bwgrid/compiler/intel/ct_4.0/Compiler/11.1/072/bin/intel64/ifort
>>>>>>>>
>>>>>>>>
>>>>>>>> FCVERSION=11.1
>>>>>>>> FDEBUG=-g $(addprefix $(FDEFINE),_DEBUG)
>>>>>>>> FDEFINE=-D
>>>>>>>> FFLAGS=-i8 -r8 -pc64 -auto -Vaxlib -vec-report0 -cxxlib -openmp
>>>>>>>> FOPT=-O3
>>>>>>>> FOPT0=-O0
>>>>>>>> FOPT1=-O1
>>>>>>>> FOPT2=-O2
>>>>>>>> FOPT3=-O3
>>>>>>>> FPP=-fpp
>>>>>>>> FPROFILE=-p
>>>>>>>> FSTATIC=
>>>>>>>> FTCFLAGS=molpro unix unix-i8 Linux lapack mpp blas
>>>>>>>> HDF5INCLDE=
>>>>>>>> HDF5LIB=
>>>>>>>> HOSTFILE_FORMAT=
>>>>>>>> INSTBIN=/bwfs/scratch/temphomes/hd/f81/molpro2010.1-install/bin
>>>>>>>> INSTHTML=/bwfs/scratch/temphomes/hd/f81/molpro2010.1-install/html
>>>>>>>> INSTLIB=/bwfs/scratch/temphomes/hd/f81/molpro2010.1-install/lib/molprop_2010_1_Linux_x86_64_i8
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> INST_PL=0
>>>>>>>> INTEGER=8
>>>>>>>> LAPACKLIB=
>>>>>>>> LATEX2HTML=
>>>>>>>> LAUNCHER=/opt/bwgrid/mpi/mpiexec/0.84/bin/mpiexec -n %n %x
>>>>>>>> LD_ENV=/opt/bwgrid/compiler/intel/ct_4.0/Compiler/11.1/072/lib/intel64:/opt/bwgrid/compiler/intel/ct_4.0/mkl/10.2.5.035/lib/em64t
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> LD_ENVNAME=LD_LIBRARY_PATH
>>>>>>>> LIBRARY_SUFFIX=a
>>>>>>>> LIBS=-lz
>>>>>>>> LIBS_FRONT=
>>>>>>>> LINKOPT=
>>>>>>>> LINKOPT_FRONT=
>>>>>>>> LN=ln -s
>>>>>>>> MACROS=MOLPRO MOLPRO_ifort MOLPRO_f2003 GA_TOOLS GA_ GA_VERSION_GE_5
>>>>>>>> _I8_ BLAS_INT=8 LAPACK_INT=8 MOLPRO_FORCE_VECTOR MOLPRO_NEXTSCALAR
>>>>>>>> MOLPRO_NO_RECURRENCE MOLPRO_NOVECTOR MOLPRO_SHORTLOOP _MOLCAS_MPP_
>>>>>>>> MAKEDEPEND_OPTIONS=
>>>>>>>> MAKEINDEX=/usr/bin/makeindex
>>>>>>>> MAPLE=
>>>>>>>> MKDIR=mkdir -p
>>>>>>>> MODULE_FLAG=-I
>>>>>>>> MODULE_SUFFIX=mod
>>>>>>>> MPILIB=-L/opt/bwgrid/compiler/intel/ct_4.0/impi/4.0.0.028/intel64/lib
>>>>>>>>
>>>>>>>> -lmpi -lmpigf -lmpigi -lpthread -L/usr/lib64 -libverbs -libcommon
>>>>>>>> -libumad -lpthread
>>>>>>>> MPPLIB=-L/bwfs/scratch/temphomes/hd/f81/ga-5-0-2-install/lib/LINUX64
>>>>>>>> -lga -larmci
>>>>>>>> OBJECT_SUFFIX=o
>>>>>>>> OPT0=copyc6.f
>>>>>>>> OPT1=nevpt2_optrpc.f explicit_util.f artwo2.f drv2el_l3ext_lmp2g.f
>>>>>>>> drv2el_l3ext_cen.f rmp2_f12_drv2.f90 ri_lmp2g.f df_llmp2.f
>>>>>>>> readdump.f
>>>>>>>> OPT2=integrals.f90 RL4gen1.f basis_integral_shells.f
>>>>>>>> OPT3=
>>>>>>>> PAPER=a4paper
>>>>>>>> PARSE=parse-Linux-x86_64-i8.o
>>>>>>>> PDFLATEX=/usr/bin/pdflatex -halt-on-error
>>>>>>>> PNAME=molprop_2010_1_Linux_x86_64_i8
>>>>>>>> PTSIZE=11
>>>>>>>> RANLIB=ranlib
>>>>>>>> RM=rm -rf
>>>>>>>> SHELL=/bin/sh
>>>>>>>> STRIP=strip
>>>>>>>> SUFFIXES=f F f90 F90 c cpp
>>>>>>>> TAR=tar -cf
>>>>>>>> UNTAR=tar -xf
>>>>>>>> VERBOSE=@
>>>>>>>> VERSION=2010.1XSD=/usr/bin/xmllint --noout --schema
>>>>>>>> XSLT=/usr/bin/xsltproc
>>>>>>>> YACC=bison -b y
>>>>>>>>
>>>>>>>> .SUFFIXES:
>>>>>>>> MAKEFLAGS+=-r
>>>>>>>> ifneq ($(LD_ENVNAME),)
>>>>>>>> $(LD_ENVNAME):=$(LD_ENV):$($(LD_ENVNAME))
>>>>>>>> endif
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> Best regards,
>>>>>>>> Evgeniy
>>>>>>>>
>>>>>>>> Manhui Wang wrote:
>>>>>>>>> Dear Evgeniy,
>>>>>>>>>
>>>>>>>>> I have not checked your output carefully, which shows 800MWord
>>>>>>>>> (6.4GB
>>>>>>>>> per process):
>>>>>>>>>
>>>>>>>>> Variable memory set to  800000000 words,  buffer space   230000
>>>>>>>>> words
>>>>>>>>>
>>>>>>>>> I run the job  with 32 processes on 4 nodes, and it might crash
>>>>>>>>> eventually due to lack of memory you mentioned.
>>>>>>>>>
>>>>>>>>> Could you please update your progress with GA? On my side, I will
>>>>>>>>> look
>>>>>>>>> into my code related to helper. It might exist a bug which is
>>>>>>>>> exposed
>>>>>>>>> when very large global data structure is used.
>>>>>>>>>
>>>>>>>>> Best wishes,
>>>>>>>>> Manhui
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> Evgeniy Gromov wrote:
>>>>>>>>>> Dear Manhui,
>>>>>>>>>>
>>>>>>>>>> Many thanks for your help with the compilation problems and
>>>>>>>>>> with the problem in the large eom-ccsd calculation.
>>>>>>>>>> Sorry I forgot to tell you about the memory. I usually specify
>>>>>>>>>> the memory in the command line using -m option. For that large
>>>>>>>>>> eom-ccsd job I used 800m (-m800m) because 200m was not enough for
>>>>>>>>>> the CCSD step. As I wrote the job passed the CCSD and crashed
>>>>>>>>>> in the beginning of the EOM-CCSD. Actually, are you running it
>>>>>>>>>> on 1 CPU? I haven't tried it to run on 1 CPU as it will likely
>>>>>>>>>> take ages cause many MOs. Well I am going to rerun it using
>>>>>>>>>> a new binary compiled with GA.
>>>>>>>>>>
>>>>>>>>>> Best wishes,
>>>>>>>>>> Evgeniy
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> Manhui Wang wrote:
>>>>>>>>>>> Dear Evgeniy,
>>>>>>>>>>>
>>>>>>>>>>> I can't see the error you mentioned with MPI-2 Molpro2010.1
>>>>>>>>>>> PL20 so
>>>>>>>>>>> far.
>>>>>>>>>>> I tried your input without any modification first, and it failed
>>>>>>>>>>> immediately because you have not set memory in input (the
>>>>>>>>>>> default is
>>>>>>>>>>> only 8MWord). I have added
>>>>>>>>>>>
>>>>>>>>>>> memory,200,m
>>>>>>>>>>>
>>>>>>>>>>> in input, it appears to work fine so far (still running).  Could
>>>>>>>>>>> you
>>>>>>>>>>> please try the new input on your side?
>>>>>>>>>>>
>>>>>>>>>>> Best wishes,
>>>>>>>>>>> Manhui
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> Evgeniy Gromov wrote:
>>>>>>>>>>>> Manhui,
>>>>>>>>>>>>
>>>>>>>>>>>> I attached wrong files in the previous email. Here are the
>>>>>>>>>>>> correct
>>>>>>>>>>>> input and output files. In addition I removed all the MO
>>>>>>>>>>>> vectors in
>>>>>>>>>>>> the output to make the size smaller.
>>>>>>>>>>>>
>>>>>>>>>>>> Best regards,
>>>>>>>>>>>> Evgeniy
>>>>>>>>>>>>
>>>>>>>>>>>> Manhui Wang wrote:
>>>>>>>>>>>>> Dear Evgeniy,
>>>>>>>>>>>>>
>>>>>>>>>>>>> Could you please provide the inputs/outputs? I will see whether
>>>>>>>>>>>>> I can
>>>>>>>>>>>>> reproduce the problem firstly.
>>>>>>>>>>>>>
>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>> Manhui
>>>>>>>>>>>>>
>>>>>>>>>>>>> Evgeniy Gromov wrote:
>>>>>>>>>>>>>> Dear Molpro Community,
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> I faced problems when running a large eom-ccsd calculation
>>>>>>>>>>>>>> (466
>>>>>>>>>>>>>> MOs,
>>>>>>>>>>>>>> no symmetry) with molpro2010.1 in parallel (MPI-2). The
>>>>>>>>>>>>>> calculation
>>>>>>>>>>>>>> crashed after the CCSD (ground state) step had been done. No
>>>>>>>>>>>>>> error
>>>>>>>>>>>>>> message was present in the output. In the err file there is
>>>>>>>>>>>>>> the
>>>>>>>>>>>>>> following diagnostic (though unclear to me):
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> ERROR in twosided_helpga_locate_server: over range!
>>>>>>>>>>>>>> ilo=1,ihigh=148114575
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Any idea what can be the problem.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Well, it seems to relate to helper process. I suspect that in
>>>>>>>>>>>>>> the
>>>>>>>>>>>>>> case
>>>>>>>>>>>>>> of GA it would work, although there might be some other
>>>>>>>>>>>>>> problems.
>>>>>>>>>>>>>> Just in case the same run with 296 MO (another basis set) ran
>>>>>>>>>>>>>> fine.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Also, I would like to thank Manhui for the response to my
>>>>>>>>>>>>>> previous
>>>>>>>>>>>>>> post:
>>>>>>>>>>>>>> "Is there always a helper process in parallel calculations
>>>>>>>>>>>>>> with
>>>>>>>>>>>>>> molpro2010.1 ?"
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Best regards,
>>>>>>>>>>>>>> Evgeniy
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>
> 

-- 
-----------
Manhui  Wang
School of Chemistry, Cardiff University,
Main Building, Park Place,
Cardiff CF10 3AT, UK



More information about the Molpro-user mailing list