[molpro-user] VMP2, VCI calculations teminated with 'segmentation Violation error' , using downloaded MOLPRO 2010.1 binary code

Andy May MayAJ1 at cardiff.ac.uk
Wed Sep 21 11:39:25 BST 2011


Gang,

This problem is hopefully fixed now with the latest patch (2010.1.24). 
Note there was also a change to the input file in the manual.

Best wishes,

Andy

On 08/08/11 13:36, Andy May wrote:
> Gang,
>
> I've managed to reproduce the problem you mention and am testing a fix
> for this.
>
> Best wishes,
>
> Andy
>
> On 05/08/11 13:30, Gang Li wrote:
>> Dear Molpro Users,
>>
>> I recently downloaded the new 2010.1 binary version. However, quick tests
>> using the VMP2 sample input in the Molpro 2010.1 Manual crash every time
>> with the following error in terminal. Similar error message always shows
>> up with the VCI sample input and the CBS frequency input in the manual.
>> The system OS is Ubuntu 10.04 LTS and the 2008 version I was using worked
>> smoothly. Anyone knows the reason? Thanks!
>>
>>
>> Creating: host=virgo, user=root,
>> file=/usr/local/bin/molprop_2010_1_Linux_x86_64_i8.exe,
>> port=53336
>> 3:Segmentation Violation error, status=: 11
>> (rank:3 hostname:virgo pid:21797):ARMCI DASSERT fail.
>> src/signaltrap.c:SigSegvHandler():312 cond:0
>> 1:Segmentation Violation error, status=: 11
>> (rank:1 hostname:virgo pid:21795):ARMCI DASSERT fail.
>> src/signaltrap.c:SigSegvHandler():312 cond:0
>> Last System Error Message from Task 3:: Bad file descriptor
>> Last System Error Message from Task 1:: Bad file descriptor
>> 3: ARMCI aborting 11 (0xb).
>> 3: ARMCI aborting 11 (0xb).
>> 1: ARMCI aborting 11 (0xb).
>> 1: ARMCI aborting 11 (0xb).
>> system error message: Bad file descriptor
>> 0:Segmentation Violation error, status=: 11
>> 2:Segmentation Violation error, status=: 11
>> system error message: Bad file descriptor
>> (rank:2 hostname:virgo pid:21796):ARMCI DASSERT fail.
>> src/signaltrap.c:SigSegvHandler():312 cond:0
>> Last System Error Message from Task 2:: Bad file descriptor
>> (rank:0 hostname:virgo pid:21794):ARMCI DASSERT fail.
>> src/signaltrap.c:SigSegvHandler():312 cond:0
>> 2: ARMCI aborting 11 (0xb).
>> 2: ARMCI aborting 11 (0xb).
>> system error message: Bad file descriptor
>> Last System Error Message from Task 0:: Inappropriate ioctl for device
>> 0: ARMCI aborting 11 (0xb).
>> 0: ARMCI aborting 11 (0xb).
>> system error message: Inappropriate ioctl for device
>> 4: interrupt(1)
>> WaitAll: Child (21794) finished, status=0x100 (exited with code 1).
>> WaitAll: Child (21797) finished, status=0x100 (exited with code 1).
>> WaitAll: Child (21795) finished, status=0x100 (exited with code 1).
>> WaitAll: No children or error in wait?
>>
>>
>>
>> For the VMP2 input, the calculation reached the SURF calculation and then
>> stopped response after 2-D coupling potential
>>
>>
>> PROGRAM * SURF (Surface generation) Authors: G. Rauhut 2004, T. Hrenar
>> 2005
>> ...
>> Harmonic and 1D-anharmonic vibrational frequencies
>>
>> Mode Harmonic Diagonal Intens
>>
>> 3 B2 3970.77 4063.06 33.06
>> 2 A1 3851.51 3765.54 5.35
>> 1 A1 1677.46 1663.36 59.75
>>
>> Calculating 2D coupling potential
>>
>> 2D: 3 2 Points: 18 Conv: 0.04825417 0.04870899
>> 2D: 3 1 Points: 18 Conv: 0.00686833 0.00693136
>> 2D: 2 1 Points: 36 Conv: 0.00654850 0.00663013
>>
>> Calculating 3D coupling potential
>> 'stopped here'.
>>
>>
>> Best wishes,
>> Gang
>>
>>
>>
>>
>>
>> -----Original Message-----
>> From: molpro-user-bounces at molpro.net
>> [mailto:molpro-user-bounces at molpro.net]
>> On Behalf Of molpro-user-request at molpro.net
>> Sent: 05 August 2011 12:00
>> To: molpro-user at molpro.net
>> Subject: Molpro-user Digest, Vol 37, Issue 2
>>
>> Send Molpro-user mailing list submissions to
>> molpro-user at molpro.net
>>
>> To subscribe or unsubscribe via the World Wide Web, visit
>> http://www.molpro.net/mailman/listinfo/molpro-user
>> or, via email, send a message with subject or body 'help' to
>> molpro-user-request at molpro.net
>>
>> You can reach the person managing the list at
>> molpro-user-owner at molpro.net
>>
>> When replying, please edit your Subject line so it is more specific
>> than "Re: Contents of Molpro-user digest..."
>>
>>
>> Today's Topics:
>>
>> 1. Re: Permanent installation of dependencies (Andy May)
>>
>>
>> ----------------------------------------------------------------------
>>
>> Message: 1
>> Date: Thu, 04 Aug 2011 21:21:11 +0100
>> From: Andy May<MayAJ1 at cardiff.ac.uk>
>> To: Gregory Magoon<gmagoon at MIT.EDU>
>> Cc: molpro-user at molpro.net
>> Subject: Re: [molpro-user] Permanent installation of dependencies
>> Message-ID:<4E3AFF37.8050603 at cardiff.ac.uk>
>> Content-Type: text/plain; charset=ISO-8859-1; format=flowed
>>
>> Greg,
>>
>> Thanks for the info, perhaps it was auto-build with openmpi which could
>> not be relocated in this way. We have a bug open about the problem and
>> I'm working on a solution.
>>
>> mpptune.com can be run with, for example:
>>
>> ./bin/molpro -n 2 mpptune.com
>>
>> and should add parameters to lib/tuning.rc when complete, but will not
>> add parameters if only 1 process is used.
>>
>> I've attempted to make modifications to close the libmol.index file when
>> not in use. The change is available in the nightly build which you can
>> download from the website. Please let us know if this does makes a
>> difference.
>>
>> Best wishes,
>>
>> Andy
>>
>> On 25/07/11 16:13, Gregory Magoon wrote:
>>> I was able to get it to work (I think) for -auto-ga-tcgmsg-mpich2 by
>>> copying
>>> /src/mpich2-install to a permanent location and editing the LAUNCHER
>>> variable
>>> in the script file.
>>>
>>> A couple more quick questions:
>>> -Is use of mpptune.com recommended or is this deprecated? I think I was
>>> able to
>>> figure out how to run it, but I wasn't sure whether this is helpful or
>> not,
>>> performance-wise.
>>> -I noticed that a text file called libmol.index is kept open by molpro
>>> during
>>> execution (and possibly read from (?))...I'm thinking this may be
>> adversely
>>> affecting performance when the file happens to be on an NFS file system.
>>> If my
>>> suspicions are correct, is there any way the file can be cached by
>>> reading into
>>> memory?
>>>
>>> Thanks,
>>> Greg
>>>
>>> Quoting Andy May<MayAJ1 at cardiff.ac.uk>:
>>>
>>>> Greg,
>>>>
>>>> Glad to hear you got Molpro running. At the moment the auto-build
>>>> options are marked as experimental, however the only known issue is
>>>> the installation problem you mention. I'm not sure one can simply move
>>>> mpich2/openmpi installation directories, but we will try to implement
>>>> a solution.
>>>>
>>>> Best wishes,
>>>>
>>>> Andy
>>>>
>>>> On 24/07/11 04:12, Gregory Magoon wrote:
>>>>> After some work I was finally able to trace this to some sort of
>>>>> issue between
>>>>> NFSv4 and MPICH2; I can get this to work properly when I mount the
>>>>> NFS drives
>>>>> as NFSv3 (as opposed to NFSv4), so the issue is now more-or-less
>>>>> resolved.
>>>>>
>>>>> A quick follow-up question: Is there a recommended approach for
>>>>> permanent
>>>>> installation of the mpich2 dependency (and maybe also GA?) when using
>>>>> the auto
>>>>> build approach? By default, it seems that the installation scripts
>>>>> leave the
>>>>> mpiexec in the compile directory. I saw that the
>>>>> makefile/installation scripts
>>>>> mention an option called REDIST which seemed like it might allow this,
>>>>> but they
>>>>> don't seem to make use of this option (REDIST=NO).
>>>>>
>>>>> Thanks,
>>>>> Greg
>>>>>
>>>>> Quoting Gregory Magoon<gmagoon at MIT.EDU>:
>>>>>
>>>>>> Hi,
>>>>>> I have successfully compiled molpro (with Global Arrays/TCGMSG;
>>>>>> mpich2 from
>>>>>> Ubuntu package) on one of our compute nodes for our new server, and
>>>>>> installed
>>>>>> it in an NFS directory on our head node. The initial tests on the
>>>>>> compute node
>>>>>> ran fine but since the installation, I've had issues with running
>>>>>> molpro on the
>>>>>> compute nodes (it seems to work fine on the head node). Sometimes
>>>>>> (sorry I can't
>>>>>> be more precise, but it does not seem to be reproducible), when
>>>>>> running on the
>>>>>> compute node, the job will get stuck in the early stages, producing a
>>>>>> lot (~14+
>>>>>> Mbps outbound to headnode and 7Mbps inbound from headnode) of NFS
>>>>>> traffic and
>>>>>> causing fairly high nfsd process CPU% usage on the head node. Molpro
>>>>>> processes
>>>>>> in the stuck state are shown in "top" command display at the bottom
>>>>>> of the
>>>>>> e-mail. I have also attached example verbose output for a case that
>>>>>> works and a
>>>>>> case that gets stuck.
>>>>>>
>>>>>> Some notes:
>>>>>> -/usr/local is mounted as NFS read-only file system; /home is mounted
>>>>>> as NFS rw
>>>>>> file system
>>>>>> -It seems like runs with fewer processors (e.g. 6) are more likely
>>>>>> to run
>>>>>> successfully
>>>>>>
>>>>>> I've tried several approaches for addressing the issue, including 1.
>>>>>> Mounting
>>>>>> /usr/local as rw file system, and 2. Changing the rsize and wsize
>>>>>> parameters
>>>>>> for the NFS filesystem but none seem to work. We also tried piping<
>>>>>> /dev/null
>>>>>> when calling the process, which seemed like it was helping at first,
>>>>>> but later
>>>>>> tests suggested that this wasn't actually helping.
>>>>>>
>>>>>> If anyone has any tips or ideas to help diagnose the issue here, it
>>>>>> would be
>>>>>> greatly appreciated. If there are any additional details I can
>>>>>> provide to help
>>>>>> describe the problem, I'd be happy to provide them.
>>>>>>
>>>>>> Thanks very much,
>>>>>> Greg
>>>>>>
>>>>>> Top processes in "top" output in stuck state:
>>>>>> PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
>>>>>> 10 root 20 0 0 0 0 S 10 0.0 0:16.50 kworker/0:1
>>>>>> 2 root 20 0 0 0 0 S 6 0.0 0:10.86 kthreadd
>>>>>> 1496 root 20 0 0 0 0 S 1 0.0 0:04.73 kworker/0:2
>>>>>> 3 root 20 0 0 0 0 S 1 0.0 0:00.93 ksoftirqd/0
>>>>>>
>>>>>> Processes in "top" output for user in stuck state:
>>>>>> PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
>>>>>> 29961 user 20 0 19452 1508 1072 R 0 0.0 0:00.05 top
>>>>>> 1176 user 20 0 91708 1824 868 S 0 0.0 0:00.01 sshd
>>>>>> 1177 user 20 0 24980 7620 1660 S 0 0.0 0:00.41 bash
>>>>>> 1289 user 20 0 91708 1824 868 S 0 0.0 0:00.00 sshd
>>>>>> 1290 user 20 0 24980 7600 1640 S 0 0.0 0:00.32 bash
>>>>>> 1386 user 20 0 4220 664 524 S 0 0.0 0:00.01 molpro
>>>>>> 1481 user 20 0 18764 1196 900 S 0 0.0 0:00.00 mpiexec
>>>>>> 1482 user 20 0 18828 1092 820 S 0 0.0 0:00.00 hydra_pmi_proxy
>>>>>> 1483 user 20 0 18860 488 212 D 0 0.0 0:00.00 hydra_pmi_proxy
>>>>>> 1484 user 20 0 18860 488 212 D 0 0.0 0:00.00 hydra_pmi_proxy
>>>>>> 1485 user 20 0 18860 488 212 D 0 0.0 0:00.00 hydra_pmi_proxy
>>>>>> 1486 user 20 0 18860 488 212 D 0 0.0 0:00.00 hydra_pmi_proxy
>>>>>> 1487 user 20 0 18860 488 212 D 0 0.0 0:00.00 hydra_pmi_proxy
>>>>>> 1488 user 20 0 18860 488 212 D 0 0.0 0:00.00 hydra_pmi_proxy
>>>>>> 1489 user 20 0 18860 488 212 D 0 0.0 0:00.00 hydra_pmi_proxy
>>>>>> 1490 user 20 0 18860 488 208 D 0 0.0 0:00.00 hydra_pmi_proxy
>>>>>> 1491 user 20 0 18860 488 208 D 0 0.0 0:00.00 hydra_pmi_proxy
>>>>>> 1492 user 20 0 18860 488 208 D 0 0.0 0:00.00 hydra_pmi_proxy
>>>>>> 1493 user 20 0 18860 488 208 D 0 0.0 0:00.00 hydra_pmi_proxy
>>>>>> 1494 user 20 0 18860 492 212 D 0 0.0 0:00.00 hydra_pmi_proxy
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> Molpro-user mailing list
>>>>> Molpro-user at molpro.net
>>>>> http://www.molpro.net/mailman/listinfo/molpro-user
>>>>
>>>
>>>
>>
>>
>> ------------------------------
>>
>> _______________________________________________
>> Molpro-user mailing list
>> Molpro-user at molpro.net
>> http://www.molpro.net/mailman/listinfo/molpro-user
>>
>>
>> End of Molpro-user Digest, Vol 37, Issue 2
>> ******************************************
>>
>> _______________________________________________
>> Molpro-user mailing list
>> Molpro-user at molpro.net
>> http://www.molpro.net/mailman/listinfo/molpro-user
> _______________________________________________
> Molpro-user mailing list
> Molpro-user at molpro.net
> http://www.molpro.net/mailman/listinfo/molpro-user



More information about the Molpro-user mailing list