[molpro-user] VMP2, VCI calculations teminated with 'segmentation Violation error' , using downloaded MOLPRO 2010.1 binary code

Gang Li gl525 at york.ac.uk
Fri Aug 5 13:30:01 BST 2011


Dear Molpro Users,

I recently downloaded the new 2010.1 binary version. However, quick tests
using the VMP2 sample input in the Molpro 2010.1 Manual crash every time
with the following error in terminal.   Similar error message always shows
up with the VCI sample input and the CBS frequency input in the manual.
The system OS is Ubuntu 10.04 LTS and the 2008 version I was using worked
smoothly.  Anyone knows the reason? Thanks!


Creating: host=virgo, user=root,
           file=/usr/local/bin/molprop_2010_1_Linux_x86_64_i8.exe,
port=53336
3:Segmentation Violation error, status=: 11
(rank:3 hostname:virgo pid:21797):ARMCI DASSERT fail.
src/signaltrap.c:SigSegvHandler():312 cond:0
1:Segmentation Violation error, status=: 11
(rank:1 hostname:virgo pid:21795):ARMCI DASSERT fail.
src/signaltrap.c:SigSegvHandler():312 cond:0
Last System Error Message from Task 3:: Bad file descriptor
Last System Error Message from Task 1:: Bad file descriptor
  3: ARMCI aborting 11 (0xb).
  3: ARMCI aborting 11 (0xb).
  1: ARMCI aborting 11 (0xb).
  1: ARMCI aborting 11 (0xb).
system error message: Bad file descriptor
0:Segmentation Violation error, status=: 11
2:Segmentation Violation error, status=: 11
system error message: Bad file descriptor
(rank:2 hostname:virgo pid:21796):ARMCI DASSERT fail.
src/signaltrap.c:SigSegvHandler():312 cond:0
Last System Error Message from Task 2:: Bad file descriptor
(rank:0 hostname:virgo pid:21794):ARMCI DASSERT fail.
src/signaltrap.c:SigSegvHandler():312 cond:0
  2: ARMCI aborting 11 (0xb).
  2: ARMCI aborting 11 (0xb).
system error message: Bad file descriptor
Last System Error Message from Task 0:: Inappropriate ioctl for device
  0: ARMCI aborting 11 (0xb).
  0: ARMCI aborting 11 (0xb).
system error message: Inappropriate ioctl for device
  4: interrupt(1)
WaitAll: Child (21794) finished, status=0x100 (exited with code 1).
WaitAll: Child (21797) finished, status=0x100 (exited with code 1).
WaitAll: Child (21795) finished, status=0x100 (exited with code 1).
WaitAll: No children or error in wait?



For the VMP2 input, the calculation reached the SURF calculation and then
stopped response after 2-D coupling potential


PROGRAM * SURF (Surface generation)   Authors: G. Rauhut 2004, T. Hrenar
2005
... 
 Harmonic and 1D-anharmonic vibrational frequencies

   Mode     Harmonic     Diagonal     Intens

   3 B2      3970.77      4063.06      33.06
   2 A1      3851.51      3765.54       5.35
   1 A1      1677.46      1663.36      59.75

 Calculating 2D coupling potential

 2D:   3   2  Points:  18 Conv: 0.04825417 0.04870899
 2D:   3   1  Points:  18 Conv: 0.00686833 0.00693136
 2D:   2   1  Points:  36 Conv: 0.00654850 0.00663013

 Calculating 3D coupling potential
 'stopped here'.


Best wishes,
Gang





-----Original Message-----
From: molpro-user-bounces at molpro.net [mailto:molpro-user-bounces at molpro.net]
On Behalf Of molpro-user-request at molpro.net
Sent: 05 August 2011 12:00
To: molpro-user at molpro.net
Subject: Molpro-user Digest, Vol 37, Issue 2

Send Molpro-user mailing list submissions to
	molpro-user at molpro.net

To subscribe or unsubscribe via the World Wide Web, visit
	http://www.molpro.net/mailman/listinfo/molpro-user
or, via email, send a message with subject or body 'help' to
	molpro-user-request at molpro.net

You can reach the person managing the list at
	molpro-user-owner at molpro.net

When replying, please edit your Subject line so it is more specific
than "Re: Contents of Molpro-user digest..."


Today's Topics:

   1. Re: Permanent installation of dependencies (Andy May)


----------------------------------------------------------------------

Message: 1
Date: Thu, 04 Aug 2011 21:21:11 +0100
From: Andy May <MayAJ1 at cardiff.ac.uk>
To: Gregory Magoon <gmagoon at MIT.EDU>
Cc: molpro-user at molpro.net
Subject: Re: [molpro-user] Permanent installation of dependencies
Message-ID: <4E3AFF37.8050603 at cardiff.ac.uk>
Content-Type: text/plain; charset=ISO-8859-1; format=flowed

Greg,

Thanks for the info, perhaps it was auto-build with openmpi which could 
not be relocated in this way. We have a bug open about the problem and 
I'm working on a solution.

mpptune.com can be run with, for example:

./bin/molpro -n 2 mpptune.com

and should add parameters to lib/tuning.rc when complete, but will not 
add parameters if only 1 process is used.

I've attempted to make modifications to close the libmol.index file when 
not in use. The change is available in the nightly build which you can 
download from the website. Please let us know if this does makes a 
difference.

Best wishes,

Andy

On 25/07/11 16:13, Gregory Magoon wrote:
> I was able to get it to work (I think) for -auto-ga-tcgmsg-mpich2 by
> copying
> /src/mpich2-install to a permanent location and editing the LAUNCHER
> variable
> in the script file.
>
> A couple more quick questions:
> -Is use of mpptune.com recommended or is this deprecated? I think I was
> able to
> figure out how to run it, but I wasn't sure whether this is helpful or
not,
> performance-wise.
> -I noticed that a text file called libmol.index is kept open by molpro
> during
> execution (and possibly read from (?))...I'm thinking this may be
adversely
> affecting performance when the file happens to be on an NFS file system.
> If my
> suspicions are correct, is there any way the file can be cached by
> reading into
> memory?
>
> Thanks,
> Greg
>
> Quoting Andy May <MayAJ1 at cardiff.ac.uk>:
>
>> Greg,
>>
>> Glad to hear you got Molpro running. At the moment the auto-build
>> options are marked as experimental, however the only known issue is
>> the installation problem you mention. I'm not sure one can simply move
>> mpich2/openmpi installation directories, but we will try to implement
>> a solution.
>>
>> Best wishes,
>>
>> Andy
>>
>> On 24/07/11 04:12, Gregory Magoon wrote:
>>> After some work I was finally able to trace this to some sort of
>>> issue between
>>> NFSv4 and MPICH2; I can get this to work properly when I mount the
>>> NFS drives
>>> as NFSv3 (as opposed to NFSv4), so the issue is now more-or-less
>>> resolved.
>>>
>>> A quick follow-up question: Is there a recommended approach for
>>> permanent
>>> installation of the mpich2 dependency (and maybe also GA?) when using
>>> the auto
>>> build approach? By default, it seems that the installation scripts
>>> leave the
>>> mpiexec in the compile directory. I saw that the
>>> makefile/installation scripts
>>> mention an option called REDIST which seemed like it might allow this,
>>> but they
>>> don't seem to make use of this option (REDIST=NO).
>>>
>>> Thanks,
>>> Greg
>>>
>>> Quoting Gregory Magoon<gmagoon at MIT.EDU>:
>>>
>>>> Hi,
>>>> I have successfully compiled molpro (with Global Arrays/TCGMSG;
>>>> mpich2 from
>>>> Ubuntu package) on one of our compute nodes for our new server, and
>>>> installed
>>>> it in an NFS directory on our head node. The initial tests on the
>>>> compute node
>>>> ran fine but since the installation, I've had issues with running
>>>> molpro on the
>>>> compute nodes (it seems to work fine on the head node). Sometimes
>>>> (sorry I can't
>>>> be more precise, but it does not seem to be reproducible), when
>>>> running on the
>>>> compute node, the job will get stuck in the early stages, producing a
>>>> lot (~14+
>>>> Mbps outbound to headnode and 7Mbps inbound from headnode) of NFS
>>>> traffic and
>>>> causing fairly high nfsd process CPU% usage on the head node. Molpro
>>>> processes
>>>> in the stuck state are shown in "top" command display at the bottom
>>>> of the
>>>> e-mail. I have also attached example verbose output for a case that
>>>> works and a
>>>> case that gets stuck.
>>>>
>>>> Some notes:
>>>> -/usr/local is mounted as NFS read-only file system; /home is mounted
>>>> as NFS rw
>>>> file system
>>>> -It seems like runs with fewer processors (e.g. 6) are more likely
>>>> to run
>>>> successfully
>>>>
>>>> I've tried several approaches for addressing the issue, including 1.
>>>> Mounting
>>>> /usr/local as rw file system, and 2. Changing the rsize and wsize
>>>> parameters
>>>> for the NFS filesystem but none seem to work. We also tried piping<
>>>> /dev/null
>>>> when calling the process, which seemed like it was helping at first,
>>>> but later
>>>> tests suggested that this wasn't actually helping.
>>>>
>>>> If anyone has any tips or ideas to help diagnose the issue here, it
>>>> would be
>>>> greatly appreciated. If there are any additional details I can
>>>> provide to help
>>>> describe the problem, I'd be happy to provide them.
>>>>
>>>> Thanks very much,
>>>> Greg
>>>>
>>>> Top processes in "top" output in stuck state:
>>>> PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
>>>> 10 root 20 0 0 0 0 S 10 0.0 0:16.50 kworker/0:1
>>>> 2 root 20 0 0 0 0 S 6 0.0 0:10.86 kthreadd
>>>> 1496 root 20 0 0 0 0 S 1 0.0 0:04.73 kworker/0:2
>>>> 3 root 20 0 0 0 0 S 1 0.0 0:00.93 ksoftirqd/0
>>>>
>>>> Processes in "top" output for user in stuck state:
>>>> PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
>>>> 29961 user 20 0 19452 1508 1072 R 0 0.0 0:00.05 top
>>>> 1176 user 20 0 91708 1824 868 S 0 0.0 0:00.01 sshd
>>>> 1177 user 20 0 24980 7620 1660 S 0 0.0 0:00.41 bash
>>>> 1289 user 20 0 91708 1824 868 S 0 0.0 0:00.00 sshd
>>>> 1290 user 20 0 24980 7600 1640 S 0 0.0 0:00.32 bash
>>>> 1386 user 20 0 4220 664 524 S 0 0.0 0:00.01 molpro
>>>> 1481 user 20 0 18764 1196 900 S 0 0.0 0:00.00 mpiexec
>>>> 1482 user 20 0 18828 1092 820 S 0 0.0 0:00.00 hydra_pmi_proxy
>>>> 1483 user 20 0 18860 488 212 D 0 0.0 0:00.00 hydra_pmi_proxy
>>>> 1484 user 20 0 18860 488 212 D 0 0.0 0:00.00 hydra_pmi_proxy
>>>> 1485 user 20 0 18860 488 212 D 0 0.0 0:00.00 hydra_pmi_proxy
>>>> 1486 user 20 0 18860 488 212 D 0 0.0 0:00.00 hydra_pmi_proxy
>>>> 1487 user 20 0 18860 488 212 D 0 0.0 0:00.00 hydra_pmi_proxy
>>>> 1488 user 20 0 18860 488 212 D 0 0.0 0:00.00 hydra_pmi_proxy
>>>> 1489 user 20 0 18860 488 212 D 0 0.0 0:00.00 hydra_pmi_proxy
>>>> 1490 user 20 0 18860 488 208 D 0 0.0 0:00.00 hydra_pmi_proxy
>>>> 1491 user 20 0 18860 488 208 D 0 0.0 0:00.00 hydra_pmi_proxy
>>>> 1492 user 20 0 18860 488 208 D 0 0.0 0:00.00 hydra_pmi_proxy
>>>> 1493 user 20 0 18860 488 208 D 0 0.0 0:00.00 hydra_pmi_proxy
>>>> 1494 user 20 0 18860 492 212 D 0 0.0 0:00.00 hydra_pmi_proxy
>>>>
>>>>
>>>>
>>>>
>>>
>>>
>>> _______________________________________________
>>> Molpro-user mailing list
>>> Molpro-user at molpro.net
>>> http://www.molpro.net/mailman/listinfo/molpro-user
>>
>
>


------------------------------

_______________________________________________
Molpro-user mailing list
Molpro-user at molpro.net
http://www.molpro.net/mailman/listinfo/molpro-user


End of Molpro-user Digest, Vol 37, Issue 2
******************************************




More information about the Molpro-user mailing list