Page 1 of 1

Project: 2669 (Run 15, Clone 62, Gen 193)

Posted: Wed Aug 19, 2009 4:41 pm
by Torin3
I keep getting this one of my Ubuntu VMs. Even after deleting the queue.dat and client.cfg files along with the work directory.
It immediately cancels out with a "Range checking error"

Code: Select all

--- Opening Log file [August 16 23:14:56 UTC] 


# Linux SMP Console Edition ###################################################
###############################################################################

                       Folding@Home Client Version 6.24beta

                          http://folding.stanford.edu

###############################################################################
###############################################################################

Launch directory: /home/torin/folding/FAH
Executable: ./fah6
Arguments: -smp -verbosity 9 

[23:14:56] Configuring Folding@Home...


[23:15:04] - Ask before connecting: No
[23:15:04] - User name: Torin3 (Team 32)
[23:15:04] - User ID: 8FE186655B6ED8B
[23:15:04] - Machine ID: 1
[23:15:04] 
[23:15:04] Could not open work queue, generating new queue...
[23:15:04] - Preparing to get new work unit...
[23:15:04] + Attempting to get work packet
[23:15:04] - Will indicate memory of 981 MB
[23:15:04] - Connecting to assignment server
[23:15:04] Connecting to http://assign.stanford.edu:8080/
[23:15:04] - Autosending finished units... [August 16 23:15:04 UTC]
[23:15:04] Trying to send all finished work units
[23:15:04] + No unsent completed units remaining.
[23:15:04] - Autosend completed
[23:15:05] Posted data.
[23:15:05] Initial: 40AB; - Successful: assigned to (171.64.65.56).
[23:15:05] + News From Folding@Home: Welcome to Folding@Home
[23:15:05] Loaded queue successfully.
[23:15:05] Connecting to http://171.64.65.56:8080/
[23:15:13] Posted data.
[23:15:13] Initial: 0000; - Receiving payload (expected size: 1513252)
[23:15:16] - Downloaded at ~492 kB/s
[23:15:16] - Averaged speed for that direction ~492 kB/s
[23:15:16] + Received work.
[23:15:16] + Closed connections
[23:15:16] 
[23:15:16] + Processing work unit
[23:15:16] At least 4 processors must be requested.Core required: FahCore_a2.exe
[23:15:16] Core found.
[23:15:16] Working on queue slot 01 [August 16 23:15:16 UTC]
[23:15:16] + Working ...
[23:15:16] - Calling './mpiexec -np 4 -host 127.0.0.1 ./FahCore_a2.exe -dir work/ -suffix 01 -checkpoint 15 -verbose -lifeline 10554 -version 624'

[23:15:16] 
[23:15:16] *------------------------------*
[23:15:16] Folding@Home Gromacs SMP Core
[23:15:16] Version 2.07 (Sun Apr 19 14:51:09 PDT 2009)
[23:15:16] 
[23:15:16] Preparing to commence simulation
[23:15:16] - Ensuring status. Please wait.
[23:15:17] Called DecompressByteArray: compressed_data_size=1512740 data_size=23979129, decompressed_data_size=23979129 diff=0
[23:15:17] - Digital signature verified
[23:15:17] 
[23:15:17] Project: 2669 (Run 15, Clone 62, Gen 193)
[23:15:17] 
[23:15:17] Assembly optimizations on if available.
[23:15:17] Entering M.D.
[23:15:27] un 15, Clone 62, Gen 193)
[23:15:27] 
[23:15:27] Entering M.D.
[23:23:19] ***** Got an Activate signal (2)
[23:23:19] Killing all core threads

Folding@Home Client Shutdown.

Code: Select all

[23:15:04] - Ask before connecting: No
[23:15:04] - User name: Torin3 (Team 32)
[23:15:04] - User ID: 8FE186655B6ED8B
[23:15:04] - Machine ID: 1
[23:15:04] 
[23:15:04] Could not open work queue, generating new queue...
[23:15:04] - Preparing to get new work unit...
[23:15:04] + Attempting to get work packet
[23:15:04] - Will indicate memory of 981 MB
[23:15:04] - Connecting to assignment server
[23:15:04] Connecting to http://assign.stanford.edu:8080/
[23:15:04] - Autosending finished units... [August 16 23:15:04 UTC]
[23:15:04] Trying to send all finished work units
[23:15:04] + No unsent completed units remaining.
[23:15:04] - Autosend completed
[23:15:05] Posted data.
[23:15:05] Initial: 40AB; - Successful: assigned to (171.64.65.56).
[23:15:05] + News From Folding@Home: Welcome to Folding@Home
[23:15:05] Loaded queue successfully.
[23:15:05] Connecting to http://171.64.65.56:8080/
[23:15:13] Posted data.
[23:15:13] Initial: 0000; - Receiving payload (expected size: 1513252)
[23:15:16] - Downloaded at ~492 kB/s
[23:15:16] - Averaged speed for that direction ~492 kB/s
[23:15:16] + Received work.
[23:15:16] + Closed connections
[23:15:16] 
[23:15:16] + Processing work unit
[23:15:16] At least 4 processors must be requested.Core required: 
FahCore_a2.exe
[23:15:16] Core found.
[23:15:16] Working on queue slot 01 [August 16 23:15:16 UTC]
[23:15:16] + Working ...
[23:15:16] - Calling './mpiexec -np 4 -host 127.0.0.1 ./FahCore_a2.exe 
-dir work/ -suffix 01 -checkpoint 15 -verbose -lifeline 10554 -version 
624'

[23:15:16] 
[23:15:16] *------------------------------*
[23:15:16] Folding@Home Gromacs SMP Core
[23:15:16] Version 2.07 (Sun Apr 19 14:51:09 PDT 2009)
[23:15:16] 
[23:15:16] Preparing to commence simulation
[23:15:16] - Ensuring status. Please wait.
[23:15:17] Called DecompressByteArray: compressed_data_size=1512740 
data_size=23979129, decompressed_data_size=23979129 diff=0
[23:15:17] - Digital signature verified
[23:15:17] 
[23:15:17] Project: 2669 (Run 15, Clone 62, Gen 193)
[23:15:17] 
[23:15:17] Assembly optimizations on if available.
[23:15:17] Entering M.D.
[23:15:27] un 15, Clone 62, Gen 193)
[23:15:27] 
[23:15:27] Entering M.D.
NNODES=4, MYRANK=1, HOSTNAME=torin-desktop
NNODES=4, MYRANK=3, HOSTNAME=torin-desktop
NNODES=4, MYRANK=2, HOSTNAME=torin-desktop
NNODES=4, MYRANK=0, HOSTNAME=torin-desktop
NODEID=2 argc=20
NODEID=3 argc=20
NODEID=0 argc=20
NODEID=1 argc=20
                         :-)  G  R  O  M  A  C  S  (-:

                   Groningen Machine for Chemical Simulation

                 :-)  VERSION 4.0.99_development_20090307  (-:


      Written by David van der Spoel, Erik Lindahl, Berk Hess, and 
others.
       Copyright (c) 1991-2000, University of Groningen, The 
Netherlands.
             Copyright (c) 2001-2008, The GROMACS development team,
            check out http://www.gromacs.org for more information.


                                :-)  mdrun  (-:

Reading file work/wudata_01.tpr, VERSION 3.3.99_development_20070618 
(single precision)
Note: tpx file_version 48, software version 64

NOTE: The tpr file used for this simulation is in an old format, for 
less memory usage and possibly more performance create a new tpr file 
with an up to date version of grompp

Making 1D domain decomposition 1 x 1 x 4
starting mdrun '22869 system'
48500000 steps,  97000.0 ps (continuing from step 48250000,  96500.0 
ps).

-------------------------------------------------------
Program mdrun, VERSION 4.0.99_development_20090307
Source code file: nsgrid.c, line: 357

Range checking error:
Explanation: During neighborsearching, we assign each particle to a grid
based on its coordinates. If your system contains collisions or 
parameter
errors that give particles very high velocities you might end up with 
some
coordinates being +-Infinity or NaN (not-a-number). Obviously, we cannot
put these on a grid, so this is usually where we detect those errors.
Make sure your system is properly energy-minimized and that the 
potential
energy seems reasonable before trying again.

Variable ci has value -2147483269. It should have been within [ 0 .. 
9464 ]

For more information and tips for trouble shooting please check the 
GROMACS Wiki at
http://wiki.gromacs.org/index.php/Errors
-------------------------------------------------------

Thanx for Using GROMACS - Have a Nice Day

Error on node 0, will try to stop all the nodes
Halting parallel program mdrun on CPU 0 out of 4

gcq#0: Thanx for Using GROMACS - Have a Nice Day

[cli_0]: aborting job:
application called MPI_Abort(MPI_COMM_WORLD, -1) - process 0

-------------------------------------------------------
Program mdrun, VERSION 4.0.99_development_20090307
Source code file: nsgrid.c, line: 357

Range checking error:
Explanation: During neighborsearching, we assign each particle to a grid
based on its coordinates. If your system contains collisions or 
parameter
errors that give particles very high velocities you might end up with 
some
coordinates being +-Infinity or NaN (not-a-number). Obviously, we cannot
put these on a grid, so this is usually where we detect those errors.
Make sure your system is properly energy-minimized and that the 
potential
energy seems reasonable before trying again.

Variable ci has value -2147483611. It should have been within [ 0 .. 256 
]

For more information and tips for trouble shooting please check the 
GROMACS Wiki at
http://wiki.gromacs.org/index.php/Errors
-------------------------------------------------------

Thanx for Using GROMACS - Have a Nice Day

Error on node 3, will try to stop all the nodes
Halting parallel program mdrun on CPU 3 out of 4

gcq#0: Thanx for Using GROMACS - Have a Nice Day

[cli_3]: aborting job:
application called MPI_Abort(MPI_COMM_WORLD, -1) - process 3
[23:23:19] ***** Got an Activate signal (2)
[23:23:19] Killing all core threads

Folding@Home Client Shutdown.

Re: Project: 2669 (Run 15, Clone 62, Gen 193)

Posted: Wed Aug 19, 2009 4:47 pm
by toTOW
This is a common issue those days.

I add your WU to the list : viewtopic.php?f=19&t=11098