[molpro-user] parallel molpro behavior
Nick Wilson
WilsonNT at Cardiff.ac.uk
Thu Sep 2 15:32:42 BST 2004
Dear Sigismondo,
The scratch files 1 (basis set, geometry, and the one and two electron
integrals), 2 (orbitals, CI coefficients, and density matrices) and 3
(auxiliary restart file) ignore the "default implementation of scratch
files" setting and are always set to direct files (df) unless explicitly
set otherwise using:
file,1,implementation=sf;
file,2,implementation=sf;
file,3,implementation=sf;
with sf for shared files, ga for global arrays, etc
I'm not sure if using the ga or sf implementation for these scratch
files will work in all cases as it has never been tested. I have run a
simple testjob within a regatta node using the sf implementation and
that appeared to work and another very simple testjob between two
regatta nodes which also appeared to work. Larger tests are waiting in
the queue.
For debugging I/O information you can use:
gprint,io
Best wishes,
Nick Wilson
Sigismondo Boschi wrote:
> I am running for the very first times molpro in parallel on an IBM
> regatta (p690) system with colony switch.
>
> The typical target systems of our users are CCSD(T) optimnization.
>
>
> Running one of them, I would like to use as much as possibile the memory
> for integrals, and then, if not possibile, the disk, that is a shared
> GPFS filesystem: there is no benefit in concurrent accesses to it, and
> it became the bottleneck for the code.
>
> With typical options I have found that the task don't use much memory,
> nor standard neither in GA.
>
> Running on 16 cpus, At the very beginning of the output I found:
>
>
> **********
> ARMCI configured for 2 cluster nodes
>
> MPP nodes nproc
> sp154 8
> sp152 8
> ga_uses_ma=false, calling ma_init with nominal heap. Any -G option will
> be ignored.
>
> Primary working directories: /scratch_ssa/abc0
> Secondary working directories: /scratch_ssa/abc0
>
> blaslib=default
>
> MPP tuning parameters: Latency= 84 Microseconds, Broadcast speed=
> 233 MB/sec
> default implementation of scratch files=ga
> **********
>
> Only if I use one task (or maybe, one node) I can find ga_uses_ma=true
> on the other side the statement: "default implementation of scratch
> files=ga" would let me think that they are "in-memory files"... however
> what happend at run-time does not correspond to it:
>
> In fact I observe a lot of I/O, and the used memory is about 200 MB (of
> 2GB) for each task.
>
> After the CCSD I get:
> DISK USED * 9.10 GB
> GA USED * 120.58 MB (max) .00 MB (current)
>
> And actually I set in the beginning:
> memory,200,M
>
> (that is not the GA memory, but the -G option is ignored... I do not
> understand why).
>
>
> Can anybody of you explain some of these facts, and give some suggestion
> for parallel runs?
>
> For istance I tried also direct calculations, but:
> 1. it was very slow
> 2. it terminates with the error:
>
> ******
> FILE 5 RECORD 1380 OFFSET= 0. NOT FOUND
>
> Records on file 5
>
> IREC NAME TYPE OFFSET LENGTH IMPLEMENTATION EXT PREV
> PARENT MPP_STATE
> 1 4000 4096. 21301. df 0 0
> 0 1
> 2 4001 25397. 166404. df 0 0
> 0 1
> 3 4002 191801. 10725. df 0 0
> 0 0
> 4 4003 202526. 178782. df 0 0
> 0 1
> 5 35020 381308. 10496. df 0 0
> 0 1
> 6 3600 391804. 273. df 0 0
> 0 1
> 7 3601 392077. 273. df 0 0
> 0 1
> 8 35000 392350. 10. df 0 0
> 0 1
> 9 35001 392360. 10. df 0 0
> 0 1
> 10 35010 392370. 320. df 0 0
> 0 1
> 11 35011 392690. 320. df 0 0
> 0 1
> 12 7005 393010. 314964. df 0 0
> 0 1
> 13 8005 707974. 314964. df 0 0
> 0 1
> 14 9101 1022938. 9567696. df 0 0
> 0 0
> 15 9103 10590634. 9567696. df 0 0
> 0 0
>
> ? Error
> ? Record not found
> ? The problem occurs in readm
>
> ERROR EXIT
> CURRENT STACK: CIPRO MAIN
> *******
>
> Many Thanks for any help
>
> Regards
>
>
> Sigismondo Boschi
>
>
>
>
More information about the Molpro-user
mailing list