option, parallelism,...
Moderators: fgoudreault, mcote
Forum rules
Please have a look at ~abinit/doc/config/build-config.ac in the source package for detailed and up-to-date information about the configuration of Abinit 8 builds.
For a video explanation on how to build Abinit 7.x for Linux, please go to:
http://www.youtube.com/watch?v=DppLQ-KQA68.
IMPORTANT: when an answer solves your problem, please check the little green V-like button on its upper-right corner to accept it.
-
arras
- Posts: 16
- Joined: Mon Aug 17, 2009 9:17 am
- Location: Munich
Post
by arras » Thu Aug 19, 2010 4:15 pm
Hi,
I'm trying to perform a parallel calculation on 4 procs using the syntax:
abinit -n 4 < ab.files > log
And I get the following message:
MPI has run out of internal group entries.
Please set the environment variable MPI_GROUP_MAX for additional space.
The current value of MPI_GROUP_MAX is 32
MPI has run out of internal group entries.
Please set the environment variable MPI_GROUP_MAX for additional space.
The current value of MPI_GROUP_MAX is 32
MPI: MPI_COMM_WORLD rank 0 has terminated without calling MPI_Finalize()
MPI: aborting job
There are 20 kpoints in my calculation and paral_kgb=0. Yet other calculations with less kpoints have passed allright.
Thus I confess I don't quite understand the complain...
I'm using Abinit 6.0.3 on SGI Altix 4700 (
http://www.lrz.de/services/compute/hlrb/)
Thank you for any help..
Emmanuel ARRAS
Physikdepartemnt E20
Technische Universität München
James-Franck Straße 1
D-85748 Garching
-
mcote
- Posts: 16
- Joined: Fri Aug 14, 2009 3:23 pm
- Location: Montreal, Canada
-
Contact:
Post
by mcote » Mon Aug 23, 2010 3:37 pm
Hi,
We had similar problems on our SGI machine some times ago. We had to set the environment variables (in bash):
export MPI_GROUP_MAX=20000
export MPI_COMM_MAX=1000
We set these variables in the submission script, PBS in our case. I know that ABINIT uses one group(maybe more) per k-point. Somehow, the default for these environment variables in the SGI implementation is lower than on other machines.
Michel
-
arras
- Posts: 16
- Joined: Mon Aug 17, 2009 9:17 am
- Location: Munich
Post
by arras » Wed Aug 25, 2010 12:15 pm
All right, I'll try that.
Thank you for the tip !
Emmanuel
Physikdepartemnt E20
Technische Universität München
James-Franck Straße 1
D-85748 Garching
-
arras
- Posts: 16
- Joined: Mon Aug 17, 2009 9:17 am
- Location: Munich
Post
by arras » Fri Dec 03, 2010 12:25 pm
Hi again,
I'm using paral_kgb parallelisation with the following parameters:
paral_kgb 1
nband 448
npband 16
npfft 1
npkpt 4
acell 19.537533 28.2 12 Angstrom
on 64 processors with Abinit 6.4.1. The calculations runs allright, but the writting of the WFK file fails. The complain is :
Please set the environment variable MPI_TYPE_MAX for additional space.
The current value of MPI_TYPE_MAX is 8192
So I increased MPI_TYPE_MAX up to 1048576 ! And it still complains (yet on less lines). Is that normal ?? Is there a limit I should not cross on the number MPI_TYPE_MAX ? Or can I increase it until it works ?
Thanks !
Emmanuel
Physikdepartemnt E20
Technische Universität München
James-Franck Straße 1
D-85748 Garching
-
mcote
- Posts: 16
- Joined: Fri Aug 14, 2009 3:23 pm
- Location: Montreal, Canada
-
Contact:
Post
by mcote » Wed Dec 15, 2010 3:45 pm
I never had this problem but I presume that is because you are using MPI IO? I searched the web and it seems to be a specific SGI problem, for example see:
linux.math.tifr.res.in/programming-doc/lam/lam-users-guide.ps
and search for "for additional space".
It looks like you are doing the right thing to increase MPI_TYPE_MAX. Maybe you should increase it more...
Michel