Hello,
What is your PartAllocFactor parameter?
Maybe that is too low
Cheers,
Antonio
On Jul 28, 2015 2:00 PM, "Sandeep Kumar Kataria" <
sandeep.kataria_at_iiap.res.in> wrote:
>
> Hello Everyone,
>
> I am getting following error while running the code with initial
> conditions generated by NEMO package. I checked initial condition can be
> read with splash and the snapread.c file.
>
> >
> This is Gadget, version `2.0'.
>
> Running on 2 processors.
>
> Allocated 25 MByte communication buffer per processor.
>
> Communication buffer has room for 504122 particles in gravity computation
> Communication buffer has room for 204800 particles in density computation
> Communication buffer has room for 163840 particles in hydro computation
> Communication buffer has room for 163840 particles in domain decomposition
>
>
> Hubble (internal units) = 0.1
> G (internal units) = 43007.1
> UnitMass_in_g = 1.989e+43
> UnitTime_in_s = 3.08568e+16
> UnitVelocity_in_cm_per_s = 100000
> UnitDensity_in_cgs = 6.76991e-22
> UnitEnergy_in_cgs = 1.989e+53
>
>
> reading file `/home/sandeep/Gadget-2.0.7/ICs/Galaxy_checkg' on task=0
> (contains 15000 particles.)
> distributing this file to tasks 0-1
> Type 0 (gas): 0 (tot= 0000000000) masstab=0
> Type 1 (halo): 10000 (tot= 0000010000) masstab=0
> too many particles
> task 1: endrun called with an error level of 1313
>
>
> Type 2 (disk): 5000 (tot= 0000005000) masstab=0.0005
> Type 3 (bulge): 0 (tot= 0000000000) masstab=0
> Type 4 (stars): 0 (tot= 0000000000) masstab=0
> Type 5 (bndry): 0 (tot= 0000000000) masstab=0
>
> too many particles
> task 0: endrun called with an error level of 1313
>
>
> --------------------------------------------------------------------------
> MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD
> with errorcode 1313.
>
> NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.
> You may or may not see output from other processes, depending on
> exactly when Open MPI kills them.
>
>
>
>
> Regards,
> Sandeep Kumar
> Indian Institute of Astrophysics
> Bangalore,India
>
>
>
>
> -----------------------------------------------------------
> If you wish to unsubscribe from this mailing, send mail to
> minimalist_at_MPA-Garching.MPG.de with a subject of: unsubscribe gadget-list
> A web-archive of this mailing list is available here:
> http://www.mpa-garching.mpg.de/gadget/gadget-list
>
Received on 2015-07-28 14:16:45