furstj/myFoam

hypersonicCylinder bug in parallel

jiaqiwang969 opened this issue · 4 comments

Hi, when running the example hypersonicCylinder through parallel, the results diverge.
It's Ok in non-parallel.

 Time = 35
LUSGS: iteration 1
[1] #0  Foam::error::printStack(Foam::Ostream&) at ??:?
[1] #1  Foam::sigFpe::sigHandler(int) at ??:?
[1] #2  ? in /lib/x86_64-linux-gnu/libpthread.so.0
[1] #3  ? in /home/ofuser/OpenFOAM/ofuser-v2006/platforms/linux64GccDPInt32Opt/bin/myLusgsFoam
[1] #4  __libc_start_main in /lib/x86_64-linux-gnu/libc.so.6
[1] #5  ? in /home/ofuser/OpenFOAM/ofuser-v2006/platforms/linux64GccDPInt32Opt/bin/myLusgsFoam
srun: error: cas643: task 1: Floating point exception
srun: Terminating job step 9600744.0
slurmstepd: error: *** STEP 9600744.0 ON cas643 CANCELLED AT 2022-02-14T11:00:02 ***
srun: error: cas643: tasks 0,2-3: Terminated
srun: Force Terminated job step 9600744.0
Execution finished
Script done

How many CPUs did you use? Can you send me your system/decomposeParDict?

It follows, this is defalt (no modified)

numberOfSubdomains  4;

method              hierarchical;

simpleCoeffs
{
    n               (2 2 1);
    delta           0.001;
}

hierarchicalCoeffs
{
    n               (2 2 1);
    delta           0.001;
    order           xyz;
}

manualCoeffs
{
    dataFile        "cellDecomposition";
}

I will try to find the bug. I did also some computations with "method scotch;" with several values of "numberOfSubdomains". It seems that the code works well for odd number of subdomains and crashes for even number of subdomains!

I did a change in the setup of the case. It seems that the original setup was too aggressive and the parallel run didn't survive it.