The issue seems to occur inside the strdep.f90 routine, which calculates the structure displacement.
The same ALE simulation runs fine with other turbulence models.
The error appears after a few (10 more or less) time steps after activating the ALE implicit coupling module (following a fluid initialization without FSI). I do not see inconsistent values in terms of mesh displacement etc before the error.
Could it be an incomptability between the EBRSM and the implicit coupling?
Thank you very much in advance.
Best regards,
Daniele
I am not aware of any known incompatibility between ALE and EBRSM.
What type of crash do you have (i.e. in the end of the "listing"or in the error_* files). A floating-point exception, a memory error (SIGSEV), something else ?
Do you have a small test case you could share on which we could reproduce/debug this ?
It seems the the code gets stuck inside the strdep.f90 routine.
I do not have a small test case right now. I can try to create one and test it first by my own.
But since I have the error only with the EBRSM, I was wondering if there could be a specific variable (for calculating the force for example) specific to the EBRSM that could cause the issue. It is not a very detailed analysis I know...
Yes, I saw the call stack, but do you have any associated error message ? Possibly in a n error_r* file ?
I do not think there is anything specific to EBRSM, so I would suspect a memory overwrite error somewhere. Or possibly a name collision. The rest of the run_solver.log may help here.
Sorry I didn't get what you meant.
The error file contains exactly the call stack details shown in my previous post.
I had a look inside the run_solver.log and actually found something more:
What I mean is "are there any other files named "error_r*" matching error files for ranks other than 0 ?" When they are present, they contain the more useful information.
If you have no such files, then you might have some error message in the terminal from which the code was run, or in cas of a batch system, in the job log and error files (this happens for example when the Fortran runtime detects an error).
Unfortunately there is no error file created beside the one describing exactly the same error found in the listing.
The output file of the batch job gives the following error message (I have no error file created):
**********************
Starting calculation
**********************
--------------------------------------------------------------------------
MPI_ABORT was invoked on rank 0 in communicator MPI_COMM_WORLD
with errorcode 1.
NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.
You may or may not see output from other processes, depending on
exactly when Open MPI kills them.
--------------------------------------------------------------------------
solver script exited with status 1.
Error running the calculation.
Check Code_Saturne log (listing) and error* files for details.
*****************************
Post-calculation operations
*****************************
Error in calculation stage.
It is probably not enough to help understanding...
If the case is not too large, running it in serial mode may help provide information about the crash.
Otherwise, a small test case would allow debugging.