Lisandro Dalcin
Lisandro Dalcin
I can repeat my local tests tomorrow with current main and then report the outcome.
Folks, I've updated the description. All my local tests are with ompi@main. @janjust My CI also failed with deadlock using [email protected], see [here](https://github.com/mpi4py/mpi4py-testing/actions/runs/7894698961/job/21545827190).
@hppritcha This is what I get from `mpiexec --display allocation ...` ``` ====================== ALLOCATED NODES ====================== kw61149: slots=1 max_slots=0 slots_inuse=0 state=UP Flags: DAEMON_LAUNCHED:LOCATION_VERIFIED aliases: kw61149 ================================================================= ====================== ALLOCATED NODES ======================...
@rhc54 I'll submit a trivial reproducer here as soon as I can. The issue is not particular of my testsuite, any spawn example in singleton init mode with a relocated...
Sorry, I mixed up issues, I was talking about #12349. Regarding spawn testsuites, what mine does that probably no other one does is to issue spawn/spawn_multiple calls in rapid succession...
@rhc54 @hppritcha Here you have a C reproducer, as simple as it can get. ```c #include #include #include int main(int argc, char *argv[]) { int maxnp = argc >= 2...
> I'll add an MCA param to select between the two methods. I hope your new method will become the default... The old code is evidently broken.
@rhc54 After your diagnosis, what would you suggest for the mpi4py test suite? Should I just skip all these spawn tests as know failures, at least until we all have...
> which is why we don't see complaints (well, plus of course the fact that hardly anyone uses comm-spawn). I have had quite a bit of emails over the years...
@jsquyres I run this PR: https://github.com/mpi4py/mpi4py-testing/actions/runs/7588222142/job/20670268020 My test is failing, but that's not necessarily Open MPI's fault, but my assumptions. This is how things go: 1. I call `Wait`, then...