[mpich-discuss] [petsc-dev] Is mpich/master:a8a2b30fd21 tested with Petsc?

Min Si msi at anl.gov
Tue Mar 27 15:59:26 CDT 2018


Hi Eric,

It will be great if you could give us a simple MPI program (not with 
PETSc) to reproduce this issue. If this is a problem happens only when 
PETSc is involved, the PETSc team can give you more suggestions.

Thanks,
Min

On 2018/03/27 15:38, Eric Chamberland wrote:
> Hi,
>
> since more than 2 weeks that the master branch of mpich is still and 
> it can be reproduced with a simple "make test" after a fresh 
> installation of PETSc...
>
> Is anyone testing it?
>
> Is it supposed to be working?
>
> Just tell me if I should "follow" another mpich branch please.
>
> Thanks,
>
> Eric
>
>
> On 14/03/18 03:35 AM, Eric Chamberland wrote:
>> Hi,
>>
>> fwiw, the actual mpich/master branch doesn't passes the PETSc "make 
>> test" after a fresh installation...  It hangs just afer the 1 MPI 
>> process test, meaning it is locked into the 2 process test:
>>
>> make 
>> PETSC_DIR=/pmi/cmpbib/compilation_BIB_dernier_mpich/COMPILE_AUTO/mpich-3.x-debug/petsc-3.8.3-debug 
>> PETSC_ARCH=arch-linux2-c-debug test
>> Running test examples to verify correct installation
>> Using 
>> PETSC_DIR=/pmi/cmpbib/compilation_BIB_dernier_mpich/COMPILE_AUTO/mpich-3.x-debug/petsc-3.8.3-debug 
>> and PETSC_ARCH=arch-linux2-c-debug
>> C/C++ example src/snes/examples/tutorials/ex19 run successfully with 
>> 1 MPI process
>>
>>
>>
>>
>> ^Cmakefile:151: recipe for target 'test' failed
>> make: [test] Interrupt (ignored)
>>
>> thanks,
>>
>> Eric
>>
>> On 13/03/18 08:07 AM, Eric Chamberland wrote:
>>>
>>> Hi,
>>>
>>> each night we are testing mpich/master with our petsc-based code. I 
>>> don't know if PETSc team is doing the same thing with 
>>> mpich/master?   (Maybe it is a good idea?)
>>>
>>> Everything was fine (except the issue 
>>> https://github.com/pmodels/mpich/issues/2892) up to commit 
>>> 7b8d64debd, but since commit mpich:a8a2b30fd21), I have a segfault 
>>> on a any parallel nightly test.
>>>
>>> For example, a 2 process test ends at almost different execution 
>>> points:
>>>
>>> rank 0:
>>>
>>> #003: /lib64/libpthread.so.0(+0xf870) [0x7f25bf908870]
>>> #004: 
>>> /pmi/cmpbib/compilation_BIB_dernier_mpich/COMPILE_AUTO/BIB/bin/BIBMEFGD.opt() 
>>> [0x64a788]
>>> #005: /lib64/libc.so.6(+0x35140) [0x7f25bca18140]
>>> #006: /lib64/libc.so.6(__poll+0x2d) [0x7f25bcabfbfd]
>>> #007: /opt/mpich-3.x_debug/lib/libmpi.so.0(+0x1e4cc9) [0x7f25bd90ccc9]
>>> #008: /opt/mpich-3.x_debug/lib/libmpi.so.0(+0x1ea55c) [0x7f25bd91255c]
>>> #009: /opt/mpich-3.x_debug/lib/libmpi.so.0(+0xba657) [0x7f25bd7e2657]
>>> #010: /opt/mpich-3.x_debug/lib/libmpi.so.0(PMPI_Waitall+0xe3) 
>>> [0x7f25bd7e3343]
>>> #011: 
>>> /opt/petsc-3.8.3_debug_mpich-3.x_debug/lib/libpetsc.so.3.8(PetscGatherMessageLengths+0x654) 
>>> [0x7f25c4bb3193]
>>> #012: 
>>> /opt/petsc-3.8.3_debug_mpich-3.x_debug/lib/libpetsc.so.3.8(VecScatterCreate_PtoS+0x859) 
>>> [0x7f25c4e82d7f]
>>> #013: 
>>> /opt/petsc-3.8.3_debug_mpich-3.x_debug/lib/libpetsc.so.3.8(VecScatterCreate+0x5684) 
>>> [0x7f25c4e4d055]
>>> #014: 
>>> /opt/petsc-3.8.3_debug_mpich-3.x_debug/lib/libpetsc.so.3.8(VecCreateGhostWithArray+0x688) 
>>> [0x7f25c4e01a39]
>>> #015: 
>>> /opt/petsc-3.8.3_debug_mpich-3.x_debug/lib/libpetsc.so.3.8(VecCreateGhost+0x179) 
>>> [0x7f25c4e020f6]
>>>
>>> rank 1:
>>>
>>> #002: 
>>> /pmi/cmpbib/compilation_BIB_dernier_mpich/COMPILE_AUTO/GIREF/lib/libgiref_opt_Util.so(traitementSignal+0x2bd0) 
>>> [0x7f62df8e7310]
>>> #003: /lib64/libc.so.6(+0x35140) [0x7f62d3bc9140]
>>> #004: /lib64/libc.so.6(__poll+0x2d) [0x7f62d3c70bfd]
>>> #005: /opt/mpich-3.x_debug/lib/libmpi.so.0(+0x1e4cc9) [0x7f62d4abdcc9]
>>> #006: /opt/mpich-3.x_debug/lib/libmpi.so.0(+0x1ea55c) [0x7f62d4ac355c]
>>> #007: /opt/mpich-3.x_debug/lib/libmpi.so.0(+0x12c9c5) [0x7f62d4a059c5]
>>> #008: /opt/mpich-3.x_debug/lib/libmpi.so.0(+0x12e102) [0x7f62d4a07102]
>>> #009: /opt/mpich-3.x_debug/lib/libmpi.so.0(+0xf17a1) [0x7f62d49ca7a1]
>>> #010: /opt/mpich-3.x_debug/lib/libmpi.so.0(+0x3facf) [0x7f62d4918acf]
>>> #011: /opt/mpich-3.x_debug/lib/libmpi.so.0(+0x3fc3d) [0x7f62d4918c3d]
>>> #012: /opt/mpich-3.x_debug/lib/libmpi.so.0(+0xf18d8) [0x7f62d49ca8d8]
>>> #013: /opt/mpich-3.x_debug/lib/libmpi.so.0(+0x3fb88) [0x7f62d4918b88]
>>> #014: /opt/mpich-3.x_debug/lib/libmpi.so.0(+0x3fc3d) [0x7f62d4918c3d]
>>> #015: /opt/mpich-3.x_debug/lib/libmpi.so.0(MPI_Barrier+0x27b) 
>>> [0x7f62d4918edb]
>>> #016: 
>>> /opt/petsc-3.8.3_debug_mpich-3.x_debug/lib/libpetsc.so.3.8(PetscCommGetNewTag+0x3ff) 
>>> [0x7f62dbceb055]
>>> #017: 
>>> /opt/petsc-3.8.3_debug_mpich-3.x_debug/lib/libpetsc.so.3.8(PetscObjectGetNewTag+0x15d) 
>>> [0x7f62dbceaadb]
>>> #018: 
>>> /opt/petsc-3.8.3_debug_mpich-3.x_debug/lib/libpetsc.so.3.8(VecScatterCreateCommon_PtoS+0x1ee) 
>>> [0x7f62dc03625c]
>>> #019: 
>>> /opt/petsc-3.8.3_debug_mpich-3.x_debug/lib/libpetsc.so.3.8(VecScatterCreate_PtoS+0x29c4) 
>>> [0x7f62dc035eea]
>>> #020: 
>>> /opt/petsc-3.8.3_debug_mpich-3.x_debug/lib/libpetsc.so.3.8(VecScatterCreate+0x5684) 
>>> [0x7f62dbffe055]
>>> #021: 
>>> /opt/petsc-3.8.3_debug_mpich-3.x_debug/lib/libpetsc.so.3.8(VecCreateGhostWithArray+0x688) 
>>> [0x7f62dbfb2a39]
>>> #022: 
>>> /opt/petsc-3.8.3_debug_mpich-3.x_debug/lib/libpetsc.so.3.8(VecCreateGhost+0x179) 
>>> [0x7f62dbfb30f6]
>>>
>>> Have some other users (PETSc users?) reported problem?
>>>
>>> Thanks,
>>>
>>> Eric
>>>
>>> ps: usual informations:
>>>
>>> mpich logs:
>>> http://www.giref.ulaval.ca/~cmpgiref/dernier_mpich/2018.03.12.05h39m54s_config.log 
>>>
>>> http://www.giref.ulaval.ca/~cmpgiref/dernier_mpich/2018.03.12.05h39m54s_config.system 
>>>
>>> http://www.giref.ulaval.ca/~cmpgiref/dernier_mpich/2018.03.12.05h39m54s_mpich_version.txt 
>>>
>>> http://www.giref.ulaval.ca/~cmpgiref/dernier_mpich/2018.03.12.05h39m54s_c.txt 
>>>
>>> http://www.giref.ulaval.ca/~cmpgiref/dernier_mpich/2018.03.12.05h39m54s_m.txt 
>>>
>>> http://www.giref.ulaval.ca/~cmpgiref/dernier_mpich/2018.03.12.05h39m54s_mi.txt 
>>>
>>> http://www.giref.ulaval.ca/~cmpgiref/dernier_mpich/2018.03.12.05h39m54s_openmpa_config.log 
>>>
>>> http://www.giref.ulaval.ca/~cmpgiref/dernier_mpich/2018.03.12.05h39m54s_mpl_config.log 
>>>
>>> http://www.giref.ulaval.ca/~cmpgiref/dernier_mpich/2018.03.12.05h39m54s_pm_hydra_config.log 
>>>
>>> http://www.giref.ulaval.ca/~cmpgiref/dernier_mpich/2018.03.12.05h39m54s_pm_hydra_tools_topo_config.log 
>>>
>>> http://www.giref.ulaval.ca/~cmpgiref/dernier_mpich/2018.03.12.05h39m54s_mpiexec_info.txt 
>>>
>>>
>>> Petsc logs:
>>> http://www.giref.ulaval.ca/~cmpgiref/dernier_mpich/2018.03.12.05h39m54s_configure.log 
>>>
>>> http://www.giref.ulaval.ca/~cmpgiref/dernier_mpich/2018.03.12.05h39m54s_make.log 
>>>
>>> http://www.giref.ulaval.ca/~cmpgiref/dernier_mpich/2018.03.12.05h39m54s_default.log 
>>>
>>> http://www.giref.ulaval.ca/~cmpgiref/dernier_mpich/2018.03.12.05h39m54s_RDict.log 
>>>
>>> http://www.giref.ulaval.ca/~cmpgiref/dernier_mpich/2018.03.12.05h39m54s_CMakeLists.txt 
>>>
>>>
>>>
>>
> _______________________________________________
> discuss mailing list     discuss at mpich.org
> To manage subscription options or unsubscribe:
> https://lists.mpich.org/mailman/listinfo/discuss

_______________________________________________
discuss mailing list     discuss at mpich.org
To manage subscription options or unsubscribe:
https://lists.mpich.org/mailman/listinfo/discuss


More information about the discuss mailing list