[Mpi-forum] propose ABI working group

Jeff Hammond jeff.science at gmail.com
Wed Nov 16 09:17:26 CST 2022



> On 16Nov 2022, at 5:11 PM, Wes Bland <wesley at wesbland.com> wrote:
> 
> The rules say you need to get four IMOVE (voting) orgs to support creating a WG at a meeting:
> 
> > Working groups can be established at MPI Forum meetings once at least four IMOVE organizations indicate support for that proposed Working Group.
> 
> So feel free to propose it at the December meeting and have some folks lined up to give it a thumbs up. Of course, in the meantime, folks are free to start getting together and discussing the topic. The old mailing list from 2008 is still around <https://lists.mpi-forum.org/mailman/listinfo/mpi3-abi>, but I’d recommend you not use it since it uses our old naming scheme. Once the group is official, I’ll make the new list and GitHub org.

Thanks, I’ll rally the votes.

Interested parties should ping me on Slack to be added to #wg-abi.

Somebody was optimistic enough to create https://github.com/mpiwg-abi/ this morning but it’s a mystery who 😉

Jeff

> Thanks,
> Wes
> 
>> On Nov 16, 2022, at 1:54 AM, Jeff Hammond via mpi-forum <mpi-forum at lists.mpi-forum.org> wrote:
>> 
>> I don't know what we do to create new working groups with the post-COVID rules, but I would like to create and chair a WG focused on ABI standardization.
>> 
>> There is strong support for this effort in many user communities, including developers and maintainers of Spack, mpi4py, Julia MPI (MPI.jl), Rust MPI (rsmpi), PETSc and NVHPC SDK, to name a few.  There are even a few implementers who have expressed support, but I won't name them for their own protection.
>> 
>> The problem is so exasperating for our users that there are at least two different projects devoted to mitigating ABI problems (not including shims built in to the aforementioned MPI wrappers):
>> 
>> https://github.com/cea-hpc/wi4mpi <https://github.com/cea-hpc/wi4mpi>
>> https://github.com/eschnett/MPItrampoline <https://github.com/eschnett/MPItrampoline>
>> 
>> I've written about this a bit already, for those who are interested.  More material will be forthcoming once I have time for more experiments.
>> 
>> https://github.com/jeffhammond/blog/blob/main/MPI_Needs_ABI.md <https://github.com/jeffhammond/blog/blob/main/MPI_Needs_ABI.md>
>> https://github.com/jeffhammond/blog/blob/main/MPI_Needs_ABI_Part_2.md <https://github.com/jeffhammond/blog/blob/main/MPI_Needs_ABI_Part_2.md>
>> https://github.com/jeffhammond/blog/blob/main/MPI_Needs_ABI_Part_3.md <https://github.com/jeffhammond/blog/blob/main/MPI_Needs_ABI_Part_3.md>
>> https://github.com/jeffhammond/blog/blob/main/MPI_Needs_ABI_Part_4.md <https://github.com/jeffhammond/blog/blob/main/MPI_Needs_ABI_Part_4.md>
>> 
>> I understand this is a controversial topic, particularly for implementers.  I hope that we can proceed objectively.
>> 
>> Thanks,
>> 
>> Jeff
>> 
>> -- 
>> Jeff Hammond
>> jeff.science at gmail.com <mailto:jeff.science at gmail.com>
>> http://jeffhammond.github.io/ <http://jeffhammond.github.io/>_______________________________________________
>> mpi-forum mailing list
>> mpi-forum at lists.mpi-forum.org
>> https://lists.mpi-forum.org/mailman/listinfo/mpi-forum
> 

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mpi-forum.org/pipermail/mpi-forum/attachments/20221116/7f8833bb/attachment-0001.html>


More information about the mpi-forum mailing list