[Mpi-22] Higher-level languages proposal
Supalov, Alexander
alexander.supalov at [hidden]
Tue Oct 14 10:15:25 CDT 2008
Hi,
The proposal is ready in draft, see
https://svn.mpi-forum.org/trac/mpi-forum-web/ticket/39 . I targeted it
to MPI-2.2 for now. As you will see, it resolves the problem of thread
addressability without any extension to the Probe/Recv calls. I bet
there are more things that will follow, too.
Here's the current text for reference:
"A collective MPI call, MPI_Comm_thread_register, with the following
syntax (in C):
int MPI_Comm_thread_register(MPI_Comm comm, int index, int num, MPI_Comm
*newcomm)
returns a newcomm for all num threads of the comm that called this
function. All threads are treated as MPI processes in the newcomm, and
their ranks are ordered according to the index argument that ranges
between 0 and num-1. This argument must be unique in every thread on in
the given MPI process of the comm.
>From this moment on, all threads contained in the newcomm are considered
as MPI processes, with all that this entails, including individual MPI
rank that makes the respective thread addressable in the usual manner.
All MPI communicator and group management calls can be applied to the
newcomm in order to produce new communicators, reorder the processes in
it, etc. (see Figure 1).
A slightly modified call MPI_Comm_free with the standard syntax (in C):
int MPI_Comm_free(MPI_Comm comm)
can be used to destroy the respective communicator comm and thus
"demote" all the threads from the status of MPI processes in the comm
back to the unnamed threads typical of the MPI standard.
This pair of calls, or their equivalent, allow threads to be addressed
directly in all MPI calls, and since the sequence of the
MPI_Comm_thread_register and MPI_Comm_free calls can be repeated as
needed, OpenMP parallel sections or any equivalent groups of threads in
the MPI program can become MPI processes for a while and then return to
their original status.
If threads use (as they usually do) joint address space with one
(former) MPI process, the MPI communication calls can certainly take
advantage of this by copying data directly from the source to the
destination buffer. This equally applies to all point-to-point,
collective, one-sided, and file I/O calls.
This call certainly makes sense only at the thread support level
MPI_THREAD_MULTIPLE."
Best regards.
Alexander
-----Original Message-----
From: Supalov, Alexander
Sent: Tuesday, October 14, 2008 2:44 PM
To: 'MPI 2.2'
Subject: RE: [Mpi-22] Higher-level languages proposal
Sure. This will most likely be a MPI-3 topic, though. I'll drop in a
link here once ready.
-----Original Message-----
From: mpi-22-bounces_at_[hidden]
[mailto:mpi-22-bounces_at_[hidden]] On Behalf Of Jeff Squyres
Sent: Tuesday, October 14, 2008 2:35 PM
To: MPI 2.2
Subject: Re: [Mpi-22] Higher-level languages proposal
On Oct 14, 2008, at 7:07 AM, Supalov, Alexander wrote:
> Thanks. I'd rather say that if the purpose of the extension is
> indeed to
> serialize the Probe/Recv pair, the better way to solve this and many
> other problems would be to make threads directly addressable, as if
> they
> were MPI processes.
>
> One way to do this might be, say, to create a call like
> MPI_Comm_thread_enroll that creates an intra-communicator out of all
> threads that call this function in a loosely synchronous fashion,
> collectively over one or several MPI processes they constitute.
I'm still not sure I follow. Can you provide more details, perhaps
with function prototypes and specific rules? (i.e., an alternate
proposal)?
> If paired with the appropriately extended MPI_Comm_free, this would
> allow, for example, all threads in an OpenMP parallel section to be
> addressed as if they were fully fledged MPI processes. Note that this
> would allow more than one parallel section during the program run.
>
> Other threading models would profit from this "opt-in/opt-out" method,
> too. This may be a more flexible way of dealing with threads than the
> one-time MPI_Init variety mentioned by George Bosilica in his
> EuroPVM/MPI keynote, by the way.
>
> Best regards.
>
> Alexander
>
> -----Original Message-----
> From: mpi-22-bounces_at_[hidden]
> [mailto:mpi-22-bounces_at_[hidden]] On Behalf Of Terry Dontje
> Sent: Tuesday, October 14, 2008 12:45 PM
> To: MPI 2.2
> Subject: Re: [Mpi-22] Higher-level languages proposal
>
> Supalov, Alexander wrote:
>> Dear Jeff,
>>
>> Unfortunately, I won't be in Chicago, so we should rather discuss
>> this
>> here. I talked to Torsten last time about this extension. As far as I
>> can remember, the main purpose of this extension is to make sure that
>> the thread that called the MPI_Probe also calls the MPI_Recv and gets
>> the message matched by the aforementioned MPI_Probe.
>>
>> If so, the main problem here is not the matching. The main problem is
>> that one cannot address threads in MPI. If we fix that, the proposed
>> extension with the message handle and such will become superfluous.
>>
>> See what I mean?
>>
>>
> Interesting, so you are basically redefining the MPI_Probe/Recv pair
> to
> guarrantee a message to go to a specific thread. Or in other words
> lowering the proposal's MPI_Mprobe/recv to be in the implementation of
> MPI_Probe/Recv. This seems reasonable to me since MPI_Probe/Recv
> itself
>
> is basically useless unless the programmer assures serialization when
> that combination is used.
>
> --td
>> Best regards.
>>
>> Alexander
>>
>> -----Original Message-----
>> From: mpi-22-bounces_at_[hidden]
>> [mailto:mpi-22-bounces_at_[hidden]] On Behalf Of Jeff Squyres
>> Sent: Monday, October 13, 2008 11:48 PM
>> To: MPI 2.2
>> Subject: Re: [Mpi-22] Higher-level languages proposal
>>
>> On Oct 13, 2008, at 10:46 AM, Supalov, Alexander wrote:
>>
>>
>>> Thanks. The 2.1.1, which was presented last time, in my opinion does
>
>>> not
>>> seem to solve the right problem. Instead of defining a way for
>>> unambiguous addressing of the threads in MPI, which would eliminate
>>> the
>>> MPI_Probe/Recv ambiguity and many other issues, it attempts to add
> yet
>>> another concept (this time, a message id) in the current situation
>>> where
>>> any thread can do what they please.
>>>
>>
>> I'm not quite sure I understand your proposal.
>>
>> <... after typing out a lengthy/rambling discourse that made very
>> little sense and was fraught with questions and ambiguities :-) ...>
>>
>> Let's discuss this in Chicago; Rich has allocated 5-7pm on Monday for
>
>> discussion of this proposal. These are exactly the kinds of larger
>> issues that we want to raise via this proposal.
>>
>>
>
> _______________________________________________
> mpi-22 mailing list
> mpi-22_at_[hidden]
> http://lists.mpi-forum.org/mailman/listinfo.cgi/mpi-22
> ---------------------------------------------------------------------
> Intel GmbH
> Dornacher Strasse 1
> 85622 Feldkirchen/Muenchen Germany
> Sitz der Gesellschaft: Feldkirchen bei Muenchen
> Geschaeftsfuehrer: Douglas Lusk, Peter Gleissner, Hannes Schwaderer
> Registergericht: Muenchen HRB 47456 Ust.-IdNr.
> VAT Registration No.: DE129385895
> Citibank Frankfurt (BLZ 502 109 00) 600119052
>
> This e-mail and any attachments may contain confidential material for
> the sole use of the intended recipient(s). Any review or distribution
> by others is strictly prohibited. If you are not the intended
> recipient, please contact the sender and delete all copies.
>
>
> _______________________________________________
> mpi-22 mailing list
> mpi-22_at_[hidden]
> http://lists.mpi-forum.org/mailman/listinfo.cgi/mpi-22
--
Jeff Squyres
Cisco Systems
_______________________________________________
mpi-22 mailing list
mpi-22_at_[hidden]
http://lists.mpi-forum.org/mailman/listinfo.cgi/mpi-22
---------------------------------------------------------------------
Intel GmbH
Dornacher Strasse 1
85622 Feldkirchen/Muenchen Germany
Sitz der Gesellschaft: Feldkirchen bei Muenchen
Geschaeftsfuehrer: Douglas Lusk, Peter Gleissner, Hannes Schwaderer
Registergericht: Muenchen HRB 47456 Ust.-IdNr.
VAT Registration No.: DE129385895
Citibank Frankfurt (BLZ 502 109 00) 600119052
This e-mail and any attachments may contain confidential material for
the sole use of the intended recipient(s). Any review or distribution
by others is strictly prohibited. If you are not the intended
recipient, please contact the sender and delete all copies.
More information about the Mpi-22
mailing list