[Mpi3-hybridpm] Reminder for the hybrid telecon tomorrow
Schulz, Martin
schulzm at llnl.gov
Mon Apr 22 17:09:16 CDT 2013
Would we actually specify what "not being initialized" anymore means, except that users are no longer allowed to call MPI routines before they call MPI_Init again? If so, the impact on the implementation should be minimal, since it could just keep MPI "alive" under the hood.
Martin
On Apr 22, 2013, at 12:08 PM, Jim Dinan <dinan at mcs.anl.gov> wrote:
> I think the new semantics for MPI_Init() would be backward compatible, but we could also add a new routine: MPI_Init_awesome(). :)
>
> On 4/22/13 11:42 AM, Jeff Squyres (jsquyres) wrote:
>> Sayantan and I talked about this on the phone today. It seems like re-initialization might be a good idea, but a good first step might be asking all the hardware/software vendors if there's a technical reason they don't allow re-initialzation today (other than "because it's not required by MPI"). I.e., I think that some API's/networks (like PSM) don't allow re-initialization -- is there a technical reason for that, or is it just an overcome-able software limitation?
>>
>> I'm not a big fan of re-defining the default behavior of MPI_INIT, however -- I think there might be a big impact on legacy applications.
>>
>>
>>
>> On Apr 19, 2013, at 3:06 PM, Jim Dinan <dinan at mcs.anl.gov> wrote:
>>
>>> Another issue that ought to consider is that MPI currently can only be initialized/finalized once. This requirement breaks the "MPI is a library" semantic and leads to some of the nastiness Jeff S. mentioned below. I think we should re-evaluate if this restriction is really required, or if it's just convenient for implementers.
>>>
>>> Another suggestion on this front -- Why not modify the semantics of MPI_Init to match what we want?
>>>
>>> MPI_Init:
>>>
>>> - Always THREAD_MULTIPLE
>>> - Always a thread safe call
>>> - Ref-counted
>>> - Can be used to initialize/finalize MPI multiple times
>>> - Cannot be combined with MPI_Init_thread
>>>
>>> If apps really care about getting rid of threading overhead, then they should use MPI_Init_thread() and use the thread level argument to give a performance hint.
>>>
>>> ~Jim.
>>>
>>> On 4/19/13 1:11 PM, Jeff Squyres (jsquyres) wrote:
>>>> Points to think about for the Monday teleconf...
>>>>
>>>> With regards to ref-counted MPI_INIT / MPI_FINALIZE (https://svn.mpi-forum.org/trac/mpi-forum-web/ticket/302):
>>>>
>>>> PROBLEMS IT SOLVES:
>>>> - multiple, separate libraries in a single process needing access to MPI
>>>> ==> works best when all entities call MPI_INIT* at the beginning of time, work for a "long" period of time, and then call MPI_FINALIZE at the end of time (i.e., there's no race condition -- see below)
>>>>
>>>> PROBLEMS IT DOES NOT SOLVE:
>>>> - Implementation not providing MPI_THREAD_MULTIPLE support (e.g., if the separate libraries are in different threads and someone already initialized MPI with THREAD_SINGLE, other threads can't know if it's safe to call MPI_INIT* or not)
>>>> - The "finalize" problem (i.e., can't guarantee to know if MPI has been finalized or not -- there's a race between calling MPI_FINALIZED, seeing that MPI is not finalized, and then calling MPI_INIT)
>>>>
>>>> PROBLEMS IT CREATES:
>>>> - Will need to change the definition of "main thread"
>>>> - Possibly also need to change the definitions of MPI_THREAD_SERIALIZED and MPI_THREAD_FUNNELED
>>>>
>>>> OPEN QUESTIONS:
>>>> - Do we still need to keep the restriction that the thread that initializes MPI is the same thread that finalizes MPI?
>>>> - Should we allow re-initialization? This effectively solves some (but not all) of the problems that have been discussed, but probably opens a new can of worms...
>>>>
>>>>
>>>>
>>>> On Apr 12, 2013, at 8:27 AM, Pavan Balaji <balaji at mcs.anl.gov> wrote:
>>>>
>>>>> All,
>>>>>
>>>>> Jim and I will be late for the April 22nd meeting. So we decided to
>>>>> move the endpoints discussion to the telecon after this one.
>>>>>
>>>>> I chatted with Jeff Squyres yesterday. He'll be driving the April 22nd
>>>>> telecon to discuss more details on the ref-counted init/finalize issue.
>>>>> He'll be sending out some notes before the call for discussion.
>>>>>
>>>>> Thanks,
>>>>>
>>>>> -- Pavan
>>>>>
>>>>> On 04/08/2013 11:47 AM US Central Time, Pavan Balaji wrote:
>>>>>>
>>>>>> The next call will be on April 22nd, 11am central. Same telecon number.
>>>>>>
>>>>>> -- Pavan
>>>>>>
>>>>>> On 04/08/2013 11:42 AM US Central Time, Jim Dinan wrote:
>>>>>>> Meeting notes are on the wiki:
>>>>>>>
>>>>>>> https://svn.mpi-forum.org/trac/mpi-forum-web/wiki/MPI3Hybrid/notes-2013-04-08
>>>>>>>
>>>>>>>
>>>>>>> ~Jim.
>>>>>>>
>>>>>>> On 4/7/13 12:45 PM, Pavan Balaji wrote:
>>>>>>>> All,
>>>>>>>>
>>>>>>>> This is a reminder that we'll have our hybrid telecon tomorrow at 11am.
>>>>>>>> Here's the telecon information:
>>>>>>>>
>>>>>>>> International dial-in number: 1-719-234-7800
>>>>>>>> Domestic dial-in number: 1-888-850-4523
>>>>>>>> Participant Passcode: 314159
>>>>>>>>
>>>>>>>> The main item we'll be discussing is Jeff Squyres' ref-count proposal.
>>>>>>>>
>>>>>>>> https://svn.mpi-forum.org/trac/mpi-forum-web/ticket/302
>>>>>>>>
>>>>>>>> -- Pavan
>>>>>>>>
>>>>>>> _______________________________________________
>>>>>>> Mpi3-hybridpm mailing list
>>>>>>> Mpi3-hybridpm at lists.mpi-forum.org
>>>>>>> http://lists.mpi-forum.org/mailman/listinfo.cgi/mpi3-hybridpm
>>>>>>
>>>>>
>>>>> --
>>>>> Pavan Balaji
>>>>> http://www.mcs.anl.gov/~balaji
>>>>
>>>>
>>> _______________________________________________
>>> Mpi3-hybridpm mailing list
>>> Mpi3-hybridpm at lists.mpi-forum.org
>>> http://lists.mpi-forum.org/mailman/listinfo.cgi/mpi3-hybridpm
>>
>>
> _______________________________________________
> Mpi3-hybridpm mailing list
> Mpi3-hybridpm at lists.mpi-forum.org
> http://lists.mpi-forum.org/mailman/listinfo.cgi/mpi3-hybridpm
________________________________________________________________________
Martin Schulz, schulzm at llnl.gov, http://people.llnl.gov/schulzm
CASC @ Lawrence Livermore National Laboratory, Livermore, USA
More information about the mpiwg-hybridpm
mailing list