[mpiwg-rma] [Mpi-forum] 3/14: Formal Readings

Jeff Hammond jeff.science at gmail.com
Thu Feb 20 09:19:28 CST 2014


On Thu, Feb 20, 2014 at 9:16 AM, Rolf Rabenseifner <rabenseifner at hlrs.de> wrote:
>>  Most users don't understand this and write code that is incorrect.
>
> Therefore such a clarification is essential.

I dispute the need for the MPI standard to educate programmers about
the pitfalls of shared memory programming.  The standard is not a
programmer manual.  Does C99 standard have text on how shared memory
works?

Jeff

> ----- Original Message -----
>> From: "William Gropp" <wgropp at illinois.edu>
>> To: "MPI WG Remote Memory Access working group" <mpiwg-rma at lists.mpi-forum.org>
>> Sent: Thursday, February 20, 2014 3:46:14 PM
>> Subject: Re: [mpiwg-rma] [Mpi-forum] 3/14: Formal Readings
>>
>>
>> I want to emphasize Jeff's point about the "eventually" - this is the
>> case not just for MPI unified but for most shared memory models (I
>> say most because you can build hardware and software to force
>> changes to be visible immediately, but at a cost in performance).
>>  Most users don't understand this and write code that is incorrect.
>>
>>
>> Bill
>>
>>
>>
>>
>>
>>
>>
>>
>> William Gropp
>> Director, Parallel Computing Institute
>> Deputy Director for Research
>> Institute for Advanced Computing Applications and Technologies Thomas
>> M. Siebel Chair in Computer Science
>>
>>
>> University of Illinois Urbana-Champaign
>>
>>
>>
>>
>>
>>
>> On Feb 20, 2014, at 5:09 AM, Jeff Hammond wrote:
>>
>>
>>
>> The critical word I think you're overlooking is "eventually".
>>
>> Win_sync induces the public-private alignment at the point of call.
>> It also syncs between diff procs doing load-store as a wrapper
>> around the processor-specific memory barrier.
>>
>> If we want to add pedantic text, that's fine though, but I still
>> don't think it's critical.
>>
>> Jeff
>>
>> Sent from my iPhone
>>
>>
>>
>> On Feb 20, 2014, at 3:55 AM, Rolf Rabenseifner < rabenseifner at hlrs.de
>> > wrote:
>>
>>
>>
>>
>>
>> Jeff,
>>
>>
>>
>>
>>
>> the problem is that we are in the unified model.
>>
>>
>>
>>
>>
>> I expect that nobody would expect that
>>
>>
>>
>>
>>
>> "the purposes of synchronizing the private and public window"
>>
>>
>>
>>
>>
>> (from your cited text) is needed if
>>
>>
>>
>>
>>
>> "public and private copies are identical",
>>
>>
>>
>>
>>
>> see MPI-3.0 p436:37-40, which say
>>
>>
>>
>>
>>
>> "In the RMA unified model, public and private copies are identical
>> and updates via put
>>
>>
>> or accumulate calls are eventually observed by load operations
>> without additional RMA
>>
>>
>> calls. A store access to a window is eventually visible to remote get
>> or accumulate calls
>>
>>
>> without additional RMA calls."
>>
>>
>>
>>
>>
>> MPI-3.0 p456:3ff say
>>
>>
>>
>>
>>
>> "In the MPI_WIN_UNIFIED memory model, the rules are much simpler
>> because the public
>>
>>
>> and private windows are the same. ..."
>>
>>
>>
>>
>>
>> and especially p456:34-36
>>
>>
>>
>>
>>
>> "This permits updates to memory with
>>
>>
>> store operations without requiring an RMA epoch."
>>
>>
>>
>>
>>
>> I read all this text and thought that I do not need any additional
>>
>>
>> synchronization besides the (empty) pt-to-pt Messages.
>>
>>
>> The members of the RMA working group convinced me
>>
>>
>> that the MPI_WIN_SYNC is needed to guarantee that a locally
>>
>>
>> visible X=13 may not be remote visible without the MPI_WIN_SYNC
>>
>>
>> although the MPI-3.0 text clearly says
>>
>>
>> "in the RMA unified model, public and private copies are identical".
>>
>>
>>
>>
>>
>> Currently, there is no example in this section showing the behavior
>>
>>
>> in the unified model with only using load/store, i.e. without any
>>
>>
>> RMA call. All existing examples use some PUT or GET.
>>
>>
>>
>>
>>
>> I tried to fill this gap to prevent any mis-interpretation
>>
>>
>> of p436:37-40 and p456:3-p457:3.
>>
>>
>>
>>
>>
>> Best regards
>>
>>
>> Rolf
>>
>>
>>
>>
>>
>>
>>
>>
>> ----- Original Message -----
>>
>>
>>
>>
>> From: "Jeff Hammond" < jeff.science at gmail.com >
>>
>>
>>
>>
>> To: "MPI WG Remote Memory Access working group" <
>> mpiwg-rma at lists.mpi-forum.org >
>>
>>
>>
>>
>> Cc: "Jeff Squyres" < jsquyres at cisco.com >
>>
>>
>>
>>
>> Sent: Wednesday, February 19, 2014 7:19:14 PM
>>
>>
>>
>>
>> Subject: Re: [mpiwg-rma] [Mpi-forum] 3/14: Formal Readings
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> Other than interactions that are unique to Fortran, I do not
>>
>>
>>
>>
>> understand what is unclear about the following text from MPI-3:
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> "For the purposes of synchronizing the private and public window,
>>
>>
>>
>>
>> MPI_WIN_SYNC has the effect of ending and reopening an access and
>>
>>
>>
>>
>> exposure epoch on the window."
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> Thus, the valid usage is prescribed by its effective equivalence to
>>
>>
>>
>>
>> "MPI_WIN_UNLOCK; MPI_WIN_LOCK;".  I apologize if the WG has been
>>
>>
>>
>>
>> sloppy in how we've discussed MPI_WIN_SYNC, but I do not feel the
>>
>>
>>
>>
>> standard is ambiguous.
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> Now, if you are arguing that Fortran is a special problem for
>>
>>
>>
>>
>> MPI_WIN_SYNC, then I will gladly support your argument that Fortran
>>
>>
>>
>>
>> is
>>
>>
>>
>>
>> a special problem for lots of things :-)
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> Jeff
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> On Wed, Feb 19, 2014 at 11:44 AM, Rolf Rabenseifner
>>
>>
>>
>>
>> < rabenseifner at hlrs.de > wrote:
>>
>>
>>
>>
>>
>>
>> Jim,
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> Yes Jim, you are fully right and I updated ticket 413 according to
>>
>>
>>
>>
>>
>>
>> your corrections.
>>
>>
>>
>>
>>
>>
>> Thank you for your carefully reading and your corrections.
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> The reason for this ticket is very simple:
>>
>>
>>
>>
>>
>>
>> Nothing about the use of MPI_Win_sync for the use-case
>>
>>
>>
>>
>>
>>
>> in this example is really explained by MPI-3.0.
>>
>>
>>
>>
>>
>>
>> I expect, that for MPI-4.0, the rules for RMA synchronization
>>
>>
>>
>>
>>
>>
>> for shared memory windows must be revisited.
>>
>>
>>
>>
>>
>>
>> But this would be another ticket.
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> Best regards
>>
>>
>>
>>
>>
>>
>> Rolf
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> ----- Original Message -----
>>
>>
>>
>>
>>
>>
>>
>>
>> From: "Jim Dinan" < james.dinan at gmail.com >
>>
>>
>>
>>
>>
>>
>>
>>
>> To: "MPI WG Remote Memory Access working group"
>>
>>
>>
>>
>>
>>
>>
>>
>> < mpiwg-rma at lists.mpi-forum.org >
>>
>>
>>
>>
>>
>>
>>
>>
>> Cc: "Rolf Rabenseifner" < rabenseifner at hlrs.de >, "Jeff Squyres"
>>
>>
>>
>>
>>
>>
>>
>>
>> < jsquyres at cisco.com >
>>
>>
>>
>>
>>
>>
>>
>>
>> Sent: Monday, February 17, 2014 11:30:42 PM
>>
>>
>>
>>
>>
>>
>>
>>
>> Subject: Re: [mpiwg-rma] [Mpi-forum] 3/14: Formal Readings
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> Rolf,
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> I think this ticket needs to be reviewed by the RMA WG before
>>
>>
>>
>>
>>
>>
>>
>>
>> moving
>>
>>
>>
>>
>>
>>
>>
>>
>> it forward.  I would suggest updating the text to incorporate the
>>
>>
>>
>>
>>
>>
>>
>>
>> following changes:
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> Example 11.13 demonstrates the proper synchronization in the
>>
>>
>>
>>
>>
>>
>>
>>
>> unified
>>
>>
>>
>>
>>
>>
>>
>>
>> memory model when a data transfer is implemented with load and
>>
>>
>>
>>
>>
>>
>>
>>
>> store
>>
>>
>>
>>
>>
>>
>>
>>
>> (instead of MPI_PUT or MPI_GET) and the synchronization between
>>
>>
>>
>>
>>
>>
>>
>>
>> processes is performed using point-to-point communication. The
>>
>>
>>
>>
>>
>>
>>
>>
>> synchronization between processes must be supplemented with a
>>
>>
>>
>>
>>
>>
>>
>>
>> memory
>>
>>
>>
>>
>>
>>
>>
>>
>> synchronization through calls to MPI_WIN_SYNC, which act locally
>>
>>
>>
>>
>>
>>
>>
>>
>> as
>>
>>
>>
>>
>>
>>
>>
>>
>> a processor-memory barrier.  In Fortran, reordering of the
>>
>>
>>
>>
>>
>>
>>
>>
>> MPI_WIN_SYNC calls must be prevented with MPI_F_SYNC_REG
>>
>>
>>
>>
>>
>>
>>
>>
>> operations.
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> The variable X is contained within a shared memory window and X
>>
>>
>>
>>
>>
>>
>>
>>
>> corresponds to the same memory location at both processes. The
>>
>>
>>
>>
>>
>>
>>
>>
>> MPI_WIN_SYNC operation performed by process A ensures completion
>>
>>
>>
>>
>>
>>
>>
>>
>> of
>>
>>
>>
>>
>>
>>
>>
>>
>> the load/store operations issued by process A. The MPI_WIN_SYNC
>>
>>
>>
>>
>>
>>
>>
>>
>> operation performed by process B ensures that process A's updates
>>
>>
>>
>>
>>
>>
>>
>>
>> to
>>
>>
>>
>>
>>
>>
>>
>>
>> X are visible to process B.
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> In the example, I don't see the reason for the second set of SYNC
>>
>>
>>
>>
>>
>>
>>
>>
>> operations after B's read of X.  If A updates X and B only reads
>>
>>
>>
>>
>>
>>
>>
>>
>> it,
>>
>>
>>
>>
>>
>>
>>
>>
>> the second send/recv synchronization should be sufficient.  That
>>
>>
>>
>>
>>
>>
>>
>>
>> is,
>>
>>
>>
>>
>>
>>
>>
>>
>> B has not made any updates to X that need to be made visible A,
>>
>>
>>
>>
>>
>>
>>
>>
>> and
>>
>>
>>
>>
>>
>>
>>
>>
>> B's read of X will be ordered because of the send operation.  The
>>
>>
>>
>>
>>
>>
>>
>>
>> F_SYNC could still be needed to preserve this ordering.
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> ~Jim.
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> On Mon, Feb 17, 2014 at 12:23 PM, Jeff Hammond <
>>
>>
>>
>>
>>
>>
>>
>>
>> jeff.science at gmail.com > wrote:
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> Switching to the WG list so that everyone is involved...
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> I do not see adding an example as so urgent that it needs to be
>>
>>
>>
>>
>>
>>
>>
>>
>> dealt
>>
>>
>>
>>
>>
>>
>>
>>
>> with at the next meeting, given how overloaded the relevant people
>>
>>
>>
>>
>>
>>
>>
>>
>> are.
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> Honestly, it is more likely to be read by users if the example and
>>
>>
>>
>>
>>
>>
>>
>>
>> commentary on it are the subject of a blog post on Squyres' blog.
>>
>>
>>
>>
>>
>>
>>
>>
>> At
>>
>>
>>
>>
>>
>>
>>
>>
>> the very least, that will ensure Google indexes it and thus
>>
>>
>>
>>
>>
>>
>>
>>
>> curious
>>
>>
>>
>>
>>
>>
>>
>>
>> people will find it (as much cannot be said for the MPI standard
>>
>>
>>
>>
>>
>>
>>
>>
>> itself).
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> Jeff
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> On Mon, Feb 17, 2014 at 10:50 AM, Rolf Rabenseifner
>>
>>
>>
>>
>>
>>
>>
>>
>> < rabenseifner at hlrs.de > wrote:
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> Pavan,
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> do you put also #413 on the list.
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> I believe, it's better to have it on the list
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> although it is only an example and therefore the RMA group
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> may put it on the errata without plenary.
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> Please can you do all what is needed
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> that it comes on the MPI-3.0 errata list.
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> Best regards
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> Rolf
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> Pavan,
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>  thank you for supporting it in the March meeting (Rajeev
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>  will
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> not
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>  be there).
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>  Is there a RMA WG Meeting at the March Forum Meeting?
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>  Will you do an MPI-3.0 errata plenary reading
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>  or will you put it into the errata by WG dicision,
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>  because it is only an example?
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>  In both cases #413 should be latest tomorrow on the agenda.
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>  Because it is one block of text at one precise location,
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>  the ticket format may be enough formalism, i.e., no extra
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>  pdf.
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> ----- Original Message -----
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> From: "Jim Dinan" < james.dinan at gmail.com >
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> To: "Main MPI Forum mailing list" <
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> mpi-forum at lists.mpi-forum.org
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> Sent: Monday, February 17, 2014 4:35:51 PM
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> Subject: [Mpi-forum] 3/14: Formal Readings
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> Hi All,
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> The RMA and Hybrid working groups would like to put forward the
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> following tickets for formal readings at the upcoming meeting:
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> #380 - Endpoints proposal
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> https://svn.mpi-forum.org/trac/mpi-forum-web/attachment/ticket/380/mpi-report.pdf
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> Read by: Pavan Balaji
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> #349, #402, #404 - Address arithmetic proposal
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> https://svn.mpi-forum.org/trac/mpi-forum-web/attachment/ticket/349/review-349-402-404.pdf
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> Read by: David Goodell
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> #369 - Add same_disp_unit info key for RMA window creation
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> https://svn.mpi-forum.org/trac/mpi-forum-web/attachment/ticket/369/mpi-report.2.pdf
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> Read by: Pavan Balaji
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> Please add these to the agenda.  Unfortunately, I will not be
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> able
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> to
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> attend this meeting, so I have included a contact person for
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> each
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> ticket.
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> Thanks!
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> ~Jim.
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> _______________________________________________
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> mpi-forum mailing list
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> mpi-forum at lists.mpi-forum.org
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> http://lists.mpi-forum.org/mailman/listinfo.cgi/mpi-forum
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> --
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> Dr. Rolf Rabenseifner . . . . . . . . . .. email
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> rabenseifner at hlrs.de
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> High Performance Computing Center (HLRS) . phone
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> ++49(0)711/685-65530
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> University of Stuttgart . . . . . . . . .. fax ++49(0)711 /
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> 685-65832
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> Head of Dpmt Parallel Computing . . .
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> www.hlrs.de/people/rabenseifner
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> Nobelstr. 19, D-70550 Stuttgart, Germany . . . . (Office: Room
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> 1.307)
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> --
>>
>>
>>
>>
>>
>>
>>
>>
>> Jeff Hammond
>>
>>
>>
>>
>>
>>
>>
>>
>> jeff.science at gmail.com
>>
>>
>>
>>
>>
>>
>>
>>
>> _______________________________________________
>>
>>
>>
>>
>>
>>
>>
>>
>> mpiwg-rma mailing list
>>
>>
>>
>>
>>
>>
>>
>>
>> mpiwg-rma at lists.mpi-forum.org
>>
>>
>>
>>
>>
>>
>>
>>
>> http://lists.mpi-forum.org/mailman/listinfo.cgi/mpiwg-rma
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> --
>>
>>
>>
>>
>>
>>
>> Dr. Rolf Rabenseifner . . . . . . . . . .. email
>>
>>
>>
>>
>>
>>
>> rabenseifner at hlrs.de
>>
>>
>>
>>
>>
>>
>> High Performance Computing Center (HLRS) . phone
>>
>>
>>
>>
>>
>>
>> ++49(0)711/685-65530
>>
>>
>>
>>
>>
>>
>> University of Stuttgart . . . . . . . . .. fax ++49(0)711 /
>>
>>
>>
>>
>>
>>
>> 685-65832
>>
>>
>>
>>
>>
>>
>> Head of Dpmt Parallel Computing . . .
>>
>>
>>
>>
>>
>>
>> www.hlrs.de/people/rabenseifner
>>
>>
>>
>>
>>
>>
>> Nobelstr. 19, D-70550 Stuttgart, Germany . . . . (Office: Room
>>
>>
>>
>>
>>
>>
>> 1.307)
>>
>>
>>
>>
>>
>>
>> _______________________________________________
>>
>>
>>
>>
>>
>>
>> mpiwg-rma mailing list
>>
>>
>>
>>
>>
>>
>> mpiwg-rma at lists.mpi-forum.org
>>
>>
>>
>>
>>
>>
>> http://lists.mpi-forum.org/mailman/listinfo.cgi/mpiwg-rma
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> --
>>
>>
>>
>>
>> Jeff Hammond
>>
>>
>>
>>
>> jeff.science at gmail.com
>>
>>
>>
>>
>> _______________________________________________
>>
>>
>>
>>
>> mpiwg-rma mailing list
>>
>>
>>
>>
>> mpiwg-rma at lists.mpi-forum.org
>>
>>
>>
>>
>> http://lists.mpi-forum.org/mailman/listinfo.cgi/mpiwg-rma
>>
>>
>>
>>
>>
>> --
>>
>>
>> Dr. Rolf Rabenseifner . . . . . . . . . .. email rabenseifner at hlrs.de
>>
>>
>> High Performance Computing Center (HLRS) . phone ++49(0)711/685-65530
>>
>>
>> University of Stuttgart . . . . . . . . .. fax ++49(0)711 / 685-65832
>>
>>
>> Head of Dpmt Parallel Computing . . . www.hlrs.de/people/rabenseifner
>>
>>
>> Nobelstr. 19, D-70550 Stuttgart, Germany . . . . (Office: Room 1.307)
>>
>>
>> _______________________________________________
>>
>>
>> mpiwg-rma mailing list
>>
>>
>> mpiwg-rma at lists.mpi-forum.org
>>
>>
>> http://lists.mpi-forum.org/mailman/listinfo.cgi/mpiwg-rma
>> _______________________________________________
>> mpiwg-rma mailing list
>> mpiwg-rma at lists.mpi-forum.org
>> http://lists.mpi-forum.org/mailman/listinfo.cgi/mpiwg-rma
>>
>>
>> _______________________________________________
>> mpiwg-rma mailing list
>> mpiwg-rma at lists.mpi-forum.org
>> http://lists.mpi-forum.org/mailman/listinfo.cgi/mpiwg-rma
>
> --
> Dr. Rolf Rabenseifner . . . . . . . . . .. email rabenseifner at hlrs.de
> High Performance Computing Center (HLRS) . phone ++49(0)711/685-65530
> University of Stuttgart . . . . . . . . .. fax ++49(0)711 / 685-65832
> Head of Dpmt Parallel Computing . . . www.hlrs.de/people/rabenseifner
> Nobelstr. 19, D-70550 Stuttgart, Germany . . . . (Office: Room 1.307)
> _______________________________________________
> mpiwg-rma mailing list
> mpiwg-rma at lists.mpi-forum.org
> http://lists.mpi-forum.org/mailman/listinfo.cgi/mpiwg-rma



-- 
Jeff Hammond
jeff.science at gmail.com



More information about the mpiwg-rma mailing list