[Mpi3-rma] RMA proposal 1 update
Underwood, Keith D
keith.d.underwood at intel.com
Wed May 19 20:25:21 CDT 2010
So, perhaps enumerating the relevant GA constructs and their semantics would be informative here...
Keith
> -----Original Message-----
> From: mpi3-rma-bounces at lists.mpi-forum.org [mailto:mpi3-rma-
> bounces at lists.mpi-forum.org] On Behalf Of Jeff Hammond
> Sent: Wednesday, May 19, 2010 7:23 PM
> To: MPI 3.0 Remote Memory Access working group
> Cc: MPI 3.0 Remote Memory Access working group
> Subject: Re: [Mpi3-rma] RMA proposal 1 update
>
> Can I mix that call with other sync mechanisms?
>
> So I implement GA by calling fence inside of GA_Create to expose the
> window and use fence+barrier for GA_Sync, but can I mix in lock and
> unlock as well as the forthcoming p2p flush (as I can do in GA/ARMCI
> now)?
>
> The standard presents three synchronization schemes. It does not
> suggest one can intermix them at will.
>
> Jeff
>
> Sent from my iPhone
>
> On May 19, 2010, at 2:58 PM, "Underwood, Keith D"
> <keith.d.underwood at intel.com
> > wrote:
>
> > Jeff,
> >
> > Another question for you: If you are going to call
> > MPI_Win_all_flush_all, why not just use active target and call
> > MPI_Win_fence?
> >
> > Keith
> >
> >> -----Original Message-----
> >> From: mpi3-rma-bounces at lists.mpi-forum.org [mailto:mpi3-rma-
> >> bounces at lists.mpi-forum.org] On Behalf Of Jeff Hammond
> >> Sent: Sunday, May 16, 2010 7:27 PM
> >> To: MPI 3.0 Remote Memory Access working group
> >> Subject: Re: [Mpi3-rma] RMA proposal 1 update
> >>
> >> Tortsten,
> >>
> >> There seemed to be decent agreement on adding MPI_Win_all_flush_all
> >> (equivalent to MPI_Win_flush_all called from every rank in the
> >> communicator associated with the window) since this function can be
> >> implemented far more efficiently as a collective than the equivalent
> >> point-wise function calls.
> >>
> >> Is there a problem with adding this to your proposal?
> >>
> >> Jeff
> >>
> >> On Sun, May 16, 2010 at 12:48 AM, Torsten Hoefler
> <htor at illinois.edu>
> >> wrote:
> >>> Hello all,
> >>>
> >>> After the discussions at the last Forum I updated the group's first
> >>> proposal.
> >>>
> >>> The proposal (one-side-2.pdf) is attached to the wiki page
> >>> https://svn.mpi-forum.org/trac/mpi-forum-web/wiki/RmaWikiPage
> >>>
> >>> The changes with regards to the last version are:
> >>>
> >>> 1) added MPI_NOOP to MPI_Get_accumulate and MPI_Accumulate_get
> >>>
> >>> 2) (re)added MPI_Win_flush and MPI_Win_flush_all to passive target
> >> mode
> >>>
> >>> Some remarks:
> >>>
> >>> 1) We didn't straw-vote on MPI_Accumulate_get, so this function
> >>> might
> >>> go. The removal would be very clean.
> >>>
> >>> 2) Should we allow MPI_NOOP in MPI_Accumulate (this does not make
> >> sense
> >>> and is incorrect in my current proposal)
> >>>
> >>> 3) Should we allow MPI_REPLACE in
> >> MPI_Get_accumulate/MPI_Accumulate_get?
> >>> (this would make sense and is allowed in the current proposal but
> >> we
> >>> didn't talk about it in the group)
> >>>
> >>>
> >>> All the Best,
> >>> Torsten
> >>>
> >>> --
> >>> bash$ :(){ :|:&};: --------------------- http://www.unixer.de/
> >>> -----
> >>> Torsten Hoefler | Research Associate
> >>> Blue Waters Directorate | University of Illinois
> >>> 1205 W Clark Street | Urbana, IL, 61801
> >>> NCSA Building | +01 (217) 244-7736
> >>> _______________________________________________
> >>> mpi3-rma mailing list
> >>> mpi3-rma at lists.mpi-forum.org
> >>> http://lists.mpi-forum.org/mailman/listinfo.cgi/mpi3-rma
> >>>
> >>
> >>
> >>
> >> --
> >> Jeff Hammond
> >> Argonne Leadership Computing Facility
> >> jhammond at mcs.anl.gov / (630) 252-5381
> >> http://www.linkedin.com/in/jeffhammond
> >>
> >> _______________________________________________
> >> mpi3-rma mailing list
> >> mpi3-rma at lists.mpi-forum.org
> >> http://lists.mpi-forum.org/mailman/listinfo.cgi/mpi3-rma
> >
> > _______________________________________________
> > mpi3-rma mailing list
> > mpi3-rma at lists.mpi-forum.org
> > http://lists.mpi-forum.org/mailman/listinfo.cgi/mpi3-rma
> _______________________________________________
> mpi3-rma mailing list
> mpi3-rma at lists.mpi-forum.org
> http://lists.mpi-forum.org/mailman/listinfo.cgi/mpi3-rma
More information about the mpiwg-rma
mailing list