[Mpi3-rma] FW: MPI_Accumulate
Pavan Balaji
balaji at mcs.anl.gov
Tue Oct 2 13:22:52 CDT 2012
Put is not byte wise atomic. It is not atomic at all. Only the accumulate-style operations are predefined datatype granularity atomic.
--
Pavan Balaji
http://www.mcs.anl.gov/~balaji
On Oct 2, 2012, at 11:48 AM, Jeff Hammond <jhammond at alcf.anl.gov> wrote:
> My understanding has always been that MPI_Accumulate was and is atomic
> at granularity of the underlying built-in datatype, as opposed to
> byte-wise atomic like MPI_Put. This is the whole reason that
> MPI_Accumulate w/ MPI_REPLACE exists. Without the additional
> atomicity guarantee, that feature would be pointless.
>
> Jeff
>
> On Tue, Oct 2, 2012 at 11:38 AM, Richard Graham <richardg at mellanox.com> wrote:
>> resending
>>
>>
>>
>> From: Richard Graham
>> Sent: Tuesday, October 02, 2012 12:22 PM
>> To: MPI 3.0 Remote Memory Access working group
>> (mpi3-rma at lists.mpi-forum.org)
>> Subject: MPI_Accumulate
>>
>>
>>
>> What are the requirements placed on MPI_Accumulate if more than one mpi
>> process tries to update the same location ? Does MPI provide any
>> consistency promises, or is it up to the application to guarantee these ? I
>> see that the get accumulate routines are defined to be atomic, but don’t see
>> the same requirement for accumulate.
>>
>>
>>
>> Thanks,
>>
>> Rich
>>
>>
>> _______________________________________________
>> mpi3-rma mailing list
>> mpi3-rma at lists.mpi-forum.org
>> http://lists.mpi-forum.org/mailman/listinfo.cgi/mpi3-rma
>
>
>
> --
> Jeff Hammond
> Argonne Leadership Computing Facility
> University of Chicago Computation Institute
> jhammond at alcf.anl.gov / (630) 252-5381
> http://www.linkedin.com/in/jeffhammond
> https://wiki.alcf.anl.gov/parts/index.php/User:Jhammond
>
> _______________________________________________
> mpi3-rma mailing list
> mpi3-rma at lists.mpi-forum.org
> http://lists.mpi-forum.org/mailman/listinfo.cgi/mpi3-rma
More information about the mpiwg-rma
mailing list