[mpi3-coll] New nonblocking collective intro text

Adam Moody moody20 at llnl.gov
Thu Jan 29 13:39:12 CST 2009


Here is the updated text after incorporating the fixes Bronis caught. 
Thanks, Bronis.
-Adam


As described in Section ?? (Section 3.7), performance on many systems 
can be improved by overlapping communication and computation. 
Nonblocking collectives combine the potential benefits of nonblocking 
point-to-point operations to exploit overlap and to avoid 
synchronization with the optimized implementation and message scheduling 
provided by collective operations [1,4]. One way of doing this would be 
to perform a blocking collective operation in a separate thread. An 
alternative mechanism that often leads to better performance (e.g., 
avoids context switching, scheduler overheads, and thread management) is 
to use nonblocking collective communication [2].

The nonblocking collective communication model is similar to the model 
used in nonblocking point-to-point communication. A nonblocking start 
call initiates the collective operation, but does not complete it. A 
separate completion call is needed to complete the operation. Once 
initiated, the operation may progress independently of any computation 
or other communication at participating processes. In this manner, 
nonblocking collectives can mitigate synchronizing effects of collective 
operations by running them in the "background." In addition to enabling 
communication-computation overlap, nonblocking collectives can perform 
collective operations on overlapping communicators that would lead to 
deadlock with blocking operations. The semantic advantages of 
nonblocking collectives can also be useful in combination with 
point-to-point communication.

As in the nonblocking point-to-point case, all start calls are local and 
return immediately irrespective of the status of other processes. The 
start call initiates the operation which indicates that the system may 
start to copy data out of the send buffer and into the receive buffer. 
Once intiated, all associated send buffers should not be modified and 
all associated receive buffers should not be accessed until the 
collective operation completes. The start call returns a request handle, 
which must be passed to a completion call to complete the operation.

All completion calls (e.g., MPI_WAIT) described in Section ?? (Section 
3.7.3) are supported for nonblocking collective operations. Similarly to 
the blocking case, collective operations are considered to be complete 
when the local part of the operation is finished, i.e., the semantics of 
the operation are guaranteed and all buffers can be safely accessed and 
modified. Completion does not imply that other processes have completed 
or even started the operation unless otherwise specified in or implied 
by the description of the operation. Completion of a particular 
nonblocking collective operation also does not imply completion of any 
other posted nonblocking collective (or send-receive) operations, 
whether they are posted before or after the completed operation.

Advice to users. Some implementations may have the effect of 
synchronizing processes during the completion of a nonblocking 
collective. A correct, portable program cannot rely on such 
synchronization side-effects, however, one must program so as to allow 
them. (End of advice to users.)

Upon returning from a completion call in which a nonblocking colletive 
completes, the MPI_ERROR field in the associated status object is set 
appropriately to indicate any errors. The values of the MPI_SOURCE and 
MPI_TAG fields are undefined. It is valid to mix different request types 
(i.e., any combination of collective requests, I/O requests, generalized 
requests, or point-to-point requests) in functions that enable multiple 
completions (e.g., MPI_WAITALL). It is erroneous to call 
MPI_REQUEST_FREE or MPI_CANCEL with a request for a nonblocking 
collective operation.

Rationale. Freeing an active nonblocking collective request could cause 
similar problems as discussed for point-to-point requests (see Section 
?? (3.7.3)). Cancelling a request is not supported because the semantics 
of this operation are not well-defined. (End of rationale.)

Multiple nonblocking collective operations can be outstanding on a 
single communicator. If the nonblocking collective causes some system 
resource to be exhausted, then it will fail and generate an MPI 
exception. Quality implementations of MPI should ensure that this 
happens only in pathological cases. That is, an MPI implementation 
should be able to support a large number of pending nonblocking 
collective operations.

Unlike point-to-point operations, nonblocking collective operations do 
not match with blocking collectives, and collective operations do not 
have a tag argument. All processes must call collective operations 
(blocking and nonblocking) in the same order per communicator. In 
particular, once a process calls a collective operation, all other 
processes in the communicator must eventually call the same collective 
operation, and no other collective operation in between. This is 
consistent with the ordering rules for blocking collective operations in 
threaded environments.

Rationale. Matching blocking and nonblocking collectives is not allowed 
because an implementation might use different communication algorithms 
for the two cases. Blocking collectives may be optimized for minimal 
time to completion, while nonblocking collectives may balance time to 
completion with CPU overhead and asynchronous progression.

The use of tags for collective operations can prevent certain hardware 
optimizations. (End of rationale.)

Advice to users. If program semantics require matching blocking and 
nonblocking collectives, then a nonblocking collective operation can be 
initiated and immediately completed with a blocking wait to emulate 
blocking behavior. (End of advice to users.)

In terms of data movements, each nonblocking collective operation has 
the same effect as its blocking counterpart for intracommunicators and 
intercommunicators after completion. The use of the “in place” option is 
allowed exactly as described for the corresponding blocking collective 
operations. Likewise, upon completion, nonblocking collective reduction 
operations have the same effect as their blocking counterparts, and the 
same restrictions and recommendations on reduction orders apply.

Progression rules for nonblocking collectives are similar to progression 
of nonblocking point-to-point operations, refer to Section ?? (Section 
3.7.4).

Advice to implementors. Nonblocking collective operations can be 
implemented with local execution schedules [3] using nonblocking 
point-to-point communication and a reserved tag-space. (End of advice to 
implementors.)



More information about the mpiwg-coll mailing list