[Mpi3-ft] MPI_UNIVERSE_NUM_NODES?

Supalov, Alexander alexander.supalov at intel.com
Wed Feb 17 08:17:17 CST 2010


Understand. Seems too cluster specific, if you wish. Not that I was totally against that, but we may want to develop a more general approach. I remember that one WG wanted to work on the topology discovery. How far did this go? We may need them now.

One potential solution could be inspired by IMPI (Interoperable MPI, see http://impi.nist.gov/) that provided colors by machine, etc. One could count those and deduce what was necessary. We may want to go this way here, too. Then a "simple" reduction op will yield a reliable result on any parallel machine.

The question is whether this attribute based approach is acceptable, and if so, what attributes we need.

-----Original Message-----
From: mpi3-ft-bounces at lists.mpi-forum.org [mailto:mpi3-ft-bounces at lists.mpi-forum.org] On Behalf Of Josh Hursey
Sent: Wednesday, February 17, 2010 2:35 PM
To: MPI 3.0 Fault Tolerance and Dynamic Process Control working Group
Subject: Re: [Mpi3-ft] MPI_UNIVERSE_NUM_NODES?

They were trying to primarily answer two questions:
  - How many nodes are in my allocation?
  - What is the maximum number of MPI processes per node that I can  
reasonably use without oversubscribing the node?

So, for example, this is useful if an application wanted to spawn 1  
process per node, then use MPI_UNIVERSE_SIZE/MPI_UNIVERSE_NUM_NODES to  
determine the maximum number of threads to start per node from each  
MPI process.

This does assume that all nodes in the system are uniform in size,  
which these applications already do.

-- Josh

On Feb 17, 2010, at 8:24 AM, Supalov, Alexander wrote:

> Thanks. Why should we care of the node number? I think users care  
> primarily of the maximum number of the MPI processes they can  
> reasonably use without oversubscribing the system.
>
> -----Original Message-----
> From: mpi3-ft-bounces at lists.mpi-forum.org [mailto:mpi3-ft-bounces at lists.mpi-forum.org 
> ] On Behalf Of Josh Hursey
> Sent: Wednesday, February 17, 2010 2:17 PM
> To: MPI 3.0 Fault Tolerance and Dynamic Process Control working Group
> Subject: [Mpi3-ft] MPI_UNIVERSE_NUM_NODES?
>
> I have been talking with a couple application groups that are
> interested in using comm_spawn for dynamic resource management
> scenarios. MPI_UNIVERSE_SIZE (MPI 10.5.1) is a good starting place for
> them to portably find a soft limit on the number of processes that can
> be started within their allocation. However, they would also really
> like to know how many machines/nodes are in their allocation in an
> equally portable way (So they can try to create N procs per node, for
> example). Currently they do some preprocessing on the runtime
> environment variables to determine how many nodes are in the
> allocation. Since this is something the MPI implementation is already
> doing, I suggested that it might be possible to expose this in a
> portable way to the application through the MPI interface.
>
> So I am proposing a MPI_UNIVERSE_NUM_NODES attribute to MPI_COMM_WORLD
> with similar wording as MPI_UNIVERSE_SIZE. This will let the
> application access the number of nodes in their allocation in a
> portable manner. As for on-node topologies (i.e., sockets, cores,
> etc.), there are other tools to help make that more portable (e.g.
> hwloc).
>
> I am interested in what others think of the idea of introducing a
> MPI_UNIVERSE_NUM_NODES attribute to the standard.
>
> Cheers,
> Josh
>
> _______________________________________________
> mpi3-ft mailing list
> mpi3-ft at lists.mpi-forum.org
> http://lists.mpi-forum.org/mailman/listinfo.cgi/mpi3-ft
> ---------------------------------------------------------------------
> Intel GmbH
> Dornacher Strasse 1
> 85622 Feldkirchen/Muenchen Germany
> Sitz der Gesellschaft: Feldkirchen bei Muenchen
> Geschaeftsfuehrer: Douglas Lusk, Peter Gleissner, Hannes Schwaderer
> Registergericht: Muenchen HRB 47456 Ust.-IdNr.
> VAT Registration No.: DE129385895
> Citibank Frankfurt (BLZ 502 109 00) 600119052
>
> This e-mail and any attachments may contain confidential material for
> the sole use of the intended recipient(s). Any review or distribution
> by others is strictly prohibited. If you are not the intended
> recipient, please contact the sender and delete all copies.
>
>
> _______________________________________________
> mpi3-ft mailing list
> mpi3-ft at lists.mpi-forum.org
> http://lists.mpi-forum.org/mailman/listinfo.cgi/mpi3-ft

_______________________________________________
mpi3-ft mailing list
mpi3-ft at lists.mpi-forum.org
http://lists.mpi-forum.org/mailman/listinfo.cgi/mpi3-ft
---------------------------------------------------------------------
Intel GmbH
Dornacher Strasse 1
85622 Feldkirchen/Muenchen Germany
Sitz der Gesellschaft: Feldkirchen bei Muenchen
Geschaeftsfuehrer: Douglas Lusk, Peter Gleissner, Hannes Schwaderer
Registergericht: Muenchen HRB 47456 Ust.-IdNr.
VAT Registration No.: DE129385895
Citibank Frankfurt (BLZ 502 109 00) 600119052

This e-mail and any attachments may contain confidential material for
the sole use of the intended recipient(s). Any review or distribution
by others is strictly prohibited. If you are not the intended
recipient, please contact the sender and delete all copies.





More information about the mpiwg-ft mailing list