9abeaad6e2
in the v1.2 series the cid's could never go above the max. allowed for a particular pml. Because of that, pml_add_comm never checked for the cid, and in fact pml_add_comm was called in comm_set, which is *before* we knew the cid. in the v1.3 series (and trunk) we check now the cid to detect overflow, and because of that pml_add_comm has been moved *after* the cid allocation routine, namely into the comm_activate routine. in the v1.2 series, the comm_activate contained a synchronization step of the old communicator in order to prevent incoming fragments on the new communicator, with the main problem being that the allreduce in the communicator allocation finished at different times on different processes, and thus, this scenario could and did really occur. in the v1.3 series, the comm_activate does not contain the synchronization step anymore, since we introduced the new queue for fragments with unknown cid. The problem is however, that whether a fragment is known or not is decided by using ompi_comm_lookup(), which will return something useful as soon as the cid allocation finished, even before pml_add_comm has been called. So there is a small time gap where we will not post a message into queue for unknown cid's, but we can also not look up the process structure belonging to the rank in that comm ( that is in pml_ob1_match_recv_frag or something like that). The current fix reintroduces the synchronization step in comm_activate, and ensures that no fragment can be received for a new communicator before the synchronization occurs , and thus comm_nextcid() and pml_add_comm has been called. It seems to be the safest and easiest way for now. Welcome back, v1.2. This commit was SVN r21970.
245 строки
9.2 KiB
C
245 строки
9.2 KiB
C
/*
|
|
* Copyright (c) 2004-2007 The Trustees of Indiana University and Indiana
|
|
* University Research and Technology
|
|
* Corporation. All rights reserved.
|
|
* Copyright (c) 2004-2008 The University of Tennessee and The University
|
|
* of Tennessee Research Foundation. All rights
|
|
* reserved.
|
|
* Copyright (c) 2004-2008 High Performance Computing Center Stuttgart,
|
|
* University of Stuttgart. All rights reserved.
|
|
* Copyright (c) 2004-2005 The Regents of the University of California.
|
|
* All rights reserved.
|
|
* Copyright (c) 2006-2007 Cisco Systems, Inc. All rights reserved.
|
|
* Copyright (c) 2006-2009 University of Houston. All rights reserved.
|
|
* $COPYRIGHT$
|
|
*
|
|
* Additional copyrights may follow
|
|
*
|
|
* $HEADER$
|
|
*/
|
|
|
|
#include "ompi_config.h"
|
|
|
|
#include "ompi/mpi/c/bindings.h"
|
|
#include "ompi/runtime/params.h"
|
|
#include "ompi/errhandler/errhandler.h"
|
|
#include "ompi/mca/pml/pml.h"
|
|
#include "ompi/communicator/communicator.h"
|
|
#include "ompi/request/request.h"
|
|
#include "ompi/memchecker.h"
|
|
|
|
#if OPAL_HAVE_WEAK_SYMBOLS && OMPI_PROFILING_DEFINES
|
|
#pragma weak MPI_Intercomm_create = PMPI_Intercomm_create
|
|
#endif
|
|
|
|
#if OMPI_PROFILING_DEFINES
|
|
#include "ompi/mpi/c/profile/defines.h"
|
|
#endif
|
|
|
|
static const char FUNC_NAME[] = "MPI_Intercomm_create";
|
|
|
|
|
|
int MPI_Intercomm_create(MPI_Comm local_comm, int local_leader,
|
|
MPI_Comm bridge_comm, int remote_leader,
|
|
int tag, MPI_Comm *newintercomm)
|
|
{
|
|
int local_size=0, local_rank=0;
|
|
int lleader=0, rleader=0;
|
|
ompi_communicator_t *newcomp=NULL;
|
|
struct ompi_proc_t **rprocs=NULL;
|
|
int rc=0, rsize=0;
|
|
ompi_proc_t **proc_list=NULL;
|
|
int j;
|
|
ompi_group_t *new_group_pointer;
|
|
|
|
MEMCHECKER(
|
|
memchecker_comm(local_comm);
|
|
memchecker_comm(bridge_comm);
|
|
);
|
|
|
|
if ( MPI_PARAM_CHECK ) {
|
|
OMPI_ERR_INIT_FINALIZE(FUNC_NAME);
|
|
|
|
if ( ompi_comm_invalid ( local_comm ) ||
|
|
( local_comm->c_flags & OMPI_COMM_INTER ) )
|
|
return OMPI_ERRHANDLER_INVOKE ( MPI_COMM_WORLD, MPI_ERR_COMM,
|
|
FUNC_NAME);
|
|
|
|
if ( NULL == newintercomm )
|
|
return OMPI_ERRHANDLER_INVOKE ( local_comm, MPI_ERR_ARG,
|
|
FUNC_NAME);
|
|
|
|
/* if ( tag < 0 || tag > MPI_TAG_UB )
|
|
return OMPI_ERRHANDLER_INVOKE ( local_comm, MPI_ERR_ARG,
|
|
FUNC_NAME);
|
|
*/
|
|
}
|
|
|
|
OPAL_CR_ENTER_LIBRARY();
|
|
|
|
local_size = ompi_comm_size ( local_comm );
|
|
local_rank = ompi_comm_rank ( local_comm );
|
|
lleader = local_leader;
|
|
rleader = remote_leader;
|
|
|
|
if ( MPI_PARAM_CHECK ) {
|
|
if ( (0 > local_leader) || (local_leader >= local_size) )
|
|
return OMPI_ERRHANDLER_INVOKE ( local_comm, MPI_ERR_ARG,
|
|
FUNC_NAME);
|
|
|
|
/* remember that the remote_leader and bridge_comm arguments
|
|
just have to be valid at the local_leader */
|
|
if ( local_rank == local_leader ) {
|
|
if ( ompi_comm_invalid ( bridge_comm ) ||
|
|
(bridge_comm->c_flags & OMPI_COMM_INTER) ) {
|
|
OPAL_CR_EXIT_LIBRARY();
|
|
return OMPI_ERRHANDLER_INVOKE ( local_comm, MPI_ERR_COMM,
|
|
FUNC_NAME);
|
|
}
|
|
if ( (remote_leader < 0) || (remote_leader >= ompi_comm_size(bridge_comm))) {
|
|
OPAL_CR_EXIT_LIBRARY();
|
|
return OMPI_ERRHANDLER_INVOKE ( local_comm, MPI_ERR_ARG,
|
|
FUNC_NAME);
|
|
}
|
|
} /* if ( local_rank == local_leader ) */
|
|
}
|
|
|
|
if ( local_rank == local_leader ) {
|
|
MPI_Request req;
|
|
|
|
/* local leader exchange group sizes lists */
|
|
rc = MCA_PML_CALL(irecv(&rsize, 1, MPI_INT, rleader, tag, bridge_comm,
|
|
&req));
|
|
if ( rc != MPI_SUCCESS ) {
|
|
goto err_exit;
|
|
}
|
|
rc = MCA_PML_CALL(send (&local_size, 1, MPI_INT, rleader, tag,
|
|
MCA_PML_BASE_SEND_STANDARD, bridge_comm));
|
|
if ( rc != MPI_SUCCESS ) {
|
|
goto err_exit;
|
|
}
|
|
rc = ompi_request_wait_all ( 1, &req, MPI_STATUS_IGNORE);
|
|
if ( rc != MPI_SUCCESS ) {
|
|
goto err_exit;
|
|
}
|
|
}
|
|
|
|
/* bcast size and list of remote processes to all processes in local_comm */
|
|
rc = local_comm->c_coll.coll_bcast ( &rsize, 1, MPI_INT, lleader,
|
|
local_comm,
|
|
local_comm->c_coll.coll_bcast_module);
|
|
if ( rc != MPI_SUCCESS ) {
|
|
goto err_exit;
|
|
}
|
|
|
|
rprocs = ompi_comm_get_rprocs ( local_comm, bridge_comm, lleader,
|
|
remote_leader, tag, rsize );
|
|
if ( NULL == rprocs ) {
|
|
goto err_exit;
|
|
}
|
|
|
|
if ( MPI_PARAM_CHECK ) {
|
|
if(OMPI_GROUP_IS_DENSE(local_comm->c_local_group)) {
|
|
rc = ompi_comm_overlapping_groups(local_comm->c_local_group->grp_proc_count,
|
|
local_comm->c_local_group->grp_proc_pointers,
|
|
rsize,
|
|
rprocs);
|
|
}
|
|
else {
|
|
proc_list = (ompi_proc_t **) calloc (local_comm->c_local_group->grp_proc_count,
|
|
sizeof (ompi_proc_t *));
|
|
for(j=0 ; j<local_comm->c_local_group->grp_proc_count ; j++) {
|
|
proc_list[j] = ompi_group_peer_lookup(local_comm->c_local_group,j);
|
|
}
|
|
rc = ompi_comm_overlapping_groups(local_comm->c_local_group->grp_proc_count,
|
|
proc_list,
|
|
rsize,
|
|
rprocs);
|
|
}
|
|
if ( OMPI_SUCCESS != rc ) {
|
|
goto err_exit;
|
|
}
|
|
}
|
|
new_group_pointer=ompi_group_allocate(rsize);
|
|
if( NULL == new_group_pointer ) {
|
|
OPAL_CR_EXIT_LIBRARY();
|
|
return MPI_ERR_GROUP;
|
|
}
|
|
|
|
/* put group elements in the list */
|
|
for (j = 0; j < rsize; j++) {
|
|
new_group_pointer->grp_proc_pointers[j] = rprocs[j];
|
|
}
|
|
|
|
ompi_group_increment_proc_count(new_group_pointer);
|
|
|
|
rc = ompi_comm_set ( &newcomp, /* new comm */
|
|
local_comm, /* old comm */
|
|
local_comm->c_local_group->grp_proc_count, /* local_size */
|
|
NULL, /* local_procs*/
|
|
rsize, /* remote_size */
|
|
NULL, /* remote_procs */
|
|
NULL, /* attrs */
|
|
local_comm->error_handler, /* error handler*/
|
|
NULL, /* topo mpodule */
|
|
local_comm->c_local_group, /* local group */
|
|
new_group_pointer /* remote group */
|
|
);
|
|
|
|
if ( NULL == newcomp ) {
|
|
rc = MPI_ERR_INTERN;
|
|
goto err_exit;
|
|
}
|
|
if ( MPI_SUCCESS != rc ) {
|
|
goto err_exit;
|
|
}
|
|
|
|
ompi_group_decrement_proc_count (new_group_pointer);
|
|
OBJ_RELEASE(new_group_pointer);
|
|
new_group_pointer = MPI_GROUP_NULL;
|
|
|
|
/* Determine context id. It is identical to f_2_c_handle */
|
|
rc = ompi_comm_nextcid ( newcomp, /* new comm */
|
|
local_comm, /* old comm */
|
|
bridge_comm, /* bridge comm */
|
|
&lleader, /* local leader */
|
|
&rleader, /* remote_leader */
|
|
OMPI_COMM_CID_INTRA_BRIDGE, /* mode */
|
|
-1 ); /* send_first */
|
|
|
|
if ( MPI_SUCCESS != rc ) {
|
|
goto err_exit;
|
|
}
|
|
|
|
/* activate comm and init coll-module */
|
|
rc = ompi_comm_activate ( &newcomp,
|
|
local_comm, /* old comm */
|
|
bridge_comm, /* bridge comm */
|
|
&lleader, /* local leader */
|
|
&rleader, /* remote_leader */
|
|
OMPI_COMM_CID_INTRA_BRIDGE, /* mode */
|
|
-1 ); /* send_first */
|
|
if ( MPI_SUCCESS != rc ) {
|
|
goto err_exit;
|
|
}
|
|
|
|
err_exit:
|
|
OPAL_CR_EXIT_LIBRARY();
|
|
|
|
if ( NULL != rprocs ) {
|
|
free ( rprocs );
|
|
}
|
|
if ( NULL != proc_list ) {
|
|
free ( proc_list );
|
|
}
|
|
if ( OMPI_SUCCESS != rc ) {
|
|
*newintercomm = MPI_COMM_NULL;
|
|
return OMPI_ERRHANDLER_INVOKE(local_comm, MPI_ERR_INTERN,
|
|
FUNC_NAME);
|
|
}
|
|
|
|
*newintercomm = newcomp;
|
|
return MPI_SUCCESS;
|
|
}
|
|
|