566a050c23
- move files out of toplevel include/ and etc/, moving it into the sub-projects - rather than including config headers with <project>/include, have them as <project> - require all headers to be included with a project prefix, with the exception of the config headers ({opal,orte,ompi}_config.h mpi.h, and mpif.h) This commit was SVN r8985.
91 строка
2.6 KiB
C
91 строка
2.6 KiB
C
/*
|
|
* Copyright (c) 2004-2005 The Trustees of Indiana University and Indiana
|
|
* University Research and Technology
|
|
* Corporation. All rights reserved.
|
|
* Copyright (c) 2004-2005 The University of Tennessee and The University
|
|
* of Tennessee Research Foundation. All rights
|
|
* reserved.
|
|
* Copyright (c) 2004-2005 High Performance Computing Center Stuttgart,
|
|
* University of Stuttgart. All rights reserved.
|
|
* Copyright (c) 2004-2005 The Regents of the University of California.
|
|
* All rights reserved.
|
|
* $COPYRIGHT$
|
|
*
|
|
* Additional copyrights may follow
|
|
*
|
|
* $HEADER$
|
|
*/
|
|
#include "ompi_config.h"
|
|
#include <stdio.h>
|
|
|
|
#include "ompi/mpi/c/bindings.h"
|
|
#include "ompi/datatype/datatype.h"
|
|
|
|
#if OMPI_HAVE_WEAK_SYMBOLS && OMPI_PROFILING_DEFINES
|
|
#pragma weak MPI_Bcast = PMPI_Bcast
|
|
#endif
|
|
|
|
#if OMPI_PROFILING_DEFINES
|
|
#include "ompi/mpi/c/profile/defines.h"
|
|
#endif
|
|
|
|
static const char FUNC_NAME[] = "MPI_Bcast";
|
|
|
|
|
|
int MPI_Bcast(void *buffer, int count, MPI_Datatype datatype,
|
|
int root, MPI_Comm comm)
|
|
{
|
|
int err;
|
|
|
|
if (MPI_PARAM_CHECK) {
|
|
err = MPI_SUCCESS;
|
|
OMPI_ERR_INIT_FINALIZE(FUNC_NAME);
|
|
if (ompi_comm_invalid(comm)) {
|
|
return OMPI_ERRHANDLER_INVOKE(MPI_COMM_WORLD, MPI_ERR_COMM,
|
|
FUNC_NAME);
|
|
}
|
|
|
|
/* Errors for all ranks */
|
|
|
|
OMPI_CHECK_DATATYPE_FOR_SEND(err, datatype, count);
|
|
OMPI_ERRHANDLER_CHECK(err, comm, err, FUNC_NAME);
|
|
if (MPI_IN_PLACE == buffer) {
|
|
return OMPI_ERRHANDLER_INVOKE(comm, MPI_ERR_ARG, FUNC_NAME);
|
|
}
|
|
|
|
/* Errors for intracommunicators */
|
|
|
|
if (OMPI_COMM_IS_INTRA(comm)) {
|
|
if ((root >= ompi_comm_size(comm)) || (root < 0)) {
|
|
return OMPI_ERRHANDLER_INVOKE(comm, MPI_ERR_ROOT, FUNC_NAME);
|
|
}
|
|
}
|
|
|
|
/* Errors for intercommunicators */
|
|
|
|
else {
|
|
if (! ((root >= 0 && root < ompi_comm_remote_size(comm)) ||
|
|
MPI_ROOT == root || MPI_PROC_NULL == root)) {
|
|
return OMPI_ERRHANDLER_INVOKE(comm, MPI_ERR_ROOT, FUNC_NAME);
|
|
}
|
|
}
|
|
}
|
|
|
|
/* If there's only one node, we're done */
|
|
|
|
if (OMPI_COMM_IS_INTRA(comm) && ompi_comm_size(comm) <= 1) {
|
|
return MPI_SUCCESS;
|
|
}
|
|
|
|
/* Can we optimize? */
|
|
|
|
if (count == 0) {
|
|
return MPI_SUCCESS;
|
|
}
|
|
|
|
/* Invoke the coll component to perform the back-end operation */
|
|
|
|
err = comm->c_coll.coll_bcast(buffer, count, datatype, root, comm);
|
|
OMPI_ERRHANDLER_RETURN(err, comm, err, FUNC_NAME);
|
|
}
|