
This commit rewrites parts of libnbc to fix issues identified by coverity and myself. The changes are as follows: - libnbc function would return invalid error codes (internal to libnbc) to the mpi layer. These codes names are of the form NBC_. They do not match up with the error codes expected by the mpi layer. I purged the use of all these error codes with the exception of NBC_OK and NBC_CONTINUE in progress. These codes are used to identify when a request handle is complete. - Handles and schedules were leaked by all collective routines on error. A new routine was added to return a collective handle (NBC_Return_handle). - Temporary buffers containting in/out neighbors for neighborhood collectives were always leaked. - Neigborhood collectives contained code to handle MPI_IN_PLACE which is never a valid input for the send or receive buffer. Stipped this code out. - Files were inconsistently named. Most are nbc_isomething.c but one was named coll_libnbc_ireduce_scatter_block.c. - Made the NBC_Schedule "structure" and object so it can be retained/released. This may enable the use of schedule caching at a later time. More testing will be needed to ensure the caching code works. If it doesn't the code should be stripped out completely. - Added code to simply common case of scheduling send/recv + barrier. - Code cleanup for readability. The code now passes the clang static analyzer. Signed-off-by: Nathan Hjelm <hjelmn@lanl.gov>
183 строки
5.5 KiB
C
183 строки
5.5 KiB
C
/* -*- Mode: C; c-basic-offset:2 ; indent-tabs-mode:nil -*- */
|
|
/*
|
|
* Copyright (c) 2006 The Trustees of Indiana University and Indiana
|
|
* University Research and Technology
|
|
* Corporation. All rights reserved.
|
|
* Copyright (c) 2006 The Technical University of Chemnitz. All
|
|
* rights reserved.
|
|
*
|
|
* Author(s): Torsten Hoefler <htor@cs.indiana.edu>
|
|
*
|
|
* Copyright (c) 2012 Oracle and/or its affiliates. All rights reserved.
|
|
* Copyright (c) 2013-2015 Los Alamos National Security, LLC. All rights
|
|
* reserved.
|
|
* Copyright (c) 2014 Research Organization for Information Science
|
|
* and Technology (RIST). All rights reserved.
|
|
*
|
|
*/
|
|
#include "nbc_internal.h"
|
|
|
|
/* an allgatherv schedule can not be cached easily because the contents
|
|
* ot the recvcounts array may change, so a comparison of the address
|
|
* would not be sufficient ... we simply do not cache it */
|
|
|
|
/* simple linear MPI_Iallgatherv
|
|
* the algorithm uses p-1 rounds
|
|
* first round:
|
|
* each node sends to it's left node (rank+1)%p sendcount elements
|
|
* each node begins with it's right node (rank-11)%p and receives from it recvcounts[(rank+1)%p] elements
|
|
* second round:
|
|
* each node sends to node (rank+2)%p sendcount elements
|
|
* each node receives from node (rank-2)%p recvcounts[(rank+2)%p] elements */
|
|
int ompi_coll_libnbc_iallgatherv(void* sendbuf, int sendcount, MPI_Datatype sendtype, void* recvbuf, int *recvcounts, int *displs,
|
|
MPI_Datatype recvtype, struct ompi_communicator_t *comm, ompi_request_t ** request,
|
|
struct mca_coll_base_module_2_1_0_t *module)
|
|
{
|
|
int rank, p, res, speer, rpeer;
|
|
MPI_Aint rcvext;
|
|
NBC_Schedule *schedule;
|
|
char *rbuf, *sbuf, inplace;
|
|
NBC_Handle *handle;
|
|
ompi_coll_libnbc_module_t *libnbc_module = (ompi_coll_libnbc_module_t*) module;
|
|
|
|
NBC_IN_PLACE(sendbuf, recvbuf, inplace);
|
|
|
|
rank = ompi_comm_rank (comm);
|
|
p = ompi_comm_size (comm);
|
|
|
|
res = MPI_Type_extent (recvtype, &rcvext);
|
|
if (OPAL_UNLIKELY(MPI_SUCCESS != res)) {
|
|
NBC_Error ("MPI Error in MPI_Type_extent() (%i)", res);
|
|
return res;
|
|
}
|
|
|
|
if (inplace) {
|
|
sendtype = recvtype;
|
|
sendcount = recvcounts[rank];
|
|
} else {
|
|
/* copy my data to receive buffer */
|
|
rbuf = (char *) recvbuf + displs[rank] * rcvext;
|
|
res = NBC_Copy (sendbuf, sendcount, sendtype, rbuf, recvcounts[rank], recvtype, comm);
|
|
if (OPAL_UNLIKELY(OMPI_SUCCESS != res)) {
|
|
return res;
|
|
}
|
|
}
|
|
|
|
schedule = OBJ_NEW(NBC_Schedule);
|
|
if (NULL == schedule) {
|
|
return OMPI_ERR_OUT_OF_RESOURCE;
|
|
}
|
|
|
|
sbuf = (char *) recvbuf + displs[rank] * rcvext;
|
|
|
|
/* do p-1 rounds */
|
|
for (int r = 1 ; r < p ; ++r) {
|
|
speer = (rank + r) % p;
|
|
rpeer = (rank - r + p) % p;
|
|
rbuf = (char *)recvbuf + displs[rpeer] * rcvext;
|
|
|
|
res = NBC_Sched_recv (rbuf, false, recvcounts[rpeer], recvtype, rpeer, schedule, false);
|
|
if (OPAL_UNLIKELY(OMPI_SUCCESS != res)) {
|
|
OBJ_RELEASE(schedule);
|
|
return res;
|
|
}
|
|
|
|
res = NBC_Sched_send (sbuf, false, sendcount, sendtype, speer, schedule, false);
|
|
if (OPAL_UNLIKELY(OMPI_SUCCESS != res)) {
|
|
OBJ_RELEASE(schedule);
|
|
return res;
|
|
}
|
|
}
|
|
|
|
res = NBC_Sched_commit (schedule);
|
|
if (OPAL_UNLIKELY(OMPI_SUCCESS != res)) {
|
|
OBJ_RELEASE(schedule);
|
|
return res;
|
|
}
|
|
|
|
res = NBC_Init_handle (comm, &handle, libnbc_module);
|
|
if (OPAL_UNLIKELY(OMPI_SUCCESS != res)) {
|
|
OBJ_RELEASE(schedule);
|
|
return res;
|
|
}
|
|
|
|
res = NBC_Start (handle, schedule);
|
|
if (OPAL_UNLIKELY(OMPI_SUCCESS != res)) {
|
|
NBC_Return_handle (handle);
|
|
return res;
|
|
}
|
|
|
|
*request = (ompi_request_t *) handle;
|
|
|
|
return OMPI_SUCCESS;
|
|
}
|
|
|
|
int ompi_coll_libnbc_iallgatherv_inter(void* sendbuf, int sendcount, MPI_Datatype sendtype, void* recvbuf, int *recvcounts, int *displs,
|
|
MPI_Datatype recvtype, struct ompi_communicator_t *comm, ompi_request_t ** request,
|
|
struct mca_coll_base_module_2_1_0_t *module)
|
|
{
|
|
int res, rsize;
|
|
MPI_Aint rcvext;
|
|
NBC_Schedule *schedule;
|
|
NBC_Handle *handle;
|
|
ompi_coll_libnbc_module_t *libnbc_module = (ompi_coll_libnbc_module_t*) module;
|
|
|
|
rsize = ompi_comm_remote_size (comm);
|
|
|
|
res = MPI_Type_extent(recvtype, &rcvext);
|
|
if (OPAL_UNLIKELY(MPI_SUCCESS != res)) {
|
|
NBC_Error ("MPI Error in MPI_Type_extent() (%i)", res);
|
|
return res;
|
|
}
|
|
|
|
schedule = OBJ_NEW(NBC_Schedule);
|
|
if (NULL == schedule) {
|
|
return OMPI_ERR_OUT_OF_RESOURCE;
|
|
}
|
|
|
|
/* do rsize rounds */
|
|
for (int r = 0 ; r < rsize ; ++r) {
|
|
char *rbuf = (char *) recvbuf + displs[r] * rcvext;
|
|
|
|
if (recvcounts[r]) {
|
|
res = NBC_Sched_recv (rbuf, false, recvcounts[r], recvtype, r, schedule, false);
|
|
if (OPAL_UNLIKELY(OMPI_SUCCESS != res)) {
|
|
OBJ_RELEASE(schedule);
|
|
return res;
|
|
}
|
|
}
|
|
}
|
|
|
|
if (sendcount) {
|
|
for (int r = 0 ; r < rsize ; ++r) {
|
|
res = NBC_Sched_send (sendbuf, false, sendcount, sendtype, r, schedule, false);
|
|
if (OPAL_UNLIKELY(OMPI_SUCCESS != res)) {
|
|
OBJ_RELEASE(schedule);
|
|
return res;
|
|
}
|
|
}
|
|
}
|
|
|
|
res = NBC_Sched_commit (schedule);
|
|
if (OPAL_UNLIKELY(OMPI_SUCCESS != res)) {
|
|
OBJ_RELEASE(schedule);
|
|
return res;
|
|
}
|
|
|
|
res = NBC_Init_handle (comm, &handle, libnbc_module);
|
|
if (OPAL_UNLIKELY(OMPI_SUCCESS != res)) {
|
|
OBJ_RELEASE(schedule);
|
|
return res;
|
|
}
|
|
|
|
res = NBC_Start (handle, schedule);
|
|
if (OPAL_UNLIKELY(OMPI_SUCCESS != res)) {
|
|
NBC_Return_handle (handle);
|
|
return res;
|
|
}
|
|
|
|
*request = (ompi_request_t *) handle;
|
|
|
|
return OMPI_SUCCESS;
|
|
}
|