1
1
openmpi/ompi/mca/coll/base/coll_base_scan.c
Mikhail Kurnosov 177c6ce51f Move algorithms from coll/spacc to coll/base and remove coll/spacc
Signed-off-by: Mikhail Kurnosov <mkurnosov@gmail.com>
2018-04-04 10:21:06 +07:00

131 строка
4.9 KiB
C

/* -*- Mode: C; c-basic-offset:4 ; indent-tabs-mode:nil -*- */
/*
* Copyright (c) 2018 Siberian State University of Telecommunications
* and Information Science. All rights reserved.
* $COPYRIGHT$
*
* Additional copyrights may follow
*
* $HEADER$
*/
#include "ompi_config.h"
#include "mpi.h"
#include "ompi/constants.h"
#include "ompi/datatype/ompi_datatype.h"
#include "ompi/communicator/communicator.h"
#include "ompi/mca/coll/coll.h"
#include "ompi/mca/coll/base/coll_base_functions.h"
#include "ompi/mca/coll/base/coll_tags.h"
#include "ompi/mca/coll/base/coll_base_util.h"
#include "ompi/mca/pml/pml.h"
#include "ompi/op/op.h"
/*
* ompi_coll_base_scan_intra_recursivedoubling
*
* Function: Recursive doubling algorithm for inclusive scan.
* Accepts: Same as MPI_Scan
* Returns: MPI_SUCCESS or error code
*
* Description: Implements recursive doubling algorithm for MPI_Scan.
* The algorithm preserves order of operations so it can
* be used both by commutative and non-commutative operations.
*
* Example for 5 processes and commutative operation MPI_SUM:
* Process: 0 1 2 3 4
* recvbuf: [0] [1] [2] [3] [4]
* psend: [0] [1] [2] [3] [4]
*
* Step 1:
* recvbuf: [0] [0+1] [2] [2+3] [4]
* psend: [1+0] [0+1] [3+2] [2+3] [4]
*
* Step 2:
* recvbuf: [0] [0+1] [(1+0)+2] [(1+0)+(2+3)] [4]
* psend: [(3+2)+(1+0)] [(2+3)+(0+1)] [(1+0)+(3+2)] [(1+0)+(2+3)] [4]
*
* Step 3:
* recvbuf: [0] [0+1] [(1+0)+2] [(1+0)+(2+3)] [((3+2)+(1+0))+4]
* psend: [4+((3+2)+(1+0))] [((3+2)+(1+0))+4]
*
* Time complexity (worst case): \ceil(\log_2(p))(2\alpha + 2m\beta + 2m\gamma)
* Memory requirements (per process): 2 * count * typesize = O(count)
* Limitations: intra-communicators only
*/
int ompi_coll_base_scan_intra_recursivedoubling(
const void *sendbuf, void *recvbuf, int count, struct ompi_datatype_t *datatype,
struct ompi_op_t *op, struct ompi_communicator_t *comm,
mca_coll_base_module_t *module)
{
int err = MPI_SUCCESS;
char *tmpsend_raw = NULL, *tmprecv_raw = NULL;
int comm_size = ompi_comm_size(comm);
int rank = ompi_comm_rank(comm);
OPAL_OUTPUT((ompi_coll_base_framework.framework_output,
"coll:base:scan_intra_recursivedoubling: rank %d/%d",
rank, comm_size));
if (count == 0)
return MPI_SUCCESS;
if (sendbuf != MPI_IN_PLACE) {
err = ompi_datatype_copy_content_same_ddt(datatype, count, recvbuf, (char *)sendbuf);
if (MPI_SUCCESS != err) { goto cleanup_and_return; }
}
if (comm_size < 2)
return MPI_SUCCESS;
ptrdiff_t dsize, gap;
dsize = opal_datatype_span(&datatype->super, count, &gap);
tmpsend_raw = malloc(dsize);
tmprecv_raw = malloc(dsize);
if (NULL == tmpsend_raw || NULL == tmprecv_raw) {
err = OMPI_ERR_OUT_OF_RESOURCE;
goto cleanup_and_return;
}
char *psend = tmpsend_raw - gap;
char *precv = tmprecv_raw - gap;
err = ompi_datatype_copy_content_same_ddt(datatype, count, psend, recvbuf);
if (MPI_SUCCESS != err) { goto cleanup_and_return; }
int is_commute = ompi_op_is_commute(op);
for (int mask = 1; mask < comm_size; mask <<= 1) {
int remote = rank ^ mask;
if (remote < comm_size) {
err = ompi_coll_base_sendrecv(psend, count, datatype, remote,
MCA_COLL_BASE_TAG_SCAN,
precv, count, datatype, remote,
MCA_COLL_BASE_TAG_SCAN, comm,
MPI_STATUS_IGNORE, rank);
if (MPI_SUCCESS != err) { goto cleanup_and_return; }
if (rank > remote) {
/* Accumulate prefix reduction: recvbuf = precv <op> recvbuf */
ompi_op_reduce(op, precv, recvbuf, count, datatype);
/* Partial result: psend = precv <op> psend */
ompi_op_reduce(op, precv, psend, count, datatype);
} else {
if (is_commute) {
/* psend = precv <op> psend */
ompi_op_reduce(op, precv, psend, count, datatype);
} else {
/* precv = psend <op> precv */
ompi_op_reduce(op, psend, precv, count, datatype);
char *tmp = psend;
psend = precv;
precv = tmp;
}
}
}
}
cleanup_and_return:
if (NULL != tmpsend_raw)
free(tmpsend_raw);
if (NULL != tmprecv_raw)
free(tmprecv_raw);
return err;
}