1
1
openmpi/ompi/mca/osc/rdma/osc_rdma_peer.c
2016-11-22 15:03:20 -08:00

343 строки
12 KiB
C

/* -*- Mode: C; c-basic-offset:4 ; indent-tabs-mode:nil -*- */
/*
* Copyright (c) 2007-2016 Los Alamos National Security, LLC. All rights
* reserved.
* Copyright (c) 2015 Research Organization for Information Science
* and Technology (RIST). All rights reserved.
* Copyright (c) 2016 Cisco Systems, Inc. All rights reserved.
* $COPYRIGHT$
*
* Additional copyrights may follow
*
* $HEADER$
*/
#include "ompi_config.h"
#ifdef HAVE_ALLOCA_H
#include <alloca.h>
#endif
#include "osc_rdma_comm.h"
#include "ompi/mca/bml/base/base.h"
#define NODE_ID_TO_RANK(module, peer_data, node_id) ((int)(peer_data)->len)
/**
* @brief find the btl endpoint for a process
*
* @param[in] module osc rdma module
* @param[in] peer_id process rank in the module communicator
*
* @returns NULL on error
* @returns btl endpoint on success
*/
struct mca_btl_base_endpoint_t *ompi_osc_rdma_peer_btl_endpoint (struct ompi_osc_rdma_module_t *module, int peer_id)
{
ompi_proc_t *proc = ompi_comm_peer_lookup (module->comm, peer_id);
mca_bml_base_endpoint_t *bml_endpoint;
int num_btls;
/* for not just use the bml to get the btl endpoint */
bml_endpoint = mca_bml_base_get_endpoint (proc);
num_btls = mca_bml_base_btl_array_get_size (&bml_endpoint->btl_rdma);
for (int btl_index = 0 ; btl_index < num_btls ; ++btl_index) {
if (bml_endpoint->btl_rdma.bml_btls[btl_index].btl == module->selected_btl) {
return bml_endpoint->btl_rdma.bml_btls[btl_index].btl_endpoint;
}
}
/* very unlikely. if this happened the btl section process is broken */
return NULL;
}
int ompi_osc_rdma_new_peer (struct ompi_osc_rdma_module_t *module, int peer_id, ompi_osc_rdma_peer_t **peer_out) {
struct mca_btl_base_endpoint_t *endpoint;
ompi_osc_rdma_peer_t *peer;
*peer_out = NULL;
endpoint = ompi_osc_rdma_peer_btl_endpoint (module, peer_id);
if (OPAL_UNLIKELY(NULL == endpoint)) {
return OMPI_ERR_UNREACH;
}
if (MPI_WIN_FLAVOR_DYNAMIC == module->flavor) {
peer = (ompi_osc_rdma_peer_t *) OBJ_NEW(ompi_osc_rdma_peer_dynamic_t);
} else if (module->same_size && module->same_disp_unit) {
/* use a smaller peer object when same_size and same_disp_unit are set */
peer = (ompi_osc_rdma_peer_t *) OBJ_NEW(ompi_osc_rdma_peer_basic_t);
} else {
peer = (ompi_osc_rdma_peer_t *) OBJ_NEW(ompi_osc_rdma_peer_extended_t);
}
peer->data_endpoint = endpoint;
peer->rank = peer_id;
*peer_out = peer;
return OMPI_SUCCESS;
}
/**
* @brief finish initializing a peer object
*
* @param[in] module osc rdma module
* @param[in] peer peer object to set up
*
* This function reads the registration handle and state pointer from the peer that holds that data. If necessary
* it will then ready information about the peer from its state data structure. This information includes the
* displacement unit, base pointer, window size, and registation handle (if applicable).
*/
static int ompi_osc_rdma_peer_setup (ompi_osc_rdma_module_t *module, ompi_osc_rdma_peer_t *peer)
{
ompi_osc_rdma_peer_extended_t *ex_peer = (ompi_osc_rdma_peer_extended_t *) peer;
uint64_t peer_data_size;
uint64_t peer_data_offset, array_pointer;
struct mca_btl_base_endpoint_t *array_endpoint;
ompi_osc_rdma_region_t *array_peer_data, *node_peer_data;
ompi_osc_rdma_rank_data_t rank_data;
int registration_handle_size = 0;
int node_id, node_rank, array_index;
int ret, disp_unit, comm_size;
char *peer_data;
OSC_RDMA_VERBOSE(MCA_BASE_VERBOSE_DEBUG, "configuring peer for rank %d", peer->rank);
if (module->selected_btl->btl_register_mem) {
registration_handle_size = module->selected_btl->btl_registration_handle_size;
}
comm_size = ompi_comm_size (module->comm);
/* each node is responsible for holding a part of the rank -> node/local rank mapping array. this code
* calculates the node and offset the mapping can be found. once the mapping has been read the state
* part of the peer structure can be initialized. */
node_id = (peer->rank * module->node_count) / comm_size;
array_peer_data = (ompi_osc_rdma_region_t *) ((intptr_t) module->node_comm_info + node_id * module->region_size);
/* the node leader rank is stored in the length field */
node_rank = NODE_ID_TO_RANK(module, array_peer_data, node_id);
array_index = peer->rank % ((comm_size + module->node_count - 1) / module->node_count);
array_pointer = array_peer_data->base + array_index * sizeof (rank_data);
/* lookup the btl endpoint needed to retrieve the mapping */
array_endpoint = ompi_osc_rdma_peer_btl_endpoint (module, node_rank);
if (OPAL_UNLIKELY(NULL == array_endpoint)) {
return OMPI_ERR_UNREACH;
}
OSC_RDMA_VERBOSE(MCA_BASE_VERBOSE_DEBUG, "reading region data for %d from rank: %d, index: %d, pointer: 0x%" PRIx64
", size: %lu", peer->rank, node_rank, array_index, array_pointer, sizeof (rank_data));
ret = ompi_osc_get_data_blocking (module, array_endpoint, array_pointer, (mca_btl_base_registration_handle_t *) array_peer_data->btl_handle_data,
&rank_data, sizeof (rank_data));
if (OPAL_UNLIKELY(OMPI_SUCCESS != ret)) {
return ret;
}
/* initialize the state part of the peer object. NTH: for now the state data is for every node is stored on
* every node. this gives a good balance of code complexity and memory usage at this time. we take advantage
* of this by re-using the endpoint and pointer stored in the node_comm_info array. */
node_peer_data = (ompi_osc_rdma_region_t *) ((intptr_t) module->node_comm_info + rank_data.node_id * module->region_size);
peer->state = node_peer_data->base + module->state_offset + module->state_size * rank_data.rank;
if (registration_handle_size) {
peer->state_handle = (mca_btl_base_registration_handle_t *) node_peer_data->btl_handle_data;
}
peer->state_endpoint = ompi_osc_rdma_peer_btl_endpoint (module, NODE_ID_TO_RANK(module, node_peer_data, rank_data.node_id));
if (OPAL_UNLIKELY(NULL == peer->state_endpoint)) {
return OPAL_ERR_UNREACH;
}
/* nothing more to do for dynamic memory windows */
if (MPI_WIN_FLAVOR_DYNAMIC == module->flavor) {
return OMPI_SUCCESS;
}
/* read window data from the target rank */
if (module->same_disp_unit) {
/* do not bother reading the displacement unit as it is already known */
peer_data_offset = offsetof (ompi_osc_rdma_state_t, regions);
} else {
peer_data_offset = offsetof (ompi_osc_rdma_state_t, disp_unit);
}
peer_data_size = module->state_size - peer_data_offset;
peer_data = alloca (peer_data_size);
/* read window data from the end of the target's state structure */
ret = ompi_osc_get_data_blocking (module, peer->state_endpoint, peer->state + peer_data_offset, peer->state_handle,
peer_data, peer_data_size);
if (OPAL_UNLIKELY(OMPI_SUCCESS != ret)) {
return ret;
}
if (!module->same_disp_unit) {
/* unpack displacement */
memcpy (&ex_peer->disp_unit, peer_data, sizeof (ex_peer->disp_unit));
peer_data += offsetof (ompi_osc_rdma_state_t, regions) - offsetof (ompi_osc_rdma_state_t, disp_unit);
disp_unit = ex_peer->disp_unit;
} else {
disp_unit = module->disp_unit;
}
ompi_osc_rdma_region_t *base_region = (ompi_osc_rdma_region_t *) peer_data;
OSC_RDMA_VERBOSE(MCA_BASE_VERBOSE_DEBUG, "peer %d: remote base region: 0x%" PRIx64 ", size: %" PRId64
", flags: 0x%x, disp_unit: %d", peer->rank, base_region->base, base_region->len,
peer->flags, disp_unit);
(void)disp_unit; // silence compiler warning
if (ompi_osc_rdma_peer_local_base (peer)) {
/* for now we store the local address in the standard place. do no overwrite it */
return OMPI_SUCCESS;
}
ex_peer->super.base = base_region->base;
/* save size and base */
if (!module->same_size) {
ex_peer->size = base_region->len;
}
if (base_region->len) {
if (registration_handle_size) {
ex_peer->super.base_handle = malloc (registration_handle_size);
if (OPAL_UNLIKELY(NULL == ex_peer->super.base_handle)) {
return OMPI_ERR_OUT_OF_RESOURCE;
}
peer->flags |= OMPI_OSC_RDMA_PEER_BASE_FREE;
memcpy (ex_peer->super.base_handle, base_region->btl_handle_data, registration_handle_size);
}
if (MPI_WIN_FLAVOR_ALLOCATE == module->flavor) {
ex_peer->super.super.data_endpoint = ex_peer->super.super.state_endpoint;
}
}
return OMPI_SUCCESS;
}
/**
* @brief lookup (or allocate) a peer for a rank (internal)
*
* @param[in] module osc rdma module
* @param[in] peer_id rank of remote peer (in module communicator)
*
* @returns peer object on success
* @returns NULL on error
*
* This is an internal function for looking up or allocating a peer object for a window rank. This
* function requires the peer lock to be held and is only expected to be called from itself or
* the ompi_osc_rdma_peer_lookup() helper function.
*/
static struct ompi_osc_rdma_peer_t *ompi_osc_rdma_peer_lookup_internal (struct ompi_osc_rdma_module_t *module, int peer_id)
{
ompi_osc_rdma_peer_t *peer;
int ret;
OSC_RDMA_VERBOSE(MCA_BASE_VERBOSE_DEBUG, "looking up peer data for rank %d", peer_id);
peer = ompi_osc_module_get_peer (module, peer_id);
if (NULL != peer) {
return peer;
}
ret = ompi_osc_rdma_new_peer (module, peer_id, &peer);
if (OPAL_UNLIKELY(OMPI_SUCCESS != ret)) {
return NULL;
}
ret = ompi_osc_rdma_peer_setup (module, peer);
if (OPAL_UNLIKELY(OMPI_SUCCESS != ret)) {
OBJ_RELEASE(peer);
return NULL;
}
ret = ompi_osc_module_add_peer (module, peer);
if (OPAL_SUCCESS != ret) {
/* out of memory */
OBJ_RELEASE(peer);
return NULL;
}
/* ensure the peer hash is updated before we drop the lock */
opal_atomic_wmb ();
return peer;
}
struct ompi_osc_rdma_peer_t *ompi_osc_rdma_peer_lookup (struct ompi_osc_rdma_module_t *module, int peer_id)
{
struct ompi_osc_rdma_peer_t *peer;
opal_mutex_lock (&module->peer_lock);
peer = ompi_osc_rdma_peer_lookup_internal (module, peer_id);
opal_mutex_unlock (&module->peer_lock);
return peer;
}
/******* peer objects *******/
static void ompi_osc_rdma_peer_construct (ompi_osc_rdma_peer_t *peer)
{
memset ((char *) peer + sizeof (peer->super), 0, sizeof (*peer) - sizeof (peer->super));
}
static void ompi_osc_rdma_peer_destruct (ompi_osc_rdma_peer_t *peer)
{
if (peer->state_handle && (peer->flags & OMPI_OSC_RDMA_PEER_STATE_FREE)) {
free (peer->state_handle);
}
}
OBJ_CLASS_INSTANCE(ompi_osc_rdma_peer_t, opal_object_t,
ompi_osc_rdma_peer_construct,
ompi_osc_rdma_peer_destruct);
static void ompi_osc_rdma_peer_basic_construct (ompi_osc_rdma_peer_basic_t *peer)
{
memset ((char *) peer + sizeof (peer->super), 0, sizeof (*peer) - sizeof (peer->super));
}
static void ompi_osc_rdma_peer_basic_destruct (ompi_osc_rdma_peer_basic_t *peer)
{
if (peer->base_handle && (peer->super.flags & OMPI_OSC_RDMA_PEER_BASE_FREE)) {
free (peer->base_handle);
}
}
OBJ_CLASS_INSTANCE(ompi_osc_rdma_peer_basic_t, ompi_osc_rdma_peer_t,
ompi_osc_rdma_peer_basic_construct,
ompi_osc_rdma_peer_basic_destruct);
OBJ_CLASS_INSTANCE(ompi_osc_rdma_peer_extended_t, ompi_osc_rdma_peer_basic_t,
NULL, NULL);
static void ompi_osc_rdma_peer_dynamic_construct (ompi_osc_rdma_peer_dynamic_t *peer)
{
memset ((char *) peer + sizeof (peer->super), 0, sizeof (*peer) - sizeof (peer->super));
}
static void ompi_osc_rdma_peer_dynamic_destruct (ompi_osc_rdma_peer_dynamic_t *peer)
{
if (peer->regions) {
free (peer->regions);
}
}
OBJ_CLASS_INSTANCE(ompi_osc_rdma_peer_dynamic_t, ompi_osc_rdma_peer_t,
ompi_osc_rdma_peer_dynamic_construct,
ompi_osc_rdma_peer_dynamic_destruct);