772772b944
This commit was SVN r17813.
713 строки
23 KiB
C
713 строки
23 KiB
C
/*
|
|
* Copyright (c) 2004-2007 The Trustees of Indiana University and Indiana
|
|
* University Research and Technology
|
|
* Corporation. All rights reserved.
|
|
* Copyright (c) 2004-2008 The University of Tennessee and The University
|
|
* of Tennessee Research Foundation. All rights
|
|
* reserved.
|
|
* Copyright (c) 2004-2007 High Performance Computing Center Stuttgart,
|
|
* University of Stuttgart. All rights reserved.
|
|
* Copyright (c) 2004-2005 The Regents of the University of California.
|
|
* All rights reserved.
|
|
* Copyright (c) 2006-2007 Voltaire. All rights reserved.
|
|
* $COPYRIGHT$
|
|
*
|
|
* Additional copyrights may follow
|
|
*
|
|
* $HEADER$
|
|
*/
|
|
|
|
#include "ompi_config.h"
|
|
|
|
#include <string.h>
|
|
#include <stdlib.h>
|
|
#include <sys/types.h>
|
|
#include <sys/stat.h>
|
|
#include <fcntl.h>
|
|
#include <errno.h>
|
|
|
|
#include "opal/threads/mutex.h"
|
|
#include "ompi/datatype/convertor.h"
|
|
#include "opal/sys/atomic.h"
|
|
#include "opal/util/output.h"
|
|
#include "opal/util/if.h"
|
|
#include "orte/util/proc_info.h"
|
|
#include "opal/util/printf.h"
|
|
#include "orte/util/sys_info.h"
|
|
#include "ompi/class/ompi_fifo.h"
|
|
#include "ompi/class/ompi_free_list.h"
|
|
#include "ompi/mca/pml/pml.h"
|
|
#include "ompi/mca/btl/btl.h"
|
|
#include "ompi/mca/mpool/base/base.h"
|
|
#include "ompi/mca/common/sm/common_sm_mmap.h"
|
|
#include "btl_sm.h"
|
|
#include "btl_sm_endpoint.h"
|
|
#include "btl_sm_frag.h"
|
|
#include "btl_sm_fifo.h"
|
|
#include "ompi/proc/proc.h"
|
|
|
|
mca_btl_sm_t mca_btl_sm = {
|
|
{
|
|
&mca_btl_sm_component.super,
|
|
0, /* btl_eager_limit */
|
|
0, /* btl_rndv_eager_limit */
|
|
0, /* btl_max_send_size */
|
|
0, /* btl_rdma_pipeline_send_length */
|
|
0, /* btl_rdma_pipeline_frag_size */
|
|
0, /* btl_min_rdma_pipeline_size */
|
|
0, /* btl_exclusivity */
|
|
0, /* btl_latency */
|
|
0, /* btl_bandwidth */
|
|
0, /* btl flags */
|
|
mca_btl_sm_add_procs,
|
|
mca_btl_sm_del_procs,
|
|
NULL,
|
|
mca_btl_sm_finalize,
|
|
mca_btl_sm_alloc,
|
|
mca_btl_sm_free,
|
|
mca_btl_sm_prepare_src,
|
|
NULL,
|
|
mca_btl_sm_send,
|
|
NULL, /* put */
|
|
NULL, /* get */
|
|
mca_btl_base_dump,
|
|
NULL, /* mpool */
|
|
mca_btl_sm_register_error_cb, /* register error */
|
|
mca_btl_sm_ft_event
|
|
}
|
|
};
|
|
|
|
/* track information needed to synchronise a Shared Memory BTL module */
|
|
mca_btl_sm_module_resource_t mca_btl_sm_module_resource;
|
|
|
|
/*
|
|
* calculate offset of an address from the beginning of a shared memory segment
|
|
*/
|
|
#define ADDR2OFFSET(ADDR, BASE) ((char*)(ADDR) - (char*)(BASE))
|
|
|
|
/*
|
|
* calculate an absolute address in a local address space given an offset and
|
|
* a base address of a shared memory segment
|
|
*/
|
|
#define OFFSET2ADDR(OFFSET, BASE) ((ptrdiff_t)(OFFSET) + (char*)(BASE))
|
|
|
|
|
|
static void *mpool_calloc(size_t nmemb, size_t size)
|
|
{
|
|
void *buf;
|
|
size_t bsize = nmemb * size;
|
|
|
|
buf = mca_btl_sm_component.sm_mpool->mpool_alloc(
|
|
mca_btl_sm_component.sm_mpool, bsize, CACHE_LINE_SIZE, 0, NULL);
|
|
if (NULL == buf)
|
|
return NULL;
|
|
|
|
memset(buf, 0, bsize);
|
|
return buf;
|
|
}
|
|
|
|
static int init_fifos(ompi_fifo_t *f, int n)
|
|
{
|
|
int j;
|
|
for(j=0; j < n; j++) {
|
|
f[j].head = (ompi_cb_fifo_wrapper_t*)OMPI_CB_FREE;
|
|
f[j].tail = (ompi_cb_fifo_wrapper_t*)OMPI_CB_FREE;
|
|
if(opal_using_threads()) {
|
|
char *buf = mpool_calloc(2, CACHE_LINE_SIZE);
|
|
/* allocate head and tail locks on different cache lines */
|
|
if(NULL == buf)
|
|
return OMPI_ERROR;
|
|
|
|
f[j].head_lock = (opal_atomic_lock_t*)buf;
|
|
f[j].tail_lock = (opal_atomic_lock_t*)(buf + CACHE_LINE_SIZE);
|
|
opal_atomic_init(f[j].head_lock, OPAL_ATOMIC_UNLOCKED);
|
|
opal_atomic_init(f[j].tail_lock, OPAL_ATOMIC_UNLOCKED);
|
|
} else {
|
|
f[j].head_lock = NULL;
|
|
f[j].tail_lock = NULL;
|
|
}
|
|
}
|
|
|
|
return OMPI_SUCCESS;
|
|
}
|
|
|
|
static int sm_btl_first_time_init(mca_btl_sm_t *sm_btl, int n)
|
|
{
|
|
size_t size, length, length_payload;
|
|
char *sm_ctl_file;
|
|
mca_btl_sm_module_resource_t *sm_ctl_hdr;
|
|
ompi_fifo_t * volatile *rel_fifosp;
|
|
volatile char **rel_ptr;
|
|
ompi_fifo_t *my_fifos;
|
|
|
|
/* lookup shared memory pool */
|
|
mca_btl_sm_component.sm_mpool =
|
|
mca_mpool_base_module_lookup(mca_btl_sm_component.sm_mpool_name);
|
|
if(NULL == mca_btl_sm_component.sm_mpool) {
|
|
mca_btl_sm_component.sm_mpool =
|
|
mca_mpool_base_module_create(mca_btl_sm_component.sm_mpool_name,
|
|
sm_btl, NULL);
|
|
}
|
|
|
|
/* Sanity check to ensure that we found it */
|
|
if(NULL == mca_btl_sm_component.sm_mpool)
|
|
return OMPI_ERR_OUT_OF_RESOURCE;
|
|
|
|
mca_btl_sm_component.sm_mpool_base =
|
|
mca_btl_sm_component.sm_mpool->mpool_base(mca_btl_sm_component.sm_mpool);
|
|
|
|
/* set the shared memory offset */
|
|
mca_btl_sm_component.sm_offset = (ptrdiff_t*)calloc(n, sizeof(ptrdiff_t));
|
|
if(NULL == mca_btl_sm_component.sm_offset)
|
|
return OMPI_ERR_OUT_OF_RESOURCE;
|
|
|
|
/* create a list of peers */
|
|
mca_btl_sm_component.sm_peers = (struct mca_btl_base_endpoint_t**)
|
|
calloc(n, sizeof(struct mca_btl_base_endpoint_t*));
|
|
if(NULL == mca_btl_sm_component.sm_peers)
|
|
return OMPI_ERR_OUT_OF_RESOURCE;
|
|
|
|
/* Allocate Shared Memory BTL process coordination
|
|
* data structure. This will reside in shared memory */
|
|
|
|
/* set file name */
|
|
if(asprintf(&sm_ctl_file, "%s"OPAL_PATH_SEP"shared_mem_btl_module.%s",
|
|
orte_process_info.job_session_dir,
|
|
orte_system_info.nodename) < 0)
|
|
return OMPI_ERR_OUT_OF_RESOURCE;
|
|
|
|
/* Pass in a data segment alignment of 0 to get no data
|
|
segment (only the shared control structure) */
|
|
size = sizeof(mca_btl_sm_module_resource_t);
|
|
if(NULL == (mca_btl_sm_component.mmap_file =
|
|
mca_common_sm_mmap_init(size, sm_ctl_file, size, 0))) {
|
|
opal_output(0, "mca_btl_sm_add_procs: unable to create shared memory "
|
|
"BTL coordinating strucure :: size %lu \n",
|
|
(unsigned long)size);
|
|
free(sm_ctl_file);
|
|
return OMPI_ERROR;
|
|
}
|
|
|
|
free(sm_ctl_file);
|
|
|
|
/* set the pointer to the shared memory control structure */
|
|
sm_ctl_hdr = (mca_btl_sm_module_resource_t *)
|
|
mca_btl_sm_component.mmap_file->map_seg;
|
|
mca_btl_sm_component.sm_ctl_header = sm_ctl_hdr;
|
|
|
|
/* Allocate a fixed size pointer array for the 2-D Shared memory queues.
|
|
* Excess slots will be allocated for future growth. One could
|
|
* make this array growable, but then one would need to uses mutexes
|
|
* for any access to these queues to ensure data consistancy when
|
|
* the array is grown */
|
|
if(0 == mca_btl_sm_component.my_smp_rank) {
|
|
/* allocate ompi_fifo_t strucutes for each fifo of the queue
|
|
* pairs - one per pair of local processes */
|
|
/* check to make sure number of local procs is within the
|
|
* specified limits */
|
|
if(mca_btl_sm_component.sm_max_procs > 0 &&
|
|
mca_btl_sm_component.num_smp_procs + n >
|
|
mca_btl_sm_component.sm_max_procs) {
|
|
return OMPI_ERROR;
|
|
}
|
|
|
|
/* allocate array of ompi_fifo_t* elements -
|
|
* offset relative to base segement is stored, so that
|
|
* this can be used by other procs */
|
|
sm_ctl_hdr->fifo =
|
|
(volatile ompi_fifo_t**)mpool_calloc(n, sizeof(ompi_fifo_t*));
|
|
if(NULL == sm_ctl_hdr->fifo)
|
|
return OMPI_ERR_OUT_OF_RESOURCE;
|
|
|
|
/* allocate and initialize the array to hold the virtual address
|
|
* of the shared memory base */
|
|
sm_ctl_hdr->segment_header.base_shared_mem_segment =
|
|
(volatile char **)mpool_calloc(n, sizeof(char *));
|
|
if(NULL == sm_ctl_hdr->segment_header.base_shared_mem_segment)
|
|
return OMPI_ERR_OUT_OF_RESOURCE;
|
|
|
|
/* set the addresses to be a relative, so that
|
|
* they can be used by other procs */
|
|
sm_ctl_hdr->fifo = (volatile ompi_fifo_t **)
|
|
ADDR2OFFSET(sm_ctl_hdr->fifo, mca_btl_sm_component.sm_mpool_base);
|
|
|
|
sm_ctl_hdr->segment_header.base_shared_mem_segment = (volatile char **)
|
|
ADDR2OFFSET(sm_ctl_hdr->segment_header.base_shared_mem_segment,
|
|
mca_btl_sm_component.sm_mpool_base);
|
|
|
|
/* allow other procs to use this shared memory map */
|
|
mca_btl_sm_component.mmap_file->map_seg->seg_inited = true;
|
|
|
|
/* memory barrier to ensure this flag is set before other
|
|
* flags are set */
|
|
opal_atomic_mb();
|
|
} else {
|
|
/* Note: Need to make sure that proc 0 initializes control
|
|
* structures before any of the other procs can progress.
|
|
* Spin unitl local proc 0 initializes the segment */
|
|
while(!mca_btl_sm_component.mmap_file->map_seg->seg_inited) {
|
|
opal_atomic_rmb();
|
|
opal_progress();
|
|
}
|
|
}
|
|
|
|
/* set the base of the shared memory segment, and flag
|
|
* indicating that it is set */
|
|
rel_ptr = (volatile char **)
|
|
OFFSET2ADDR(sm_ctl_hdr->segment_header.base_shared_mem_segment,
|
|
mca_btl_sm_component.sm_mpool_base);
|
|
rel_ptr[mca_btl_sm_component.my_smp_rank] =
|
|
(char*)mca_btl_sm_component.sm_mpool_base;
|
|
|
|
/*
|
|
* initialize the array of fifo's "owned" by this process
|
|
* The virtual addresses are valid only in the sender's
|
|
* address space - unless the base of the shared memory
|
|
* segment is mapped at the same location in the reader's
|
|
* virtual address space.
|
|
*/
|
|
if(NULL == (my_fifos = (ompi_fifo_t*)mpool_calloc(n, sizeof(ompi_fifo_t))))
|
|
return OMPI_ERR_OUT_OF_RESOURCE;
|
|
|
|
if(init_fifos(my_fifos, n) != OMPI_SUCCESS)
|
|
return OMPI_ERR_OUT_OF_RESOURCE;
|
|
|
|
rel_fifosp = (ompi_fifo_t * volatile *)
|
|
OFFSET2ADDR(sm_ctl_hdr->fifo, mca_btl_sm_component.sm_mpool_base);
|
|
rel_fifosp[mca_btl_sm_component.my_smp_rank] = my_fifos;
|
|
|
|
opal_atomic_wmb();
|
|
|
|
/* cache the pointer to the 2d fifo array. These addresses
|
|
* are valid in the current process space */
|
|
mca_btl_sm_component.fifo = (ompi_fifo_t**)malloc(sizeof(ompi_fifo_t*) * n);
|
|
|
|
if(NULL == mca_btl_sm_component.fifo)
|
|
return OMPI_ERR_OUT_OF_RESOURCE;
|
|
|
|
mca_btl_sm_component.fifo[mca_btl_sm_component.my_smp_rank] = my_fifos;
|
|
|
|
/* initialize fragment descriptor free lists */
|
|
|
|
/* allocation will be for the fragment descriptor and payload buffer */
|
|
length = sizeof(mca_btl_sm_frag1_t);
|
|
length_payload =
|
|
sizeof(mca_btl_sm_hdr_t) + mca_btl_sm_component.eager_limit;
|
|
ompi_free_list_init_new(&mca_btl_sm_component.sm_frags1, length,
|
|
CACHE_LINE_SIZE, OBJ_CLASS(mca_btl_sm_frag1_t),
|
|
length_payload, CACHE_LINE_SIZE,
|
|
mca_btl_sm_component.sm_free_list_num,
|
|
mca_btl_sm_component.sm_free_list_max,
|
|
mca_btl_sm_component.sm_free_list_inc,
|
|
mca_btl_sm_component.sm_mpool);
|
|
|
|
length = sizeof(mca_btl_sm_frag2_t);
|
|
length_payload =
|
|
sizeof(mca_btl_sm_hdr_t) + mca_btl_sm_component.max_frag_size;
|
|
ompi_free_list_init_new(&mca_btl_sm_component.sm_frags2, length,
|
|
CACHE_LINE_SIZE, OBJ_CLASS(mca_btl_sm_frag2_t),
|
|
length_payload, CACHE_LINE_SIZE,
|
|
mca_btl_sm_component.sm_free_list_num,
|
|
mca_btl_sm_component.sm_free_list_max,
|
|
mca_btl_sm_component.sm_free_list_inc,
|
|
mca_btl_sm_component.sm_mpool);
|
|
|
|
ompi_free_list_init_new(&mca_btl_sm_component.sm_frags,
|
|
sizeof(mca_btl_sm_frag_t), CACHE_LINE_SIZE,
|
|
OBJ_CLASS(mca_btl_sm_frag_t), 0, CACHE_LINE_SIZE,
|
|
mca_btl_sm_component.sm_free_list_num,
|
|
-1,
|
|
mca_btl_sm_component.sm_free_list_inc,
|
|
NULL);
|
|
|
|
opal_free_list_init(&mca_btl_sm_component.pending_send_fl,
|
|
sizeof(btl_sm_pending_send_item_t),
|
|
OBJ_CLASS(opal_free_list_item_t),
|
|
16, -1, 32);
|
|
|
|
/* set flag indicating btl has been inited */
|
|
sm_btl->btl_inited = true;
|
|
|
|
return OMPI_SUCCESS;
|
|
}
|
|
|
|
static struct mca_btl_base_endpoint_t *
|
|
create_sm_endpoint(int local_proc, struct ompi_proc_t *proc)
|
|
{
|
|
struct mca_btl_base_endpoint_t *ep;
|
|
#if OMPI_ENABLE_PROGRESS_THREADS == 1
|
|
char path[PATH_MAX];
|
|
#endif
|
|
|
|
ep = (struct mca_btl_base_endpoint_t*)
|
|
malloc(sizeof(struct mca_btl_base_endpoint_t));
|
|
if(NULL == ep)
|
|
return NULL;
|
|
ep->peer_smp_rank = local_proc + mca_btl_sm_component.num_smp_procs;
|
|
|
|
OBJ_CONSTRUCT(&ep->pending_sends, opal_list_t);
|
|
#if OMPI_ENABLE_PROGRESS_THREADS == 1
|
|
sprintf(path, "%s"OPAL_PATH_SEP"sm_fifo.%lu",
|
|
orte_process_info.job_session_dir,
|
|
(unsigned long)proc->proc_name.vpid);
|
|
ep->fifo_fd = open(path, O_WRONLY);
|
|
if(ep->fifo_fd < 0) {
|
|
opal_output(0, "mca_btl_sm_add_procs: open(%s) failed with errno=%d\n",
|
|
path, errno);
|
|
free(ep);
|
|
return NULL;
|
|
}
|
|
#endif
|
|
return ep;
|
|
}
|
|
|
|
static void calc_sm_max_procs(int n)
|
|
{
|
|
/* see if need to allocate space for extra procs */
|
|
if(0 > mca_btl_sm_component.sm_max_procs) {
|
|
/* no limit */
|
|
if(0 <= mca_btl_sm_component.sm_extra_procs) {
|
|
/* limit */
|
|
mca_btl_sm_component.sm_max_procs =
|
|
n + mca_btl_sm_component.sm_extra_procs;
|
|
} else {
|
|
/* no limit */
|
|
mca_btl_sm_component.sm_max_procs = 2 * n;
|
|
}
|
|
}
|
|
}
|
|
|
|
int mca_btl_sm_add_procs(
|
|
struct mca_btl_base_module_t* btl,
|
|
size_t nprocs,
|
|
struct ompi_proc_t **procs,
|
|
struct mca_btl_base_endpoint_t **peers,
|
|
ompi_bitmap_t* reachability)
|
|
{
|
|
int return_code = OMPI_SUCCESS;
|
|
int32_t n_local_procs = 0, proc, j,
|
|
my_smp_rank = mca_btl_sm_component.my_smp_rank;
|
|
ompi_proc_t* my_proc; /* pointer to caller's proc structure */
|
|
mca_btl_sm_t *sm_btl;
|
|
ompi_fifo_t * volatile *fifo_tmp;
|
|
volatile char **tmp_ptr;
|
|
bool have_connected_peer = false;
|
|
mca_btl_sm_module_resource_t *sm_ctl_hdr;
|
|
/* initializion */
|
|
|
|
sm_btl = (mca_btl_sm_t *)btl;
|
|
|
|
/* get pointer to my proc structure */
|
|
if(NULL == (my_proc = ompi_proc_local()))
|
|
return OMPI_ERR_OUT_OF_RESOURCE;
|
|
|
|
/* Get unique host identifier for each process in the list,
|
|
* and idetify procs that are on this host. Add procs on this
|
|
* host to shared memory reachbility list. Also, get number
|
|
* of local procs in the procs list. */
|
|
for(proc = 0; proc < (int32_t)nprocs; proc++) {
|
|
/* check to see if this proc can be reached via shmem (i.e.,
|
|
if they're on my local host and in my job) */
|
|
if (procs[proc]->proc_name.jobid != my_proc->proc_name.jobid ||
|
|
0 == (procs[proc]->proc_flags & OMPI_PROC_FLAG_LOCAL)) {
|
|
peers[proc] = NULL;
|
|
continue;
|
|
}
|
|
|
|
/* check to see if this is me */
|
|
if(my_proc == procs[proc]) {
|
|
my_smp_rank = mca_btl_sm_component.my_smp_rank = n_local_procs++;
|
|
continue;
|
|
}
|
|
|
|
/* we have someone to talk to */
|
|
have_connected_peer = true;
|
|
|
|
if(!(peers[proc] = create_sm_endpoint(n_local_procs, procs[proc]))) {
|
|
return_code = OMPI_ERROR;
|
|
goto CLEANUP;
|
|
}
|
|
n_local_procs++;
|
|
|
|
/* add this proc to shared memory accessibility list */
|
|
return_code = ompi_bitmap_set_bit(reachability, proc);
|
|
if(OMPI_SUCCESS != return_code)
|
|
goto CLEANUP;
|
|
}
|
|
|
|
/* jump out if there's not someone we can talk to */
|
|
if (!have_connected_peer)
|
|
goto CLEANUP;
|
|
|
|
/* make sure that my_smp_rank has been defined */
|
|
if(-1 == my_smp_rank) {
|
|
return_code = OMPI_ERROR;
|
|
goto CLEANUP;
|
|
}
|
|
|
|
calc_sm_max_procs(n_local_procs);
|
|
|
|
if (!sm_btl->btl_inited) {
|
|
return_code =
|
|
sm_btl_first_time_init(sm_btl, mca_btl_sm_component.sm_max_procs);
|
|
if(return_code != OMPI_SUCCESS)
|
|
goto CLEANUP;
|
|
}
|
|
|
|
/* set local proc's smp rank in the peers structure for
|
|
* rapid access and calulcate reachebility */
|
|
for(proc = 0; proc < (int32_t)nprocs; proc++) {
|
|
if(NULL == peers[proc])
|
|
continue;
|
|
mca_btl_sm_component.sm_peers[peers[proc]->peer_smp_rank] = peers[proc];
|
|
peers[proc]->my_smp_rank = my_smp_rank;
|
|
}
|
|
|
|
sm_ctl_hdr = mca_btl_sm_component.sm_ctl_header;
|
|
fifo_tmp = (ompi_fifo_t * volatile *)
|
|
OFFSET2ADDR(sm_ctl_hdr->fifo, mca_btl_sm_component.sm_mpool_base);
|
|
tmp_ptr = (volatile char **)
|
|
OFFSET2ADDR(sm_ctl_hdr->segment_header.base_shared_mem_segment,
|
|
mca_btl_sm_component.sm_mpool_base);
|
|
|
|
for(j = mca_btl_sm_component.num_smp_procs;
|
|
j < mca_btl_sm_component.num_smp_procs + n_local_procs; j++) {
|
|
ptrdiff_t diff;
|
|
|
|
if(j == my_smp_rank)
|
|
continue;
|
|
|
|
/* spin until this element is allocated */
|
|
while(NULL == fifo_tmp[j]) {
|
|
opal_atomic_rmb();
|
|
opal_progress();
|
|
}
|
|
|
|
/* Calculate the difference as (my_base - their_base) */
|
|
diff = ADDR2OFFSET(tmp_ptr[my_smp_rank], tmp_ptr[j]);
|
|
mca_btl_sm_component.sm_offset[j] = diff;
|
|
|
|
/* store local address of remote fifos */
|
|
mca_btl_sm_component.fifo[j] =
|
|
(ompi_fifo_t*)OFFSET2ADDR(diff, fifo_tmp[j]);
|
|
|
|
/* don't forget to update the head_lock if allocated because this
|
|
* address is also in the remote process */
|
|
if(mca_btl_sm_component.fifo[j][my_smp_rank].head_lock != NULL) {
|
|
mca_btl_sm_component.fifo[j][my_smp_rank].head_lock =
|
|
(opal_atomic_lock_t*)OFFSET2ADDR(diff, mca_btl_sm_component.fifo[j][my_smp_rank].head_lock);
|
|
}
|
|
|
|
/* Initialize fifo for use. Note that sender does initialization */
|
|
return_code = ompi_fifo_init(mca_btl_sm_component.size_of_cb_queue,
|
|
mca_btl_sm_component.cb_lazy_free_freq,
|
|
mca_btl_sm_component.cb_max_num,
|
|
0,0,0,
|
|
&mca_btl_sm_component.fifo[j][my_smp_rank],
|
|
mca_btl_sm_component.sm_offset[j],
|
|
mca_btl_sm_component.sm_mpool);
|
|
|
|
if(return_code != OMPI_SUCCESS)
|
|
goto CLEANUP;
|
|
}
|
|
|
|
/* update the local smp process count */
|
|
mca_btl_sm_component.num_smp_procs += n_local_procs;
|
|
|
|
/* make sure we have enough eager fragmnents for each process */
|
|
return_code = ompi_free_list_resize(&mca_btl_sm_component.sm_frags1,
|
|
mca_btl_sm_component.num_smp_procs * 2);
|
|
if (OMPI_SUCCESS != return_code)
|
|
goto CLEANUP;
|
|
|
|
CLEANUP:
|
|
return return_code;
|
|
}
|
|
|
|
int mca_btl_sm_del_procs(
|
|
struct mca_btl_base_module_t* btl,
|
|
size_t nprocs,
|
|
struct ompi_proc_t **procs,
|
|
struct mca_btl_base_endpoint_t **peers)
|
|
{
|
|
return OMPI_SUCCESS;
|
|
}
|
|
|
|
|
|
/**
|
|
* MCA->BTL Clean up any resources held by BTL module
|
|
* before the module is unloaded.
|
|
*
|
|
* @param btl (IN) BTL module.
|
|
*
|
|
* Prior to unloading a BTL module, the MCA framework will call
|
|
* the BTL finalize method of the module. Any resources held by
|
|
* the BTL should be released and if required the memory corresponding
|
|
* to the BTL module freed.
|
|
*
|
|
*/
|
|
|
|
int mca_btl_sm_finalize(struct mca_btl_base_module_t* btl)
|
|
{
|
|
return OMPI_SUCCESS;
|
|
}
|
|
|
|
|
|
/*
|
|
* Register callback function for error handling..
|
|
*/
|
|
int mca_btl_sm_register_error_cb(
|
|
struct mca_btl_base_module_t* btl,
|
|
mca_btl_base_module_error_cb_fn_t cbfunc)
|
|
{
|
|
mca_btl_sm_t *sm_btl = (mca_btl_sm_t *)btl;
|
|
sm_btl->error_cb = cbfunc;
|
|
return OMPI_SUCCESS;
|
|
}
|
|
|
|
/**
|
|
* Allocate a segment.
|
|
*
|
|
* @param btl (IN) BTL module
|
|
* @param size (IN) Request segment size.
|
|
*/
|
|
extern mca_btl_base_descriptor_t* mca_btl_sm_alloc(
|
|
struct mca_btl_base_module_t* btl,
|
|
struct mca_btl_base_endpoint_t* endpoint,
|
|
uint8_t order,
|
|
size_t size,
|
|
uint32_t flags)
|
|
{
|
|
mca_btl_sm_frag_t* frag;
|
|
int rc;
|
|
if(size <= mca_btl_sm_component.eager_limit) {
|
|
MCA_BTL_SM_FRAG_ALLOC1(frag,rc);
|
|
} else if (size <= mca_btl_sm_component.max_frag_size) {
|
|
MCA_BTL_SM_FRAG_ALLOC2(frag,rc);
|
|
} else {
|
|
return NULL;
|
|
}
|
|
|
|
if (frag != NULL) {
|
|
frag->segment.seg_len = size;
|
|
frag->base.des_flags = flags;
|
|
}
|
|
|
|
return (mca_btl_base_descriptor_t*)frag;
|
|
}
|
|
|
|
/**
|
|
* Return a segment allocated by this BTL.
|
|
*
|
|
* @param btl (IN) BTL module
|
|
* @param segment (IN) Allocated segment.
|
|
*/
|
|
extern int mca_btl_sm_free(
|
|
struct mca_btl_base_module_t* btl,
|
|
mca_btl_base_descriptor_t* des)
|
|
{
|
|
mca_btl_sm_frag_t* frag = (mca_btl_sm_frag_t*)des;
|
|
MCA_BTL_SM_FRAG_RETURN(frag);
|
|
|
|
return OMPI_SUCCESS;
|
|
}
|
|
|
|
|
|
/**
|
|
* Pack data
|
|
*
|
|
* @param btl (IN) BTL module
|
|
*/
|
|
struct mca_btl_base_descriptor_t* mca_btl_sm_prepare_src(
|
|
struct mca_btl_base_module_t* btl,
|
|
struct mca_btl_base_endpoint_t* endpoint,
|
|
mca_mpool_base_registration_t* registration,
|
|
struct ompi_convertor_t* convertor,
|
|
uint8_t order,
|
|
size_t reserve,
|
|
size_t* size,
|
|
uint32_t flags)
|
|
{
|
|
mca_btl_sm_frag_t* frag;
|
|
struct iovec iov;
|
|
uint32_t iov_count = 1;
|
|
size_t max_data = *size;
|
|
int rc;
|
|
|
|
MCA_BTL_SM_FRAG_ALLOC2(frag, rc);
|
|
if(NULL == frag) {
|
|
return NULL;
|
|
}
|
|
|
|
if(reserve + max_data > frag->size) {
|
|
max_data = frag->size - reserve;
|
|
}
|
|
iov.iov_len = max_data;
|
|
iov.iov_base =
|
|
(IOVBASE_TYPE*)(((unsigned char*)(frag->segment.seg_addr.pval)) +
|
|
reserve);
|
|
|
|
rc = ompi_convertor_pack(convertor, &iov, &iov_count, &max_data );
|
|
if(rc < 0) {
|
|
MCA_BTL_SM_FRAG_RETURN(frag);
|
|
return NULL;
|
|
}
|
|
frag->segment.seg_len = reserve + max_data;
|
|
frag->base.des_flags = flags;
|
|
*size = max_data;
|
|
return &frag->base;
|
|
}
|
|
|
|
|
|
/**
|
|
* Initiate a send to the peer.
|
|
*
|
|
* @param btl (IN) BTL module
|
|
* @param peer (IN) BTL peer addressing
|
|
*/
|
|
|
|
int mca_btl_sm_send(
|
|
struct mca_btl_base_module_t* btl,
|
|
struct mca_btl_base_endpoint_t* endpoint,
|
|
struct mca_btl_base_descriptor_t* descriptor,
|
|
mca_btl_base_tag_t tag)
|
|
{
|
|
mca_btl_sm_frag_t* frag = (mca_btl_sm_frag_t*)descriptor;
|
|
int rc;
|
|
|
|
/* available header space */
|
|
frag->hdr->len = frag->segment.seg_len;
|
|
/* type of message, pt-2-pt, one-sided, etc */
|
|
frag->hdr->tag = tag;
|
|
|
|
frag->endpoint = endpoint;
|
|
|
|
/*
|
|
* post the descriptor in the queue - post with the relative
|
|
* address
|
|
*/
|
|
MCA_BTL_SM_FIFO_WRITE(endpoint, endpoint->my_smp_rank,
|
|
endpoint->peer_smp_rank, frag->hdr, false, rc);
|
|
return OMPI_SUCCESS;
|
|
}
|
|
|
|
int mca_btl_sm_ft_event(int state) {
|
|
if(OPAL_CRS_CHECKPOINT == state) {
|
|
;
|
|
}
|
|
else if(OPAL_CRS_CONTINUE == state) {
|
|
;
|
|
}
|
|
else if(OPAL_CRS_RESTART == state) {
|
|
;
|
|
}
|
|
else if(OPAL_CRS_TERM == state ) {
|
|
;
|
|
}
|
|
else {
|
|
;
|
|
}
|
|
|
|
return OMPI_SUCCESS;
|
|
}
|