/* * Copyright (c) 2004-2006 The Trustees of Indiana University and Indiana * University Research and Technology * Corporation. All rights reserved. * Copyright (c) 2004-2005 The University of Tennessee and The University * of Tennessee Research Foundation. All rights * reserved. * Copyright (c) 2004-2005 High Performance Computing Center Stuttgart, * University of Stuttgart. All rights reserved. * Copyright (c) 2004-2005 The Regents of the University of California. * All rights reserved. * Copyright (c) 2006 Sandia National Laboratories. All rights * reserved. * Copyright (c) 2006 Sun Microsystems, Inc. All rights reserved. * Copyright (c) 2007 Cisco Systems, Inc. All rights reserved. * * $COPYRIGHT$ * * Additional copyrights may follow * * $HEADER$ */ #include "ompi_config.h" #include "ompi/constants.h" #include "opal/event/event.h" #include "opal/util/if.h" #include "opal/util/argv.h" #include "opal/util/output.h" #include "ompi/mca/pml/pml.h" #include "ompi/mca/btl/btl.h" #include "ompi/runtime/params.h" #include "opal/mca/base/mca_base_param.h" #include "orte/mca/errmgr/errmgr.h" #include "ompi/mca/mpool/base/base.h" #include "ompi/mca/mpool/rdma/mpool_rdma.h" #include "btl_udapl.h" #include "btl_udapl_frag.h" #include "btl_udapl_endpoint.h" #include "btl_udapl_proc.h" #include "ompi/mca/btl/base/base.h" #include "ompi/mca/btl/base/btl_base_error.h" #include "ompi/datatype/convertor.h" #include "btl_udapl_endpoint.h" #include "orte/util/proc_info.h" #include "ompi/mca/pml/base/pml_base_module_exchange.h" /* * Local Functions */ static inline int mca_btl_udapl_frag_progress_one(mca_btl_udapl_module_t* udapl_btl, mca_btl_udapl_frag_t* frag); void mca_btl_udapl_frag_progress_pending(mca_btl_udapl_module_t* udapl_btl, mca_btl_base_endpoint_t* endpoint, const int connection); mca_btl_udapl_component_t mca_btl_udapl_component = { { /* First, the mca_base_component_t struct containing meta information about the component itself */ { /* Indicate that we are a pml v1.0.0 component (which also implies a specific MCA version) */ MCA_BTL_BASE_VERSION_1_0_0, "udapl", /* MCA component name */ OMPI_MAJOR_VERSION, /* MCA component major version */ OMPI_MINOR_VERSION, /* MCA component minor version */ OMPI_RELEASE_VERSION, /* MCA component release version */ mca_btl_udapl_component_open, /* component open */ mca_btl_udapl_component_close /* component close */ }, /* Next the MCA v1.0.0 component meta data */ { /* Whether the component is checkpointable or not */ false }, mca_btl_udapl_component_init, mca_btl_udapl_component_progress, } }; /** * Report a uDAPL error - for debugging */ #if OMPI_ENABLE_DEBUG void mca_btl_udapl_error(DAT_RETURN ret, char* str) { char* major; char* minor; if(DAT_SUCCESS != dat_strerror(ret, (const char**)&major, (const char**)&minor)) { printf("dat_strerror failed! ret is %d\n", ret); exit(-1); } OPAL_OUTPUT((0, "ERROR: %s %s %s\n", str, major, minor)); } #endif /* * Utility routines for parameter registration */ static inline char* mca_btl_udapl_param_register_string( const char* param_name, const char* default_value) { char *param_value; int id = mca_base_param_register_string("btl","udapl",param_name,NULL,default_value); mca_base_param_lookup_string(id, ¶m_value); return param_value; } static inline int mca_btl_udapl_param_register_int( const char* param_name, int default_value) { int id = mca_base_param_register_int("btl","udapl",param_name,NULL,default_value); int param_value = default_value; mca_base_param_lookup_int(id,¶m_value); return param_value; } /* * Called by MCA framework to open the component, registers * component parameters. */ int mca_btl_udapl_component_open(void) { /* initialize state */ mca_btl_udapl_component.udapl_num_btls=0; mca_btl_udapl_component.udapl_btls=NULL; /* initialize objects */ OBJ_CONSTRUCT(&mca_btl_udapl_component.udapl_procs, opal_list_t); OBJ_CONSTRUCT(&mca_btl_udapl_component.udapl_lock, opal_mutex_t); /* register uDAPL component parameters */ mca_btl_udapl_component.udapl_free_list_num = mca_btl_udapl_param_register_int("free_list_num", 8); mca_btl_udapl_component.udapl_free_list_max = mca_btl_udapl_param_register_int("free_list_max", -1); mca_btl_udapl_component.udapl_free_list_inc = mca_btl_udapl_param_register_int("free_list_inc", 8); mca_btl_udapl_component.udapl_mpool_name = mca_btl_udapl_param_register_string("mpool", "rdma"); mca_btl_udapl_component.udapl_max_btls = mca_btl_udapl_param_register_int("max_modules", 8); mca_btl_udapl_component.udapl_evd_qlen = mca_btl_udapl_param_register_int("evd_qlen", 32); mca_btl_udapl_component.udapl_max_request_dtos = mca_btl_udapl_param_register_int("max_request_dtos", 18); mca_btl_udapl_component.udapl_max_recv_dtos = mca_btl_udapl_param_register_int("max_recv_dtos", 18); mca_btl_udapl_component.udapl_num_recvs = mca_btl_udapl_param_register_int("num_recvs", 8); mca_btl_udapl_component.udapl_num_sends = mca_btl_udapl_param_register_int("num_sends", 7); mca_btl_udapl_component.udapl_sr_win = mca_btl_udapl_param_register_int("sr_win", 4); mca_btl_udapl_component.udapl_eager_rdma_num = mca_btl_udapl_param_register_int("eager_rdma_num", 8); mca_btl_udapl_component.udapl_max_eager_rdma_peers = mca_btl_udapl_param_register_int("max_eager_rdma_peers", 16); mca_btl_udapl_component.udapl_eager_rdma_win = mca_btl_udapl_param_register_int("eager_rdma_win", 4); mca_btl_udapl_component.udapl_timeout = mca_btl_udapl_param_register_int("timeout", 10000000); mca_btl_udapl_component.udapl_eager_rdma_guarantee = mca_btl_udapl_param_register_int("eager_rdma_guarantee", 0); /* register uDAPL module parameters */ mca_btl_udapl_module.super.btl_exclusivity = mca_btl_udapl_param_register_int ("exclusivity", MCA_BTL_EXCLUSIVITY_DEFAULT - 10); mca_btl_udapl_module.super.btl_eager_limit = mca_btl_udapl_param_register_int ("eager_limit", 32*1024); mca_btl_udapl_module.super.btl_min_send_size = mca_btl_udapl_param_register_int ("min_send_size", 16*1024); mca_btl_udapl_module.super.btl_max_send_size = mca_btl_udapl_param_register_int ("max_send_size", 64*1024); mca_btl_udapl_module.super.btl_min_rdma_size = mca_btl_udapl_param_register_int("min_rdma_size", 512*1024); mca_btl_udapl_module.super.btl_max_rdma_size = mca_btl_udapl_param_register_int("max_rdma_size", 128*1024); mca_btl_udapl_module.super.btl_flags = mca_btl_udapl_param_register_int("flags", MCA_BTL_FLAGS_PUT); mca_btl_udapl_module.super.btl_bandwidth = mca_btl_udapl_param_register_int("bandwidth", 225); /* compute udapl_eager_frag_size and udapl_max_frag_size */ mca_btl_udapl_component.udapl_eager_frag_size = mca_btl_udapl_module.super.btl_eager_limit; mca_btl_udapl_module.super.btl_eager_limit -= (sizeof(mca_btl_udapl_footer_t) + sizeof(mca_btl_udapl_rdma_footer_t)); mca_btl_udapl_component.udapl_max_frag_size = mca_btl_udapl_module.super.btl_max_send_size; mca_btl_udapl_module.super.btl_max_send_size -= (sizeof(mca_btl_udapl_footer_t) + sizeof(mca_btl_udapl_rdma_footer_t)); /* compute udapl_eager_rdma_frag_size */ mca_btl_udapl_component.udapl_eager_rdma_frag_size = sizeof(mca_btl_udapl_frag_eager_rdma_t) + mca_btl_udapl_component.udapl_eager_frag_size; return OMPI_SUCCESS; } /* * component cleanup - sanity checking of queue lengths */ int mca_btl_udapl_component_close(void) { /* TODO - what needs to be done here? */ return OMPI_SUCCESS; } /* * Register uDAPL component addressing information. The MCA framework * will make this available to all peers. */ static int mca_btl_udapl_modex_send(void) { int rc; size_t i; size_t size; mca_btl_udapl_addr_t *addrs = NULL; size = sizeof(mca_btl_udapl_addr_t) * mca_btl_udapl_component.udapl_num_btls; if (0 != size) { addrs = (mca_btl_udapl_addr_t*)malloc(size); if (NULL == addrs) { return OMPI_ERR_OUT_OF_RESOURCE; } for (i = 0; i < mca_btl_udapl_component.udapl_num_btls; i++) { mca_btl_udapl_module_t* btl = mca_btl_udapl_component.udapl_btls[i]; addrs[i] = btl->udapl_addr; } } rc = mca_pml_base_modex_send( &mca_btl_udapl_component.super.btl_version, addrs, size); if (NULL != addrs) { free (addrs); } return rc; } /* * Callback function used for udapl btl internal control messages. * * @param btl (IN) BTL module * @param tag (IN) Not used but part of callback interface * @param descriptor (IN) Description of the data that was just transferred * @param cbdata (IN) Data used by call back function. Not used. * */ static void mca_btl_udapl_receive_control(struct mca_btl_base_module_t* btl, mca_btl_base_tag_t tag, mca_btl_base_descriptor_t* descriptor, void* cbdata) { mca_btl_udapl_frag_t* frag = (mca_btl_udapl_frag_t*)descriptor; mca_btl_udapl_endpoint_t* endpoint = frag->endpoint; mca_btl_udapl_control_header_t* ctl_hdr = frag->segment.seg_addr.pval; switch (ctl_hdr->type) { case MCA_BTL_UDAPL_CONTROL_RDMA_CONNECT: { mca_btl_udapl_eager_rdma_connect_t* rdma_connect = frag->segment.seg_addr.pval; if (endpoint->endpoint_eager_rdma_remote.base.pval) { BTL_ERROR(("ERROR: Received RDMA connect twice!")); return; } endpoint->endpoint_eager_rdma_remote.rkey = rdma_connect->rkey; endpoint->endpoint_eager_rdma_remote.base.pval = rdma_connect->rdma_start.pval; OPAL_THREAD_ADD32(&(endpoint->endpoint_eager_rdma_remote.tokens), mca_btl_udapl_component.udapl_eager_rdma_num); break; } case MCA_BTL_UDAPL_CONTROL_RDMA_CREDIT: { mca_btl_udapl_eager_rdma_credit_t* rdma_credit = frag->segment.seg_addr.pval; /* don't return credits used for rdma credit control message */ OPAL_THREAD_ADD32( &(endpoint->endpoint_sr_credits[BTL_UDAPL_EAGER_CONNECTION]), -1); OPAL_THREAD_ADD32(&(endpoint->endpoint_eager_rdma_remote.tokens), rdma_credit->credits); break; } case MCA_BTL_UDAPL_CONTROL_SR_CREDIT: { mca_btl_udapl_sr_credit_t* sr_credit = frag->segment.seg_addr.pval; /* don't return credits used for sr credit control message */ OPAL_THREAD_ADD32( &(endpoint->endpoint_sr_credits[sr_credit->connection]), -1); OPAL_THREAD_ADD32( &(endpoint->endpoint_sr_tokens[sr_credit->connection]), sr_credit->credits); break; } default: BTL_ERROR(("ERROR: Unknown contrl message type received by BTL")); break; } } /* * Initialize the uDAPL component, * check how many interfaces are available and create a btl module for each. */ mca_btl_base_module_t ** mca_btl_udapl_component_init (int *num_btl_modules, bool enable_progress_threads, bool enable_mpi_threads) { DAT_PROVIDER_INFO* datinfo; DAT_PROVIDER_INFO** datinfoptr; mca_btl_base_module_t **btls; mca_btl_udapl_module_t *btl; DAT_COUNT num_ias; int32_t i; /* enumerate uDAPL interfaces */ /* Have to do weird pointer stuff to make uDAPL happy - just an array of DAT_PROVIDER_INFO isn't good enough. */ datinfo = malloc(sizeof(DAT_PROVIDER_INFO) * mca_btl_udapl_component.udapl_max_btls); datinfoptr = malloc(sizeof(DAT_PROVIDER_INFO*) * mca_btl_udapl_component.udapl_max_btls); if(NULL == datinfo || NULL == datinfoptr) { return NULL; } for(i = 0; i < (int32_t)mca_btl_udapl_component.udapl_max_btls; i++) { datinfoptr[i] = &datinfo[i]; } if(DAT_SUCCESS != dat_registry_list_providers( mca_btl_udapl_component.udapl_max_btls, (DAT_COUNT*)&num_ias, datinfoptr)) { free(datinfo); free(datinfoptr); return NULL; } free(datinfoptr); /* allocate space for the each possible BTL */ mca_btl_udapl_component.udapl_btls = (mca_btl_udapl_module_t **) malloc(num_ias * sizeof(mca_btl_udapl_module_t *)); if(NULL == mca_btl_udapl_component.udapl_btls) { free(datinfo); return NULL; } /* create a BTL module for each interface */ for(mca_btl_udapl_component.udapl_num_btls = i = 0; i < num_ias; i++) { btl = malloc(sizeof(mca_btl_udapl_module_t)); if(NULL == btl) { free(datinfo); free(mca_btl_udapl_component.udapl_btls); return NULL; } /* copy default values into the new BTL */ memcpy(btl, &mca_btl_udapl_module, sizeof(mca_btl_udapl_module_t)); /* initialize this BTL */ /* TODO - make use of the thread-safety info in datinfo also */ if(OMPI_SUCCESS != mca_btl_udapl_init(datinfo[i].ia_name, btl)) { free(btl); continue; } /* register internal control message callback */ btl->udapl_reg[MCA_BTL_TAG_BTL].cbfunc = mca_btl_udapl_receive_control; btl->udapl_reg[MCA_BTL_TAG_BTL].cbdata = NULL; /* successful btl creation */ mca_btl_udapl_component.udapl_btls[mca_btl_udapl_component.udapl_num_btls] = btl; if(++mca_btl_udapl_component.udapl_num_btls >= mca_btl_udapl_component.udapl_max_btls) { break; } } /* finished with datinfo */ free(datinfo); /* Make sure we have some interfaces */ if(0 == mca_btl_udapl_component.udapl_num_btls) { mca_btl_base_error_no_nics("uDAPL", "NIC"); free(mca_btl_udapl_component.udapl_btls); return NULL; } /* publish uDAPL parameters with the MCA framework */ if (OMPI_SUCCESS != mca_btl_udapl_modex_send()) { free(mca_btl_udapl_component.udapl_btls); return NULL; } /* Post OOB receive */ mca_btl_udapl_endpoint_post_oob_recv(); /* return array of BTLs */ btls = (mca_btl_base_module_t**) malloc(sizeof(mca_btl_base_module_t *) * mca_btl_udapl_component.udapl_num_btls); if (NULL == btls) { free(mca_btl_udapl_component.udapl_btls); return NULL; } memcpy(btls, mca_btl_udapl_component.udapl_btls, mca_btl_udapl_component.udapl_num_btls * sizeof(mca_btl_udapl_module_t *)); *num_btl_modules = mca_btl_udapl_component.udapl_num_btls; return btls; } static int mca_btl_udapl_accept_connect(mca_btl_udapl_module_t* btl, DAT_CR_HANDLE cr_handle) { DAT_EP_HANDLE endpoint; int rc; rc = mca_btl_udapl_endpoint_create(btl, &endpoint); if(OMPI_SUCCESS != rc) { BTL_ERROR(("ERROR: mca_btl_udapl_endpoint_create")); return OMPI_ERROR; } rc = dat_cr_accept(cr_handle, endpoint, 0, NULL); if(DAT_SUCCESS != rc) { char* major; char* minor; dat_strerror(rc, (const char**)&major, (const char**)&minor); BTL_ERROR(("ERROR: %s %s %s\n", "dat_cr_accept", major, minor)); return OMPI_ERROR; } return OMPI_SUCCESS; } static inline int mca_btl_udapl_sendrecv(mca_btl_udapl_module_t* btl, DAT_EP_HANDLE* endpoint) { mca_btl_udapl_frag_t* frag; DAT_DTO_COOKIE cookie; static int32_t connection_seq = 1; int rc; /* Post a receive to get the peer's address data */ frag = (mca_btl_udapl_frag_t*)mca_btl_udapl_alloc( (mca_btl_base_module_t*)btl, sizeof(mca_btl_udapl_addr_t) + sizeof(int32_t)); cookie.as_ptr = frag; frag->type = MCA_BTL_UDAPL_CONN_RECV; rc = dat_ep_post_recv(endpoint, 1, &frag->triplet, cookie, DAT_COMPLETION_DEFAULT_FLAG); if(DAT_SUCCESS != rc) { char* major; char* minor; dat_strerror(rc, (const char**)&major, (const char**)&minor); BTL_ERROR(("ERROR: %s %s %s\n", "dat_ep_post_recv", major, minor)); return OMPI_ERROR; } /* Send our local address data over this EP */ frag = (mca_btl_udapl_frag_t*)mca_btl_udapl_alloc( (mca_btl_base_module_t*)btl, sizeof(mca_btl_udapl_addr_t) + sizeof(int32_t)); cookie.as_ptr = frag; memcpy(frag->segment.seg_addr.pval, &btl->udapl_addr, sizeof(mca_btl_udapl_addr_t)); memcpy((char *)frag->segment.seg_addr.pval + sizeof(mca_btl_udapl_addr_t), &connection_seq, sizeof(int32_t)); connection_seq++; frag->type = MCA_BTL_UDAPL_CONN_SEND; rc = dat_ep_post_send(endpoint, 1, &frag->triplet, cookie, DAT_COMPLETION_DEFAULT_FLAG); if(DAT_SUCCESS != rc) { char* major; char* minor; dat_strerror(rc, (const char**)&major, (const char**)&minor); BTL_ERROR(("ERROR: %s %s %s\n", "dat_ep_post_send", major, minor)); return OMPI_ERROR; } return OMPI_SUCCESS; } static inline int mca_btl_udapl_frag_progress_one( mca_btl_udapl_module_t* udapl_btl, mca_btl_udapl_frag_t* frag) { int rc; switch(frag->type) { case MCA_BTL_UDAPL_SEND: rc = mca_btl_udapl_endpoint_send(frag->endpoint, frag); break; case MCA_BTL_UDAPL_PUT: rc = mca_btl_udapl_put((mca_btl_base_module_t*)udapl_btl, frag->endpoint, (mca_btl_base_descriptor_t*)frag); break; default: rc = OMPI_ERROR; BTL_ERROR(("Error : Progressing pending operation, invalid type %d\n", frag->type)); break; } return rc; } void mca_btl_udapl_frag_progress_pending(mca_btl_udapl_module_t* udapl_btl, mca_btl_base_endpoint_t* endpoint, const int connection) { int len; int i; mca_btl_udapl_frag_t* frag; if (BTL_UDAPL_EAGER_CONNECTION == connection) { len = opal_list_get_size(&endpoint->endpoint_eager_frags); /* progress eager frag queue as needed */ for(i = 0; i < len && BTL_UDAPL_TOKENS(endpoint, connection) > 0; i++) { OPAL_THREAD_LOCK(&endpoint->endpoint_lock); frag = (mca_btl_udapl_frag_t*)opal_list_remove_first(&(endpoint->endpoint_eager_frags)); OPAL_THREAD_UNLOCK(&endpoint->endpoint_lock); if(NULL == frag) { return; } if(mca_btl_udapl_frag_progress_one(udapl_btl, frag) != OMPI_SUCCESS) { BTL_ERROR(("ERROR: Not able to progress on connection(%d)\n", BTL_UDAPL_EAGER_CONNECTION)); return; } } } else if (BTL_UDAPL_MAX_CONNECTION == connection) { len = opal_list_get_size(&endpoint->endpoint_max_frags); /* progress max frag queue as needed */ for(i = 0; i < len && BTL_UDAPL_TOKENS(endpoint, connection) > 0; i++) { OPAL_THREAD_LOCK(&endpoint->endpoint_lock); frag = (mca_btl_udapl_frag_t*)opal_list_remove_first(&(endpoint->endpoint_max_frags)); OPAL_THREAD_UNLOCK(&endpoint->endpoint_lock); if(NULL == frag) { return; } if(mca_btl_udapl_frag_progress_one(udapl_btl, frag) != OMPI_SUCCESS) { BTL_ERROR(("ERROR: Not able to progress on connection(%d)\n", BTL_UDAPL_MAX_CONNECTION)); return; } } } else { BTL_ERROR(("ERROR: Can not progress pending fragment on unknown connection\n")); } return; } /* * uDAPL component progress. */ int mca_btl_udapl_component_progress() { mca_btl_udapl_module_t* btl; static int32_t inprogress = 0; DAT_EVENT event; #if defined(__SVR4) && defined(__sun) DAT_COUNT nmore; /* used by dat_evd_wait, see comment below */ #endif size_t i; int32_t j, rdma_ep_count; int count = 0; /* prevent deadlock - only one thread should be 'progressing' at a time */ if(OPAL_THREAD_ADD32(&inprogress, 1) > 1) { OPAL_THREAD_ADD32(&inprogress, -1); return OMPI_SUCCESS; } /* check for work to do on each uDAPL btl */ OPAL_THREAD_LOCK(&mca_btl_udapl_component.udapl_lock); for(i = 0; i < mca_btl_udapl_component.udapl_num_btls; i++) { btl = mca_btl_udapl_component.udapl_btls[i]; /* Check DTO EVD */ while(DAT_SUCCESS == dat_evd_dequeue(btl->udapl_evd_dto, &event)) { DAT_DTO_COMPLETION_EVENT_DATA* dto; mca_btl_udapl_frag_t* frag; switch(event.event_number) { case DAT_DTO_COMPLETION_EVENT: dto = &event.event_data.dto_completion_event_data; frag = dto->user_cookie.as_ptr; /* if we are using the "guarantee" rdma code path * the extra write sets cookie to NULL, when this * happens we ignore it because the completion * write event is coming */ if (frag == NULL) break; /* Was the DTO successful? */ if(DAT_DTO_SUCCESS != dto->status) { OPAL_OUTPUT((0, "btl_udapl ***** DTO error %d %d %d %p*****\n", dto->status, frag->type, frag->size, dto->ep_handle)); break; } switch(frag->type) { case MCA_BTL_UDAPL_RDMA_WRITE: { assert(frag->base.des_src == &frag->segment); assert(frag->base.des_src_cnt == 1); assert(frag->base.des_dst == NULL); assert(frag->base.des_dst_cnt == 0); assert(frag->type == MCA_BTL_UDAPL_RDMA_WRITE); frag->base.des_cbfunc(&btl->super, frag->endpoint, &frag->base, OMPI_SUCCESS); mca_btl_udapl_frag_progress_pending(btl, frag->endpoint, BTL_UDAPL_EAGER_CONNECTION); break; } case MCA_BTL_UDAPL_SEND: { assert(frag->base.des_src == &frag->segment); assert(frag->base.des_src_cnt == 1); assert(frag->base.des_dst == NULL); assert(frag->base.des_dst_cnt == 0); assert(frag->type == MCA_BTL_UDAPL_SEND); frag->base.des_cbfunc(&btl->super, frag->endpoint, &frag->base, OMPI_SUCCESS); if(frag->size == mca_btl_udapl_component.udapl_eager_frag_size) { mca_btl_udapl_frag_progress_pending(btl, frag->endpoint, BTL_UDAPL_EAGER_CONNECTION); } else { assert(frag->size == mca_btl_udapl_component.udapl_max_frag_size); mca_btl_udapl_frag_progress_pending(btl, frag->endpoint, BTL_UDAPL_MAX_CONNECTION); } break; } case MCA_BTL_UDAPL_RECV: { mca_btl_base_recv_reg_t* reg; int cntrl_msg = -1; assert(frag->base.des_dst == &frag->segment); assert(frag->base.des_dst_cnt == 1); assert(frag->base.des_src == NULL); assert(frag->base.des_src_cnt == 0); assert(frag->type == MCA_BTL_UDAPL_RECV); assert(frag->triplet.virtual_address == (DAT_VADDR)frag->segment.seg_addr.pval); assert(frag->triplet.segment_length == frag->size); assert(frag->btl == btl); /* setup frag ftr location and do callback */ frag->segment.seg_len = dto->transfered_length - sizeof(mca_btl_udapl_footer_t); frag->ftr = (mca_btl_udapl_footer_t *) ((char *)frag->segment.seg_addr.pval + frag->segment.seg_len); cntrl_msg = frag->ftr->tag; reg = &btl->udapl_reg[frag->ftr->tag]; OPAL_THREAD_UNLOCK(&mca_btl_udapl_component.udapl_lock); reg->cbfunc(&btl->super, frag->ftr->tag, &frag->base, reg->cbdata); OPAL_THREAD_LOCK(&mca_btl_udapl_component.udapl_lock); /* Repost the frag */ frag->ftr = frag->segment.seg_addr.pval; frag->segment.seg_len = (frag->size - sizeof(mca_btl_udapl_footer_t) - sizeof(mca_btl_udapl_rdma_footer_t)); frag->base.des_flags = 0; if(frag->size == mca_btl_udapl_component.udapl_eager_frag_size) { OPAL_THREAD_ADD32(&(frag->endpoint->endpoint_sr_credits[BTL_UDAPL_EAGER_CONNECTION]), 1); dat_ep_post_recv(frag->endpoint->endpoint_eager, 1, &frag->triplet, dto->user_cookie, DAT_COMPLETION_DEFAULT_FLAG); if (frag->endpoint->endpoint_sr_credits[BTL_UDAPL_EAGER_CONNECTION] >= mca_btl_udapl_component.udapl_sr_win) { mca_btl_udapl_endpoint_send_sr_credits(frag->endpoint, BTL_UDAPL_EAGER_CONNECTION); } if (MCA_BTL_TAG_BTL == cntrl_msg) { mca_btl_udapl_frag_progress_pending(btl, frag->endpoint, BTL_UDAPL_EAGER_CONNECTION); } } else { assert(frag->size == mca_btl_udapl_component.udapl_max_frag_size); OPAL_THREAD_ADD32(&(frag->endpoint->endpoint_sr_credits[BTL_UDAPL_MAX_CONNECTION]), 1); dat_ep_post_recv(frag->endpoint->endpoint_max, 1, &frag->triplet, dto->user_cookie, DAT_COMPLETION_DEFAULT_FLAG); if (frag->endpoint->endpoint_sr_credits[BTL_UDAPL_MAX_CONNECTION] >= mca_btl_udapl_component.udapl_sr_win) { mca_btl_udapl_endpoint_send_sr_credits(frag->endpoint, BTL_UDAPL_MAX_CONNECTION); } if (MCA_BTL_TAG_BTL == cntrl_msg) { mca_btl_udapl_frag_progress_pending(btl, frag->endpoint, BTL_UDAPL_MAX_CONNECTION); } } break; } case MCA_BTL_UDAPL_PUT: { assert(frag->base.des_src == &frag->segment); assert(frag->base.des_src_cnt == 1); assert(frag->base.des_dst_cnt == 1); assert(frag->type == MCA_BTL_UDAPL_PUT); frag->base.des_cbfunc(&btl->super, frag->endpoint, &frag->base, OMPI_SUCCESS); OPAL_THREAD_ADD32(&(frag->endpoint->endpoint_sr_tokens[BTL_UDAPL_MAX_CONNECTION]), 1); mca_btl_udapl_frag_progress_pending(btl, frag->endpoint, BTL_UDAPL_MAX_CONNECTION); break; } case MCA_BTL_UDAPL_CONN_RECV: mca_btl_udapl_endpoint_finish_connect(btl, frag->segment.seg_addr.pval, (int32_t *)((char *)frag->segment.seg_addr.pval + sizeof(mca_btl_udapl_addr_t)), event.event_data.connect_event_data.ep_handle); /* No break - fall through to free */ case MCA_BTL_UDAPL_CONN_SEND: frag->segment.seg_len = mca_btl_udapl_module.super.btl_eager_limit; mca_btl_udapl_free((mca_btl_base_module_t*)btl, (mca_btl_base_descriptor_t*)frag); break; default: OPAL_OUTPUT((0, "WARNING unknown frag type: %d\n", frag->type)); } count++; break; default: OPAL_OUTPUT((0, "WARNING unknown dto event: %d\n", event.event_number)); } } /* Check connection EVD */ while(DAT_SUCCESS == #if defined(__SVR4) && defined(__sun) /* There is a bug is Solaris udapl implementation * such that dat_evd_dequeue does not dequeue * DAT_CONNECTION_REQUEST_EVENT. Workaround is to use * wait. This should be removed when fix available. */ dat_evd_wait(btl->udapl_evd_conn, 0, 1, &event, &nmore)) { #else dat_evd_dequeue(btl->udapl_evd_conn, &event)) { #endif switch(event.event_number) { case DAT_CONNECTION_REQUEST_EVENT: /* Accept a new connection */ mca_btl_udapl_accept_connect(btl, event.event_data.cr_arrival_event_data.cr_handle); count++; break; case DAT_CONNECTION_EVENT_ESTABLISHED: /* Both the client and server side of a connection generate this event */ mca_btl_udapl_sendrecv(btl, event.event_data.connect_event_data.ep_handle); count++; break; case DAT_CONNECTION_EVENT_PEER_REJECTED: case DAT_CONNECTION_EVENT_NON_PEER_REJECTED: case DAT_CONNECTION_EVENT_ACCEPT_COMPLETION_ERROR: case DAT_CONNECTION_EVENT_DISCONNECTED: case DAT_CONNECTION_EVENT_BROKEN: case DAT_CONNECTION_EVENT_TIMED_OUT: /* handle this case specially? if we have finite timeout, we might want to try connecting again here. */ case DAT_CONNECTION_EVENT_UNREACHABLE: /* Need to set the BTL endpoint to MCA_BTL_UDAPL_FAILED See dat_ep_connect documentation pdf pg 198 */ BTL_OUTPUT(("WARNING : Connection event not handled : %d\n", event.event_number)); break; default: BTL_ERROR(("ERROR: unknown connection event : %d", event.event_number)); } } /* Check async EVD */ while(DAT_SUCCESS == dat_evd_dequeue(btl->udapl_evd_async, &event)) { switch(event.event_number) { case DAT_ASYNC_ERROR_EVD_OVERFLOW: case DAT_ASYNC_ERROR_IA_CATASTROPHIC: case DAT_ASYNC_ERROR_EP_BROKEN: case DAT_ASYNC_ERROR_TIMED_OUT: case DAT_ASYNC_ERROR_PROVIDER_INTERNAL_ERROR: BTL_OUTPUT(("WARNING: async event ignored : %d", event.event_number)); break; default: BTL_OUTPUT(("WARNING unknown async event: %d\n", event.event_number)); } } /* * Check eager rdma segments */ /* find the number of endpoints with rdma buffers */ rdma_ep_count = btl->udapl_eager_rdma_endpoint_count; for (j = 0; j < rdma_ep_count; j++) { mca_btl_udapl_endpoint_t* endpoint; mca_btl_udapl_frag_t *local_rdma_frag; DAT_LMR_TRIPLET local_rdma_segment; endpoint = orte_pointer_array_get_item(btl->udapl_eager_rdma_endpoints, j); OPAL_THREAD_LOCK(&endpoint->endpoint_eager_rdma_local.lock); local_rdma_frag = MCA_BTL_UDAPL_GET_LOCAL_RDMA_FRAG(endpoint, endpoint->endpoint_eager_rdma_local.head); /* sync local memory before checking if active * Question, will narrowing sync area to just the active byte * one, work and two, improve performance */ local_rdma_segment.lmr_context = local_rdma_frag->triplet.lmr_context; local_rdma_segment.virtual_address = (DAT_VADDR)local_rdma_frag->segment.seg_addr.pval; local_rdma_segment.segment_length = local_rdma_frag->size; dat_lmr_sync_rdma_write(endpoint->endpoint_btl->udapl_ia, &local_rdma_segment, 1); if (local_rdma_frag->rdma_ftr->active == 1) { int pad = 0; mca_btl_base_recv_reg_t* reg; MCA_BTL_UDAPL_RDMA_NEXT_INDEX(endpoint->endpoint_eager_rdma_local.head); OPAL_THREAD_UNLOCK(&endpoint->endpoint_eager_rdma_local.lock); /* compute pad as needed */ MCA_BTL_UDAPL_FRAG_CALC_ALIGNMENT_PAD(pad, (local_rdma_frag->rdma_ftr->size + sizeof(mca_btl_udapl_footer_t))); /* set fragment information */ local_rdma_frag->ftr = (mca_btl_udapl_footer_t *) ((char *)local_rdma_frag->rdma_ftr - pad - sizeof(mca_btl_udapl_footer_t)); local_rdma_frag->segment.seg_len = local_rdma_frag->rdma_ftr->size; local_rdma_frag->segment.seg_addr.pval = (unsigned char *) ((char *)local_rdma_frag->ftr - local_rdma_frag->segment.seg_len); /* retrieve callback and callback */ reg = &btl->udapl_reg[local_rdma_frag->ftr->tag]; reg->cbfunc(&btl->super, local_rdma_frag->ftr->tag, &local_rdma_frag->base, reg->cbdata); /* repost */ local_rdma_frag->rdma_ftr->active = 0; local_rdma_frag->segment.seg_addr.pval = (unsigned char*)(local_rdma_frag + 1); local_rdma_frag->segment.seg_len = mca_btl_udapl_module.super.btl_eager_limit; local_rdma_frag->base.des_flags = 0; /* increment local rdma credits */ OPAL_THREAD_ADD32(&(endpoint->endpoint_eager_rdma_local.credits), 1); if (endpoint->endpoint_eager_rdma_local.credits >= mca_btl_udapl_component.udapl_eager_rdma_win) { mca_btl_udapl_endpoint_send_eager_rdma_credits(endpoint); } count++; } else { OPAL_THREAD_UNLOCK(&endpoint->endpoint_eager_rdma_local.lock); } } /* end of rdma_count loop */ } /* unlock and return */ OPAL_THREAD_UNLOCK(&mca_btl_udapl_component.udapl_lock); OPAL_THREAD_ADD32(&inprogress, -1); return count; }