/* * Copyright (c) 2004-2005 The Trustees of Indiana University and Indiana * University Research and Technology * Corporation. All rights reserved. * Copyright (c) 2004-2011 The University of Tennessee and The University * of Tennessee Research Foundation. All rights * reserved. * Copyright (c) 2004-2005 High Performance Computing Center Stuttgart, * University of Stuttgart. All rights reserved. * Copyright (c) 2004-2005 The Regents of the University of California. * All rights reserved. * Copyright (c) 2006 Sandia National Laboratories. All rights * reserved. * Copyright (c) 2013-2016 Cisco Systems, Inc. All rights reserved. * Copyright (c) 2013-2014 Intel, Inc. All rights reserved * $COPYRIGHT$ * * Additional copyrights may follow * * $HEADER$ */ #include #include "opal_config.h" #include "opal_stdint.h" #include "opal/util/arch.h" #include "opal/util/show_help.h" #include "opal/constants.h" #include "btl_usnic_compat.h" #include "btl_usnic.h" #include "btl_usnic_proc.h" #include "btl_usnic_endpoint.h" #include "btl_usnic_module.h" #include "btl_usnic_util.h" #include "btl_usnic_graph.h" /* larger weight values are more desirable (i.e., worth, not cost) */ enum { WEIGHT_UNREACHABLE = -1 }; /* Helper macros for "match_modex" and friends for translating between array * indices and vertex IDs. Module vertices always come first in the graph, * followed by proc (endpoint) vertices. */ #define PROC_VERTEX(modex_idx) (mca_btl_usnic_component.num_modules + modex_idx) #define MODULE_VERTEX(module_idx) (module_idx) #define PROC_INDEX(proc_vertex) ((proc_vertex) - mca_btl_usnic_component.num_modules) #define MODULE_INDEX(module_vertex) (module_vertex) static void proc_construct(opal_btl_usnic_proc_t* proc) { proc->proc_opal = 0; proc->proc_modex = NULL; proc->proc_modex_count = 0; proc->proc_modex_claimed = NULL; proc->proc_endpoints = NULL; proc->proc_endpoint_count = 0; proc->proc_ep_match_table = NULL; proc->proc_match_exists = false; /* add to list of all proc instance */ opal_list_append(&mca_btl_usnic_component.usnic_procs, &proc->super); } static void proc_destruct(opal_btl_usnic_proc_t* proc) { /* remove from list of all proc instances */ opal_list_remove_item(&mca_btl_usnic_component.usnic_procs, &proc->super); /* release resources */ if (NULL != proc->proc_modex) { free(proc->proc_modex); proc->proc_modex = NULL; } if (NULL != proc->proc_modex_claimed) { free(proc->proc_modex_claimed); proc->proc_modex_claimed = NULL; } if (NULL != proc->proc_ep_match_table) { free(proc->proc_ep_match_table); proc->proc_ep_match_table = NULL; } /* Release all endpoints associated with this proc */ if (NULL != proc->proc_endpoints) { free(proc->proc_endpoints); proc->proc_endpoints = NULL; } } OBJ_CLASS_INSTANCE(opal_btl_usnic_proc_t, opal_list_item_t, proc_construct, proc_destruct); /* * Look for an existing usnic process instance based on the * associated opal_proc_t instance. */ opal_btl_usnic_proc_t * opal_btl_usnic_proc_lookup_ompi(opal_proc_t* opal_proc) { opal_btl_usnic_proc_t* usnic_proc; for (usnic_proc = (opal_btl_usnic_proc_t*) opal_list_get_first(&mca_btl_usnic_component.usnic_procs); usnic_proc != (opal_btl_usnic_proc_t*) opal_list_get_end(&mca_btl_usnic_component.usnic_procs); usnic_proc = (opal_btl_usnic_proc_t*) opal_list_get_next(usnic_proc)) { if (usnic_proc->proc_opal == opal_proc) { return usnic_proc; } } return NULL; } /* * Look for an existing usnic proc based on a hashed RTE process * name. */ opal_btl_usnic_endpoint_t * opal_btl_usnic_proc_lookup_endpoint(opal_btl_usnic_module_t *receiver, uint64_t sender_proc_name) { opal_btl_usnic_proc_t *proc; opal_btl_usnic_endpoint_t *endpoint; opal_list_item_t *item; MSGDEBUG1_OUT("lookup_endpoint: recvmodule=%p sendhash=0x%" PRIx64, (void *)receiver, sender_proc_name); opal_mutex_lock(&receiver->all_endpoints_lock); for (item = opal_list_get_first(&receiver->all_endpoints); item != opal_list_get_end(&receiver->all_endpoints); item = opal_list_get_next(item)) { endpoint = container_of(item, opal_btl_usnic_endpoint_t, endpoint_endpoint_li); proc = endpoint->endpoint_proc; /* Note that this works today because opal_proc_t->proc_name is unique across the universe. George is potentially working to give handles instead of proc names, and then have a function pointer to perform comparisons. This would be bad here in the critical path, though... */ if (usnic_compat_rte_hash_name(&(proc->proc_opal->proc_name)) == sender_proc_name) { MSGDEBUG1_OUT("lookup_endpoint: matched endpoint=%p", (void *)endpoint); opal_mutex_unlock(&receiver->all_endpoints_lock); return endpoint; } } opal_mutex_unlock(&receiver->all_endpoints_lock); /* Didn't find it */ return NULL; } /* * Create an opal_btl_usnic_proc_t and initialize it with modex info * and an empty array of endpoints. * * Returns OPAL_ERR_UNREACH if we can't reach the peer (i.e., we can't * find their modex data). */ static int create_proc(opal_proc_t *opal_proc, opal_btl_usnic_proc_t **usnic_proc) { opal_btl_usnic_proc_t *proc = NULL; size_t size; int rc; *usnic_proc = NULL; /* Create the proc if it doesn't already exist */ proc = OBJ_NEW(opal_btl_usnic_proc_t); if (NULL == proc) { return OPAL_ERR_OUT_OF_RESOURCE; } /* Initialize number of peers */ proc->proc_endpoint_count = 0; proc->proc_opal = opal_proc; /* query for the peer address info */ usnic_compat_modex_recv(&rc, &mca_btl_usnic_component.super.btl_version, opal_proc, &proc->proc_modex, &size); /* If this proc simply doesn't have this key, then they're not running the usnic BTL -- just ignore them. Otherwise, show an error message. */ if (OPAL_ERR_NOT_FOUND == rc) { OBJ_RELEASE(proc); return OPAL_ERR_UNREACH; } else if (OPAL_SUCCESS != rc) { opal_show_help("help-mpi-btl-usnic.txt", "internal error during init", true, opal_process_info.nodename, "", "", "opal_modex_recv() failed", __FILE__, __LINE__, opal_strerror(rc)); OBJ_RELEASE(proc); return OPAL_ERROR; } if ((size % sizeof(opal_btl_usnic_modex_t)) != 0) { char msg[1024]; snprintf(msg, sizeof(msg), "sizeof(modex for peer %s data) == %d, expected multiple of %d", usnic_compat_proc_name_print(&opal_proc->proc_name), (int) size, (int) sizeof(opal_btl_usnic_modex_t)); opal_show_help("help-mpi-btl-usnic.txt", "internal error during init", true, opal_process_info.nodename, "", 0, "invalid modex data", __FILE__, __LINE__, msg); OBJ_RELEASE(proc); return OPAL_ERR_VALUE_OUT_OF_BOUNDS; } /* See if the peer has the same underlying wire protocol as me. If not, then print an error and ignore this peer. */ // RFXXX - things are weird when i force this to fail if (mca_btl_usnic_component.transport_protocol != proc->proc_modex->protocol) { uint64_t proto; char protostr[32]; proto = mca_btl_usnic_component.transport_protocol; memset(protostr, 0, sizeof(protostr)); strncpy(protostr, fi_tostr(&proto, FI_TYPE_PROTOCOL), sizeof(protostr) - 1); proto = proc->proc_modex->protocol; opal_show_help("help-mpi-btl-usnic.txt", "transport mismatch", true, opal_process_info.nodename, protostr, "peer", fi_tostr(&proto, FI_TYPE_PROTOCOL)); OBJ_RELEASE(proc); return OPAL_ERR_UNREACH; } proc->proc_modex_count = size / sizeof(opal_btl_usnic_modex_t); if (0 == proc->proc_modex_count) { proc->proc_endpoints = NULL; OBJ_RELEASE(proc); return OPAL_ERR_UNREACH; } proc->proc_modex_claimed = (bool*) calloc(proc->proc_modex_count, sizeof(bool)); if (NULL == proc->proc_modex_claimed) { OPAL_ERROR_LOG(OPAL_ERR_OUT_OF_RESOURCE); OBJ_RELEASE(proc); return OPAL_ERR_OUT_OF_RESOURCE; } proc->proc_endpoints = (mca_btl_base_endpoint_t**) calloc(proc->proc_modex_count, sizeof(mca_btl_base_endpoint_t*)); if (NULL == proc->proc_endpoints) { OPAL_ERROR_LOG(OPAL_ERR_OUT_OF_RESOURCE); OBJ_RELEASE(proc); return OPAL_ERR_OUT_OF_RESOURCE; } *usnic_proc = proc; return OPAL_SUCCESS; } /* Compare the addresses of the local interface corresponding to module and the * remote interface corresponding to proc_modex_addr. Returns a weight value * (higher values indicate more desirable connections). */ static uint64_t compute_weight( opal_btl_usnic_module_t *module, opal_btl_usnic_modex_t *proc_modex_addr) { char my_ip_string[INET_ADDRSTRLEN], peer_ip_string[INET_ADDRSTRLEN]; struct sockaddr_in sin; struct sockaddr_in *sinp; struct fi_usnic_info *uip; uint32_t mynet, peernet; int err; int metric; uint32_t min_link_speed_gbps; uip = &module->usnic_info; sinp = module->fabric_info->src_addr; inet_ntop(AF_INET, &sinp->sin_addr, my_ip_string, sizeof(my_ip_string)); inet_ntop(AF_INET, &proc_modex_addr->ipv4_addr, peer_ip_string, sizeof(peer_ip_string)); /* Just compare the CIDR-masked IP address to see if they're on the same network. If so, we're good. */ mynet = sinp->sin_addr.s_addr & uip->ui.v1.ui_netmask_be; peernet = proc_modex_addr->ipv4_addr & proc_modex_addr->netmask; opal_output_verbose(5, USNIC_OUT, "btl:usnic:%s: checking my IP address/subnet (%s/%d) vs. peer (%s/%d): %s", __func__, my_ip_string, usnic_netmask_to_cidrlen(uip->ui.v1.ui_netmask_be), peer_ip_string, usnic_netmask_to_cidrlen(proc_modex_addr->netmask), (mynet == peernet ? "match" : "DO NOT match")); min_link_speed_gbps = MIN(module->super.btl_bandwidth, proc_modex_addr->link_speed_mbps) / 1000; /* Returned metric is: * 0 - same VLAN * 1..MAXINT - relative distance metric * -1 - unreachable */ metric = 0; memset(&sin, 0, sizeof(sin)); sin.sin_family = AF_INET; sin.sin_addr.s_addr = proc_modex_addr->ipv4_addr; err = module->usnic_av_ops->get_distance(module->av, &sin, &metric); if (0 != err || (0 == err && -1 == metric)) { return 0; /* no connectivity */ } else { /* Format in binary MSB LSB * most sig. 32-bits: 00000000 0000000A BBBBBBBB 00000001 * least sig. 32-bits: CCCCCCCC CCCCCCCC CCCCCCCC CCCCCCCC * * A = 1 iff same subnet * B = min link speed (in Gbps) between iface pair * C = metric from routing table * * That is, this prioritizes interfaces in the same subnet first, * followed by having the same link speed. The extra literal "1" is in * there to help prioritize over any zero-cost links that might * otherwise make their way into the graph. It is not strictly * necessary and could be eliminated if the extra byte is needed. * * TODO add an MCA parameter to optionally swap the offsets of A and * B, thereby prioritizing link speed over same subnet reachability. */ /* FIXME how can we check that the metric is the same before we have * communication with this host? Mismatched metrics could cause the * remote peer to make a different pairing decision... */ if (min_link_speed_gbps > 0xff) { opal_output_verbose(20, USNIC_OUT, "clamping min_link_speed_gbps=%u to 255", min_link_speed_gbps); min_link_speed_gbps = 0xff; } return ((uint64_t)(mynet == peernet) << 48) | ((uint64_t)(min_link_speed_gbps & 0xff) << 40) | ((uint64_t)0x1 << 32) | (/*metric=*/0); } } /* Populate the given proc's match table from an array of (u,v) edge pairs. * * (DJG: this unfortunately knows a bit too much about the internals of * "match_modex") */ static void edge_pairs_to_match_table( opal_btl_usnic_proc_t *proc, bool proc_is_left, int nme, int *me) { int i; int left, right; int module_idx, proc_idx; int num_modules; num_modules = (int)mca_btl_usnic_component.num_modules; assert(nme >= 0); for (i = 0; i < nme; ++i) { left = me[2*i+0]; right = me[2*i+1]; if (proc_is_left) { proc_idx = PROC_INDEX(left); module_idx = MODULE_INDEX(right); } else { module_idx = MODULE_INDEX(left); proc_idx = PROC_INDEX(right); } assert(module_idx >= 0 && module_idx < num_modules); assert(proc_idx >= 0 && proc_idx < (int)proc->proc_modex_count); proc->proc_ep_match_table[module_idx] = proc_idx; proc->proc_match_exists = true; } /* emit match summary for debugging purposes */ for (i = 0; i < num_modules; ++i) { if (-1 != proc->proc_ep_match_table[i]) { opal_output_verbose(5, USNIC_OUT, "btl:usnic:%s: module[%d] (%p) should claim endpoint[%d] on proc %p", __func__, i, (void *)mca_btl_usnic_component.usnic_active_modules[i], proc->proc_ep_match_table[i], (void *)proc); } else { opal_output_verbose(5, USNIC_OUT, "btl:usnic:%s: module[%d] (%p) will NOT claim an endpoint on proc %p", __func__, i, (void *)mca_btl_usnic_component.usnic_active_modules[i], (void *)proc); } } } /** * Constructs an interface graph from all local modules and the given proc's * remote interfaces. The resulting vertices will always have the module * vertices appear before the proc vertices. */ static int create_proc_module_graph( opal_btl_usnic_proc_t *proc, bool proc_is_left, opal_btl_usnic_graph_t **g_out) { int err; int i, j; int u, v; int num_modules; opal_btl_usnic_graph_t *g = NULL; if (NULL == g_out) { return OPAL_ERR_BAD_PARAM; } *g_out = NULL; num_modules = (int)mca_btl_usnic_component.num_modules; /* Construct a bipartite graph with remote interfaces on the one side and * local interfaces (modules) on the other. */ err = opal_btl_usnic_gr_create(NULL, NULL, &g); if (OPAL_SUCCESS != err) { OPAL_ERROR_LOG(err); goto out; } /* create vertices for each interface (local and remote) */ for (i = 0; i < num_modules; ++i) { int idx = -1; err = opal_btl_usnic_gr_add_vertex(g, mca_btl_usnic_component.usnic_active_modules[i], &idx); if (OPAL_SUCCESS != err) { OPAL_ERROR_LOG(err); goto out_free_graph; } assert(idx == MODULE_VERTEX(i)); } for (i = 0; i < (int)proc->proc_modex_count; ++i) { int idx = -1; err = opal_btl_usnic_gr_add_vertex(g, &proc->proc_modex[i], &idx); if (OPAL_SUCCESS != err) { OPAL_ERROR_LOG(err); goto out_free_graph; } assert(idx == (int)PROC_VERTEX(i)); } /* now add edges between interfaces that can communicate */ for (i = 0; i < num_modules; ++i) { for (j = 0; j < (int)proc->proc_modex_count; ++j) { int64_t weight, cost; /* assumption: compute_weight returns the same weight on the * remote process with these arguments (effectively) transposed */ weight = compute_weight(mca_btl_usnic_component.usnic_active_modules[i], &proc->proc_modex[j]); opal_output_verbose(20, USNIC_OUT, "btl:usnic:%s: weight=0x%016" PRIx64 " for edge module[%d] (%p) <--> endpoint[%d] on proc %p", __func__, weight, i, (void *)mca_btl_usnic_component.usnic_active_modules[i], j, (void *)proc); if (WEIGHT_UNREACHABLE == weight) { continue; } else { /* the graph code optimizes for minimum *cost*, but we have * been computing weights (negative costs) */ cost = -weight; } assert(INT64_MAX != cost); assert(INT64_MIN != cost); if (proc_is_left) { u = PROC_VERTEX(j); v = MODULE_VERTEX(i); } else { u = MODULE_VERTEX(i); v = PROC_VERTEX(j); } opal_output_verbose(20, USNIC_OUT, "btl:usnic:%s: adding edge (%d,%d) with cost=%" PRIi64 " for edge module[%d] <--> endpoint[%d]", __func__, u, v, cost, i, j); err = opal_btl_usnic_gr_add_edge(g, u, v, cost, /*capacity=*/1, /*e_data=*/NULL); if (OPAL_SUCCESS != err) { OPAL_ERROR_LOG(err); goto out_free_graph; } } } *g_out = g; return OPAL_SUCCESS; out_free_graph: opal_btl_usnic_gr_free(g); out: return err; } /* * For a specific module, see if this proc has matching address/modex * info. If so, create an endpoint and return it. * * Implementation note: This code relies on the order of modules on a local * side matching the order of the modex entries that we send around, otherwise * both sides may not agree on a bidirectional connection. It also assumes * that add_procs will be invoked on the local modules in that same order, for * the same reason. If those assumptions do not hold, we will need to * canonicalize this match ordering somehow, probably by (jobid,vpid) pair or * by the interface MAC or IP address. */ static int match_modex(opal_btl_usnic_module_t *module, opal_btl_usnic_proc_t *proc, int *index_out) { int err = OPAL_SUCCESS; size_t i; uint32_t num_modules; opal_btl_usnic_graph_t *g = NULL; bool proc_is_left; if (NULL == index_out) { return OPAL_ERR_BAD_PARAM; } *index_out = -1; num_modules = mca_btl_usnic_component.num_modules; opal_output_verbose(20, USNIC_OUT, "btl:usnic:%s: module=%p proc=%p with dimensions %d x %d", __func__, (void *)module, (void *)proc, num_modules, (int)proc->proc_modex_count); /* We compute an interface match-up table once for each (module,proc) pair * and cache it in the proc. Store per-proc instead of per-module, since * MPI dynamic process routines can add procs but not new modules. */ if (NULL == proc->proc_ep_match_table) { proc->proc_ep_match_table = malloc(num_modules * sizeof(*proc->proc_ep_match_table)); if (NULL == proc->proc_ep_match_table) { OPAL_ERROR_LOG(OPAL_ERR_OUT_OF_RESOURCE); return OPAL_ERR_OUT_OF_RESOURCE; } /* initialize to "no matches" */ for (i = 0; i < num_modules; ++i) { proc->proc_ep_match_table[i] = -1; } /* For graphs where all edges are equal (and even for some other * graphs), two peers making matching calculations with "mirror image" * graphs might not end up with the same matching. Ensure that both * sides are always setting up the exact same graph by always putting * the process with the lower (jobid,vpid) on the "left". */ #if 0 proc_is_left = (proc->proc_opal->proc_name < opal_proc_local_get()->proc_name); #else proc_is_left = usnic_compat_proc_name_compare(proc->proc_opal->proc_name, opal_proc_local_get()->proc_name); #endif err = create_proc_module_graph(proc, proc_is_left, &g); if (OPAL_SUCCESS != err) { goto out_free_table; } int nme = 0; int *me = NULL; err = opal_btl_usnic_solve_bipartite_assignment(g, &nme, &me); if (OPAL_SUCCESS != err) { OPAL_ERROR_LOG(err); goto out_free_graph; } edge_pairs_to_match_table(proc, proc_is_left, nme, me); free(me); err = opal_btl_usnic_gr_free(g); if (OPAL_SUCCESS != err) { OPAL_ERROR_LOG(err); return err; } } if (!proc->proc_match_exists) { opal_output_verbose(5, USNIC_OUT, "btl:usnic:%s: unable to find any valid interface pairs for proc %s", __func__, usnic_compat_proc_name_print(&proc->proc_opal->proc_name)); return OPAL_ERR_NOT_FOUND; } /* assuming no strange failure cases, this should always be present */ if (NULL != proc->proc_ep_match_table && proc->proc_match_exists) { for (i = 0; i < num_modules; ++i) { if (module == mca_btl_usnic_component.usnic_active_modules[i]) { *index_out = proc->proc_ep_match_table[i]; break; } } } /* If MTU does not match, throw an error */ /* TODO with UDP, do we still want to enforce this restriction or just take * the min of the two MTUs? Another choice is to disqualify this pairing * before running the matching algorithm on it. */ if (*index_out >= 0 && proc->proc_modex[*index_out].max_msg_size != (uint16_t) module->fabric_info->ep_attr->max_msg_size) { opal_show_help("help-mpi-btl-usnic.txt", "MTU mismatch", true, opal_process_info.nodename, module->linux_device_name, module->fabric_info->ep_attr->max_msg_size, (NULL == proc->proc_opal->proc_hostname) ? "unknown" : proc->proc_opal->proc_hostname, proc->proc_modex[*index_out].max_msg_size); *index_out = -1; return OPAL_ERR_UNREACH; } return (*index_out == -1 ? OPAL_ERR_NOT_FOUND : OPAL_SUCCESS); out_free_graph: opal_btl_usnic_gr_free(g); out_free_table: free(proc->proc_ep_match_table); proc->proc_ep_match_table = NULL; proc->proc_match_exists = false; return err; } /* * Initiate the process to create a USD dest. * It will be polled for completion later. */ static int start_av_insert(opal_btl_usnic_module_t *module, opal_btl_usnic_endpoint_t *endpoint, int channel) { int ret; opal_btl_usnic_modex_t *modex = &endpoint->endpoint_remote_modex; opal_btl_usnic_addr_context_t *context; struct sockaddr_in sin; context = calloc(1, sizeof(*context)); context->endpoint = endpoint; context->channel_id = channel; char str[IPV4STRADDRLEN]; opal_btl_usnic_snprintf_ipv4_addr(str, sizeof(str), modex->ipv4_addr, modex->netmask); opal_output_verbose(5, USNIC_OUT, "btl:usnic:start_av_insert: to channel %d at %s:%d", channel, str, modex->ports[channel]); /* build remote address */ memset(&sin, 0, sizeof(sin)); sin.sin_family = AF_INET; sin.sin_port = htons(modex->ports[channel]); sin.sin_addr.s_addr = modex->ipv4_addr; ret = fi_av_insert(module->av, &sin, 1, &endpoint->endpoint_remote_addrs[channel], 0, context); /* Did an error occur? */ if (0 != ret) { opal_show_help("help-mpi-btl-usnic.txt", "libfabric API failed", true, opal_process_info.nodename, module->linux_device_name, "fi_av_insert()", __FILE__, __LINE__, ret, "Failed to initiate AV insert"); free(context); return OPAL_ERROR; } return OPAL_SUCCESS; } /* * Create an endpoint and claim the matched modex slot */ int opal_btl_usnic_create_endpoint(opal_btl_usnic_module_t *module, opal_btl_usnic_proc_t *proc, opal_btl_usnic_endpoint_t **endpoint_o) { int rc; int modex_index; opal_btl_usnic_endpoint_t *endpoint; /* look for matching modex info */ rc = match_modex(module, proc, &modex_index); if (OPAL_SUCCESS != rc) { opal_output_verbose(5, USNIC_OUT, "btl:usnic:create_endpoint: did not match usnic modex info for peer %s", usnic_compat_proc_name_print(&proc->proc_opal->proc_name)); return rc; } endpoint = OBJ_NEW(opal_btl_usnic_endpoint_t); if (NULL == endpoint) { return OPAL_ERR_OUT_OF_RESOURCE; } /* Initalize the endpoint */ endpoint->endpoint_module = module; assert(modex_index >= 0 && modex_index < (int)proc->proc_modex_count); endpoint->endpoint_remote_modex = proc->proc_modex[modex_index]; /* Start creating destinations; one for each channel. These progress in the background.a */ for (int i = 0; i < USNIC_NUM_CHANNELS; ++i) { rc = start_av_insert(module, endpoint, i); if (OPAL_SUCCESS != rc) { OBJ_RELEASE(endpoint); return rc; } } /* Initialize endpoint sequence number info */ endpoint->endpoint_next_seq_to_send = module->local_modex.isn; endpoint->endpoint_ack_seq_rcvd = endpoint->endpoint_next_seq_to_send - 1; endpoint->endpoint_next_contig_seq_to_recv = endpoint->endpoint_remote_modex.isn; endpoint->endpoint_highest_seq_rcvd = endpoint->endpoint_next_contig_seq_to_recv - 1; endpoint->endpoint_rfstart = WINDOW_SIZE_MOD(endpoint->endpoint_next_contig_seq_to_recv); /* Now claim that modex slot */ proc->proc_modex_claimed[modex_index] = true; MSGDEBUG1_OUT("create_endpoint: module=%p claimed endpoint=%p on proc=%p (hash=0x%" PRIx64 ")\n", (void *)module, (void *)endpoint, (void *)proc, proc->proc_opal->proc_name); /* Save the endpoint on this proc's array of endpoints */ proc->proc_endpoints[proc->proc_endpoint_count] = endpoint; endpoint->endpoint_proc_index = proc->proc_endpoint_count; endpoint->endpoint_proc = proc; ++proc->proc_endpoint_count; OBJ_RETAIN(proc); /* also add endpoint to module's list of endpoints (done here and not in the endpoint constructor because we aren't able to pass the module as a constructor argument -- doh!). */ opal_mutex_lock(&module->all_endpoints_lock); opal_list_append(&(module->all_endpoints), &(endpoint->endpoint_endpoint_li)); endpoint->endpoint_on_all_endpoints = true; opal_mutex_unlock(&module->all_endpoints_lock); *endpoint_o = endpoint; return OPAL_SUCCESS; } /* * If we haven't done so already, receive the modex info for the * specified opal_proc. Search that proc's modex info; if we can find * matching address info, then create an endpoint. * * If we don't find a match, it's not an error: just return "not * found". * * This routine transfers ownership of an object reference to the caller, who * is eventually responsible for transferring or releasing that reference. * * There is a one-to-one correspondence between a opal_proc_t and a * opal_btl_usnic_proc_t instance. We cache additional data on the * opal_btl_usnic_proc_t: specifically, the list of * opal_btl_usnic_endpoint_t instances, and published addresses/modex * info. */ int opal_btl_usnic_proc_match(opal_proc_t *opal_proc, opal_btl_usnic_module_t *module, opal_btl_usnic_proc_t **proc) { /* Check if we have already created a proc structure for this peer ompi process */ *proc = opal_btl_usnic_proc_lookup_ompi(opal_proc); if (*proc != NULL) { OBJ_RETAIN(*proc); return OPAL_SUCCESS; } else { /* If not, go make one */ return create_proc(opal_proc, proc); } }