1
1
openmpi/ompi/mca/pml/csum/pml_csum_component.c
Rolf vandeVaart 65e8277add Mostly fixes for tabs, spaces and indentations.
Also, some other changes to bring the csum PML up
to date with changes that happened in ob1 over the
last two years. This includes a few bug
fixes and some minor refactoring.  

This commit was SVN r23757.
2010-09-15 18:48:06 +00:00

252 строки
8.7 KiB
C

/* -*- Mode: C; c-basic-offset:4 ; -*- */
/*
* Copyright (c) 2004-2007 The Trustees of Indiana University and Indiana
* University Research and Technology
* Corporation. All rights reserved.
* Copyright (c) 2004-2009 The University of Tennessee and The University
* of Tennessee Research Foundation. All rights
* reserved.
* Copyright (c) 2004-2005 High Performance Computing Center Stuttgart,
* University of Stuttgart. All rights reserved.
* Copyright (c) 2004-2005 The Regents of the University of California.
* All rights reserved.
* Copyright (c) 2009 IBM Corporation. All rights reserved.
* Copyright (c) 2009 Los Alamos National Security, LLC. All rights
* reserved.
* Copyright (c) 2007-2010 Cisco Systems, Inc. All rights reserved.
* Copyright (c) 2010 Oracle and/or its affiliates. All rights reserved.
* $COPYRIGHT$
*
* Additional copyrights may follow
*
* $HEADER$
*/
#include "ompi_config.h"
#include "opal/event/event.h"
#include "mpi.h"
#include "ompi/runtime/params.h"
#include "ompi/mca/pml/pml.h"
#include "opal/mca/base/mca_base_param.h"
#include "ompi/mca/pml/base/pml_base_bsend.h"
#include "pml_csum.h"
#include "pml_csum_hdr.h"
#include "pml_csum_sendreq.h"
#include "pml_csum_recvreq.h"
#include "pml_csum_rdmafrag.h"
#include "pml_csum_recvfrag.h"
#include "ompi/mca/bml/base/base.h"
#include "pml_csum_component.h"
#include "ompi/mca/allocator/base/base.h"
OBJ_CLASS_INSTANCE( mca_pml_csum_pckt_pending_t,
ompi_free_list_item_t,
NULL,
NULL );
static int mca_pml_csum_component_open(void);
static int mca_pml_csum_component_close(void);
static mca_pml_base_module_t*
mca_pml_csum_component_init( int* priority, bool enable_progress_threads,
bool enable_mpi_threads );
static int mca_pml_csum_component_fini(void);
int mca_pml_csum_output = 0;
mca_pml_base_component_2_0_0_t mca_pml_csum_component = {
/* First, the mca_base_component_t struct containing meta
information about the component itself */
{
MCA_PML_BASE_VERSION_2_0_0,
"csum", /* MCA component name */
OMPI_MAJOR_VERSION, /* MCA component major version */
OMPI_MINOR_VERSION, /* MCA component minor version */
OMPI_RELEASE_VERSION, /* MCA component release version */
mca_pml_csum_component_open, /* component open */
mca_pml_csum_component_close /* component close */
},
{
/* The component is checkpoint ready */
MCA_BASE_METADATA_PARAM_CHECKPOINT
},
mca_pml_csum_component_init, /* component init */
mca_pml_csum_component_fini /* component finalize */
};
void *mca_pml_csum_seg_alloc( struct mca_mpool_base_module_t* mpool,
size_t* size,
mca_mpool_base_registration_t** registration);
void mca_pml_csum_seg_free( struct mca_mpool_base_module_t* mpool,
void* segment );
static inline int mca_pml_csum_param_register_int(
const char* param_name,
int default_value)
{
int id = mca_base_param_register_int("pml","csum",param_name,NULL,default_value);
int param_value = default_value;
mca_base_param_lookup_int(id,&param_value);
return param_value;
}
static int mca_pml_csum_component_open(void)
{
int value;
mca_allocator_base_component_t* allocator_component;
value = mca_pml_csum_param_register_int("verbose", 0);
mca_pml_csum_output = opal_output_open(NULL);
opal_output_set_verbosity(mca_pml_csum_output, value);
mca_pml_csum.free_list_num =
mca_pml_csum_param_register_int("free_list_num", 4);
mca_pml_csum.free_list_max =
mca_pml_csum_param_register_int("free_list_max", -1);
mca_pml_csum.free_list_inc =
mca_pml_csum_param_register_int("free_list_inc", 64);
mca_pml_csum.send_pipeline_depth =
mca_pml_csum_param_register_int("send_pipeline_depth", 3);
mca_pml_csum.recv_pipeline_depth =
mca_pml_csum_param_register_int("recv_pipeline_depth", 4);
mca_pml_csum.rdma_put_retries_limit =
mca_pml_csum_param_register_int("rdma_put_retries_limit", 5);
mca_pml_csum.max_rdma_per_request =
mca_pml_csum_param_register_int("max_rdma_per_request", 4);
mca_pml_csum.max_send_per_range =
mca_pml_csum_param_register_int("max_send_per_range", 4);
mca_pml_csum.unexpected_limit =
mca_pml_csum_param_register_int("unexpected_limit", 128);
mca_base_param_reg_string(&mca_pml_csum_component.pmlm_version,
"allocator",
"Name of allocator component for unexpected messages",
false, false,
"bucket",
&mca_pml_csum.allocator_name);
allocator_component = mca_allocator_component_lookup( mca_pml_csum.allocator_name );
if(NULL == allocator_component) {
opal_output(0, "mca_pml_csum_component_open: can't find allocator: %s\n", mca_pml_csum.allocator_name);
return OMPI_ERROR;
}
mca_pml_csum.allocator = allocator_component->allocator_init(true,
mca_pml_csum_seg_alloc,
mca_pml_csum_seg_free, NULL);
if(NULL == mca_pml_csum.allocator) {
opal_output(0, "mca_pml_csum_component_open: unable to initialize allocator\n");
return OMPI_ERROR;
}
return mca_bml_base_open();
}
static int mca_pml_csum_component_close(void)
{
int rc;
if (OMPI_SUCCESS != (rc = mca_bml_base_close())) {
return rc;
}
if (NULL != mca_pml_csum.allocator_name) {
free(mca_pml_csum.allocator_name);
}
return OMPI_SUCCESS;
}
static mca_pml_base_module_t*
mca_pml_csum_component_init( int* priority,
bool enable_progress_threads,
bool enable_mpi_threads )
{
opal_output_verbose( 10, mca_pml_csum_output,
"in csum, my priority is 0\n");
/* select us only if we are specified */
if((*priority) > 0) {
*priority = 0;
return NULL;
}
*priority = 0;
if(OMPI_SUCCESS != mca_bml_base_init( enable_progress_threads,
enable_mpi_threads)) {
return NULL;
}
/* Set this here (vs in component_open()) because
ompi_mpi_leave_pinned* may have been set after MCA params were
read (e.g., by the openib btl) */
mca_pml_csum.leave_pinned = (1 == ompi_mpi_leave_pinned);
mca_pml_csum.leave_pinned_pipeline = (int) ompi_mpi_leave_pinned_pipeline;
return &mca_pml_csum.super;
}
int mca_pml_csum_component_fini(void)
{
int rc;
/* Shutdown BML */
if(OMPI_SUCCESS != (rc = mca_bml.bml_finalize()))
return rc;
if(!mca_pml_csum.enabled)
return OMPI_SUCCESS; /* never selected.. return success.. */
mca_pml_csum.enabled = false; /* not anymore */
OBJ_DESTRUCT(&mca_pml_csum.rdma_pending);
OBJ_DESTRUCT(&mca_pml_csum.pckt_pending);
OBJ_DESTRUCT(&mca_pml_csum.recv_pending);
OBJ_DESTRUCT(&mca_pml_csum.send_pending);
OBJ_DESTRUCT(&mca_pml_csum.non_existing_communicator_pending);
OBJ_DESTRUCT(&mca_pml_csum.buffers);
OBJ_DESTRUCT(&mca_pml_csum.pending_pckts);
OBJ_DESTRUCT(&mca_pml_csum.recv_frags);
OBJ_DESTRUCT(&mca_pml_csum.rdma_frags);
OBJ_DESTRUCT(&mca_pml_csum.lock);
if(OMPI_SUCCESS != (rc = mca_pml_csum.allocator->alc_finalize(mca_pml_csum.allocator))) {
return rc;
}
#if 0
if (mca_pml_base_send_requests.fl_num_allocated !=
mca_pml_base_send_requests.super.opal_list_length) {
opal_output(0, "csum send requests: %d allocated %d returned\n",
mca_pml_base_send_requests.fl_num_allocated,
mca_pml_base_send_requests.super.opal_list_length);
}
if (mca_pml_base_recv_requests.fl_num_allocated !=
mca_pml_base_recv_requests.super.opal_list_length) {
opal_output(0, "csum recv requests: %d allocated %d returned\n",
mca_pml_base_recv_requests.fl_num_allocated,
mca_pml_base_recv_requests.super.opal_list_length);
}
#endif
return OMPI_SUCCESS;
}
void *mca_pml_csum_seg_alloc( struct mca_mpool_base_module_t* mpool,
size_t* size,
mca_mpool_base_registration_t** registration) {
return malloc(*size);
}
void mca_pml_csum_seg_free( struct mca_mpool_base_module_t* mpool,
void* segment ) {
free(segment);
}