6c5532072a
OMPI and a language agnostic part in OPAL. The convertor is completely moved into OPAL. This offers several benefits as described in RFC http://www.open-mpi.org/community/lists/devel/2009/07/6387.php namely: - Fewer basic types (int* and float* types, boolean and wchar - Fixing naming scheme to ompi-nomenclature. - Usability outside of the ompi-layer. - Due to the fixed nature of simple opal types, their information is completely known at compile time and therefore constified - With fewer datatypes (22), the actual sizes of bit-field types may be reduced from 64 to 32 bits, allowing reorganizing the opal_datatype structure, eliminating holes and keeping data required in convertor (upon send/recv) in one cacheline... This has implications to the convertor-datastructure and other parts of the code. - Several performance tests have been run, the netpipe latency does not change with this patch on Linux/x86-64 on the smoky cluster. - Extensive tests have been done to verify correctness (no new regressions) using: 1. mpi_test_suite on linux/x86-64 using clean ompi-trunk and ompi-ddt: a. running both trunk and ompi-ddt resulted in no differences (except for MPI_SHORT_INT and MPI_TYPE_MIX_LB_UB do now run correctly). b. with --enable-memchecker and running under valgrind (one buglet when run with static found in test-suite, commited) 2. ibm testsuite on linux/x86-64 using clean ompi-trunk and ompi-ddt: all passed (except for the dynamic/ tests failed!! as trunk/MTT) 3. compilation and usage of HDF5 tests on Jaguar using PGI and PathScale compilers. 4. compilation and usage on Scicortex. - Please note, that for the heterogeneous case, (-m32 compiled binaries/ompi), neither ompi-trunk, nor ompi-ddt branch would successfully launch. This commit was SVN r21641.
318 строки
12 KiB
C
318 строки
12 KiB
C
/*
|
|
* Copyright (c) 2004-2005 The Trustees of Indiana University and Indiana
|
|
* University Research and Technology
|
|
* Corporation. All rights reserved.
|
|
* Copyright (c) 2004-2007 The University of Tennessee and The University
|
|
* of Tennessee Research Foundation. All rights
|
|
* reserved.
|
|
* Copyright (c) 2004-2005 High Performance Computing Center Stuttgart,
|
|
* University of Stuttgart. All rights reserved.
|
|
* Copyright (c) 2004-2005 The Regents of the University of California.
|
|
* All rights reserved.
|
|
* Copyright (c) 2009 IBM Corporation. All rights reserved.
|
|
* Copyright (c) 2009 Los Alamos National Security, LLC. All rights
|
|
* reserved.
|
|
* $COPYRIGHT$
|
|
*
|
|
* Additional copyrights may follow
|
|
*
|
|
* $HEADER$
|
|
*/
|
|
/**
|
|
* @file
|
|
*/
|
|
|
|
#ifndef MCA_PML_CSUM_H
|
|
#define MCA_PML_CSUM_H
|
|
|
|
#include "ompi_config.h"
|
|
#include "ompi/class/ompi_free_list.h"
|
|
#include "ompi/request/request.h"
|
|
#include "ompi/mca/pml/pml.h"
|
|
#include "ompi/mca/pml/base/pml_base_request.h"
|
|
#include "ompi/mca/pml/base/pml_base_bsend.h"
|
|
#include "ompi/mca/pml/base/pml_base_sendreq.h"
|
|
#include "ompi/datatype/ompi_datatype.h"
|
|
#include "pml_csum_hdr.h"
|
|
#include "ompi/mca/bml/base/base.h"
|
|
#include "ompi/proc/proc.h"
|
|
#include "ompi/mca/allocator/base/base.h"
|
|
|
|
BEGIN_C_DECLS
|
|
|
|
/**
|
|
* CSUM PML module
|
|
*/
|
|
|
|
struct mca_pml_csum_t {
|
|
mca_pml_base_module_t super;
|
|
|
|
int free_list_num; /* initial size of free list */
|
|
int free_list_max; /* maximum size of free list */
|
|
int free_list_inc; /* number of elements to grow free list */
|
|
size_t send_pipeline_depth;
|
|
size_t recv_pipeline_depth;
|
|
size_t rdma_put_retries_limit;
|
|
int max_rdma_per_request;
|
|
int max_send_per_range;
|
|
bool leave_pinned;
|
|
int leave_pinned_pipeline;
|
|
|
|
/* lock queue access */
|
|
opal_mutex_t lock;
|
|
|
|
/* free lists */
|
|
ompi_free_list_t rdma_frags;
|
|
ompi_free_list_t recv_frags;
|
|
ompi_free_list_t pending_pckts;
|
|
ompi_free_list_t buffers;
|
|
ompi_free_list_t send_ranges;
|
|
|
|
/* list of pending operations */
|
|
opal_list_t pckt_pending;
|
|
opal_list_t send_pending;
|
|
opal_list_t recv_pending;
|
|
opal_list_t rdma_pending;
|
|
/* List of pending fragments without a matching communicator */
|
|
opal_list_t non_existing_communicator_pending;
|
|
bool enabled;
|
|
char* allocator_name;
|
|
mca_allocator_base_module_t* allocator;
|
|
uint32_t unexpected_limit;
|
|
};
|
|
typedef struct mca_pml_csum_t mca_pml_csum_t;
|
|
|
|
extern mca_pml_csum_t mca_pml_csum;
|
|
|
|
/*
|
|
* PML interface functions.
|
|
*/
|
|
|
|
extern int mca_pml_csum_add_comm(
|
|
struct ompi_communicator_t* comm
|
|
);
|
|
|
|
extern int mca_pml_csum_del_comm(
|
|
struct ompi_communicator_t* comm
|
|
);
|
|
|
|
extern int mca_pml_csum_add_procs(
|
|
struct ompi_proc_t **procs,
|
|
size_t nprocs
|
|
);
|
|
|
|
extern int mca_pml_csum_del_procs(
|
|
struct ompi_proc_t **procs,
|
|
size_t nprocs
|
|
);
|
|
|
|
extern int mca_pml_csum_enable( bool enable );
|
|
|
|
extern int mca_pml_csum_progress(void);
|
|
|
|
extern int mca_pml_csum_iprobe( int dst,
|
|
int tag,
|
|
struct ompi_communicator_t* comm,
|
|
int *matched,
|
|
ompi_status_public_t* status );
|
|
|
|
extern int mca_pml_csum_probe( int dst,
|
|
int tag,
|
|
struct ompi_communicator_t* comm,
|
|
ompi_status_public_t* status );
|
|
|
|
extern int mca_pml_csum_isend_init( void *buf,
|
|
size_t count,
|
|
ompi_datatype_t *datatype,
|
|
int dst,
|
|
int tag,
|
|
mca_pml_base_send_mode_t mode,
|
|
struct ompi_communicator_t* comm,
|
|
struct ompi_request_t **request );
|
|
|
|
extern int mca_pml_csum_isend( void *buf,
|
|
size_t count,
|
|
ompi_datatype_t *datatype,
|
|
int dst,
|
|
int tag,
|
|
mca_pml_base_send_mode_t mode,
|
|
struct ompi_communicator_t* comm,
|
|
struct ompi_request_t **request );
|
|
|
|
extern int mca_pml_csum_send( void *buf,
|
|
size_t count,
|
|
ompi_datatype_t *datatype,
|
|
int dst,
|
|
int tag,
|
|
mca_pml_base_send_mode_t mode,
|
|
struct ompi_communicator_t* comm );
|
|
|
|
extern int mca_pml_csum_irecv_init( void *buf,
|
|
size_t count,
|
|
ompi_datatype_t *datatype,
|
|
int src,
|
|
int tag,
|
|
struct ompi_communicator_t* comm,
|
|
struct ompi_request_t **request );
|
|
|
|
extern int mca_pml_csum_irecv( void *buf,
|
|
size_t count,
|
|
ompi_datatype_t *datatype,
|
|
int src,
|
|
int tag,
|
|
struct ompi_communicator_t* comm,
|
|
struct ompi_request_t **request );
|
|
|
|
extern int mca_pml_csum_recv( void *buf,
|
|
size_t count,
|
|
ompi_datatype_t *datatype,
|
|
int src,
|
|
int tag,
|
|
struct ompi_communicator_t* comm,
|
|
ompi_status_public_t* status );
|
|
|
|
extern int mca_pml_csum_dump( struct ompi_communicator_t* comm,
|
|
int verbose );
|
|
|
|
extern int mca_pml_csum_start( size_t count,
|
|
ompi_request_t** requests );
|
|
|
|
extern int mca_pml_csum_ft_event( int state );
|
|
|
|
END_C_DECLS
|
|
|
|
struct mca_pml_csum_pckt_pending_t {
|
|
ompi_free_list_item_t super;
|
|
ompi_proc_t* proc;
|
|
mca_pml_csum_hdr_t hdr;
|
|
struct mca_bml_base_btl_t *bml_btl;
|
|
uint8_t order;
|
|
};
|
|
typedef struct mca_pml_csum_pckt_pending_t mca_pml_csum_pckt_pending_t;
|
|
OBJ_CLASS_DECLARATION(mca_pml_csum_pckt_pending_t);
|
|
|
|
#define MCA_PML_CSUM_PCKT_PENDING_ALLOC(pckt,rc) \
|
|
do { \
|
|
ompi_free_list_item_t* item; \
|
|
OMPI_FREE_LIST_WAIT(&mca_pml_csum.pending_pckts, item, rc); \
|
|
pckt = (mca_pml_csum_pckt_pending_t*)item; \
|
|
} while (0)
|
|
|
|
#define MCA_PML_CSUM_PCKT_PENDING_RETURN(pckt) \
|
|
do { \
|
|
/* return packet */ \
|
|
OMPI_FREE_LIST_RETURN(&mca_pml_csum.pending_pckts, \
|
|
(ompi_free_list_item_t*)pckt); \
|
|
} while(0)
|
|
|
|
#define MCA_PML_CSUM_ADD_FIN_TO_PENDING(P, D, B, O, S) \
|
|
do { \
|
|
mca_pml_csum_pckt_pending_t *_pckt; \
|
|
int _rc; \
|
|
\
|
|
MCA_PML_CSUM_PCKT_PENDING_ALLOC(_pckt,_rc); \
|
|
_pckt->hdr.hdr_common.hdr_type = MCA_PML_CSUM_HDR_TYPE_FIN; \
|
|
_pckt->hdr.hdr_fin.hdr_des.pval = (D); \
|
|
_pckt->hdr.hdr_fin.hdr_fail = (S); \
|
|
_pckt->proc = (P); \
|
|
_pckt->bml_btl = (B); \
|
|
_pckt->order = (O); \
|
|
OPAL_THREAD_LOCK(&mca_pml_csum.lock); \
|
|
opal_list_append(&mca_pml_csum.pckt_pending, \
|
|
(opal_list_item_t*)_pckt); \
|
|
OPAL_THREAD_UNLOCK(&mca_pml_csum.lock); \
|
|
} while(0)
|
|
|
|
|
|
int mca_pml_csum_send_fin(ompi_proc_t* proc, mca_bml_base_btl_t* bml_btl,
|
|
void *hdr_des, uint8_t order, uint32_t status);
|
|
|
|
/* This function tries to resend FIN/ACK packets from pckt_pending queue.
|
|
* Packets are added to the queue when sending of FIN or ACK is failed due to
|
|
* resource unavailability. bml_btl passed to the function doesn't represents
|
|
* packet's destination, it represents BTL on which resource was freed, so only
|
|
* this BTL should be considered for resending packets */
|
|
void mca_pml_csum_process_pending_packets(mca_bml_base_btl_t* bml_btl);
|
|
|
|
/* This function retries failed PUT/GET operations on frag. When RDMA operation
|
|
* cannot be accomplished for some reason, frag is put on the rdma_pending list.
|
|
* Later the operation is retried. The destination of RDMA operation is stored
|
|
* inside the frag structure */
|
|
void mca_pml_csum_process_pending_rdma(void);
|
|
|
|
#define MCA_PML_CSUM_PROGRESS_PENDING(bml_btl) \
|
|
do { \
|
|
if(opal_list_get_size(&mca_pml_csum.pckt_pending)) \
|
|
mca_pml_csum_process_pending_packets(bml_btl); \
|
|
if(opal_list_get_size(&mca_pml_csum.recv_pending)) \
|
|
mca_pml_csum_recv_request_process_pending(); \
|
|
if(opal_list_get_size(&mca_pml_csum.send_pending)) \
|
|
mca_pml_csum_send_request_process_pending(bml_btl); \
|
|
if(opal_list_get_size(&mca_pml_csum.rdma_pending)) \
|
|
mca_pml_csum_process_pending_rdma(); \
|
|
} while (0)
|
|
|
|
/*
|
|
* Compute the total number of bytes on supplied descriptor
|
|
*/
|
|
#define MCA_PML_CSUM_COMPUTE_SEGMENT_LENGTH(segments, count, hdrlen, length) \
|
|
do { \
|
|
size_t i; \
|
|
\
|
|
for( i = 0; i < count; i++ ) { \
|
|
length += segments[i].seg_len; \
|
|
} \
|
|
length -= hdrlen; \
|
|
} while(0)
|
|
|
|
/* represent BTL chosen for sending request */
|
|
struct mca_pml_csum_com_btl_t {
|
|
mca_bml_base_btl_t *bml_btl;
|
|
struct mca_mpool_base_registration_t* btl_reg;
|
|
size_t length;
|
|
};
|
|
typedef struct mca_pml_csum_com_btl_t mca_pml_csum_com_btl_t;
|
|
|
|
int mca_pml_csum_com_btl_comp(const void *v1, const void *v2);
|
|
|
|
/* Calculate what percentage of a message to send through each BTL according to
|
|
* relative weight */
|
|
static inline void
|
|
mca_pml_csum_calc_weighted_length( mca_pml_csum_com_btl_t *btls, int num_btls, size_t size,
|
|
double weight_total )
|
|
{
|
|
int i;
|
|
size_t length_left;
|
|
|
|
/* shortcut for common case for only one BTL */
|
|
if( OPAL_LIKELY(1 == num_btls) ) {
|
|
btls[0].length = size;
|
|
return;
|
|
}
|
|
|
|
/* sort BTLs according of their weights so BTLs with smaller weight will
|
|
* not hijack all of the traffic */
|
|
qsort( btls, num_btls, sizeof(mca_pml_csum_com_btl_t),
|
|
mca_pml_csum_com_btl_comp );
|
|
|
|
for(length_left = size, i = 0; i < num_btls; i++) {
|
|
mca_bml_base_btl_t* bml_btl = btls[i].bml_btl;
|
|
size_t length = 0;
|
|
if( OPAL_UNLIKELY(0 != length_left) ) {
|
|
length = (length_left > bml_btl->btl->btl_eager_limit)?
|
|
((size_t)(size * (bml_btl->btl_weight / weight_total))) :
|
|
length_left;
|
|
|
|
if(length > length_left)
|
|
length = length_left;
|
|
length_left -= length;
|
|
}
|
|
btls[i].length = length;
|
|
}
|
|
|
|
/* account for rounding errors */
|
|
btls[0].length += length_left;
|
|
}
|
|
|
|
#endif
|