2007-02-28 04:11:01 +03:00
/*
* Copyright ( c ) 2004 - 2005 The Trustees of Indiana University and Indiana
* University Research and Technology
* Corporation . All rights reserved .
2009-08-15 01:06:23 +04:00
* Copyright ( c ) 2004 - 2009 The University of Tennessee and The University
2007-02-28 04:11:01 +03:00
* of Tennessee Research Foundation . All rights
* reserved .
* Copyright ( c ) 2004 - 2005 High Performance Computing Center Stuttgart ,
* University of Stuttgart . All rights reserved .
* Copyright ( c ) 2004 - 2005 The Regents of the University of California .
* All rights reserved .
* $ COPYRIGHT $
*
* Additional copyrights may follow
*
* $ HEADER $
*/
# include "ompi_config.h"
# include "mpi.h"
# include "ompi/constants.h"
- Split the datatype engine into two parts: an MPI specific part in
OMPI
and a language agnostic part in OPAL. The convertor is completely
moved into OPAL. This offers several benefits as described in RFC
http://www.open-mpi.org/community/lists/devel/2009/07/6387.php
namely:
- Fewer basic types (int* and float* types, boolean and wchar
- Fixing naming scheme to ompi-nomenclature.
- Usability outside of the ompi-layer.
- Due to the fixed nature of simple opal types, their information is
completely
known at compile time and therefore constified
- With fewer datatypes (22), the actual sizes of bit-field types may be
reduced
from 64 to 32 bits, allowing reorganizing the opal_datatype
structure, eliminating holes and keeping data required in convertor
(upon send/recv) in one cacheline...
This has implications to the convertor-datastructure and other parts
of the code.
- Several performance tests have been run, the netpipe latency does not
change with
this patch on Linux/x86-64 on the smoky cluster.
- Extensive tests have been done to verify correctness (no new
regressions) using:
1. mpi_test_suite on linux/x86-64 using clean ompi-trunk and
ompi-ddt:
a. running both trunk and ompi-ddt resulted in no differences
(except for MPI_SHORT_INT and MPI_TYPE_MIX_LB_UB do now run
correctly).
b. with --enable-memchecker and running under valgrind (one buglet
when run with static found in test-suite, commited)
2. ibm testsuite on linux/x86-64 using clean ompi-trunk and ompi-ddt:
all passed (except for the dynamic/ tests failed!! as trunk/MTT)
3. compilation and usage of HDF5 tests on Jaguar using PGI and
PathScale compilers.
4. compilation and usage on Scicortex.
- Please note, that for the heterogeneous case, (-m32 compiled
binaries/ompi), neither
ompi-trunk, nor ompi-ddt branch would successfully launch.
This commit was SVN r21641.
2009-07-13 08:56:31 +04:00
# include "ompi/datatype/ompi_datatype.h"
2007-02-28 04:11:01 +03:00
# include "ompi/communicator/communicator.h"
# include "ompi/mca/coll/coll.h"
# include "ompi/mca/coll/base/coll_tags.h"
# include "ompi/mca/pml/pml.h"
# include "coll_tuned.h"
# include "coll_tuned_topo.h"
# include "coll_tuned_util.h"
2007-03-02 20:02:29 +03:00
/* Todo: gather_intra_generic, gather_intra_binary, gather_intra_chain,
* gather_intra_pipeline , segmentation ? */
2007-02-28 04:11:01 +03:00
int
ompi_coll_tuned_gather_intra_binomial ( void * sbuf , int scount ,
struct ompi_datatype_t * sdtype ,
void * rbuf , int rcount ,
struct ompi_datatype_t * rdtype ,
2007-08-19 07:37:49 +04:00
int root ,
struct ompi_communicator_t * comm ,
2008-07-29 02:40:57 +04:00
mca_coll_base_module_t * module )
2007-02-28 04:11:01 +03:00
{
int line = - 1 ;
int i ;
int rank ;
int vrank ;
int size ;
int total_recv = 0 ;
char * ptmp = NULL ;
char * tempbuf = NULL ;
int err ;
ompi_coll_tree_t * bmtree ;
MPI_Status status ;
MPI_Aint sextent , slb , strue_lb , strue_extent ;
MPI_Aint rextent , rlb , rtrue_lb , rtrue_extent ;
2007-08-19 07:37:49 +04:00
mca_coll_tuned_module_t * tuned_module = ( mca_coll_tuned_module_t * ) module ;
mca_coll_tuned_comm_t * data = tuned_module - > tuned_data ;
2007-02-28 04:11:01 +03:00
size = ompi_comm_size ( comm ) ;
rank = ompi_comm_rank ( comm ) ;
2008-06-09 18:53:58 +04:00
OPAL_OUTPUT ( ( ompi_coll_tuned_stream ,
2007-08-19 07:37:49 +04:00
" ompi_coll_tuned_gather_intra_binomial rank %d " , rank ) ) ;
2007-02-28 04:11:01 +03:00
/* create the binomial tree */
2007-08-19 07:37:49 +04:00
COLL_TUNED_UPDATE_IN_ORDER_BMTREE ( comm , tuned_module , root ) ;
bmtree = data - > cached_in_order_bmtree ;
2007-02-28 04:11:01 +03:00
- Split the datatype engine into two parts: an MPI specific part in
OMPI
and a language agnostic part in OPAL. The convertor is completely
moved into OPAL. This offers several benefits as described in RFC
http://www.open-mpi.org/community/lists/devel/2009/07/6387.php
namely:
- Fewer basic types (int* and float* types, boolean and wchar
- Fixing naming scheme to ompi-nomenclature.
- Usability outside of the ompi-layer.
- Due to the fixed nature of simple opal types, their information is
completely
known at compile time and therefore constified
- With fewer datatypes (22), the actual sizes of bit-field types may be
reduced
from 64 to 32 bits, allowing reorganizing the opal_datatype
structure, eliminating holes and keeping data required in convertor
(upon send/recv) in one cacheline...
This has implications to the convertor-datastructure and other parts
of the code.
- Several performance tests have been run, the netpipe latency does not
change with
this patch on Linux/x86-64 on the smoky cluster.
- Extensive tests have been done to verify correctness (no new
regressions) using:
1. mpi_test_suite on linux/x86-64 using clean ompi-trunk and
ompi-ddt:
a. running both trunk and ompi-ddt resulted in no differences
(except for MPI_SHORT_INT and MPI_TYPE_MIX_LB_UB do now run
correctly).
b. with --enable-memchecker and running under valgrind (one buglet
when run with static found in test-suite, commited)
2. ibm testsuite on linux/x86-64 using clean ompi-trunk and ompi-ddt:
all passed (except for the dynamic/ tests failed!! as trunk/MTT)
3. compilation and usage of HDF5 tests on Jaguar using PGI and
PathScale compilers.
4. compilation and usage on Scicortex.
- Please note, that for the heterogeneous case, (-m32 compiled
binaries/ompi), neither
ompi-trunk, nor ompi-ddt branch would successfully launch.
This commit was SVN r21641.
2009-07-13 08:56:31 +04:00
ompi_datatype_get_extent ( sdtype , & slb , & sextent ) ;
ompi_datatype_get_true_extent ( sdtype , & strue_lb , & strue_extent ) ;
ompi_datatype_get_extent ( rdtype , & rlb , & rextent ) ;
ompi_datatype_get_true_extent ( rdtype , & rtrue_lb , & rtrue_extent ) ;
2007-02-28 04:11:01 +03:00
vrank = ( rank - root + size ) % size ;
if ( rank = = root ) {
if ( 0 = = root ) {
/* root on 0, just use the recv buffer */
2007-09-12 19:29:58 +04:00
ptmp = ( char * ) rbuf ;
2007-02-28 04:11:01 +03:00
if ( sbuf ! = MPI_IN_PLACE ) {
- Split the datatype engine into two parts: an MPI specific part in
OMPI
and a language agnostic part in OPAL. The convertor is completely
moved into OPAL. This offers several benefits as described in RFC
http://www.open-mpi.org/community/lists/devel/2009/07/6387.php
namely:
- Fewer basic types (int* and float* types, boolean and wchar
- Fixing naming scheme to ompi-nomenclature.
- Usability outside of the ompi-layer.
- Due to the fixed nature of simple opal types, their information is
completely
known at compile time and therefore constified
- With fewer datatypes (22), the actual sizes of bit-field types may be
reduced
from 64 to 32 bits, allowing reorganizing the opal_datatype
structure, eliminating holes and keeping data required in convertor
(upon send/recv) in one cacheline...
This has implications to the convertor-datastructure and other parts
of the code.
- Several performance tests have been run, the netpipe latency does not
change with
this patch on Linux/x86-64 on the smoky cluster.
- Extensive tests have been done to verify correctness (no new
regressions) using:
1. mpi_test_suite on linux/x86-64 using clean ompi-trunk and
ompi-ddt:
a. running both trunk and ompi-ddt resulted in no differences
(except for MPI_SHORT_INT and MPI_TYPE_MIX_LB_UB do now run
correctly).
b. with --enable-memchecker and running under valgrind (one buglet
when run with static found in test-suite, commited)
2. ibm testsuite on linux/x86-64 using clean ompi-trunk and ompi-ddt:
all passed (except for the dynamic/ tests failed!! as trunk/MTT)
3. compilation and usage of HDF5 tests on Jaguar using PGI and
PathScale compilers.
4. compilation and usage on Scicortex.
- Please note, that for the heterogeneous case, (-m32 compiled
binaries/ompi), neither
ompi-trunk, nor ompi-ddt branch would successfully launch.
This commit was SVN r21641.
2009-07-13 08:56:31 +04:00
err = ompi_datatype_sndrcv ( sbuf , scount , sdtype ,
2007-02-28 04:11:01 +03:00
ptmp , rcount , rdtype ) ;
if ( MPI_SUCCESS ! = err ) { line = __LINE__ ; goto err_hndl ; }
}
} else {
/* root is not on 0, allocate temp buffer for recv,
* rotate data at the end */
2007-09-12 19:29:58 +04:00
tempbuf = ( char * ) malloc ( rtrue_extent + ( rcount * size - 1 ) * rextent ) ;
2007-02-28 04:11:01 +03:00
if ( NULL = = tempbuf ) {
err = OMPI_ERR_OUT_OF_RESOURCE ; line = __LINE__ ; goto err_hndl ;
}
ptmp = tempbuf - rlb ;
if ( sbuf ! = MPI_IN_PLACE ) {
/* copy from sbuf to temp buffer */
- Split the datatype engine into two parts: an MPI specific part in
OMPI
and a language agnostic part in OPAL. The convertor is completely
moved into OPAL. This offers several benefits as described in RFC
http://www.open-mpi.org/community/lists/devel/2009/07/6387.php
namely:
- Fewer basic types (int* and float* types, boolean and wchar
- Fixing naming scheme to ompi-nomenclature.
- Usability outside of the ompi-layer.
- Due to the fixed nature of simple opal types, their information is
completely
known at compile time and therefore constified
- With fewer datatypes (22), the actual sizes of bit-field types may be
reduced
from 64 to 32 bits, allowing reorganizing the opal_datatype
structure, eliminating holes and keeping data required in convertor
(upon send/recv) in one cacheline...
This has implications to the convertor-datastructure and other parts
of the code.
- Several performance tests have been run, the netpipe latency does not
change with
this patch on Linux/x86-64 on the smoky cluster.
- Extensive tests have been done to verify correctness (no new
regressions) using:
1. mpi_test_suite on linux/x86-64 using clean ompi-trunk and
ompi-ddt:
a. running both trunk and ompi-ddt resulted in no differences
(except for MPI_SHORT_INT and MPI_TYPE_MIX_LB_UB do now run
correctly).
b. with --enable-memchecker and running under valgrind (one buglet
when run with static found in test-suite, commited)
2. ibm testsuite on linux/x86-64 using clean ompi-trunk and ompi-ddt:
all passed (except for the dynamic/ tests failed!! as trunk/MTT)
3. compilation and usage of HDF5 tests on Jaguar using PGI and
PathScale compilers.
4. compilation and usage on Scicortex.
- Please note, that for the heterogeneous case, (-m32 compiled
binaries/ompi), neither
ompi-trunk, nor ompi-ddt branch would successfully launch.
This commit was SVN r21641.
2009-07-13 08:56:31 +04:00
err = ompi_datatype_sndrcv ( sbuf , scount , sdtype ,
2007-02-28 04:11:01 +03:00
ptmp , rcount , rdtype ) ;
if ( MPI_SUCCESS ! = err ) { line = __LINE__ ; goto err_hndl ; }
} else {
/* copy from rbuf to temp buffer */
- Split the datatype engine into two parts: an MPI specific part in
OMPI
and a language agnostic part in OPAL. The convertor is completely
moved into OPAL. This offers several benefits as described in RFC
http://www.open-mpi.org/community/lists/devel/2009/07/6387.php
namely:
- Fewer basic types (int* and float* types, boolean and wchar
- Fixing naming scheme to ompi-nomenclature.
- Usability outside of the ompi-layer.
- Due to the fixed nature of simple opal types, their information is
completely
known at compile time and therefore constified
- With fewer datatypes (22), the actual sizes of bit-field types may be
reduced
from 64 to 32 bits, allowing reorganizing the opal_datatype
structure, eliminating holes and keeping data required in convertor
(upon send/recv) in one cacheline...
This has implications to the convertor-datastructure and other parts
of the code.
- Several performance tests have been run, the netpipe latency does not
change with
this patch on Linux/x86-64 on the smoky cluster.
- Extensive tests have been done to verify correctness (no new
regressions) using:
1. mpi_test_suite on linux/x86-64 using clean ompi-trunk and
ompi-ddt:
a. running both trunk and ompi-ddt resulted in no differences
(except for MPI_SHORT_INT and MPI_TYPE_MIX_LB_UB do now run
correctly).
b. with --enable-memchecker and running under valgrind (one buglet
when run with static found in test-suite, commited)
2. ibm testsuite on linux/x86-64 using clean ompi-trunk and ompi-ddt:
all passed (except for the dynamic/ tests failed!! as trunk/MTT)
3. compilation and usage of HDF5 tests on Jaguar using PGI and
PathScale compilers.
4. compilation and usage on Scicortex.
- Please note, that for the heterogeneous case, (-m32 compiled
binaries/ompi), neither
ompi-trunk, nor ompi-ddt branch would successfully launch.
This commit was SVN r21641.
2009-07-13 08:56:31 +04:00
err = ompi_datatype_copy_content_same_ddt ( rdtype , rcount , ptmp ,
2007-02-28 04:11:01 +03:00
( char * ) rbuf + rank * rextent * rcount ) ;
if ( MPI_SUCCESS ! = err ) { line = __LINE__ ; goto err_hndl ; }
}
}
total_recv = rcount ;
} else if ( ! ( vrank % 2 ) ) {
/* other non-leaf nodes, allocate temp buffer for data received from
* children , the most we need is half of the total data elements due
* to the property of binimoal tree */
2007-09-12 19:29:58 +04:00
tempbuf = ( char * ) malloc ( strue_extent + ( scount * size - 1 ) * sextent ) ;
2007-02-28 04:11:01 +03:00
if ( NULL = = tempbuf ) {
err = OMPI_ERR_OUT_OF_RESOURCE ; line = __LINE__ ; goto err_hndl ;
}
ptmp = tempbuf - slb ;
/* local copy to tempbuf */
- Split the datatype engine into two parts: an MPI specific part in
OMPI
and a language agnostic part in OPAL. The convertor is completely
moved into OPAL. This offers several benefits as described in RFC
http://www.open-mpi.org/community/lists/devel/2009/07/6387.php
namely:
- Fewer basic types (int* and float* types, boolean and wchar
- Fixing naming scheme to ompi-nomenclature.
- Usability outside of the ompi-layer.
- Due to the fixed nature of simple opal types, their information is
completely
known at compile time and therefore constified
- With fewer datatypes (22), the actual sizes of bit-field types may be
reduced
from 64 to 32 bits, allowing reorganizing the opal_datatype
structure, eliminating holes and keeping data required in convertor
(upon send/recv) in one cacheline...
This has implications to the convertor-datastructure and other parts
of the code.
- Several performance tests have been run, the netpipe latency does not
change with
this patch on Linux/x86-64 on the smoky cluster.
- Extensive tests have been done to verify correctness (no new
regressions) using:
1. mpi_test_suite on linux/x86-64 using clean ompi-trunk and
ompi-ddt:
a. running both trunk and ompi-ddt resulted in no differences
(except for MPI_SHORT_INT and MPI_TYPE_MIX_LB_UB do now run
correctly).
b. with --enable-memchecker and running under valgrind (one buglet
when run with static found in test-suite, commited)
2. ibm testsuite on linux/x86-64 using clean ompi-trunk and ompi-ddt:
all passed (except for the dynamic/ tests failed!! as trunk/MTT)
3. compilation and usage of HDF5 tests on Jaguar using PGI and
PathScale compilers.
4. compilation and usage on Scicortex.
- Please note, that for the heterogeneous case, (-m32 compiled
binaries/ompi), neither
ompi-trunk, nor ompi-ddt branch would successfully launch.
This commit was SVN r21641.
2009-07-13 08:56:31 +04:00
err = ompi_datatype_sndrcv ( sbuf , scount , sdtype ,
2007-02-28 04:11:01 +03:00
ptmp , scount , sdtype ) ;
if ( MPI_SUCCESS ! = err ) { line = __LINE__ ; goto err_hndl ; }
/* use sdtype,scount as rdtype,rdcount since they are ignored on
* non - root procs */
rdtype = sdtype ;
rcount = scount ;
rextent = sextent ;
total_recv = rcount ;
} else {
/* leaf nodes, no temp buffer needed, use sdtype,scount as
* rdtype , rdcount since they are ignored on non - root procs */
2007-09-12 19:29:58 +04:00
ptmp = ( char * ) sbuf ;
2007-02-28 04:11:01 +03:00
total_recv = scount ;
}
if ( ! ( vrank % 2 ) ) {
/* all non-leaf nodes recv from children */
for ( i = 0 ; i < bmtree - > tree_nextsize ; i + + ) {
2007-03-03 01:57:50 +03:00
int mycount = 0 , vkid ;
/* figure out how much data I have to send to this child */
vkid = ( bmtree - > tree_next [ i ] - root + size ) % size ;
mycount = vkid - vrank ;
if ( mycount > ( size - vkid ) )
mycount = size - vkid ;
mycount * = rcount ;
2007-02-28 04:11:01 +03:00
2008-06-09 18:53:58 +04:00
OPAL_OUTPUT ( ( ompi_coll_tuned_stream ,
2007-03-03 01:57:50 +03:00
" ompi_coll_tuned_gather_intra_binomial rank %d recv %d mycount = %d " ,
rank , bmtree - > tree_next [ i ] , mycount ) ) ;
2007-02-28 04:11:01 +03:00
err = MCA_PML_CALL ( recv ( ptmp + total_recv * rextent , rcount * size - total_recv , rdtype ,
bmtree - > tree_next [ i ] , MCA_COLL_BASE_TAG_GATHER ,
comm , & status ) ) ;
if ( MPI_SUCCESS ! = err ) { line = __LINE__ ; goto err_hndl ; }
2007-03-03 01:57:50 +03:00
total_recv + = mycount ;
2007-02-28 04:11:01 +03:00
}
}
if ( rank ! = root ) {
/* all nodes except root send to parents */
2008-06-09 18:53:58 +04:00
OPAL_OUTPUT ( ( ompi_coll_tuned_stream ,
2007-02-28 04:11:01 +03:00
" ompi_coll_tuned_gather_intra_binomial rank %d send %d count %d \n " ,
rank , bmtree - > tree_prev , total_recv ) ) ;
err = MCA_PML_CALL ( send ( ptmp , total_recv , sdtype ,
bmtree - > tree_prev ,
MCA_COLL_BASE_TAG_GATHER ,
MCA_PML_BASE_SEND_STANDARD , comm ) ) ;
if ( MPI_SUCCESS ! = err ) { line = __LINE__ ; goto err_hndl ; }
}
if ( rank = = root ) {
if ( root ! = 0 ) {
/* rotate received data on root if root != 0 */
- Split the datatype engine into two parts: an MPI specific part in
OMPI
and a language agnostic part in OPAL. The convertor is completely
moved into OPAL. This offers several benefits as described in RFC
http://www.open-mpi.org/community/lists/devel/2009/07/6387.php
namely:
- Fewer basic types (int* and float* types, boolean and wchar
- Fixing naming scheme to ompi-nomenclature.
- Usability outside of the ompi-layer.
- Due to the fixed nature of simple opal types, their information is
completely
known at compile time and therefore constified
- With fewer datatypes (22), the actual sizes of bit-field types may be
reduced
from 64 to 32 bits, allowing reorganizing the opal_datatype
structure, eliminating holes and keeping data required in convertor
(upon send/recv) in one cacheline...
This has implications to the convertor-datastructure and other parts
of the code.
- Several performance tests have been run, the netpipe latency does not
change with
this patch on Linux/x86-64 on the smoky cluster.
- Extensive tests have been done to verify correctness (no new
regressions) using:
1. mpi_test_suite on linux/x86-64 using clean ompi-trunk and
ompi-ddt:
a. running both trunk and ompi-ddt resulted in no differences
(except for MPI_SHORT_INT and MPI_TYPE_MIX_LB_UB do now run
correctly).
b. with --enable-memchecker and running under valgrind (one buglet
when run with static found in test-suite, commited)
2. ibm testsuite on linux/x86-64 using clean ompi-trunk and ompi-ddt:
all passed (except for the dynamic/ tests failed!! as trunk/MTT)
3. compilation and usage of HDF5 tests on Jaguar using PGI and
PathScale compilers.
4. compilation and usage on Scicortex.
- Please note, that for the heterogeneous case, (-m32 compiled
binaries/ompi), neither
ompi-trunk, nor ompi-ddt branch would successfully launch.
This commit was SVN r21641.
2009-07-13 08:56:31 +04:00
err = ompi_datatype_copy_content_same_ddt ( rdtype , rcount * ( size - root ) ,
2007-02-28 04:11:01 +03:00
( char * ) rbuf + rextent * root * rcount , ptmp ) ;
if ( MPI_SUCCESS ! = err ) { line = __LINE__ ; goto err_hndl ; }
- Split the datatype engine into two parts: an MPI specific part in
OMPI
and a language agnostic part in OPAL. The convertor is completely
moved into OPAL. This offers several benefits as described in RFC
http://www.open-mpi.org/community/lists/devel/2009/07/6387.php
namely:
- Fewer basic types (int* and float* types, boolean and wchar
- Fixing naming scheme to ompi-nomenclature.
- Usability outside of the ompi-layer.
- Due to the fixed nature of simple opal types, their information is
completely
known at compile time and therefore constified
- With fewer datatypes (22), the actual sizes of bit-field types may be
reduced
from 64 to 32 bits, allowing reorganizing the opal_datatype
structure, eliminating holes and keeping data required in convertor
(upon send/recv) in one cacheline...
This has implications to the convertor-datastructure and other parts
of the code.
- Several performance tests have been run, the netpipe latency does not
change with
this patch on Linux/x86-64 on the smoky cluster.
- Extensive tests have been done to verify correctness (no new
regressions) using:
1. mpi_test_suite on linux/x86-64 using clean ompi-trunk and
ompi-ddt:
a. running both trunk and ompi-ddt resulted in no differences
(except for MPI_SHORT_INT and MPI_TYPE_MIX_LB_UB do now run
correctly).
b. with --enable-memchecker and running under valgrind (one buglet
when run with static found in test-suite, commited)
2. ibm testsuite on linux/x86-64 using clean ompi-trunk and ompi-ddt:
all passed (except for the dynamic/ tests failed!! as trunk/MTT)
3. compilation and usage of HDF5 tests on Jaguar using PGI and
PathScale compilers.
4. compilation and usage on Scicortex.
- Please note, that for the heterogeneous case, (-m32 compiled
binaries/ompi), neither
ompi-trunk, nor ompi-ddt branch would successfully launch.
This commit was SVN r21641.
2009-07-13 08:56:31 +04:00
err = ompi_datatype_copy_content_same_ddt ( rdtype , rcount * root ,
2007-09-12 19:29:58 +04:00
( char * ) rbuf , ptmp + rextent * rcount * ( size - root ) ) ;
2007-02-28 04:11:01 +03:00
if ( MPI_SUCCESS ! = err ) { line = __LINE__ ; goto err_hndl ; }
free ( tempbuf ) ;
}
} else if ( ! ( vrank % 2 ) ) {
/* other non-leaf nodes */
free ( tempbuf ) ;
}
return MPI_SUCCESS ;
err_hndl :
if ( NULL ! = tempbuf )
free ( tempbuf ) ;
2008-06-09 18:53:58 +04:00
OPAL_OUTPUT ( ( ompi_coll_tuned_stream , " %s:%4d \t Error occurred %d, rank %2d " ,
2007-02-28 04:11:01 +03:00
__FILE__ , line , err , rank ) ) ;
2007-08-19 07:37:49 +04:00
return err ;
2007-02-28 04:11:01 +03:00
}
2007-06-22 00:00:36 +04:00
/*
* gather_intra_linear_sync
*
* Function : - synchronized gather operation with
* Accepts : - same arguments as MPI_Gather ( ) , first segment size
* Returns : - MPI_SUCCESS or error code
*/
int
ompi_coll_tuned_gather_intra_linear_sync ( void * sbuf , int scount ,
struct ompi_datatype_t * sdtype ,
void * rbuf , int rcount ,
struct ompi_datatype_t * rdtype ,
int root ,
struct ompi_communicator_t * comm ,
2008-07-29 02:40:57 +04:00
mca_coll_base_module_t * module ,
2007-06-22 00:00:36 +04:00
int first_segment_size )
{
int i ;
int ret , line ;
int rank , size ;
int first_segment_count ;
size_t typelng ;
MPI_Aint extent ;
MPI_Aint lb ;
size = ompi_comm_size ( comm ) ;
rank = ompi_comm_rank ( comm ) ;
2008-06-09 18:53:58 +04:00
OPAL_OUTPUT ( ( ompi_coll_tuned_stream ,
2007-08-19 07:37:49 +04:00
" ompi_coll_tuned_gather_intra_linear_sync rank %d, segment %d " , rank , first_segment_size ) ) ;
2007-06-22 00:00:36 +04:00
if ( rank ! = root ) {
/* Non-root processes:
- receive zero byte message from the root ,
- send the first segment of the data synchronously ,
- send the second segment of the data .
*/
- Split the datatype engine into two parts: an MPI specific part in
OMPI
and a language agnostic part in OPAL. The convertor is completely
moved into OPAL. This offers several benefits as described in RFC
http://www.open-mpi.org/community/lists/devel/2009/07/6387.php
namely:
- Fewer basic types (int* and float* types, boolean and wchar
- Fixing naming scheme to ompi-nomenclature.
- Usability outside of the ompi-layer.
- Due to the fixed nature of simple opal types, their information is
completely
known at compile time and therefore constified
- With fewer datatypes (22), the actual sizes of bit-field types may be
reduced
from 64 to 32 bits, allowing reorganizing the opal_datatype
structure, eliminating holes and keeping data required in convertor
(upon send/recv) in one cacheline...
This has implications to the convertor-datastructure and other parts
of the code.
- Several performance tests have been run, the netpipe latency does not
change with
this patch on Linux/x86-64 on the smoky cluster.
- Extensive tests have been done to verify correctness (no new
regressions) using:
1. mpi_test_suite on linux/x86-64 using clean ompi-trunk and
ompi-ddt:
a. running both trunk and ompi-ddt resulted in no differences
(except for MPI_SHORT_INT and MPI_TYPE_MIX_LB_UB do now run
correctly).
b. with --enable-memchecker and running under valgrind (one buglet
when run with static found in test-suite, commited)
2. ibm testsuite on linux/x86-64 using clean ompi-trunk and ompi-ddt:
all passed (except for the dynamic/ tests failed!! as trunk/MTT)
3. compilation and usage of HDF5 tests on Jaguar using PGI and
PathScale compilers.
4. compilation and usage on Scicortex.
- Please note, that for the heterogeneous case, (-m32 compiled
binaries/ompi), neither
ompi-trunk, nor ompi-ddt branch would successfully launch.
This commit was SVN r21641.
2009-07-13 08:56:31 +04:00
ompi_datatype_type_size ( sdtype , & typelng ) ;
ompi_datatype_get_extent ( sdtype , & lb , & extent ) ;
2007-06-22 00:00:36 +04:00
first_segment_count = scount ;
COLL_TUNED_COMPUTED_SEGCOUNT ( ( size_t ) first_segment_size , typelng ,
first_segment_count ) ;
ret = MCA_PML_CALL ( recv ( sbuf , 0 , MPI_BYTE , root ,
MCA_COLL_BASE_TAG_GATHER ,
comm , MPI_STATUS_IGNORE ) ) ;
if ( ret ! = MPI_SUCCESS ) { line = __LINE__ ; goto error_hndl ; }
ret = MCA_PML_CALL ( send ( sbuf , first_segment_count , sdtype , root ,
MCA_COLL_BASE_TAG_GATHER ,
MCA_PML_BASE_SEND_STANDARD , comm ) ) ;
if ( ret ! = MPI_SUCCESS ) { line = __LINE__ ; goto error_hndl ; }
ret = MCA_PML_CALL ( send ( ( char * ) sbuf + extent * first_segment_count ,
( scount - first_segment_count ) , sdtype ,
root , MCA_COLL_BASE_TAG_GATHER ,
MCA_PML_BASE_SEND_STANDARD , comm ) ) ;
if ( ret ! = MPI_SUCCESS ) { line = __LINE__ ; goto error_hndl ; }
}
else {
/* Root process,
- For every non - root node :
2007-08-19 07:37:49 +04:00
- post irecv for the first segment of the message
- send zero byte message to signal node to send the message
- post irecv for the second segment of the message
- wait for the first segment to complete
2007-06-22 00:00:36 +04:00
- Copy local data if necessary
- Waitall for all the second segments to complete .
2007-08-19 07:37:49 +04:00
*/
2007-06-22 00:00:36 +04:00
char * ptmp ;
ompi_request_t * * reqs = NULL , * first_segment_req ;
reqs = ( ompi_request_t * * ) calloc ( size , sizeof ( ompi_request_t * ) ) ;
if ( NULL = = reqs ) { ret = - 1 ; line = __LINE__ ; goto error_hndl ; }
- Split the datatype engine into two parts: an MPI specific part in
OMPI
and a language agnostic part in OPAL. The convertor is completely
moved into OPAL. This offers several benefits as described in RFC
http://www.open-mpi.org/community/lists/devel/2009/07/6387.php
namely:
- Fewer basic types (int* and float* types, boolean and wchar
- Fixing naming scheme to ompi-nomenclature.
- Usability outside of the ompi-layer.
- Due to the fixed nature of simple opal types, their information is
completely
known at compile time and therefore constified
- With fewer datatypes (22), the actual sizes of bit-field types may be
reduced
from 64 to 32 bits, allowing reorganizing the opal_datatype
structure, eliminating holes and keeping data required in convertor
(upon send/recv) in one cacheline...
This has implications to the convertor-datastructure and other parts
of the code.
- Several performance tests have been run, the netpipe latency does not
change with
this patch on Linux/x86-64 on the smoky cluster.
- Extensive tests have been done to verify correctness (no new
regressions) using:
1. mpi_test_suite on linux/x86-64 using clean ompi-trunk and
ompi-ddt:
a. running both trunk and ompi-ddt resulted in no differences
(except for MPI_SHORT_INT and MPI_TYPE_MIX_LB_UB do now run
correctly).
b. with --enable-memchecker and running under valgrind (one buglet
when run with static found in test-suite, commited)
2. ibm testsuite on linux/x86-64 using clean ompi-trunk and ompi-ddt:
all passed (except for the dynamic/ tests failed!! as trunk/MTT)
3. compilation and usage of HDF5 tests on Jaguar using PGI and
PathScale compilers.
4. compilation and usage on Scicortex.
- Please note, that for the heterogeneous case, (-m32 compiled
binaries/ompi), neither
ompi-trunk, nor ompi-ddt branch would successfully launch.
This commit was SVN r21641.
2009-07-13 08:56:31 +04:00
ompi_datatype_type_size ( rdtype , & typelng ) ;
ompi_datatype_get_extent ( rdtype , & lb , & extent ) ;
2007-06-22 00:00:36 +04:00
first_segment_count = rcount ;
COLL_TUNED_COMPUTED_SEGCOUNT ( ( size_t ) first_segment_size , typelng ,
first_segment_count ) ;
ptmp = ( char * ) rbuf ;
2007-07-06 00:40:50 +04:00
for ( i = 0 ; i < size ; + + i ) {
2007-06-22 00:00:36 +04:00
if ( i = = rank ) {
/* skip myself */
reqs [ i ] = MPI_REQUEST_NULL ;
continue ;
}
/* irecv for the first segment from i */
ptmp = ( char * ) rbuf + i * rcount * extent ;
ret = MCA_PML_CALL ( irecv ( ptmp , first_segment_count , rdtype , i ,
MCA_COLL_BASE_TAG_GATHER , comm ,
& first_segment_req ) ) ;
if ( ret ! = MPI_SUCCESS ) { line = __LINE__ ; goto error_hndl ; }
/* send sync message */
ret = MCA_PML_CALL ( send ( rbuf , 0 , MPI_BYTE , i ,
MCA_COLL_BASE_TAG_GATHER ,
MCA_PML_BASE_SEND_STANDARD , comm ) ) ;
if ( ret ! = MPI_SUCCESS ) { line = __LINE__ ; goto error_hndl ; }
/* irecv for the second segment */
ptmp = ( char * ) rbuf + ( i * rcount + first_segment_count ) * extent ;
ret = MCA_PML_CALL ( irecv ( ptmp , ( rcount - first_segment_count ) ,
rdtype , i , MCA_COLL_BASE_TAG_GATHER , comm ,
& reqs [ i ] ) ) ;
if ( ret ! = MPI_SUCCESS ) { line = __LINE__ ; goto error_hndl ; }
/* wait on the first segment to complete */
ret = ompi_request_wait ( & first_segment_req , MPI_STATUS_IGNORE ) ;
if ( ret ! = MPI_SUCCESS ) { line = __LINE__ ; goto error_hndl ; }
}
/* copy local data if necessary */
if ( MPI_IN_PLACE ! = sbuf ) {
- Split the datatype engine into two parts: an MPI specific part in
OMPI
and a language agnostic part in OPAL. The convertor is completely
moved into OPAL. This offers several benefits as described in RFC
http://www.open-mpi.org/community/lists/devel/2009/07/6387.php
namely:
- Fewer basic types (int* and float* types, boolean and wchar
- Fixing naming scheme to ompi-nomenclature.
- Usability outside of the ompi-layer.
- Due to the fixed nature of simple opal types, their information is
completely
known at compile time and therefore constified
- With fewer datatypes (22), the actual sizes of bit-field types may be
reduced
from 64 to 32 bits, allowing reorganizing the opal_datatype
structure, eliminating holes and keeping data required in convertor
(upon send/recv) in one cacheline...
This has implications to the convertor-datastructure and other parts
of the code.
- Several performance tests have been run, the netpipe latency does not
change with
this patch on Linux/x86-64 on the smoky cluster.
- Extensive tests have been done to verify correctness (no new
regressions) using:
1. mpi_test_suite on linux/x86-64 using clean ompi-trunk and
ompi-ddt:
a. running both trunk and ompi-ddt resulted in no differences
(except for MPI_SHORT_INT and MPI_TYPE_MIX_LB_UB do now run
correctly).
b. with --enable-memchecker and running under valgrind (one buglet
when run with static found in test-suite, commited)
2. ibm testsuite on linux/x86-64 using clean ompi-trunk and ompi-ddt:
all passed (except for the dynamic/ tests failed!! as trunk/MTT)
3. compilation and usage of HDF5 tests on Jaguar using PGI and
PathScale compilers.
4. compilation and usage on Scicortex.
- Please note, that for the heterogeneous case, (-m32 compiled
binaries/ompi), neither
ompi-trunk, nor ompi-ddt branch would successfully launch.
This commit was SVN r21641.
2009-07-13 08:56:31 +04:00
ret = ompi_datatype_sndrcv ( sbuf , scount , sdtype ,
2007-06-22 00:00:36 +04:00
( char * ) rbuf + rank * rcount * extent ,
rcount , rdtype ) ;
if ( ret ! = MPI_SUCCESS ) { line = __LINE__ ; goto error_hndl ; }
}
/* wait all second segments to complete */
ret = ompi_request_wait_all ( size , reqs , MPI_STATUSES_IGNORE ) ;
if ( ret ! = MPI_SUCCESS ) { line = __LINE__ ; goto error_hndl ; }
free ( reqs ) ;
}
/* All done */
return MPI_SUCCESS ;
error_hndl :
2008-06-09 18:53:58 +04:00
OPAL_OUTPUT ( ( ompi_coll_tuned_stream ,
2007-06-22 00:00:36 +04:00
" ERROR_HNDL: node %d file %s line %d error %d \n " ,
rank , __FILE__ , line , ret ) ) ;
return ret ;
}
2007-02-28 04:11:01 +03:00
/*
* Linear functions are copied from the BASIC coll module
* they do not segment the message and are simple implementations
* but for some small number of nodes and / or small data sizes they
* are just as fast as tuned / tree based segmenting operations
* and as such may be selected by the decision functions
* These are copied into this module due to the way we select modules
* in V1 . i . e . in V2 we will handle this differently and so will not
* have to duplicate code .
* JPG following the examples from other coll_tuned implementations . Dec06 .
*/
/* copied function (with appropriate renaming) starts here */
/*
* gather_intra
*
* Function : - basic gather operation
* Accepts : - same arguments as MPI_Gather ( )
* Returns : - MPI_SUCCESS or error code
*/
int
ompi_coll_tuned_gather_intra_basic_linear ( void * sbuf , int scount ,
struct ompi_datatype_t * sdtype ,
void * rbuf , int rcount ,
struct ompi_datatype_t * rdtype ,
2007-08-19 07:37:49 +04:00
int root ,
struct ompi_communicator_t * comm ,
2008-07-29 02:40:57 +04:00
mca_coll_base_module_t * module )
2007-02-28 04:11:01 +03:00
{
int i ;
int err ;
int rank ;
int size ;
char * ptmp ;
MPI_Aint incr ;
MPI_Aint extent ;
MPI_Aint lb ;
size = ompi_comm_size ( comm ) ;
rank = ompi_comm_rank ( comm ) ;
/* Everyone but root sends data and returns. */
2008-06-09 18:53:58 +04:00
OPAL_OUTPUT ( ( ompi_coll_tuned_stream ,
2007-08-19 07:37:49 +04:00
" ompi_coll_tuned_gather_intra_basic_linear rank %d " , rank ) ) ;
2007-02-28 04:11:01 +03:00
if ( rank ! = root ) {
return MCA_PML_CALL ( send ( sbuf , scount , sdtype , root ,
MCA_COLL_BASE_TAG_GATHER ,
MCA_PML_BASE_SEND_STANDARD , comm ) ) ;
}
/* I am the root, loop receiving the data. */
- Split the datatype engine into two parts: an MPI specific part in
OMPI
and a language agnostic part in OPAL. The convertor is completely
moved into OPAL. This offers several benefits as described in RFC
http://www.open-mpi.org/community/lists/devel/2009/07/6387.php
namely:
- Fewer basic types (int* and float* types, boolean and wchar
- Fixing naming scheme to ompi-nomenclature.
- Usability outside of the ompi-layer.
- Due to the fixed nature of simple opal types, their information is
completely
known at compile time and therefore constified
- With fewer datatypes (22), the actual sizes of bit-field types may be
reduced
from 64 to 32 bits, allowing reorganizing the opal_datatype
structure, eliminating holes and keeping data required in convertor
(upon send/recv) in one cacheline...
This has implications to the convertor-datastructure and other parts
of the code.
- Several performance tests have been run, the netpipe latency does not
change with
this patch on Linux/x86-64 on the smoky cluster.
- Extensive tests have been done to verify correctness (no new
regressions) using:
1. mpi_test_suite on linux/x86-64 using clean ompi-trunk and
ompi-ddt:
a. running both trunk and ompi-ddt resulted in no differences
(except for MPI_SHORT_INT and MPI_TYPE_MIX_LB_UB do now run
correctly).
b. with --enable-memchecker and running under valgrind (one buglet
when run with static found in test-suite, commited)
2. ibm testsuite on linux/x86-64 using clean ompi-trunk and ompi-ddt:
all passed (except for the dynamic/ tests failed!! as trunk/MTT)
3. compilation and usage of HDF5 tests on Jaguar using PGI and
PathScale compilers.
4. compilation and usage on Scicortex.
- Please note, that for the heterogeneous case, (-m32 compiled
binaries/ompi), neither
ompi-trunk, nor ompi-ddt branch would successfully launch.
This commit was SVN r21641.
2009-07-13 08:56:31 +04:00
ompi_datatype_get_extent ( rdtype , & lb , & extent ) ;
2007-02-28 04:11:01 +03:00
incr = extent * rcount ;
for ( i = 0 , ptmp = ( char * ) rbuf ; i < size ; + + i , ptmp + = incr ) {
if ( i = = rank ) {
if ( MPI_IN_PLACE ! = sbuf ) {
- Split the datatype engine into two parts: an MPI specific part in
OMPI
and a language agnostic part in OPAL. The convertor is completely
moved into OPAL. This offers several benefits as described in RFC
http://www.open-mpi.org/community/lists/devel/2009/07/6387.php
namely:
- Fewer basic types (int* and float* types, boolean and wchar
- Fixing naming scheme to ompi-nomenclature.
- Usability outside of the ompi-layer.
- Due to the fixed nature of simple opal types, their information is
completely
known at compile time and therefore constified
- With fewer datatypes (22), the actual sizes of bit-field types may be
reduced
from 64 to 32 bits, allowing reorganizing the opal_datatype
structure, eliminating holes and keeping data required in convertor
(upon send/recv) in one cacheline...
This has implications to the convertor-datastructure and other parts
of the code.
- Several performance tests have been run, the netpipe latency does not
change with
this patch on Linux/x86-64 on the smoky cluster.
- Extensive tests have been done to verify correctness (no new
regressions) using:
1. mpi_test_suite on linux/x86-64 using clean ompi-trunk and
ompi-ddt:
a. running both trunk and ompi-ddt resulted in no differences
(except for MPI_SHORT_INT and MPI_TYPE_MIX_LB_UB do now run
correctly).
b. with --enable-memchecker and running under valgrind (one buglet
when run with static found in test-suite, commited)
2. ibm testsuite on linux/x86-64 using clean ompi-trunk and ompi-ddt:
all passed (except for the dynamic/ tests failed!! as trunk/MTT)
3. compilation and usage of HDF5 tests on Jaguar using PGI and
PathScale compilers.
4. compilation and usage on Scicortex.
- Please note, that for the heterogeneous case, (-m32 compiled
binaries/ompi), neither
ompi-trunk, nor ompi-ddt branch would successfully launch.
This commit was SVN r21641.
2009-07-13 08:56:31 +04:00
err = ompi_datatype_sndrcv ( sbuf , scount , sdtype ,
2007-02-28 04:11:01 +03:00
ptmp , rcount , rdtype ) ;
} else {
err = MPI_SUCCESS ;
}
} else {
err = MCA_PML_CALL ( recv ( ptmp , rcount , rdtype , i ,
MCA_COLL_BASE_TAG_GATHER ,
comm , MPI_STATUS_IGNORE ) ) ;
}
if ( MPI_SUCCESS ! = err ) {
return err ;
}
}
/* All done */
return MPI_SUCCESS ;
}
/* copied function (with appropriate renaming) ends here */
/* The following are used by dynamic and forced rules */
/* publish details of each algorithm and if its forced/fixed/locked in */
/* as you add methods/algorithms you must update this and the query/map
routines */
/* this routine is called by the component only */
/* this makes sure that the mca parameters are set to their initial values
and perms */
/* module does not call this they call the forced_getvalues routine instead */
int
ompi_coll_tuned_gather_intra_check_forced_init ( coll_tuned_force_algorithm_mca_param_indices_t * mca_param_indices )
{
2007-06-22 00:00:36 +04:00
int rc , max_alg = 3 , requested_alg ;
2007-02-28 04:11:01 +03:00
ompi_coll_tuned_forced_max_algorithms [ GATHER ] = max_alg ;
rc = mca_base_param_reg_int ( & mca_coll_tuned_component . super . collm_version ,
2009-08-15 01:06:23 +04:00
" gather_algorithm_count " ,
" Number of gather algorithms available " ,
false , true , max_alg , NULL ) ;
2007-02-28 04:11:01 +03:00
mca_param_indices - > algorithm_param_index
2009-08-15 01:06:23 +04:00
= mca_base_param_reg_int ( & mca_coll_tuned_component . super . collm_version ,
" gather_algorithm " ,
" Which gather algorithm is used. Can be locked down to choice of: 0 ignore, 1 basic linear, 2 binomial, 3 linear with synchronization. " ,
false , false , 0 , NULL ) ;
2009-01-03 18:56:25 +03:00
if ( mca_param_indices - > algorithm_param_index < 0 ) {
return mca_param_indices - > algorithm_param_index ;
}
2007-02-28 04:11:01 +03:00
mca_base_param_lookup_int ( mca_param_indices - > algorithm_param_index ,
& ( requested_alg ) ) ;
2008-08-11 15:21:04 +04:00
if ( 0 > requested_alg | | requested_alg > max_alg ) {
2009-08-15 01:06:23 +04:00
if ( 0 = = ompi_comm_rank ( MPI_COMM_WORLD ) ) {
opal_output ( 0 , " Gather algorithm #%d is not available (range [0..%d]). Switching back to default(0) \n " ,
requested_alg , max_alg ) ;
}
mca_base_param_set_int ( mca_param_indices - > algorithm_param_index , 0 ) ;
2007-02-28 04:11:01 +03:00
}
mca_param_indices - > segsize_param_index
2009-08-15 01:06:23 +04:00
= mca_base_param_reg_int ( & mca_coll_tuned_component . super . collm_version ,
" gather_algorithm_segmentsize " ,
" Segment size in bytes used by default for gather algorithms. Only has meaning if algorithm is forced and supports segmenting. 0 bytes means no segmentation. Currently, available algorithms do not support segmentation. " ,
false , false , 0 , NULL ) ;
2007-02-28 04:11:01 +03:00
mca_param_indices - > tree_fanout_param_index
2009-08-15 01:06:23 +04:00
= mca_base_param_reg_int ( & mca_coll_tuned_component . super . collm_version ,
" gather_algorithm_tree_fanout " ,
" Fanout for n-tree used for gather algorithms. Only has meaning if algorithm is forced and supports n-tree topo based operation. Currently, available algorithms do not support n-tree topologies. " ,
false , false ,
ompi_coll_tuned_init_tree_fanout , /* get system wide default */
NULL ) ;
2007-02-28 04:11:01 +03:00
mca_param_indices - > chain_fanout_param_index
2009-08-15 01:06:23 +04:00
= mca_base_param_reg_int ( & mca_coll_tuned_component . super . collm_version ,
" gather_algorithm_chain_fanout " ,
" Fanout for chains used for gather algorithms. Only has meaning if algorithm is forced and supports chain topo based operation. Currently, available algorithms do not support chain topologies. " ,
false , false ,
ompi_coll_tuned_init_chain_fanout , /* get system wide default */
NULL ) ;
2007-02-28 04:11:01 +03:00
return ( MPI_SUCCESS ) ;
}
int
ompi_coll_tuned_gather_intra_do_forced ( void * sbuf , int scount ,
struct ompi_datatype_t * sdtype ,
void * rbuf , int rcount ,
struct ompi_datatype_t * rdtype ,
2007-08-19 07:37:49 +04:00
int root ,
struct ompi_communicator_t * comm ,
2008-07-29 02:40:57 +04:00
mca_coll_base_module_t * module )
2007-02-28 04:11:01 +03:00
{
2007-08-19 07:37:49 +04:00
mca_coll_tuned_module_t * tuned_module = ( mca_coll_tuned_module_t * ) module ;
mca_coll_tuned_comm_t * data = tuned_module - > tuned_data ;
2008-06-09 18:53:58 +04:00
OPAL_OUTPUT ( ( ompi_coll_tuned_stream ,
2007-02-28 04:11:01 +03:00
" coll:tuned:gather_intra_do_forced selected algorithm %d " ,
2007-08-19 07:37:49 +04:00
data - > user_forced [ GATHER ] . algorithm ) ) ;
2007-02-28 04:11:01 +03:00
2007-08-19 07:37:49 +04:00
switch ( data - > user_forced [ GATHER ] . algorithm ) {
2007-02-28 04:11:01 +03:00
case ( 0 ) :
return ompi_coll_tuned_gather_intra_dec_fixed ( sbuf , scount , sdtype ,
rbuf , rcount , rdtype ,
2007-08-19 07:37:49 +04:00
root , comm , module ) ;
2007-02-28 04:11:01 +03:00
case ( 1 ) :
return ompi_coll_tuned_gather_intra_basic_linear ( sbuf , scount , sdtype ,
rbuf , rcount , rdtype ,
2007-08-19 07:37:49 +04:00
root , comm , module ) ;
2007-06-22 00:00:36 +04:00
case ( 2 ) :
return ompi_coll_tuned_gather_intra_binomial ( sbuf , scount , sdtype ,
rbuf , rcount , rdtype ,
2007-08-19 07:37:49 +04:00
root , comm , module ) ;
2007-06-22 00:00:36 +04:00
case ( 3 ) :
{
2007-08-19 07:37:49 +04:00
const int first_segment_size = data - > user_forced [ GATHER ] . segsize ;
return ompi_coll_tuned_gather_intra_linear_sync ( sbuf , scount , sdtype ,
rbuf , rcount , rdtype ,
root , comm , module ,
2007-06-22 00:00:36 +04:00
first_segment_size ) ;
}
2007-08-19 07:37:49 +04:00
default :
2008-06-09 18:53:58 +04:00
OPAL_OUTPUT ( ( ompi_coll_tuned_stream ,
2007-08-19 07:37:49 +04:00
" coll:tuned:gather_intra_do_forced attempt to select algorithm %d when only 0-%d is valid? " ,
data - > user_forced [ GATHER ] . algorithm ,
ompi_coll_tuned_forced_max_algorithms [ GATHER ] ) ) ;
return ( MPI_ERR_ARG ) ;
2007-02-28 04:11:01 +03:00
} /* switch */
}
int
ompi_coll_tuned_gather_intra_do_this ( void * sbuf , int scount ,
struct ompi_datatype_t * sdtype ,
void * rbuf , int rcount ,
struct ompi_datatype_t * rdtype ,
2007-08-19 07:37:49 +04:00
int root ,
struct ompi_communicator_t * comm ,
2008-07-29 02:40:57 +04:00
mca_coll_base_module_t * module ,
2007-02-28 04:11:01 +03:00
int algorithm , int faninout , int segsize )
{
2008-06-09 18:53:58 +04:00
OPAL_OUTPUT ( ( ompi_coll_tuned_stream ,
2007-02-28 04:11:01 +03:00
" coll:tuned:gather_intra_do_this selected algorithm %d topo faninout %d segsize %d " ,
algorithm , faninout , segsize ) ) ;
switch ( algorithm ) {
case ( 0 ) :
return ompi_coll_tuned_gather_intra_dec_fixed ( sbuf , scount , sdtype ,
rbuf , rcount , rdtype ,
2007-08-19 07:37:49 +04:00
root , comm , module ) ;
2007-02-28 04:11:01 +03:00
case ( 1 ) :
2007-08-19 07:37:49 +04:00
return ompi_coll_tuned_gather_intra_basic_linear ( sbuf , scount , sdtype ,
rbuf , rcount , rdtype ,
root , comm , module ) ;
2007-02-28 04:11:01 +03:00
case ( 2 ) :
return ompi_coll_tuned_gather_intra_binomial ( sbuf , scount , sdtype ,
rbuf , rcount , rdtype ,
2007-08-19 07:37:49 +04:00
root , comm , module ) ;
2007-06-22 00:00:36 +04:00
case ( 3 ) :
return ompi_coll_tuned_gather_intra_linear_sync ( sbuf , scount , sdtype ,
rbuf , rcount , rdtype ,
2007-08-19 07:37:49 +04:00
root , comm , module ,
segsize ) ;
default :
2008-06-09 18:53:58 +04:00
OPAL_OUTPUT ( ( ompi_coll_tuned_stream ,
2007-08-19 07:37:49 +04:00
" coll:tuned:gather_intra_do_this attempt to select algorithm %d when only 0-%d is valid? " ,
algorithm ,
ompi_coll_tuned_forced_max_algorithms [ GATHER ] ) ) ;
return ( MPI_ERR_ARG ) ;
} /* switch */
2007-02-28 04:11:01 +03:00
}