1
1
openmpi/ompi/debuggers/ompi_common_dll.c

531 строка
24 KiB
C
Исходник Обычный вид История

/*
* Copyright (c) 2007-2008 Cisco Systems, Inc. All rights reserved.
* Copyright (c) 2004-2007 The University of Tennessee and The University
* of Tennessee Research Foundation. All rights
* reserved.
* Copyright (c) 2008-2009 Sun Microsystems, Inc. All rights reserved.
* $COPYRIGHT$
*
* Additional copyrights may follow
*
* $HEADER$
*/
/**********************************************************************
* Copyright (C) 2000-2004 by Etnus, LLC.
* Copyright (C) 1999 by Etnus, Inc.
* Copyright (C) 1997-1998 Dolphin Interconnect Solutions Inc.
*
* Permission is hereby granted to use, reproduce, prepare derivative
* works, and to redistribute to others.
*
* DISCLAIMER
*
* Neither Dolphin Interconnect Solutions, Etnus LLC, nor any of their
* employees, makes any warranty express or implied, or assumes any
* legal liability or responsibility for the accuracy, completeness,
* or usefulness of any information, apparatus, product, or process
* disclosed, or represents that its use would not infringe privately
* owned rights.
*
* This code was written by
* James Cownie: Dolphin Interconnect Solutions. <jcownie@dolphinics.com>
* Etnus LLC <jcownie@etnus.com>
**********************************************************************/
#include "ompi_config.h"
#include "ompi_common_dll_defs.h"
- Split the datatype engine into two parts: an MPI specific part in OMPI and a language agnostic part in OPAL. The convertor is completely moved into OPAL. This offers several benefits as described in RFC http://www.open-mpi.org/community/lists/devel/2009/07/6387.php namely: - Fewer basic types (int* and float* types, boolean and wchar - Fixing naming scheme to ompi-nomenclature. - Usability outside of the ompi-layer. - Due to the fixed nature of simple opal types, their information is completely known at compile time and therefore constified - With fewer datatypes (22), the actual sizes of bit-field types may be reduced from 64 to 32 bits, allowing reorganizing the opal_datatype structure, eliminating holes and keeping data required in convertor (upon send/recv) in one cacheline... This has implications to the convertor-datastructure and other parts of the code. - Several performance tests have been run, the netpipe latency does not change with this patch on Linux/x86-64 on the smoky cluster. - Extensive tests have been done to verify correctness (no new regressions) using: 1. mpi_test_suite on linux/x86-64 using clean ompi-trunk and ompi-ddt: a. running both trunk and ompi-ddt resulted in no differences (except for MPI_SHORT_INT and MPI_TYPE_MIX_LB_UB do now run correctly). b. with --enable-memchecker and running under valgrind (one buglet when run with static found in test-suite, commited) 2. ibm testsuite on linux/x86-64 using clean ompi-trunk and ompi-ddt: all passed (except for the dynamic/ tests failed!! as trunk/MTT) 3. compilation and usage of HDF5 tests on Jaguar using PGI and PathScale compilers. 4. compilation and usage on Scicortex. - Please note, that for the heterogeneous case, (-m32 compiled binaries/ompi), neither ompi-trunk, nor ompi-ddt branch would successfully launch. This commit was SVN r21641.
2009-07-13 08:56:31 +04:00
#include "ompi/datatype/ompi_datatype.h"
/* Basic callbacks into the debugger */
const mqs_basic_callbacks *mqs_basic_entrypoints;
#if defined(WORDS_BIGENDIAN)
static int host_is_big_endian = 1;
#else
static int host_is_big_endian = 0;
#endif
/*
* For sanity checking to try to help keep the code in this DLL in
* sync with the real structs out in the main OMPI code base. If
* we're not compiling this file inside ompi_debugger_sanity.c, then
* ompi_field_offset() won't be defined. So we define it here to be a
* call to the real function mqs_field_offset.
*/
#ifndef ompi_field_offset
#define ompi_field_offset(out_name, qh_type, struct_name, field_name) \
{ \
out_name = mqs_field_offset((qh_type), #field_name); \
if (out_name < 0) { \
fprintf(stderr, "WARNING: Field " #field_name " of type " #struct_name " not found!\n"); \
} \
}
#endif
/*
* Open MPI use a bunch of lists in order to keep track of the
* internal objects. We have to make sure we're able to find all of
* them in the image and compute their ofset in order to be able to
* parse them later. We need to find the opal_list_item_t, the
* opal_list_t, the ompi_free_list_item_t, and the ompi_free_list_t.
*
* Once we have these offsets, we should make sure that we have access
* to all requests lists and types. We're looking here only at the
* basic type for the requests as they hold all the information we
* need to export to the debugger.
*/
int ompi_fill_in_type_info(mqs_image *image, char **message)
{
char* missing_in_action;
mpi_image_info * i_info = (mpi_image_info *)mqs_get_image_info (image);
{
mqs_type* qh_type = mqs_find_type( image, "opal_list_item_t", mqs_lang_c );
if( !qh_type ) {
missing_in_action = "opal_list_item_t";
goto type_missing;
}
i_info->opal_list_item_t.type = qh_type;
i_info->opal_list_item_t.size = mqs_sizeof(qh_type);
ompi_field_offset(i_info->opal_list_item_t.offset.opal_list_next,
qh_type, opal_list_item_t, opal_list_next);
}
{
mqs_type* qh_type = mqs_find_type( image, "opal_list_t", mqs_lang_c );
if( !qh_type ) {
missing_in_action = "opal_list_t";
goto type_missing;
}
i_info->opal_list_t.type = qh_type;
i_info->opal_list_t.size = mqs_sizeof(qh_type);
ompi_field_offset(i_info->opal_list_t.offset.opal_list_sentinel,
qh_type, opal_list_t, opal_list_sentinel);
}
{
mqs_type* qh_type = mqs_find_type( image, "ompi_free_list_item_t", mqs_lang_c );
if( !qh_type ) {
missing_in_action = "ompi_free_list_item_t";
goto type_missing;
}
/* This is just an overloaded opal_list_item_t */
i_info->ompi_free_list_item_t.type = qh_type;
i_info->ompi_free_list_item_t.size = mqs_sizeof(qh_type);
}
{
mqs_type* qh_type = mqs_find_type( image, "ompi_free_list_t", mqs_lang_c );
if( !qh_type ) {
missing_in_action = "ompi_free_list_t";
goto type_missing;
}
i_info->ompi_free_list_t.type = qh_type;
i_info->ompi_free_list_t.size = mqs_sizeof(qh_type);
ompi_field_offset(i_info->ompi_free_list_t.offset.fl_mpool,
qh_type, ompi_free_list_t, fl_mpool);
ompi_field_offset(i_info->ompi_free_list_t.offset.fl_allocations,
qh_type, ompi_free_list_t, fl_allocations);
ompi_field_offset(i_info->ompi_free_list_t.offset.fl_frag_class,
qh_type, ompi_free_list_t, fl_frag_class);
ompi_field_offset(i_info->ompi_free_list_t.offset.fl_frag_size,
qh_type, ompi_free_list_t, fl_frag_size);
ompi_field_offset(i_info->ompi_free_list_t.offset.fl_frag_alignment,
qh_type, ompi_free_list_t, fl_frag_alignment);
ompi_field_offset(i_info->ompi_free_list_t.offset.fl_max_to_alloc,
qh_type, ompi_free_list_t, fl_max_to_alloc);
ompi_field_offset(i_info->ompi_free_list_t.offset.fl_num_per_alloc,
qh_type, ompi_free_list_t, fl_num_per_alloc);
ompi_field_offset(i_info->ompi_free_list_t.offset.fl_num_allocated,
qh_type, ompi_free_list_t, fl_num_allocated);
}
{
mqs_type* qh_type = mqs_find_type( image, "opal_hash_table_t", mqs_lang_c );
if( !qh_type ) {
missing_in_action = "opal_hash_table_t";
goto type_missing;
}
i_info->opal_hash_table_t.type = qh_type;
i_info->opal_hash_table_t.size = mqs_sizeof(qh_type);
ompi_field_offset(i_info->opal_hash_table_t.offset.ht_table,
qh_type, opal_hash_table_t, ht_table);
ompi_field_offset(i_info->opal_hash_table_t.offset.ht_table_size,
qh_type, opal_hash_table_t, ht_table_size);
ompi_field_offset(i_info->opal_hash_table_t.offset.ht_size,
qh_type, opal_hash_table_t, ht_size);
ompi_field_offset(i_info->opal_hash_table_t.offset.ht_mask,
qh_type, opal_hash_table_t, ht_mask);
}
/*
* Now let's look for all types required for reading the requests.
*/
{
mqs_type* qh_type = mqs_find_type( image, "ompi_request_t", mqs_lang_c );
if( !qh_type ) {
missing_in_action = "ompi_request_t";
goto type_missing;
}
i_info->ompi_request_t.type = qh_type;
i_info->ompi_request_t.size = mqs_sizeof(qh_type);
ompi_field_offset(i_info->ompi_request_t.offset.req_type,
qh_type, ompi_request_t, req_type);
ompi_field_offset(i_info->ompi_request_t.offset.req_status,
qh_type, ompi_request_t, req_status);
ompi_field_offset(i_info->ompi_request_t.offset.req_complete,
qh_type, ompi_request_t, req_complete);
ompi_field_offset(i_info->ompi_request_t.offset.req_state,
qh_type, ompi_request_t, req_state);
ompi_field_offset(i_info->ompi_request_t.offset.req_f_to_c_index,
qh_type, ompi_request_t, req_f_to_c_index);
}
{
mqs_type* qh_type = mqs_find_type( image, "mca_pml_base_request_t", mqs_lang_c );
if( !qh_type ) {
missing_in_action = "mca_pml_base_request_t";
goto type_missing;
}
i_info->mca_pml_base_request_t.type = qh_type;
i_info->mca_pml_base_request_t.size = mqs_sizeof(qh_type);
ompi_field_offset(i_info->mca_pml_base_request_t.offset.req_addr,
qh_type, mca_pml_base_request_t, req_addr);
ompi_field_offset(i_info->mca_pml_base_request_t.offset.req_count,
qh_type, mca_pml_base_request_t, req_count);
ompi_field_offset(i_info->mca_pml_base_request_t.offset.req_peer,
qh_type, mca_pml_base_request_t, req_peer);
ompi_field_offset(i_info->mca_pml_base_request_t.offset.req_tag,
qh_type, mca_pml_base_request_t, req_tag);
ompi_field_offset(i_info->mca_pml_base_request_t.offset.req_comm,
qh_type, mca_pml_base_request_t, req_comm);
ompi_field_offset(i_info->mca_pml_base_request_t.offset.req_datatype,
qh_type, mca_pml_base_request_t, req_datatype);
ompi_field_offset(i_info->mca_pml_base_request_t.offset.req_proc,
qh_type, mca_pml_base_request_t, req_proc);
ompi_field_offset(i_info->mca_pml_base_request_t.offset.req_sequence,
qh_type, mca_pml_base_request_t, req_sequence);
ompi_field_offset(i_info->mca_pml_base_request_t.offset.req_type,
qh_type, mca_pml_base_request_t, req_type);
ompi_field_offset(i_info->mca_pml_base_request_t.offset.req_pml_complete,
qh_type, mca_pml_base_request_t, req_pml_complete);
}
{
mqs_type* qh_type = mqs_find_type( image, "mca_pml_base_send_request_t", mqs_lang_c );
if( !qh_type ) {
missing_in_action = "mca_pml_base_send_request_t";
goto type_missing;
}
i_info->mca_pml_base_send_request_t.type = qh_type;
i_info->mca_pml_base_send_request_t.size = mqs_sizeof(qh_type);
ompi_field_offset(i_info->mca_pml_base_send_request_t.offset.req_addr,
qh_type, mca_pml_base_send_request_t, req_addr);
ompi_field_offset(i_info->mca_pml_base_send_request_t.offset.req_bytes_packed,
qh_type, mca_pml_base_send_request_t, req_bytes_packed);
ompi_field_offset(i_info->mca_pml_base_send_request_t.offset.req_send_mode,
qh_type, mca_pml_base_send_request_t, req_send_mode);
}
{
mqs_type* qh_type = mqs_find_type( image, "mca_pml_base_recv_request_t", mqs_lang_c );
if( !qh_type ) {
missing_in_action = "mca_pml_base_recv_request_t";
goto type_missing;
}
i_info->mca_pml_base_recv_request_t.type = qh_type;
i_info->mca_pml_base_recv_request_t.size = mqs_sizeof(qh_type);
ompi_field_offset(i_info->mca_pml_base_recv_request_t.offset.req_bytes_packed,
qh_type, mca_pml_base_recv_request_t, req_bytes_packed);
}
/*
* Gather information about the received fragments and theirs headers.
*/
#if 0 /* Disabled until I find a better way */
{
mqs_type* qh_type = mqs_find_type( image, "mca_pml_ob1_common_hdr_t", mqs_lang_c );
if( !qh_type ) {
missing_in_action = "mca_pml_ob1_common_hdr_t";
goto type_missing;
}
i_info->mca_pml_ob1_common_hdr_t.type = qh_type;
i_info->mca_pml_ob1_common_hdr_t.size = mqs_sizeof(qh_type);
ompi_field_offset(i_info->mca_pml_ob1_common_hdr_t.offset.hdr_type,
qh_type, mca_pml_ob1_common_hdr_t, hdr_type);
ompi_field_offset(i_info->mca_pml_ob1_common_hdr_t.offset.hdr_flags,
qh_type, mca_pml_ob1_common_hdr_t, hdr_flags);
}
{
mqs_type* qh_type = mqs_find_type( image, "mca_pml_ob1_match_hdr_t", mqs_lang_c );
if( !qh_type ) {
missing_in_action = "mca_pml_ob1_match_hdr_t";
goto type_missing;
}
i_info->mca_pml_ob1_match_hdr_t.type = qh_type;
i_info->mca_pml_ob1_match_hdr_t.size = mqs_sizeof(qh_type);
ompi_field_offset(i_info->mca_pml_ob1_match_hdr_t.offset.hdr_common,
qh_type, mca_pml_ob1_match_hdr_t, hdr_common);
ompi_field_offset(i_info->mca_pml_ob1_match_hdr_t.offset.hdr_ctx,
qh_type, mca_pml_ob1_match_hdr_t, hdr_ctx);
ompi_field_offset(i_info->mca_pml_ob1_match_hdr_t.offset.hdr_src,
qh_type, mca_pml_ob1_match_hdr_t, hdr_src);
ompi_field_offset(i_info->mca_pml_ob1_match_hdr_t.offset.hdr_tag,
qh_type, mca_pml_ob1_match_hdr_t, hdr_tag);
ompi_field_offset(i_info->mca_pml_ob1_match_hdr_t.offset.hdr_seq,
qh_type, mca_pml_ob1_match_hdr_t, hdr_seq);
}
{
mqs_type* qh_type = mqs_find_type( image, "mca_pml_ob1_recv_frag_t", mqs_lang_c );
if( !qh_type ) {
missing_in_action = "mca_pml_ob1_recv_frag_t";
goto type_missing;
}
i_info->mca_pml_ob1_recv_frag_t.type = qh_type;
i_info->mca_pml_ob1_recv_frag_t.size = mqs_sizeof(qh_type);
ompi_field_offset(i_info->mca_pml_ob1_recv_frag_t.offset.hdr,
qh_type, mca_pml_ob1_recv_frag_t, hdr);
ompi_field_offset(i_info->mca_pml_ob1_recv_frag_t.offset.request,
qh_type, mca_pml_ob1_recv_frag_t, request);
}
#endif
/*
* And now let's look at the communicator and group structures.
*/
{
mqs_type* qh_type = mqs_find_type( image, "opal_pointer_array_t", mqs_lang_c );
if( !qh_type ) {
missing_in_action = "opal_pointer_array_t";
goto type_missing;
}
i_info->opal_pointer_array_t.type = qh_type;
i_info->opal_pointer_array_t.size = mqs_sizeof(qh_type);
ompi_field_offset(i_info->opal_pointer_array_t.offset.lowest_free,
qh_type, opal_pointer_array_t, lowest_free);
ompi_field_offset(i_info->opal_pointer_array_t.offset.number_free,
qh_type, opal_pointer_array_t, number_free);
ompi_field_offset(i_info->opal_pointer_array_t.offset.size,
qh_type, opal_pointer_array_t, size);
ompi_field_offset(i_info->opal_pointer_array_t.offset.addr,
qh_type, opal_pointer_array_t, addr);
}
{
mqs_type* qh_type = mqs_find_type( image, "ompi_communicator_t", mqs_lang_c );
if( !qh_type ) {
missing_in_action = "ompi_communicator_t";
goto type_missing;
}
i_info->ompi_communicator_t.type = qh_type;
i_info->ompi_communicator_t.size = mqs_sizeof(qh_type);
ompi_field_offset(i_info->ompi_communicator_t.offset.c_name,
qh_type, ompi_communicator_t, c_name);
ompi_field_offset(i_info->ompi_communicator_t.offset.c_contextid,
qh_type, ompi_communicator_t, c_contextid);
ompi_field_offset(i_info->ompi_communicator_t.offset.c_my_rank,
qh_type, ompi_communicator_t, c_my_rank);
ompi_field_offset(i_info->ompi_communicator_t.offset.c_local_group,
qh_type, ompi_communicator_t, c_local_group);
ompi_field_offset(i_info->ompi_communicator_t.offset.c_remote_group,
qh_type, ompi_communicator_t, c_remote_group);
ompi_field_offset(i_info->ompi_communicator_t.offset.c_flags,
qh_type, ompi_communicator_t, c_flags);
ompi_field_offset(i_info->ompi_communicator_t.offset.c_f_to_c_index,
qh_type, ompi_communicator_t, c_f_to_c_index);
ompi_field_offset(i_info->ompi_communicator_t.offset.c_topo_comm,
qh_type, ompi_communicator_t, c_topo_comm);
ompi_field_offset(i_info->ompi_communicator_t.offset.c_keyhash,
qh_type, ompi_communicator_t, c_keyhash);
}
{
mqs_type* qh_type = mqs_find_type( image, "mca_topo_base_comm_1_0_0_t", mqs_lang_c );
if( !qh_type ) {
missing_in_action = "mca_topo_base_comm_1_0_0_t";
goto type_missing;
}
i_info->ompi_mca_topo_base_comm_1_0_0_t.type = qh_type;
i_info->ompi_mca_topo_base_comm_1_0_0_t.size = mqs_sizeof(qh_type);
ompi_field_offset(i_info->ompi_mca_topo_base_comm_1_0_0_t.offset.mtc_ndims_or_nnodes,
qh_type, mca_topo_base_comm_1_0_0_t, mtc_ndims_or_nnodes);
ompi_field_offset(i_info->ompi_mca_topo_base_comm_1_0_0_t.offset.mtc_dims_or_index,
qh_type, mca_topo_base_comm_1_0_0_t, mtc_dims_or_index);
ompi_field_offset(i_info->ompi_mca_topo_base_comm_1_0_0_t.offset.mtc_periods_or_edges,
qh_type, mca_topo_base_comm_1_0_0_t, mtc_periods_or_edges);
ompi_field_offset(i_info->ompi_mca_topo_base_comm_1_0_0_t.offset.mtc_reorder,
qh_type, mca_topo_base_comm_1_0_0_t, mtc_reorder);
}
{
mqs_type* qh_type = mqs_find_type( image, "ompi_group_t", mqs_lang_c );
if( !qh_type ) {
missing_in_action = "ompi_group_t";
goto type_missing;
}
i_info->ompi_group_t.type = qh_type;
i_info->ompi_group_t.size = mqs_sizeof(qh_type);
ompi_field_offset(i_info->ompi_group_t.offset.grp_proc_count,
qh_type, ompi_group_t, grp_proc_count);
ompi_field_offset(i_info->ompi_group_t.offset.grp_proc_pointers,
qh_type, ompi_group_t, grp_proc_pointers);
ompi_field_offset(i_info->ompi_group_t.offset.grp_my_rank,
qh_type, ompi_group_t, grp_my_rank);
ompi_field_offset(i_info->ompi_group_t.offset.grp_flags,
qh_type, ompi_group_t, grp_flags);
}
{
mqs_type* qh_type = mqs_find_type( image, "ompi_status_public_t", mqs_lang_c );
if( !qh_type ) {
missing_in_action = "ompi_status_public_t";
goto type_missing;
}
i_info->ompi_status_public_t.type = qh_type;
i_info->ompi_status_public_t.size = mqs_sizeof(qh_type);
ompi_field_offset(i_info->ompi_status_public_t.offset.MPI_SOURCE,
qh_type, ompi_status_public_t, MPI_SOURCE);
ompi_field_offset(i_info->ompi_status_public_t.offset.MPI_TAG,
qh_type, ompi_status_public_t, MPI_TAG);
ompi_field_offset(i_info->ompi_status_public_t.offset.MPI_ERROR,
qh_type, ompi_status_public_t, MPI_ERROR);
ompi_field_offset(i_info->ompi_status_public_t.offset._count,
qh_type, ompi_status_public_t, _count);
ompi_field_offset(i_info->ompi_status_public_t.offset._cancelled,
qh_type, ompi_status_public_t, _cancelled);
}
{
mqs_type* qh_type = mqs_find_type( image, "ompi_datatype_t", mqs_lang_c );
if( !qh_type ) {
missing_in_action = "ompi_datatype_t";
goto type_missing;
}
i_info->ompi_datatype_t.type = qh_type;
i_info->ompi_datatype_t.size = mqs_sizeof(qh_type);
- Split the datatype engine into two parts: an MPI specific part in OMPI and a language agnostic part in OPAL. The convertor is completely moved into OPAL. This offers several benefits as described in RFC http://www.open-mpi.org/community/lists/devel/2009/07/6387.php namely: - Fewer basic types (int* and float* types, boolean and wchar - Fixing naming scheme to ompi-nomenclature. - Usability outside of the ompi-layer. - Due to the fixed nature of simple opal types, their information is completely known at compile time and therefore constified - With fewer datatypes (22), the actual sizes of bit-field types may be reduced from 64 to 32 bits, allowing reorganizing the opal_datatype structure, eliminating holes and keeping data required in convertor (upon send/recv) in one cacheline... This has implications to the convertor-datastructure and other parts of the code. - Several performance tests have been run, the netpipe latency does not change with this patch on Linux/x86-64 on the smoky cluster. - Extensive tests have been done to verify correctness (no new regressions) using: 1. mpi_test_suite on linux/x86-64 using clean ompi-trunk and ompi-ddt: a. running both trunk and ompi-ddt resulted in no differences (except for MPI_SHORT_INT and MPI_TYPE_MIX_LB_UB do now run correctly). b. with --enable-memchecker and running under valgrind (one buglet when run with static found in test-suite, commited) 2. ibm testsuite on linux/x86-64 using clean ompi-trunk and ompi-ddt: all passed (except for the dynamic/ tests failed!! as trunk/MTT) 3. compilation and usage of HDF5 tests on Jaguar using PGI and PathScale compilers. 4. compilation and usage on Scicortex. - Please note, that for the heterogeneous case, (-m32 compiled binaries/ompi), neither ompi-trunk, nor ompi-ddt branch would successfully launch. This commit was SVN r21641.
2009-07-13 08:56:31 +04:00
/* XXX TODO need to check whether super.size is OK here */
ompi_field_offset(i_info->ompi_datatype_t.offset.size,
- Split the datatype engine into two parts: an MPI specific part in OMPI and a language agnostic part in OPAL. The convertor is completely moved into OPAL. This offers several benefits as described in RFC http://www.open-mpi.org/community/lists/devel/2009/07/6387.php namely: - Fewer basic types (int* and float* types, boolean and wchar - Fixing naming scheme to ompi-nomenclature. - Usability outside of the ompi-layer. - Due to the fixed nature of simple opal types, their information is completely known at compile time and therefore constified - With fewer datatypes (22), the actual sizes of bit-field types may be reduced from 64 to 32 bits, allowing reorganizing the opal_datatype structure, eliminating holes and keeping data required in convertor (upon send/recv) in one cacheline... This has implications to the convertor-datastructure and other parts of the code. - Several performance tests have been run, the netpipe latency does not change with this patch on Linux/x86-64 on the smoky cluster. - Extensive tests have been done to verify correctness (no new regressions) using: 1. mpi_test_suite on linux/x86-64 using clean ompi-trunk and ompi-ddt: a. running both trunk and ompi-ddt resulted in no differences (except for MPI_SHORT_INT and MPI_TYPE_MIX_LB_UB do now run correctly). b. with --enable-memchecker and running under valgrind (one buglet when run with static found in test-suite, commited) 2. ibm testsuite on linux/x86-64 using clean ompi-trunk and ompi-ddt: all passed (except for the dynamic/ tests failed!! as trunk/MTT) 3. compilation and usage of HDF5 tests on Jaguar using PGI and PathScale compilers. 4. compilation and usage on Scicortex. - Please note, that for the heterogeneous case, (-m32 compiled binaries/ompi), neither ompi-trunk, nor ompi-ddt branch would successfully launch. This commit was SVN r21641.
2009-07-13 08:56:31 +04:00
qh_type, ompi_datatype_t, super.size);
ompi_field_offset(i_info->ompi_datatype_t.offset.name,
qh_type, ompi_datatype_t, name);
}
/* All the types are here. Let's succesfully return. */
*message = NULL;
return mqs_ok;
type_missing:
/*
* One of the required types is missing in the image. We are
* unable to extract the information we need from the pointers. We
* did our best but here we're at our limit. Give up!
*/
*message = missing_in_action;
printf( "The following type is missing %s\n", missing_in_action );
return err_missing_type;
}
/***********************************************************************
* Functions to access the image memory. They are specialized based *
* on the type we want to access and the debugged process architecture *
***********************************************************************/
mqs_taddr_t ompi_fetch_pointer (mqs_process *proc, mqs_taddr_t addr,
mpi_process_info *p_info)
{
int isize = p_info->sizes.pointer_size;
char buffer[8]; /* ASSUME the type fits in 8 bytes */
mqs_taddr_t res = 0;
if (mqs_ok == mqs_fetch_data (proc, addr, isize, buffer))
mqs_target_to_host (proc, buffer,
((char *)&res) + (host_is_big_endian ? sizeof(mqs_taddr_t)-isize : 0),
isize);
return res;
} /* fetch_pointer */
/***********************************************************************/
mqs_tword_t ompi_fetch_int (mqs_process *proc, mqs_taddr_t addr,
mpi_process_info *p_info)
{
int isize = p_info->sizes.int_size;
char buffer[8]; /* ASSUME the type fits in 8 bytes */
mqs_tword_t res = 0;
if (mqs_ok == mqs_fetch_data (proc, addr, isize, buffer)) {
mqs_target_to_host (proc, buffer,
((char *)&res) + (host_is_big_endian ? sizeof(mqs_tword_t)-isize : 0),
isize);
}
return res;
} /* fetch_int */
/***********************************************************************/
mqs_tword_t ompi_fetch_bool(mqs_process *proc, mqs_taddr_t addr,
mpi_process_info *p_info)
{
int isize = p_info->sizes.bool_size;
mqs_tword_t res = 0;
mqs_fetch_data (proc, addr, isize, &res);
return (0 == res ? 0 : 1);
} /* fetch_bool */
/***********************************************************************/
mqs_taddr_t ompi_fetch_size_t(mqs_process *proc, mqs_taddr_t addr,
mpi_process_info *p_info)
{
int isize = p_info->sizes.size_t_size;
char buffer[8]; /* ASSUME the type fits in 8 bytes */
mqs_taddr_t res = 0;
if (mqs_ok == mqs_fetch_data (proc, addr, isize, buffer))
mqs_target_to_host (proc, buffer,
((char *)&res) + (host_is_big_endian ? sizeof(mqs_taddr_t)-isize : 0),
isize);
return res;
} /* fetch_size_t */
/***********************************************************************/
int ompi_fetch_opal_pointer_array_info(mqs_process *proc, mqs_taddr_t addr,
mpi_process_info *p_info,
int *size, int *lowest_free,
int *number_free)
{
mqs_image *image = mqs_get_image(proc);
mpi_image_info *i_info = (mpi_image_info *) mqs_get_image_info(image);
*size = ompi_fetch_int(proc,
addr + i_info->opal_pointer_array_t.offset.size,
p_info);
*lowest_free = ompi_fetch_int(proc,
addr + i_info->opal_pointer_array_t.offset.lowest_free,
p_info);
*number_free = ompi_fetch_int(proc,
addr + i_info->opal_pointer_array_t.offset.number_free,
p_info);
return mqs_ok;
}
/***********************************************************************/
int ompi_fetch_opal_pointer_array_item(mqs_process *proc, mqs_taddr_t addr,
mpi_process_info *p_info, int index,
mqs_taddr_t *item)
{
mqs_image *image = mqs_get_image(proc);
mpi_image_info *i_info = (mpi_image_info *) mqs_get_image_info(image);
int size, lowest_free, number_free;
mqs_taddr_t base;
if (index < 0) {
return mqs_no_information;
}
ompi_fetch_opal_pointer_array_info(proc, addr, p_info, &size,
&lowest_free, &number_free);
if (index >= size) {
return mqs_no_information;
}
base = ompi_fetch_pointer(proc,
addr + i_info->opal_pointer_array_t.offset.addr,
p_info);
*item = ompi_fetch_pointer(proc,
base + index * p_info->sizes.pointer_size,
p_info);
return mqs_ok;
}