439 строки
13 KiB
C
439 строки
13 KiB
C
/*
|
|
* Copyright (c) 2004-2005 The Trustees of Indiana University and Indiana
|
|
* University Research and Technology
|
|
* Corporation. All rights reserved.
|
|
* Copyright (c) 2004-2016 The University of Tennessee and The University
|
|
* of Tennessee Research Foundation. All rights
|
|
* reserved.
|
|
* Copyright (c) 2004-2005 High Performance Computing Center Stuttgart,
|
|
* University of Stuttgart. All rights reserved.
|
|
* Copyright (c) 2004-2005 The Regents of the University of California.
|
|
* All rights reserved.
|
|
* Copyright (c) 2008-2016 University of Houston. All rights reserved.
|
|
* $COPYRIGHT$
|
|
*
|
|
* Additional copyrights may follow
|
|
*
|
|
* $HEADER$
|
|
*/
|
|
|
|
#include "ompi_config.h"
|
|
|
|
#include "ompi/communicator/communicator.h"
|
|
#include "ompi/info/info.h"
|
|
#include "ompi/file/file.h"
|
|
#include "ompi/mca/fs/fs.h"
|
|
#include "ompi/mca/fs/base/base.h"
|
|
#include "ompi/mca/fcoll/fcoll.h"
|
|
#include "ompi/mca/fcoll/base/base.h"
|
|
#include "ompi/mca/fbtl/fbtl.h"
|
|
#include "ompi/mca/fbtl/base/base.h"
|
|
|
|
#include "io_ompio.h"
|
|
#include "io_ompio_request.h"
|
|
#include "math.h"
|
|
#include <unistd.h>
|
|
|
|
/* Read and write routines are split into two interfaces.
|
|
** The
|
|
** mca_io_ompio_file_read/write[_at]
|
|
**
|
|
** routines are the ones registered with the ompio modules.
|
|
** The
|
|
**
|
|
** ompio_io_ompio_file_read/write[_at]
|
|
**
|
|
** routesin are used e.g. from the shared file pointer modules.
|
|
** The main difference is, that the first one takes an ompi_file_t
|
|
** as a file pointer argument, while the second uses the ompio internal
|
|
** mca_io_ompio_file_t structure.
|
|
*/
|
|
|
|
int mca_io_ompio_file_read (ompi_file_t *fp,
|
|
void *buf,
|
|
int count,
|
|
struct ompi_datatype_t *datatype,
|
|
ompi_status_public_t *status)
|
|
{
|
|
int ret = OMPI_SUCCESS;
|
|
mca_io_ompio_data_t *data;
|
|
|
|
data = (mca_io_ompio_data_t *) fp->f_io_selected_data;
|
|
OPAL_THREAD_LOCK(&fp->f_mutex);
|
|
ret = mca_common_ompio_file_read(&data->ompio_fh,buf,count,datatype,status);
|
|
OPAL_THREAD_UNLOCK(&fp->f_mutex);
|
|
|
|
return ret;
|
|
}
|
|
|
|
int mca_io_ompio_file_read_at (ompi_file_t *fh,
|
|
OMPI_MPI_OFFSET_TYPE offset,
|
|
void *buf,
|
|
int count,
|
|
struct ompi_datatype_t *datatype,
|
|
ompi_status_public_t * status)
|
|
{
|
|
int ret = OMPI_SUCCESS;
|
|
mca_io_ompio_data_t *data;
|
|
|
|
data = (mca_io_ompio_data_t *) fh->f_io_selected_data;
|
|
OPAL_THREAD_LOCK(&fh->f_mutex);
|
|
ret = mca_common_ompio_file_read_at(&data->ompio_fh, offset,buf,count,datatype,status);
|
|
OPAL_THREAD_UNLOCK(&fh->f_mutex);
|
|
|
|
return ret;
|
|
}
|
|
|
|
int mca_io_ompio_file_iread (ompi_file_t *fh,
|
|
void *buf,
|
|
int count,
|
|
struct ompi_datatype_t *datatype,
|
|
ompi_request_t **request)
|
|
{
|
|
int ret = OMPI_SUCCESS;
|
|
mca_io_ompio_data_t *data;
|
|
|
|
data = (mca_io_ompio_data_t *) fh->f_io_selected_data;
|
|
OPAL_THREAD_LOCK(&fh->f_mutex);
|
|
ret = mca_common_ompio_file_iread(&data->ompio_fh,buf,count,datatype,request);
|
|
OPAL_THREAD_UNLOCK(&fh->f_mutex);
|
|
|
|
return ret;
|
|
}
|
|
|
|
|
|
int mca_io_ompio_file_iread_at (ompi_file_t *fh,
|
|
OMPI_MPI_OFFSET_TYPE offset,
|
|
void *buf,
|
|
int count,
|
|
struct ompi_datatype_t *datatype,
|
|
ompi_request_t **request)
|
|
{
|
|
int ret = OMPI_SUCCESS;
|
|
mca_io_ompio_data_t *data;
|
|
|
|
data = (mca_io_ompio_data_t *) fh->f_io_selected_data;
|
|
OPAL_THREAD_LOCK(&fh->f_mutex);
|
|
ret = mca_common_ompio_file_iread_at(&data->ompio_fh,offset,buf,count,datatype,request);
|
|
OPAL_THREAD_UNLOCK(&fh->f_mutex);
|
|
|
|
return ret;
|
|
}
|
|
|
|
|
|
/* Infrastructure for collective operations */
|
|
/******************************************************/
|
|
int mca_io_ompio_file_read_all (ompi_file_t *fh,
|
|
void *buf,
|
|
int count,
|
|
struct ompi_datatype_t *datatype,
|
|
ompi_status_public_t * status)
|
|
{
|
|
int ret = OMPI_SUCCESS;
|
|
mca_io_ompio_data_t *data;
|
|
|
|
data = (mca_io_ompio_data_t *) fh->f_io_selected_data;
|
|
|
|
OPAL_THREAD_LOCK(&fh->f_mutex);
|
|
ret = data->ompio_fh.
|
|
f_fcoll->fcoll_file_read_all (&data->ompio_fh,
|
|
buf,
|
|
count,
|
|
datatype,
|
|
status);
|
|
OPAL_THREAD_UNLOCK(&fh->f_mutex);
|
|
if ( MPI_STATUS_IGNORE != status ) {
|
|
size_t size;
|
|
|
|
opal_datatype_type_size (&datatype->super, &size);
|
|
status->_ucount = count * size;
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
int mca_io_ompio_file_iread_all (ompi_file_t *fh,
|
|
void *buf,
|
|
int count,
|
|
struct ompi_datatype_t *datatype,
|
|
ompi_request_t **request)
|
|
{
|
|
int ret = OMPI_SUCCESS;
|
|
mca_io_ompio_data_t *data=NULL;
|
|
mca_io_ompio_file_t *fp=NULL;
|
|
|
|
data = (mca_io_ompio_data_t *) fh->f_io_selected_data;
|
|
fp = &data->ompio_fh;
|
|
|
|
OPAL_THREAD_LOCK(&fh->f_mutex);
|
|
if ( NULL != fp->f_fcoll->fcoll_file_iread_all ) {
|
|
ret = fp->f_fcoll->fcoll_file_iread_all (&data->ompio_fh,
|
|
buf,
|
|
count,
|
|
datatype,
|
|
request);
|
|
}
|
|
else {
|
|
/* this fcoll component does not support non-blocking
|
|
collective I/O operations. WE fake it with
|
|
individual non-blocking I/O operations. */
|
|
ret = mca_common_ompio_file_iread ( fp, buf, count, datatype, request );
|
|
}
|
|
OPAL_THREAD_UNLOCK(&fh->f_mutex);
|
|
|
|
return ret;
|
|
}
|
|
|
|
|
|
int mca_io_ompio_file_read_at_all (ompi_file_t *fh,
|
|
OMPI_MPI_OFFSET_TYPE offset,
|
|
void *buf,
|
|
int count,
|
|
struct ompi_datatype_t *datatype,
|
|
ompi_status_public_t * status)
|
|
{
|
|
int ret = OMPI_SUCCESS;
|
|
mca_io_ompio_data_t *data;
|
|
|
|
data = (mca_io_ompio_data_t *) fh->f_io_selected_data;
|
|
OPAL_THREAD_LOCK(&fh->f_mutex);
|
|
ret = mca_common_ompio_file_read_at_all(&data->ompio_fh,offset,buf,count,datatype,status);
|
|
OPAL_THREAD_UNLOCK(&fh->f_mutex);
|
|
|
|
return ret;
|
|
}
|
|
|
|
int mca_io_ompio_file_iread_at_all (ompi_file_t *fh,
|
|
OMPI_MPI_OFFSET_TYPE offset,
|
|
void *buf,
|
|
int count,
|
|
struct ompi_datatype_t *datatype,
|
|
ompi_request_t **request)
|
|
{
|
|
int ret = OMPI_SUCCESS;
|
|
mca_io_ompio_data_t *data;
|
|
data = (mca_io_ompio_data_t *) fh->f_io_selected_data;
|
|
|
|
OPAL_THREAD_LOCK(&fh->f_mutex);
|
|
ret = mca_common_ompio_file_iread_at_all ( &data->ompio_fh, offset, buf, count, datatype, request );
|
|
OPAL_THREAD_UNLOCK(&fh->f_mutex);
|
|
return ret;
|
|
}
|
|
|
|
|
|
/* Infrastructure for shared file pointer operations
|
|
** (individual and ordered)*/
|
|
/******************************************************/
|
|
int mca_io_ompio_file_read_shared (ompi_file_t *fp,
|
|
void *buf,
|
|
int count,
|
|
struct ompi_datatype_t *datatype,
|
|
ompi_status_public_t * status)
|
|
{
|
|
int ret = OMPI_SUCCESS;
|
|
mca_io_ompio_data_t *data;
|
|
mca_io_ompio_file_t *fh;
|
|
mca_sharedfp_base_module_t * shared_fp_base_module;
|
|
|
|
data = (mca_io_ompio_data_t *) fp->f_io_selected_data;
|
|
fh = &data->ompio_fh;
|
|
|
|
/*get the shared fp module associated with this file*/
|
|
shared_fp_base_module = (mca_sharedfp_base_module_t *)(fh->f_sharedfp);
|
|
if ( NULL == shared_fp_base_module ){
|
|
opal_output(0, "No shared file pointer component found for the given communicator. Can not execute\n");
|
|
return OMPI_ERROR;
|
|
}
|
|
OPAL_THREAD_LOCK(&fp->f_mutex);
|
|
ret = shared_fp_base_module->sharedfp_read(fh,buf,count,datatype,status);
|
|
OPAL_THREAD_UNLOCK(&fp->f_mutex);
|
|
|
|
return ret;
|
|
}
|
|
|
|
int mca_io_ompio_file_iread_shared (ompi_file_t *fh,
|
|
void *buf,
|
|
int count,
|
|
struct ompi_datatype_t *datatype,
|
|
ompi_request_t **request)
|
|
{
|
|
int ret = OMPI_SUCCESS;
|
|
mca_io_ompio_data_t *data;
|
|
mca_io_ompio_file_t *ompio_fh;
|
|
mca_sharedfp_base_module_t * shared_fp_base_module;
|
|
|
|
data = (mca_io_ompio_data_t *) fh->f_io_selected_data;
|
|
ompio_fh = &data->ompio_fh;
|
|
|
|
/*get the shared fp module associated with this file*/
|
|
shared_fp_base_module = (mca_sharedfp_base_module_t *)(ompio_fh->f_sharedfp);
|
|
if ( NULL == shared_fp_base_module ){
|
|
opal_output(0, "No shared file pointer component found for the given communicator. Can not execute\n");
|
|
return OMPI_ERROR;
|
|
}
|
|
OPAL_THREAD_LOCK(&fh->f_mutex);
|
|
ret = shared_fp_base_module->sharedfp_iread(ompio_fh,buf,count,datatype,request);
|
|
OPAL_THREAD_UNLOCK(&fh->f_mutex);
|
|
|
|
return ret;
|
|
}
|
|
|
|
int mca_io_ompio_file_read_ordered (ompi_file_t *fh,
|
|
void *buf,
|
|
int count,
|
|
struct ompi_datatype_t *datatype,
|
|
ompi_status_public_t * status)
|
|
{
|
|
int ret = OMPI_SUCCESS;
|
|
mca_io_ompio_data_t *data;
|
|
mca_io_ompio_file_t *ompio_fh;
|
|
mca_sharedfp_base_module_t * shared_fp_base_module;
|
|
|
|
data = (mca_io_ompio_data_t *) fh->f_io_selected_data;
|
|
ompio_fh = &data->ompio_fh;
|
|
|
|
/*get the shared fp module associated with this file*/
|
|
shared_fp_base_module = (mca_sharedfp_base_module_t *)(ompio_fh->f_sharedfp);
|
|
if ( NULL == shared_fp_base_module ){
|
|
opal_output(0, "No shared file pointer component found for the given communicator. Can not execute\n");
|
|
return OMPI_ERROR;
|
|
}
|
|
OPAL_THREAD_LOCK(&fh->f_mutex);
|
|
ret = shared_fp_base_module->sharedfp_read_ordered(ompio_fh,buf,count,datatype,status);
|
|
OPAL_THREAD_UNLOCK(&fh->f_mutex);
|
|
return ret;
|
|
}
|
|
|
|
int mca_io_ompio_file_read_ordered_begin (ompi_file_t *fh,
|
|
void *buf,
|
|
int count,
|
|
struct ompi_datatype_t *datatype)
|
|
{
|
|
int ret = OMPI_SUCCESS;
|
|
mca_io_ompio_data_t *data;
|
|
mca_io_ompio_file_t *ompio_fh;
|
|
mca_sharedfp_base_module_t * shared_fp_base_module;
|
|
|
|
data = (mca_io_ompio_data_t *) fh->f_io_selected_data;
|
|
ompio_fh = &data->ompio_fh;
|
|
|
|
/*get the shared fp module associated with this file*/
|
|
shared_fp_base_module = ompio_fh->f_sharedfp;
|
|
if ( NULL == shared_fp_base_module ){
|
|
opal_output(0, "No shared file pointer component found for the given communicator. Can not execute\n");
|
|
return OMPI_ERROR;
|
|
}
|
|
OPAL_THREAD_LOCK(&fh->f_mutex);
|
|
ret = shared_fp_base_module->sharedfp_read_ordered_begin(ompio_fh,buf,count,datatype);
|
|
OPAL_THREAD_UNLOCK(&fh->f_mutex);
|
|
|
|
return ret;
|
|
}
|
|
|
|
int mca_io_ompio_file_read_ordered_end (ompi_file_t *fh,
|
|
void *buf,
|
|
ompi_status_public_t * status)
|
|
{
|
|
int ret = OMPI_SUCCESS;
|
|
mca_io_ompio_data_t *data;
|
|
mca_io_ompio_file_t *ompio_fh;
|
|
mca_sharedfp_base_module_t * shared_fp_base_module;
|
|
|
|
data = (mca_io_ompio_data_t *) fh->f_io_selected_data;
|
|
ompio_fh = &data->ompio_fh;
|
|
|
|
/*get the shared fp module associated with this file*/
|
|
shared_fp_base_module = ompio_fh->f_sharedfp;
|
|
if ( NULL == shared_fp_base_module ){
|
|
opal_output(0, "No shared file pointer component found for the given communicator. Can not execute\n");
|
|
return OMPI_ERROR;
|
|
}
|
|
OPAL_THREAD_LOCK(&fh->f_mutex);
|
|
ret = shared_fp_base_module->sharedfp_read_ordered_end(ompio_fh,buf,status);
|
|
OPAL_THREAD_UNLOCK(&fh->f_mutex);
|
|
|
|
return ret;
|
|
}
|
|
|
|
|
|
/* Split collectives . Not really used but infrastructure is in place */
|
|
/**********************************************************************/
|
|
int mca_io_ompio_file_read_all_begin (ompi_file_t *fh,
|
|
void *buf,
|
|
int count,
|
|
struct ompi_datatype_t *datatype)
|
|
{
|
|
int ret = OMPI_SUCCESS;
|
|
mca_io_ompio_file_t *fp;
|
|
mca_io_ompio_data_t *data;
|
|
|
|
data = (mca_io_ompio_data_t *) fh->f_io_selected_data;
|
|
fp = &data->ompio_fh;
|
|
if ( true == fp->f_split_coll_in_use ) {
|
|
printf("Only one split collective I/O operation allowed per file handle at any given point in time!\n");
|
|
return MPI_ERR_OTHER;
|
|
}
|
|
/* No need for locking fh->f_mutex, that is done in file_iread_all */
|
|
ret = mca_io_ompio_file_iread_all ( fh, buf, count, datatype, &fp->f_split_coll_req );
|
|
fp->f_split_coll_in_use = true;
|
|
|
|
return ret;
|
|
}
|
|
|
|
int mca_io_ompio_file_read_all_end (ompi_file_t *fh,
|
|
void *buf,
|
|
ompi_status_public_t * status)
|
|
{
|
|
int ret = OMPI_SUCCESS;
|
|
mca_io_ompio_file_t *fp;
|
|
mca_io_ompio_data_t *data;
|
|
|
|
data = (mca_io_ompio_data_t *) fh->f_io_selected_data;
|
|
fp = &data->ompio_fh;
|
|
ret = ompi_request_wait ( &fp->f_split_coll_req, status );
|
|
|
|
/* remove the flag again */
|
|
fp->f_split_coll_in_use = false;
|
|
return ret;
|
|
}
|
|
|
|
int mca_io_ompio_file_read_at_all_begin (ompi_file_t *fh,
|
|
OMPI_MPI_OFFSET_TYPE offset,
|
|
void *buf,
|
|
int count,
|
|
struct ompi_datatype_t *datatype)
|
|
{
|
|
int ret = OMPI_SUCCESS;
|
|
mca_io_ompio_data_t *data;
|
|
mca_io_ompio_file_t *fp=NULL;
|
|
data = (mca_io_ompio_data_t *) fh->f_io_selected_data;
|
|
fp = &data->ompio_fh;
|
|
|
|
if ( true == fp->f_split_coll_in_use ) {
|
|
printf("Only one split collective I/O operation allowed per file handle at any given point in time!\n");
|
|
return MPI_ERR_REQUEST;
|
|
}
|
|
OPAL_THREAD_LOCK(&fh->f_mutex);
|
|
ret = mca_common_ompio_file_iread_at_all ( fp, offset, buf, count, datatype, &fp->f_split_coll_req );
|
|
OPAL_THREAD_UNLOCK(&fh->f_mutex);
|
|
fp->f_split_coll_in_use = true;
|
|
return ret;
|
|
}
|
|
|
|
int mca_io_ompio_file_read_at_all_end (ompi_file_t *fh,
|
|
void *buf,
|
|
ompi_status_public_t * status)
|
|
{
|
|
int ret = OMPI_SUCCESS;
|
|
mca_io_ompio_data_t *data;
|
|
mca_io_ompio_file_t *fp=NULL;
|
|
|
|
data = (mca_io_ompio_data_t *) fh->f_io_selected_data;
|
|
fp = &data->ompio_fh;
|
|
ret = ompi_request_wait ( &fp->f_split_coll_req, status );
|
|
|
|
/* remove the flag again */
|
|
fp->f_split_coll_in_use = false;
|
|
return ret;
|
|
}
|