/* * Copyright (c) 2004-2005 The Trustees of Indiana University and Indiana * University Research and Technology * Corporation. All rights reserved. * Copyright (c) 2004-2016 The University of Tennessee and The University * of Tennessee Research Foundation. All rights * reserved. * Copyright (c) 2004-2005 High Performance Computing Center Stuttgart, * University of Stuttgart. All rights reserved. * Copyright (c) 2004-2005 The Regents of the University of California. * All rights reserved. * Copyright (c) 2008-2016 University of Houston. All rights reserved. * $COPYRIGHT$ * * Additional copyrights may follow * * $HEADER$ */ #include "ompi_config.h" #include "ompi/communicator/communicator.h" #include "ompi/info/info.h" #include "ompi/file/file.h" #include "ompi/mca/fs/fs.h" #include "ompi/mca/fs/base/base.h" #include "ompi/mca/fcoll/fcoll.h" #include "ompi/mca/fcoll/base/base.h" #include "ompi/mca/fbtl/fbtl.h" #include "ompi/mca/fbtl/base/base.h" #include "io_ompio.h" #include "io_ompio_request.h" #include "math.h" #include /* Read and write routines are split into two interfaces. ** The ** mca_io_ompio_file_read/write[_at] ** ** routines are the ones registered with the ompio modules. ** The ** ** ompio_io_ompio_file_read/write[_at] ** ** routesin are used e.g. from the shared file pointer modules. ** The main difference is, that the first one takes an ompi_file_t ** as a file pointer argument, while the second uses the ompio internal ** mca_io_ompio_file_t structure. */ int mca_io_ompio_file_read (ompi_file_t *fp, void *buf, int count, struct ompi_datatype_t *datatype, ompi_status_public_t *status) { int ret = OMPI_SUCCESS; mca_io_ompio_data_t *data; data = (mca_io_ompio_data_t *) fp->f_io_selected_data; OPAL_THREAD_LOCK(&fp->f_mutex); ret = mca_common_ompio_file_read(&data->ompio_fh,buf,count,datatype,status); OPAL_THREAD_UNLOCK(&fp->f_mutex); return ret; } int mca_io_ompio_file_read_at (ompi_file_t *fh, OMPI_MPI_OFFSET_TYPE offset, void *buf, int count, struct ompi_datatype_t *datatype, ompi_status_public_t * status) { int ret = OMPI_SUCCESS; mca_io_ompio_data_t *data; data = (mca_io_ompio_data_t *) fh->f_io_selected_data; OPAL_THREAD_LOCK(&fh->f_mutex); ret = mca_common_ompio_file_read_at(&data->ompio_fh, offset,buf,count,datatype,status); OPAL_THREAD_UNLOCK(&fh->f_mutex); return ret; } int mca_io_ompio_file_iread (ompi_file_t *fh, void *buf, int count, struct ompi_datatype_t *datatype, ompi_request_t **request) { int ret = OMPI_SUCCESS; mca_io_ompio_data_t *data; data = (mca_io_ompio_data_t *) fh->f_io_selected_data; OPAL_THREAD_LOCK(&fh->f_mutex); ret = mca_common_ompio_file_iread(&data->ompio_fh,buf,count,datatype,request); OPAL_THREAD_UNLOCK(&fh->f_mutex); return ret; } int mca_io_ompio_file_iread_at (ompi_file_t *fh, OMPI_MPI_OFFSET_TYPE offset, void *buf, int count, struct ompi_datatype_t *datatype, ompi_request_t **request) { int ret = OMPI_SUCCESS; mca_io_ompio_data_t *data; data = (mca_io_ompio_data_t *) fh->f_io_selected_data; OPAL_THREAD_LOCK(&fh->f_mutex); ret = mca_common_ompio_file_iread_at(&data->ompio_fh,offset,buf,count,datatype,request); OPAL_THREAD_UNLOCK(&fh->f_mutex); return ret; } /* Infrastructure for collective operations */ /******************************************************/ int mca_io_ompio_file_read_all (ompi_file_t *fh, void *buf, int count, struct ompi_datatype_t *datatype, ompi_status_public_t * status) { int ret = OMPI_SUCCESS; mca_io_ompio_data_t *data; data = (mca_io_ompio_data_t *) fh->f_io_selected_data; OPAL_THREAD_LOCK(&fh->f_mutex); ret = data->ompio_fh. f_fcoll->fcoll_file_read_all (&data->ompio_fh, buf, count, datatype, status); OPAL_THREAD_UNLOCK(&fh->f_mutex); if ( MPI_STATUS_IGNORE != status ) { size_t size; opal_datatype_type_size (&datatype->super, &size); status->_ucount = count * size; } return ret; } int mca_io_ompio_file_iread_all (ompi_file_t *fh, void *buf, int count, struct ompi_datatype_t *datatype, ompi_request_t **request) { int ret = OMPI_SUCCESS; mca_io_ompio_data_t *data=NULL; mca_io_ompio_file_t *fp=NULL; data = (mca_io_ompio_data_t *) fh->f_io_selected_data; fp = &data->ompio_fh; OPAL_THREAD_LOCK(&fh->f_mutex); if ( NULL != fp->f_fcoll->fcoll_file_iread_all ) { ret = fp->f_fcoll->fcoll_file_iread_all (&data->ompio_fh, buf, count, datatype, request); } else { /* this fcoll component does not support non-blocking collective I/O operations. WE fake it with individual non-blocking I/O operations. */ ret = mca_common_ompio_file_iread ( fp, buf, count, datatype, request ); } OPAL_THREAD_UNLOCK(&fh->f_mutex); return ret; } int mca_io_ompio_file_read_at_all (ompi_file_t *fh, OMPI_MPI_OFFSET_TYPE offset, void *buf, int count, struct ompi_datatype_t *datatype, ompi_status_public_t * status) { int ret = OMPI_SUCCESS; mca_io_ompio_data_t *data; data = (mca_io_ompio_data_t *) fh->f_io_selected_data; OPAL_THREAD_LOCK(&fh->f_mutex); ret = mca_common_ompio_file_read_at_all(&data->ompio_fh,offset,buf,count,datatype,status); OPAL_THREAD_UNLOCK(&fh->f_mutex); return ret; } int mca_io_ompio_file_iread_at_all (ompi_file_t *fh, OMPI_MPI_OFFSET_TYPE offset, void *buf, int count, struct ompi_datatype_t *datatype, ompi_request_t **request) { int ret = OMPI_SUCCESS; mca_io_ompio_data_t *data; data = (mca_io_ompio_data_t *) fh->f_io_selected_data; OPAL_THREAD_LOCK(&fh->f_mutex); ret = mca_common_ompio_file_iread_at_all ( &data->ompio_fh, offset, buf, count, datatype, request ); OPAL_THREAD_UNLOCK(&fh->f_mutex); return ret; } /* Infrastructure for shared file pointer operations ** (individual and ordered)*/ /******************************************************/ int mca_io_ompio_file_read_shared (ompi_file_t *fp, void *buf, int count, struct ompi_datatype_t *datatype, ompi_status_public_t * status) { int ret = OMPI_SUCCESS; mca_io_ompio_data_t *data; mca_io_ompio_file_t *fh; mca_sharedfp_base_module_t * shared_fp_base_module; data = (mca_io_ompio_data_t *) fp->f_io_selected_data; fh = &data->ompio_fh; /*get the shared fp module associated with this file*/ shared_fp_base_module = (mca_sharedfp_base_module_t *)(fh->f_sharedfp); if ( NULL == shared_fp_base_module ){ opal_output(0, "No shared file pointer component found for the given communicator. Can not execute\n"); return OMPI_ERROR; } OPAL_THREAD_LOCK(&fp->f_mutex); ret = shared_fp_base_module->sharedfp_read(fh,buf,count,datatype,status); OPAL_THREAD_UNLOCK(&fp->f_mutex); return ret; } int mca_io_ompio_file_iread_shared (ompi_file_t *fh, void *buf, int count, struct ompi_datatype_t *datatype, ompi_request_t **request) { int ret = OMPI_SUCCESS; mca_io_ompio_data_t *data; mca_io_ompio_file_t *ompio_fh; mca_sharedfp_base_module_t * shared_fp_base_module; data = (mca_io_ompio_data_t *) fh->f_io_selected_data; ompio_fh = &data->ompio_fh; /*get the shared fp module associated with this file*/ shared_fp_base_module = (mca_sharedfp_base_module_t *)(ompio_fh->f_sharedfp); if ( NULL == shared_fp_base_module ){ opal_output(0, "No shared file pointer component found for the given communicator. Can not execute\n"); return OMPI_ERROR; } OPAL_THREAD_LOCK(&fh->f_mutex); ret = shared_fp_base_module->sharedfp_iread(ompio_fh,buf,count,datatype,request); OPAL_THREAD_UNLOCK(&fh->f_mutex); return ret; } int mca_io_ompio_file_read_ordered (ompi_file_t *fh, void *buf, int count, struct ompi_datatype_t *datatype, ompi_status_public_t * status) { int ret = OMPI_SUCCESS; mca_io_ompio_data_t *data; mca_io_ompio_file_t *ompio_fh; mca_sharedfp_base_module_t * shared_fp_base_module; data = (mca_io_ompio_data_t *) fh->f_io_selected_data; ompio_fh = &data->ompio_fh; /*get the shared fp module associated with this file*/ shared_fp_base_module = (mca_sharedfp_base_module_t *)(ompio_fh->f_sharedfp); if ( NULL == shared_fp_base_module ){ opal_output(0, "No shared file pointer component found for the given communicator. Can not execute\n"); return OMPI_ERROR; } OPAL_THREAD_LOCK(&fh->f_mutex); ret = shared_fp_base_module->sharedfp_read_ordered(ompio_fh,buf,count,datatype,status); OPAL_THREAD_UNLOCK(&fh->f_mutex); return ret; } int mca_io_ompio_file_read_ordered_begin (ompi_file_t *fh, void *buf, int count, struct ompi_datatype_t *datatype) { int ret = OMPI_SUCCESS; mca_io_ompio_data_t *data; mca_io_ompio_file_t *ompio_fh; mca_sharedfp_base_module_t * shared_fp_base_module; data = (mca_io_ompio_data_t *) fh->f_io_selected_data; ompio_fh = &data->ompio_fh; /*get the shared fp module associated with this file*/ shared_fp_base_module = ompio_fh->f_sharedfp; if ( NULL == shared_fp_base_module ){ opal_output(0, "No shared file pointer component found for the given communicator. Can not execute\n"); return OMPI_ERROR; } OPAL_THREAD_LOCK(&fh->f_mutex); ret = shared_fp_base_module->sharedfp_read_ordered_begin(ompio_fh,buf,count,datatype); OPAL_THREAD_UNLOCK(&fh->f_mutex); return ret; } int mca_io_ompio_file_read_ordered_end (ompi_file_t *fh, void *buf, ompi_status_public_t * status) { int ret = OMPI_SUCCESS; mca_io_ompio_data_t *data; mca_io_ompio_file_t *ompio_fh; mca_sharedfp_base_module_t * shared_fp_base_module; data = (mca_io_ompio_data_t *) fh->f_io_selected_data; ompio_fh = &data->ompio_fh; /*get the shared fp module associated with this file*/ shared_fp_base_module = ompio_fh->f_sharedfp; if ( NULL == shared_fp_base_module ){ opal_output(0, "No shared file pointer component found for the given communicator. Can not execute\n"); return OMPI_ERROR; } OPAL_THREAD_LOCK(&fh->f_mutex); ret = shared_fp_base_module->sharedfp_read_ordered_end(ompio_fh,buf,status); OPAL_THREAD_UNLOCK(&fh->f_mutex); return ret; } /* Split collectives . Not really used but infrastructure is in place */ /**********************************************************************/ int mca_io_ompio_file_read_all_begin (ompi_file_t *fh, void *buf, int count, struct ompi_datatype_t *datatype) { int ret = OMPI_SUCCESS; mca_io_ompio_file_t *fp; mca_io_ompio_data_t *data; data = (mca_io_ompio_data_t *) fh->f_io_selected_data; fp = &data->ompio_fh; if ( true == fp->f_split_coll_in_use ) { printf("Only one split collective I/O operation allowed per file handle at any given point in time!\n"); return MPI_ERR_OTHER; } /* No need for locking fh->f_mutex, that is done in file_iread_all */ ret = mca_io_ompio_file_iread_all ( fh, buf, count, datatype, &fp->f_split_coll_req ); fp->f_split_coll_in_use = true; return ret; } int mca_io_ompio_file_read_all_end (ompi_file_t *fh, void *buf, ompi_status_public_t * status) { int ret = OMPI_SUCCESS; mca_io_ompio_file_t *fp; mca_io_ompio_data_t *data; data = (mca_io_ompio_data_t *) fh->f_io_selected_data; fp = &data->ompio_fh; ret = ompi_request_wait ( &fp->f_split_coll_req, status ); /* remove the flag again */ fp->f_split_coll_in_use = false; return ret; } int mca_io_ompio_file_read_at_all_begin (ompi_file_t *fh, OMPI_MPI_OFFSET_TYPE offset, void *buf, int count, struct ompi_datatype_t *datatype) { int ret = OMPI_SUCCESS; mca_io_ompio_data_t *data; mca_io_ompio_file_t *fp=NULL; data = (mca_io_ompio_data_t *) fh->f_io_selected_data; fp = &data->ompio_fh; if ( true == fp->f_split_coll_in_use ) { printf("Only one split collective I/O operation allowed per file handle at any given point in time!\n"); return MPI_ERR_REQUEST; } OPAL_THREAD_LOCK(&fh->f_mutex); ret = mca_common_ompio_file_iread_at_all ( fp, offset, buf, count, datatype, &fp->f_split_coll_req ); OPAL_THREAD_UNLOCK(&fh->f_mutex); fp->f_split_coll_in_use = true; return ret; } int mca_io_ompio_file_read_at_all_end (ompi_file_t *fh, void *buf, ompi_status_public_t * status) { int ret = OMPI_SUCCESS; mca_io_ompio_data_t *data; mca_io_ompio_file_t *fp=NULL; data = (mca_io_ompio_data_t *) fh->f_io_selected_data; fp = &data->ompio_fh; ret = ompi_request_wait ( &fp->f_split_coll_req, status ); /* remove the flag again */ fp->f_split_coll_in_use = false; return ret; }