Remove stale code
This commit was SVN r26251.
Этот коммит содержится в:
родитель
b6cde9a8d1
Коммит
93bbeabc55
@ -1,13 +0,0 @@
|
||||
#
|
||||
# Copyright (c) 2009-2010 High Performance Computing Center Stuttgart,
|
||||
# University of Stuttgart. All rights reserved.
|
||||
# $COPYRIGHT$
|
||||
#
|
||||
# Additional copyrights may follow
|
||||
#
|
||||
# $HEADER$
|
||||
#
|
||||
|
||||
# Specific to this module
|
||||
mca_link_libraries=libopen-rte
|
||||
required_check=find_ccp
|
@ -1,56 +0,0 @@
|
||||
#
|
||||
# Copyright (c) 2004-2005 The Trustees of Indiana University and Indiana
|
||||
# University Research and Technology
|
||||
# Corporation. All rights reserved.
|
||||
# Copyright (c) 2004-2005 The University of Tennessee and The University
|
||||
# of Tennessee Research Foundation. All rights
|
||||
# reserved.
|
||||
# Copyright (c) 2004-2009 High Performance Computing Center Stuttgart,
|
||||
# University of Stuttgart. All rights reserved.
|
||||
# Copyright (c) 2004-2005 The Regents of the University of California.
|
||||
# All rights reserved.
|
||||
# Copyright (c) 2008-2010 Cisco Systems, Inc. All rights reserved.
|
||||
# $COPYRIGHT$
|
||||
#
|
||||
# Additional copyrights may follow
|
||||
#
|
||||
# $HEADER$
|
||||
#
|
||||
|
||||
EXTRA_DIST = .windows
|
||||
|
||||
AM_CPPFLAGS = $(plm_ccp_CPPFLAGS)
|
||||
|
||||
dist_pkgdata_DATA = help-plm-ccp.txt
|
||||
|
||||
sources = \
|
||||
plm_ccp.h \
|
||||
plm_ccp_component.c \
|
||||
plm_ccp_module.c
|
||||
|
||||
# Make the output library in this directory, and name it either
|
||||
# mca_<type>_<name>.la (for DSO builds) or libmca_<type>_<name>.la
|
||||
# (for static builds).
|
||||
|
||||
if MCA_BUILD_orte_plm_ccp_DSO
|
||||
lib =
|
||||
lib_sources =
|
||||
component = mca_plm_ccp.la
|
||||
component_sources = $(sources)
|
||||
else
|
||||
lib = libmca_plm_ccp.la
|
||||
lib_sources = $(sources)
|
||||
component =
|
||||
component_sources =
|
||||
endif
|
||||
|
||||
mcacomponentdir = $(pkglibdir)
|
||||
mcacomponent_LTLIBRARIES = $(component)
|
||||
mca_plm_ccp_la_SOURCES = $(component_sources)
|
||||
mca_plm_ccp_la_LDFLAGS = -module -avoid-version $(plm_ccp_LDFLAGS)
|
||||
mca_plm_ccp_la_LIBADD = $(plm_ccp_LIBS)
|
||||
|
||||
noinst_LTLIBRARIES = $(lib)
|
||||
libmca_plm_ccp_la_SOURCES = $(lib_sources)
|
||||
libmca_plm_ccp_la_LDFLAGS = -module -avoid-version $(plm_ccp_LDFLAGS)
|
||||
libmca_plm_ccp_la_LIBADD = $(plm_ccp_LIBS)
|
@ -1,29 +0,0 @@
|
||||
# -*- shell-script -*-
|
||||
#
|
||||
# Copyright (c) 2004-2005 The Trustees of Indiana University and Indiana
|
||||
# University Research and Technology
|
||||
# Corporation. All rights reserved.
|
||||
# Copyright (c) 2004-2005 The University of Tennessee and The University
|
||||
# of Tennessee Research Foundation. All rights
|
||||
# reserved.
|
||||
# Copyright (c) 2004-2005 High Performance Computing Center Stuttgart,
|
||||
# University of Stuttgart. All rights reserved.
|
||||
# Copyright (c) 2004-2005 The Regents of the University of California.
|
||||
# All rights reserved.
|
||||
dnl Copyright (c) 2010 Cisco Systems, Inc. All rights reserved.
|
||||
# $COPYRIGHT$
|
||||
#
|
||||
# Additional copyrights may follow
|
||||
#
|
||||
# $HEADER$
|
||||
#
|
||||
|
||||
# MCA_plm_ccp_CONFIG([action-if-found], [action-if-not-found])
|
||||
# -----------------------------------------------------------
|
||||
AC_DEFUN([MCA_orte_plm_ccp_CONFIG],[
|
||||
AC_CONFIG_FILES([orte/mca/plm/ccp/Makefile])
|
||||
|
||||
plm_ccp_good=0
|
||||
# CCP does never exist under Unix
|
||||
[$2]
|
||||
])dnl
|
@ -1,57 +0,0 @@
|
||||
# -*- text -*-
|
||||
#
|
||||
# Copyright (c) 2004-2005 The Trustees of Indiana University and Indiana
|
||||
# University Research and Technology
|
||||
# Corporation. All rights reserved.
|
||||
# Copyright (c) 2004-2005 The University of Tennessee and The University
|
||||
# of Tennessee Research Foundation. All rights
|
||||
# reserved.
|
||||
# Copyright (c) 2004-2005 High Performance Computing Center Stuttgart,
|
||||
# University of Stuttgart. All rights reserved.
|
||||
# Copyright (c) 2004-2005 The Regents of the University of California.
|
||||
# All rights reserved.
|
||||
# $COPYRIGHT$
|
||||
#
|
||||
# Additional copyrights may follow
|
||||
#
|
||||
# $HEADER$
|
||||
#
|
||||
[ccp-bad-launchid]
|
||||
The CCP process starter cannot spawn the specified
|
||||
application on a remote node due to an invalid launch_id.
|
||||
|
||||
Node name: %s
|
||||
Launch id: %d
|
||||
|
||||
This is most likely due to use of the "--hostfile" option to the
|
||||
command line. At this time, Open MPI/OpenRTE do not support this
|
||||
method of operation. Instead, the system expects to directly read
|
||||
information regarding the nodes to be used from the environment.
|
||||
|
||||
Removing "--hostfile" from the command line will likely allow the
|
||||
application to be launched. This will be fixed in a future release
|
||||
to support the use of "--hostfile" on the command line.
|
||||
#
|
||||
[multiple-prefixes]
|
||||
Multiple different --prefix options were specified to mpirun for the
|
||||
same node. This is a fatal error for the ccp process
|
||||
starter in Open MPI.
|
||||
|
||||
The first two prefix values supplied for node %s were:
|
||||
%s
|
||||
and %s
|
||||
#
|
||||
[ccp-spawn-failed]
|
||||
The CCP process starter failed to spawn a daemon (orted)
|
||||
on a remote node.
|
||||
|
||||
Command line: %s
|
||||
Node name: %s
|
||||
Launch id: %d
|
||||
|
||||
If you do not understand this error mesage, please try the following:
|
||||
|
||||
1. Ensure that the executable "orted" is in your PATH
|
||||
2. Use the --prefix option to indicate where we can
|
||||
find that executable
|
||||
3. Talk to your local system administrator
|
@ -1,43 +0,0 @@
|
||||
/*
|
||||
* Copyright (c) 2004-2006 The University of Tennessee and The University
|
||||
* of Tennessee Research Foundation. All rights
|
||||
* reserved.
|
||||
* Copyright (c) 2004-2008 High Performance Computing Center Stuttgart,
|
||||
* University of Stuttgart. All rights reserved.
|
||||
* $COPYRIGHT$
|
||||
*
|
||||
* Additional copyrights may follow
|
||||
*
|
||||
* $HEADER$
|
||||
*/
|
||||
|
||||
#ifndef ORTE_plm_CCP_EXPORT_H
|
||||
#define ORTE_plm_CCP_EXPORT_H
|
||||
|
||||
#include "orte_config.h"
|
||||
|
||||
#include "opal/mca/mca.h"
|
||||
#include "orte/mca/plm/plm.h"
|
||||
|
||||
BEGIN_C_DECLS
|
||||
|
||||
struct orte_plm_ccp_component_t {
|
||||
orte_plm_base_component_t super;
|
||||
int priority;
|
||||
int debug;
|
||||
int verbose;
|
||||
bool want_path_check;
|
||||
char **checked_paths;
|
||||
char *stdout_file;
|
||||
char *stderr_file;
|
||||
char *job_name;
|
||||
bool timing;
|
||||
};
|
||||
typedef struct orte_plm_ccp_component_t orte_plm_ccp_component_t;
|
||||
/* Globally exported variables */
|
||||
ORTE_DECLSPEC extern orte_plm_ccp_component_t mca_plm_ccp_component;
|
||||
extern orte_plm_base_module_t orte_plm_ccp_module;
|
||||
|
||||
END_C_DECLS
|
||||
|
||||
#endif /* ORTE_plm_CCP_EXPORT_H */
|
@ -1,172 +0,0 @@
|
||||
/*
|
||||
* Copyright (c) 2004-2005 The University of Tennessee and The University
|
||||
* of Tennessee Research Foundation. All rights
|
||||
* reserved.
|
||||
* Copyright (c) 2004-2008 High Performance Computing Center Stuttgart,
|
||||
* University of Stuttgart. All rights reserved.
|
||||
* Copyright (c) 2004-2008 The Trustees of Indiana University.
|
||||
* All rights reserved.
|
||||
* $COPYRIGHT$
|
||||
*
|
||||
* Additional copyrights may follow
|
||||
*
|
||||
* $HEADER$
|
||||
*
|
||||
*/
|
||||
|
||||
#include "orte_config.h"
|
||||
|
||||
#include "opal/mca/base/mca_base_param.h"
|
||||
#include "orte/constants.h"
|
||||
|
||||
#include "orte/util/proc_info.h"
|
||||
|
||||
#include "orte/mca/plm/plm.h"
|
||||
#include "orte/mca/plm/base/base.h"
|
||||
#include "orte/mca/plm/base/plm_private.h"
|
||||
#include "plm_ccp.h"
|
||||
|
||||
/* Import the Windows CCP API. */
|
||||
#import "ccpapi.tlb" named_guids no_namespace raw_interfaces_only \
|
||||
rename("SetEnvironmentVariable","SetEnvVar") \
|
||||
rename("GetJob", "GetSingleJob") \
|
||||
rename("AddJob", "AddSingleJob")
|
||||
|
||||
|
||||
/*
|
||||
* Public string showing the plm ompi_ccp component version number
|
||||
*/
|
||||
const char *mca_plm_ccp_component_version_string =
|
||||
"Open MPI ccp plm MCA component version " ORTE_VERSION;
|
||||
|
||||
|
||||
|
||||
/*
|
||||
* Local function
|
||||
*/
|
||||
static int plm_ccp_open(void);
|
||||
static int plm_ccp_close(void);
|
||||
static int orte_plm_ccp_component_query(mca_base_module_t **module, int *priority);
|
||||
|
||||
|
||||
/*
|
||||
* Instantiate the public struct with all of our public information
|
||||
* and pointers to our public functions in it
|
||||
*/
|
||||
|
||||
orte_plm_ccp_component_t mca_plm_ccp_component = {
|
||||
{
|
||||
/* First, the mca_component_t struct containing meta information
|
||||
about the component itself */
|
||||
|
||||
{
|
||||
ORTE_PLM_BASE_VERSION_2_0_0,
|
||||
|
||||
/* Component name and version */
|
||||
"ccp",
|
||||
ORTE_MAJOR_VERSION,
|
||||
ORTE_MINOR_VERSION,
|
||||
ORTE_RELEASE_VERSION,
|
||||
|
||||
/* Component open and close functions */
|
||||
plm_ccp_open,
|
||||
plm_ccp_close,
|
||||
orte_plm_ccp_component_query
|
||||
},
|
||||
{
|
||||
/* The component is checkpoint ready */
|
||||
MCA_BASE_METADATA_PARAM_CHECKPOINT
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
|
||||
static int plm_ccp_open(void)
|
||||
{
|
||||
int tmp, value;
|
||||
mca_base_component_t *comp = &mca_plm_ccp_component.super.base_version;
|
||||
|
||||
mca_base_param_reg_int(comp, "debug", "Enable debugging of the CCP plm",
|
||||
false, false, 0, &mca_plm_ccp_component.debug);
|
||||
mca_base_param_reg_int(comp, "verbose", "Enable verbose output of the ccp plm",
|
||||
false, false, 0, &mca_plm_ccp_component.verbose);
|
||||
|
||||
mca_base_param_reg_int(comp, "priority", "Default selection priority",
|
||||
false, false, 75, &mca_plm_ccp_component.priority);
|
||||
|
||||
mca_base_param_reg_int(comp, "want_path_check",
|
||||
"Whether the launching process should check for "
|
||||
"the plm_ccp_orted executable in the PATH before "
|
||||
"launching (the CCP API does not give an indication "
|
||||
"of failure; this is a somewhat-lame workaround; "
|
||||
"non-zero values enable this check)",
|
||||
false, false, (int) true, &tmp);
|
||||
mca_plm_ccp_component.want_path_check = OPAL_INT_TO_BOOL(tmp);
|
||||
|
||||
mca_base_param_reg_string(comp, "stdout_file",
|
||||
"Path and file name for stdout on cluster nodes. "
|
||||
"By default, stdout will be sent to Job Scheduler. "
|
||||
"If no path specified, the user home path will be used. "
|
||||
"UNC path will not work for this param. ",
|
||||
false, false, NULL,
|
||||
&mca_plm_ccp_component.stdout_file);
|
||||
|
||||
mca_base_param_reg_string(comp, "stderr_file",
|
||||
"Path and file name for stderr on cluster nodes. "
|
||||
"By default, stderr will be sent to Job Scheduler. "
|
||||
"If no path specified, the user home path will be used. "
|
||||
"UNC path will not work for this param. ",
|
||||
false, false, NULL,
|
||||
&mca_plm_ccp_component.stderr_file);
|
||||
|
||||
mca_base_param_reg_string(comp, "job_name",
|
||||
"The job name for displaying in the scheduler. "
|
||||
"It is set to the application name by default.",
|
||||
false, false, NULL,
|
||||
&mca_plm_ccp_component.job_name);
|
||||
|
||||
mca_plm_ccp_component.checked_paths = NULL;
|
||||
|
||||
return ORTE_SUCCESS;
|
||||
}
|
||||
|
||||
|
||||
static int plm_ccp_close(void)
|
||||
{
|
||||
return ORTE_SUCCESS;
|
||||
}
|
||||
|
||||
|
||||
static int orte_plm_ccp_component_query(mca_base_module_t **module, int *priority)
|
||||
{
|
||||
ICluster* pCluster = NULL;
|
||||
HRESULT hr = S_OK;
|
||||
|
||||
/* CCP is not thread safe. Use the apartment model. */
|
||||
CoInitializeEx(NULL, COINIT_APARTMENTTHREADED);
|
||||
|
||||
/* Try to create the Cluster object. */
|
||||
hr = CoCreateInstance( __uuidof(Cluster),
|
||||
NULL,
|
||||
CLSCTX_INPROC_SERVER,
|
||||
__uuidof(ICluster),
|
||||
reinterpret_cast<void **> (&pCluster) );
|
||||
if (FAILED(hr)) {
|
||||
/* We are not Windows clusters, don't select us.*/
|
||||
*module = NULL;
|
||||
return ORTE_ERROR;
|
||||
}
|
||||
|
||||
/* if we are NOT an HNP, then don't select us */
|
||||
if (!ORTE_PROC_IS_HNP) {
|
||||
pCluster->Release();
|
||||
*module = NULL;
|
||||
return ORTE_ERROR;
|
||||
}
|
||||
|
||||
/* We are Windows clusters and this is HNP. */
|
||||
pCluster->Release();
|
||||
*priority = mca_plm_ccp_component.priority;
|
||||
*module = (mca_base_module_t *) &orte_plm_ccp_module;
|
||||
return ORTE_SUCCESS;
|
||||
}
|
@ -1,761 +0,0 @@
|
||||
/*
|
||||
* Copyright (c) 2004-2007 The University of Tennessee and The University
|
||||
* of Tennessee Research Foundation. All rights
|
||||
* reserved.
|
||||
* Copyright (c) 2004-2010 High Performance Computing Center Stuttgart,
|
||||
* University of Stuttgart. All rights reserved.
|
||||
* Copyright (c) 2011 Los Alamos National Security, LLC. All rights
|
||||
* reserved.
|
||||
* $COPYRIGHT$
|
||||
*
|
||||
* Additional copyrights may follow
|
||||
*
|
||||
* $HEADER$
|
||||
*
|
||||
*/
|
||||
|
||||
#include "orte_config.h"
|
||||
#include "orte/constants.h"
|
||||
#include "orte/types.h"
|
||||
|
||||
#ifdef HAVE_UNISTD_H
|
||||
#include <unistd.h>
|
||||
#endif
|
||||
#include <signal.h>
|
||||
#ifdef HAVE_SYS_TYPES_H
|
||||
#include <sys/types.h>
|
||||
#endif
|
||||
#ifdef HAVE_SYS_STAT_H
|
||||
#include <sys/stat.h>
|
||||
#endif
|
||||
#ifdef HAVE_SYS_WAIT_H
|
||||
#include <sys/wait.h>
|
||||
#endif
|
||||
#ifdef HAVE_SCHED_H
|
||||
#include <sched.h>
|
||||
#endif
|
||||
#ifdef HAVE_SYS_TIME_H
|
||||
#include <sys/time.h>
|
||||
#endif
|
||||
#include <errno.h>
|
||||
#include <comutil.h>
|
||||
|
||||
#include "opal/mca/installdirs/installdirs.h"
|
||||
#include "opal/mca/event/event.h"
|
||||
#include "opal/util/argv.h"
|
||||
#include "opal/util/output.h"
|
||||
#include "opal/util/opal_environ.h"
|
||||
#include "opal/util/basename.h"
|
||||
#include "opal/mca/base/mca_base_param.h"
|
||||
|
||||
#include "orte/util/name_fns.h"
|
||||
#include "orte/runtime/orte_globals.h"
|
||||
#include "orte/runtime/orte_wait.h"
|
||||
#include "orte/mca/errmgr/errmgr.h"
|
||||
#include "orte/mca/rmaps/rmaps.h"
|
||||
|
||||
#include "orte/mca/plm/plm.h"
|
||||
#include "orte/mca/plm/base/plm_private.h"
|
||||
#include "plm_ccp.h"
|
||||
|
||||
|
||||
/* Import the Windows CCP API. */
|
||||
#import "ccpapi.tlb" named_guids no_namespace raw_interfaces_only \
|
||||
rename("SetEnvironmentVariable","SetEnvVar") \
|
||||
rename("GetJob", "GetSingleJob") \
|
||||
rename("AddJob", "AddSingleJob")
|
||||
|
||||
/* Include the library for ::ConvertBSTRToString */
|
||||
#pragma comment(lib, "comsuppw.lib")
|
||||
|
||||
/*
|
||||
* Local functions
|
||||
*/
|
||||
static int plm_ccp_init(void);
|
||||
static int plm_ccp_launch_job(orte_job_t *jdata);
|
||||
static int plm_ccp_terminate_orteds();
|
||||
static int plm_ccp_signal_job(orte_jobid_t jobid, int32_t signal);
|
||||
static int plm_ccp_finalize(void);
|
||||
|
||||
static int plm_ccp_connect(ICluster* pCluster);
|
||||
static int plm_ccp_disconnect(void);
|
||||
|
||||
void plm_get_cluster_message(ICluster* pCluster);
|
||||
static char *plm_ccp_commandline(char *prefix, char *node_name, int argc, char **argv);
|
||||
|
||||
/*
|
||||
* Global variable
|
||||
*/
|
||||
orte_plm_base_module_t orte_plm_ccp_module = {
|
||||
plm_ccp_init,
|
||||
orte_plm_base_set_hnp_name,
|
||||
plm_ccp_launch_job,
|
||||
NULL,
|
||||
orte_plm_base_orted_terminate_job,
|
||||
plm_ccp_terminate_orteds,
|
||||
orte_plm_base_orted_kill_local_procs,
|
||||
plm_ccp_signal_job,
|
||||
plm_ccp_finalize
|
||||
};
|
||||
|
||||
|
||||
/**
|
||||
* Init the module
|
||||
*/
|
||||
static int plm_ccp_init(void)
|
||||
{
|
||||
int rc;
|
||||
|
||||
if (ORTE_SUCCESS != (rc = orte_plm_base_comm_start())) {
|
||||
ORTE_ERROR_LOG(rc);
|
||||
}
|
||||
|
||||
/* we assign daemon nodes at launch */
|
||||
orte_plm_globals.daemon_nodes_assigned_at_launch = true;
|
||||
|
||||
return rc;
|
||||
}
|
||||
|
||||
|
||||
/* When working in this function, ALWAYS jump to "cleanup" if
|
||||
* you encounter an error so that orterun will be woken up and
|
||||
* the job can cleanly terminate
|
||||
*/
|
||||
static int plm_ccp_launch_job(orte_job_t *jdata)
|
||||
{
|
||||
orte_app_context_t *app;
|
||||
orte_node_t *node;
|
||||
orte_std_cntr_t launched = 0, i;
|
||||
|
||||
orte_job_map_t *map = NULL;
|
||||
int argc, rc, proc_vpid_index;
|
||||
char *param, **env = NULL, *var, **argv = NULL;
|
||||
bool connected = false;
|
||||
char *bin_base = NULL, *lib_base = NULL, *command_line;
|
||||
|
||||
struct timeval completionstop, launchstart, launchstop;
|
||||
struct timeval jobstart, jobstop;
|
||||
int maxtime=0, mintime=99999999, maxiter = 0, miniter = 0, deltat;
|
||||
float avgtime=0.0;
|
||||
bool failed_launch = true;
|
||||
mode_t current_umask;
|
||||
IClusterEnumerable* pNodesCollection = NULL;
|
||||
IEnumVARIANT* pNodes = NULL;
|
||||
VARIANT v;
|
||||
|
||||
INode* pNode = NULL;
|
||||
HRESULT hr = S_OK;
|
||||
ICluster* pCluster = NULL;
|
||||
IJob* pJob = NULL;
|
||||
long job_id, num_processors = 0, idle_processors = 0;
|
||||
IClusterCounter* pClusterCounter = NULL;
|
||||
ITask* pTask = NULL;
|
||||
JobPriority job_priority = JobPriority_Normal;
|
||||
|
||||
orte_jobid_t failed_job;
|
||||
orte_job_state_t job_state = ORTE_JOB_STATE_NEVER_LAUNCHED;
|
||||
orte_job_t *daemons;
|
||||
|
||||
/* default to declaring the daemon launch failed */
|
||||
failed_job = ORTE_PROC_MY_NAME->jobid;
|
||||
|
||||
/* check for timing request - get start time if so */
|
||||
if (orte_timing) {
|
||||
if (0 != gettimeofday(&jobstart, NULL)) {
|
||||
OPAL_OUTPUT_VERBOSE((1, orte_plm_globals.output,
|
||||
"plm_ccp: could not obtain job start time"));
|
||||
}
|
||||
}
|
||||
|
||||
/* if we don't want to launch, then don't attempt to
|
||||
* launch the daemons - the user really wants to just
|
||||
* look at the proposed process map
|
||||
*/
|
||||
if (orte_do_not_launch) {
|
||||
goto launch_apps;
|
||||
}
|
||||
|
||||
/* start by launching the virtual machine */
|
||||
daemons = orte_get_job_data_object(ORTE_PROC_MY_NAME->jobid);
|
||||
if (ORTE_SUCCESS != (rc = orte_plm_base_setup_virtual_machine(jdata))) {
|
||||
ORTE_ERROR_LOG(rc);
|
||||
goto cleanup;
|
||||
}
|
||||
|
||||
OPAL_OUTPUT_VERBOSE((1, orte_plm_globals.output,
|
||||
"%s plm:rsh: launching vm",
|
||||
ORTE_NAME_PRINT(ORTE_PROC_MY_NAME)));
|
||||
|
||||
/* Get the map for this job */
|
||||
if (NULL == (map = daemons->map)) {
|
||||
ORTE_ERROR_LOG(ORTE_ERR_NOT_FOUND);
|
||||
rc = ORTE_ERR_NOT_FOUND;
|
||||
goto cleanup;
|
||||
}
|
||||
|
||||
if (0 == map->num_new_daemons) {
|
||||
/* have all the daemons we need - launch app */
|
||||
goto launch_apps;
|
||||
}
|
||||
|
||||
/* add the daemon command (as specified by user) */
|
||||
argc = 0;
|
||||
argv = NULL;
|
||||
orte_plm_base_setup_orted_cmd(&argc, &argv);
|
||||
|
||||
/* Add basic orted command line options */
|
||||
orte_plm_base_orted_append_basic_args(&argc, &argv, "env",
|
||||
&proc_vpid_index,
|
||||
NULL);
|
||||
|
||||
if (0 < opal_output_get_verbosity(orte_plm_globals.output)) {
|
||||
param = opal_argv_join(argv, ' ');
|
||||
OPAL_OUTPUT_VERBOSE((1, orte_plm_globals.output,
|
||||
"%s plm:ccp: final top-level argv:\n\t%s",
|
||||
ORTE_NAME_PRINT(ORTE_PROC_MY_NAME),
|
||||
(NULL == param) ? "NULL" : param));
|
||||
if (NULL != param) free(param);
|
||||
}
|
||||
|
||||
/* CCP is not thread safe. Use the apartment model. */
|
||||
CoInitializeEx(NULL, COINIT_APARTMENTTHREADED);
|
||||
|
||||
/* Create the Cluster object. */
|
||||
hr = CoCreateInstance( __uuidof(Cluster),
|
||||
NULL,
|
||||
CLSCTX_INPROC_SERVER,
|
||||
__uuidof(ICluster),
|
||||
reinterpret_cast<void **> (&pCluster) );
|
||||
if (FAILED(hr)) {
|
||||
opal_output(orte_plm_globals.output,
|
||||
"plm:ccp: failed to create cluster object!");
|
||||
goto cleanup;
|
||||
}
|
||||
|
||||
/* Connect to the head node. */
|
||||
rc = plm_ccp_connect(pCluster);
|
||||
if (ORTE_SUCCESS != rc) {
|
||||
goto cleanup;
|
||||
}
|
||||
connected = true;
|
||||
|
||||
hr = pCluster->CreateJob(&pJob);
|
||||
if (FAILED(hr)) {
|
||||
plm_get_cluster_message(pCluster);
|
||||
opal_output(orte_plm_globals.output,
|
||||
"plm:ccp:failed to create cluster object!");
|
||||
goto cleanup;
|
||||
}
|
||||
/* Figure out the basenames for the libdir and bindir. There is a
|
||||
lengthy comment about this in plm_rsh_module.c explaining all
|
||||
the rationale for how / why we're doing this. */
|
||||
lib_base = opal_basename(opal_install_dirs.libdir);
|
||||
bin_base = opal_basename(opal_install_dirs.bindir);
|
||||
|
||||
/* setup environment */
|
||||
env = opal_argv_copy(orte_launch_environ);
|
||||
|
||||
/* add our umask -- see big note in orted.c */
|
||||
current_umask = _umask(0);
|
||||
_umask(current_umask);
|
||||
asprintf(&var, "0%o", current_umask);
|
||||
opal_setenv("ORTE_DAEMON_UMASK_VALUE", var, true, &env);
|
||||
free(var);
|
||||
|
||||
/* If we have a prefix, then modify the PATH and
|
||||
LD_LIBRARY_PATH environment variables. We only allow
|
||||
a single prefix to be specified. Since there will
|
||||
always be at least one app_context, we take it from
|
||||
there
|
||||
*/
|
||||
app = (orte_app_context_t*)opal_pointer_array_get_item(jdata->apps, 0);
|
||||
if (NULL != app->prefix_dir) {
|
||||
char *newenv;
|
||||
|
||||
for (i = 0; NULL != env && NULL != env[i]; ++i) {
|
||||
/* Reset PATH */
|
||||
if (0 == strncmp("PATH=", env[i], 5)) {
|
||||
asprintf(&newenv, "%s/%s:%s",
|
||||
app->prefix_dir, bin_base, env[i] + 5);
|
||||
OPAL_OUTPUT_VERBOSE((1, orte_plm_globals.output,
|
||||
"%s plm:ccp: resetting PATH: %s",
|
||||
ORTE_NAME_PRINT(ORTE_PROC_MY_NAME),
|
||||
newenv));
|
||||
opal_setenv("PATH", newenv, true, &env);
|
||||
free(newenv);
|
||||
}
|
||||
|
||||
/* Reset LD_LIBRARY_PATH */
|
||||
else if (0 == strncmp("LD_LIBRARY_PATH=", env[i], 16)) {
|
||||
asprintf(&newenv, "%s/%s:%s",
|
||||
app->prefix_dir, lib_base, env[i] + 16);
|
||||
OPAL_OUTPUT_VERBOSE((1, orte_plm_globals.output,
|
||||
"%s plm:ccp: resetting LD_LIBRARY_PATH: %s",
|
||||
ORTE_NAME_PRINT(ORTE_PROC_MY_NAME),
|
||||
newenv));
|
||||
opal_setenv("LD_LIBRARY_PATH", newenv, true, &env);
|
||||
free(newenv);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/* This has already been done in RAS, but I have to do it again here.
|
||||
* Because the node structure doesn't have num_processor member. */
|
||||
|
||||
/* Get the collection of nodes. */
|
||||
hr = pCluster->get_ComputeNodes(&pNodesCollection);
|
||||
|
||||
/* Get the enumerator used to iterate through the collection. */
|
||||
hr = pNodesCollection->GetEnumerator(&pNodes);
|
||||
|
||||
VariantInit(&v);
|
||||
|
||||
int *num_procs;
|
||||
num_procs = (int *) malloc(sizeof(int)*map->num_nodes);
|
||||
|
||||
/* Loop through the collection. */
|
||||
while (hr = pNodes->Next(1, &v, NULL) == S_OK) {
|
||||
v.pdispVal->QueryInterface(IID_INode, reinterpret_cast<void **> (&pNode));
|
||||
|
||||
/* Iterate through each of the nodes and check to sum up all the processors. */
|
||||
for (i = 0; i < map->nodes->size; i++) {
|
||||
if (NULL == (node = (orte_node_t*)opal_pointer_array_get_item(map->nodes, i))) {
|
||||
continue;
|
||||
}
|
||||
|
||||
BSTR node_name;
|
||||
hr = pNode->get_Name(&node_name);
|
||||
|
||||
if( 0 == strcmp(_com_util::ConvertBSTRToString(node_name), node->name)) {
|
||||
/* Get available number of processors on required node. */
|
||||
hr = pNode->get_NumberOfIdleProcessors(&idle_processors);
|
||||
num_procs[i] = idle_processors;
|
||||
num_processors += idle_processors;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
if(NULL != mca_plm_ccp_component.job_name){
|
||||
pJob->put_Name(_bstr_t(mca_plm_ccp_component.job_name));
|
||||
} else {
|
||||
pJob->put_Name(_bstr_t((*app).app));
|
||||
}
|
||||
|
||||
pJob->put_MinimumNumberOfProcessors(num_processors);
|
||||
if (FAILED(hr)) {
|
||||
OPAL_OUTPUT_VERBOSE((1, orte_plm_globals.output,
|
||||
"plm:ccp:failed to put min num of processors!"));
|
||||
goto cleanup;
|
||||
}
|
||||
|
||||
pJob->put_MaximumNumberOfProcessors(num_processors);
|
||||
if (FAILED(hr)) {
|
||||
OPAL_OUTPUT_VERBOSE((1, orte_plm_globals.output,
|
||||
"plm:ccp:failed to put max num of processors!"));
|
||||
goto cleanup;
|
||||
}
|
||||
|
||||
hr = pJob->put_Priority(job_priority);
|
||||
if (FAILED(hr)) {
|
||||
OPAL_OUTPUT_VERBOSE((1, orte_plm_globals.output,
|
||||
"plm:ccp:failed to set proiority!"));
|
||||
goto cleanup;
|
||||
}
|
||||
|
||||
hr = pJob->SetExtendedJobTerm(_bstr_t(L"extended terms"), _bstr_t(L"TermValue"));
|
||||
|
||||
/* set the job state to indicate we attempted to launch */
|
||||
job_state = ORTE_JOB_STATE_FAILED_TO_START;
|
||||
|
||||
/* Iterate through each of the nodes and spin
|
||||
* up a daemon.
|
||||
*/
|
||||
for (i = 0; i < map->nodes->size; i++) {
|
||||
char* vpid_string;
|
||||
if (NULL == (node = (orte_node_t*)opal_pointer_array_get_item(map->nodes, i))) {
|
||||
continue;
|
||||
}
|
||||
|
||||
/* if this daemon already exists, don't launch it! */
|
||||
if (node->daemon_launched) {
|
||||
continue;
|
||||
}
|
||||
|
||||
OPAL_OUTPUT_VERBOSE((1, orte_plm_globals.output,
|
||||
"%s plm:ccp: launching on node %s",
|
||||
ORTE_NAME_PRINT(ORTE_PROC_MY_NAME),
|
||||
node->name));
|
||||
|
||||
/* setup process name */
|
||||
rc = orte_util_convert_vpid_to_string(&vpid_string, node->daemon->name.vpid);
|
||||
if (ORTE_SUCCESS != rc) {
|
||||
OPAL_OUTPUT_VERBOSE((1, orte_plm_globals.output,
|
||||
"plm:ccp: unable to get daemon vpid as string"));
|
||||
exit(-1);
|
||||
}
|
||||
free(argv[proc_vpid_index]);
|
||||
argv[proc_vpid_index] = strdup(vpid_string);
|
||||
free(vpid_string);
|
||||
|
||||
/* exec the daemon */
|
||||
if (0 < opal_output_get_verbosity(orte_plm_globals.output)) {
|
||||
param = opal_argv_join(argv, ' ');
|
||||
OPAL_OUTPUT_VERBOSE((1, orte_plm_globals.output,
|
||||
"%s plm:ccp: executing:\n\t%s",
|
||||
ORTE_NAME_PRINT(ORTE_PROC_MY_NAME),
|
||||
(NULL == param) ? "NULL" : param));
|
||||
if (NULL != param) free(param);
|
||||
}
|
||||
|
||||
/* check for timing request - get start time if so */
|
||||
if (orte_timing) {
|
||||
if (0 != gettimeofday(&launchstart, NULL)) {
|
||||
OPAL_OUTPUT_VERBOSE((1, orte_plm_globals.output,
|
||||
"plm_ccp: could not obtain start time"));
|
||||
launchstart.tv_sec = 0;
|
||||
launchstart.tv_usec = 0;
|
||||
}
|
||||
}
|
||||
/* Set terms for task. */
|
||||
hr = pCluster->CreateTask(&pTask);
|
||||
if (FAILED(hr)) {
|
||||
plm_get_cluster_message(pCluster);
|
||||
OPAL_OUTPUT_VERBOSE((1, orte_plm_globals.output,
|
||||
"plm:ccp:failed to create task object!"));
|
||||
goto cleanup;
|
||||
}
|
||||
|
||||
pTask->put_MinimumNumberOfProcessors(num_procs[i]);
|
||||
if (FAILED(hr)) {
|
||||
OPAL_OUTPUT_VERBOSE((1, orte_plm_globals.output,
|
||||
"plm:ccp:failed to create task object!"));
|
||||
goto cleanup;
|
||||
}
|
||||
|
||||
pTask->put_MaximumNumberOfProcessors(num_procs[i]);
|
||||
if (FAILED(hr)) {
|
||||
OPAL_OUTPUT_VERBOSE((1, orte_plm_globals.output,
|
||||
"plm:ccp:failed to create task object!"));
|
||||
goto cleanup;
|
||||
}
|
||||
|
||||
pTask->put_RequiredNodes(_bstr_t(node->name));
|
||||
if (FAILED(hr)) {
|
||||
OPAL_OUTPUT_VERBOSE((1, orte_plm_globals.output,
|
||||
"plm:ccp:failed to set required nodes!"));
|
||||
goto cleanup;
|
||||
}
|
||||
|
||||
/* Prepare the command line a little bit. */
|
||||
command_line = plm_ccp_commandline(app->prefix_dir, node->name, argc, argv);
|
||||
|
||||
hr = pTask->put_CommandLine(_bstr_t(command_line));
|
||||
if (FAILED(hr)) {
|
||||
OPAL_OUTPUT_VERBOSE((1, orte_plm_globals.output,
|
||||
"plm:ccp:failed to put command line!"));
|
||||
goto cleanup;
|
||||
}
|
||||
|
||||
if( NULL != mca_plm_ccp_component.stdout_file ) {
|
||||
hr = pTask->put_Stdout(_bstr_t(mca_plm_ccp_component.stdout_file));
|
||||
if (FAILED(hr)) {
|
||||
OPAL_OUTPUT_VERBOSE((1, orte_plm_globals.output,
|
||||
"plm:ccp:failed to set stdout!"));
|
||||
goto cleanup;
|
||||
}
|
||||
}
|
||||
|
||||
if( NULL != mca_plm_ccp_component.stderr_file) {
|
||||
hr = pTask->put_Stderr(_bstr_t(mca_plm_ccp_component.stderr_file));
|
||||
if (FAILED(hr)) {
|
||||
OPAL_OUTPUT_VERBOSE((1, orte_plm_globals.output,
|
||||
"plm:ccp:failed to set stderr!"));
|
||||
goto cleanup;
|
||||
}
|
||||
}
|
||||
|
||||
hr = pJob->AddTask(pTask);
|
||||
if (FAILED(hr)) {
|
||||
OPAL_OUTPUT_VERBOSE((1, orte_plm_globals.output,
|
||||
"plm:ccp:failed to add task!"));
|
||||
goto cleanup;
|
||||
}
|
||||
|
||||
|
||||
/* Allow some progress to occur */
|
||||
opal_event_loop(orte_event_base, OPAL_EVLOOP_NONBLOCK);
|
||||
|
||||
launched++;
|
||||
|
||||
pTask->Release();
|
||||
}
|
||||
|
||||
/* Add job to the queue. */
|
||||
hr = pCluster->QueueJob(pJob, NULL, NULL, VARIANT_TRUE, 0, &job_id);
|
||||
if (SUCCEEDED(hr)) {
|
||||
OPAL_OUTPUT_VERBOSE((1, orte_plm_globals.output,
|
||||
"Added job %d to scheduling queue.\n", job_id));
|
||||
}else {
|
||||
plm_get_cluster_message(pCluster);
|
||||
}
|
||||
|
||||
OPAL_OUTPUT_VERBOSE((1, orte_plm_globals.output,
|
||||
"%s plm:ccp:launch: finished spawning orteds",
|
||||
ORTE_NAME_PRINT(ORTE_PROC_MY_NAME)));
|
||||
|
||||
/* wait for daemons to callback */
|
||||
if (ORTE_SUCCESS != (rc = orte_plm_base_daemon_callback(map->num_new_daemons))) {
|
||||
OPAL_OUTPUT_VERBOSE((1, orte_plm_globals.output,
|
||||
"%s plm:ccp: daemon launch failed for job %s on error %s",
|
||||
ORTE_NAME_PRINT(ORTE_PROC_MY_NAME),
|
||||
ORTE_JOBID_PRINT(jdata->jobid), ORTE_ERROR_NAME(rc)));
|
||||
goto cleanup;
|
||||
}
|
||||
|
||||
launch_apps:
|
||||
/* setup the job */
|
||||
if (ORTE_SUCCESS != (rc = orte_plm_base_setup_job(jdata))) {
|
||||
ORTE_ERROR_LOG(rc);
|
||||
failed_job = jdata->jobid;
|
||||
goto cleanup;
|
||||
}
|
||||
failed_job = jdata->jobid;
|
||||
if (ORTE_SUCCESS != (rc = orte_plm_base_launch_apps(jdata->jobid))) {
|
||||
OPAL_OUTPUT_VERBOSE((1, orte_plm_globals.output,
|
||||
"%s plm:ccp: launch of apps failed for job %s on error %s",
|
||||
ORTE_NAME_PRINT(ORTE_PROC_MY_NAME),
|
||||
ORTE_JOBID_PRINT(jdata->jobid), ORTE_ERROR_NAME(rc)));
|
||||
goto cleanup;
|
||||
}
|
||||
|
||||
/* if we get here, then everything launched okay - record that fact */
|
||||
failed_launch = false;
|
||||
|
||||
/* check for timing request - get stop time for launch completion and report */
|
||||
if (orte_timing) {
|
||||
if (0 != gettimeofday(&completionstop, NULL)) {
|
||||
OPAL_OUTPUT_VERBOSE((1, orte_plm_globals.output,
|
||||
"plm_ccp: could not obtain completion stop time"));
|
||||
} else {
|
||||
deltat = (launchstop.tv_sec - launchstart.tv_sec)*1000000 +
|
||||
(launchstop.tv_usec - launchstart.tv_usec);
|
||||
OPAL_OUTPUT_VERBOSE((1, orte_plm_globals.output,
|
||||
"plm_ccp: launch completion required %d usec", deltat));
|
||||
}
|
||||
OPAL_OUTPUT_VERBOSE((1, orte_plm_globals.output,
|
||||
"plm_ccp: Launch statistics:"));
|
||||
OPAL_OUTPUT_VERBOSE((1, orte_plm_globals.output,
|
||||
"plm_ccp: Average time to launch an orted: %f usec", avgtime));
|
||||
OPAL_OUTPUT_VERBOSE((1, orte_plm_globals.output,
|
||||
"plm_ccp: Max time to launch an orted: %d usec at iter %d", maxtime, maxiter));
|
||||
OPAL_OUTPUT_VERBOSE((1, orte_plm_globals.output,
|
||||
"plm_ccp: Min time to launch an orted: %d usec at iter %d", mintime, miniter));
|
||||
}
|
||||
|
||||
|
||||
cleanup:
|
||||
if (NULL != argv) {
|
||||
opal_argv_free(argv);
|
||||
}
|
||||
|
||||
if (NULL != env) {
|
||||
opal_argv_free(env);
|
||||
}
|
||||
|
||||
if (connected) {
|
||||
plm_ccp_disconnect();
|
||||
}
|
||||
|
||||
if (NULL != lib_base) {
|
||||
free(lib_base);
|
||||
}
|
||||
|
||||
if (NULL != bin_base) {
|
||||
free(bin_base);
|
||||
}
|
||||
|
||||
/* check for failed launch - if so, force terminate */
|
||||
if (failed_launch) {
|
||||
if (ORTE_ERR_SILENT == rc) {
|
||||
orte_errmgr.update_state(failed_job, ORTE_JOB_STATE_SILENT_ABORT,
|
||||
NULL, ORTE_PROC_STATE_UNDEF,
|
||||
0, ORTE_ERROR_DEFAULT_EXIT_CODE);
|
||||
} else {
|
||||
orte_errmgr.update_state(failed_job, job_state,
|
||||
NULL, ORTE_PROC_STATE_UNDEF,
|
||||
0, ORTE_ERROR_DEFAULT_EXIT_CODE);
|
||||
}
|
||||
}
|
||||
|
||||
/* check for timing request - get stop time and process if so */
|
||||
if (orte_timing) {
|
||||
if (0 != gettimeofday(&jobstop, NULL)) {
|
||||
OPAL_OUTPUT_VERBOSE((1, orte_plm_globals.output,
|
||||
"plm_ccp: could not obtain stop time"));
|
||||
} else {
|
||||
deltat = (jobstop.tv_sec - jobstart.tv_sec)*1000000 +
|
||||
(jobstop.tv_usec - jobstart.tv_usec);
|
||||
OPAL_OUTPUT_VERBOSE((1, orte_plm_globals.output,
|
||||
"plm_ccp: launch of entire job required %d usec", deltat));
|
||||
}
|
||||
}
|
||||
|
||||
OPAL_OUTPUT_VERBOSE((1, orte_plm_globals.output,
|
||||
"%s plm:ccp:launch: finished",
|
||||
ORTE_NAME_PRINT(ORTE_PROC_MY_NAME)));
|
||||
|
||||
return rc;
|
||||
}
|
||||
|
||||
|
||||
/**
|
||||
* Terminate the orteds for a given job
|
||||
*/
|
||||
int plm_ccp_terminate_orteds()
|
||||
{
|
||||
int rc;
|
||||
|
||||
/* now tell them to die */
|
||||
if (orte_abnormal_term_ordered) {
|
||||
/* cannot know if a daemon is able to
|
||||
* tell us it died, so just ensure they
|
||||
* all terminate
|
||||
*/
|
||||
if (ORTE_SUCCESS != (rc = orte_plm_base_orted_exit(ORTE_DAEMON_HALT_VM_CMD))) {
|
||||
ORTE_ERROR_LOG(rc);
|
||||
}
|
||||
} else {
|
||||
/* we need them to "phone home", though,
|
||||
* so we can know that they have exited
|
||||
*/
|
||||
if (ORTE_SUCCESS != (rc = orte_plm_base_orted_exit(ORTE_DAEMON_EXIT_CMD))) {
|
||||
ORTE_ERROR_LOG(rc);
|
||||
}
|
||||
}
|
||||
|
||||
return rc;
|
||||
}
|
||||
|
||||
|
||||
static int plm_ccp_signal_job(orte_jobid_t jobid, int32_t signal)
|
||||
{
|
||||
int rc;
|
||||
|
||||
/* order them to pass this signal to their local procs */
|
||||
if (ORTE_SUCCESS != (rc = orte_plm_base_orted_signal_local_procs(jobid, signal))) {
|
||||
ORTE_ERROR_LOG(rc);
|
||||
}
|
||||
|
||||
return rc;
|
||||
}
|
||||
|
||||
|
||||
/*
|
||||
* Free stuff
|
||||
*/
|
||||
static int plm_ccp_finalize(void)
|
||||
{
|
||||
int rc;
|
||||
|
||||
/* cleanup any pending recvs */
|
||||
if (ORTE_SUCCESS != (rc = orte_plm_base_comm_stop())) {
|
||||
ORTE_ERROR_LOG(rc);
|
||||
}
|
||||
|
||||
return ORTE_SUCCESS;
|
||||
}
|
||||
|
||||
|
||||
static int plm_ccp_connect(ICluster* pCluster)
|
||||
{
|
||||
size_t i, len;
|
||||
char *cluster_head = NULL;
|
||||
HRESULT hr = S_OK;
|
||||
|
||||
if (NULL == orte_ccp_headnode) {
|
||||
/* Get the cluster head nodes name */
|
||||
_dupenv_s(&cluster_head, &len, "LOGONSERVER");
|
||||
|
||||
if(cluster_head == NULL) {
|
||||
OPAL_OUTPUT_VERBOSE((1, orte_plm_globals.output,
|
||||
"plm:ccp:allocate: connot find cluster head node!"));
|
||||
return ORTE_ERROR;
|
||||
}
|
||||
|
||||
/* Get rid of the beginning '//'. */
|
||||
for( i = 0; i < len; i++){
|
||||
cluster_head[i] = cluster_head[i+2];
|
||||
cluster_head[i+2] = '\0';
|
||||
}
|
||||
} else {
|
||||
cluster_head = orte_ccp_headnode;
|
||||
}
|
||||
|
||||
/* Connect to the cluster's head node */
|
||||
hr = pCluster->Connect(_bstr_t(cluster_head));
|
||||
if (FAILED(hr)) {
|
||||
OPAL_OUTPUT_VERBOSE((1, orte_plm_globals.output,
|
||||
"plm:ccp:allocate: connection failed!"));
|
||||
return ORTE_ERROR;
|
||||
}
|
||||
OPAL_OUTPUT_VERBOSE((1, orte_plm_globals.output,
|
||||
"Connected to Cluster: %s. \n", cluster_head));
|
||||
return ORTE_SUCCESS;
|
||||
}
|
||||
|
||||
|
||||
static int plm_ccp_disconnect(void)
|
||||
{
|
||||
return ORTE_SUCCESS;
|
||||
}
|
||||
|
||||
|
||||
/* Generate the proper command line according to the env. */
|
||||
static char *plm_ccp_commandline(char *prefix, char *node_name, int argc, char **argv)
|
||||
{
|
||||
char *commandline;
|
||||
size_t i, len = 0;
|
||||
|
||||
for( i = 0; i < argc; i++ ) {
|
||||
len += strlen(argv[i]) + 1;
|
||||
}
|
||||
|
||||
if(NULL != prefix) {
|
||||
commandline = (char*)malloc(len + strlen(prefix) + 8);
|
||||
memset(commandline, 0, len + strlen(prefix) + 8);
|
||||
commandline[0] = '"';
|
||||
strcat(commandline, prefix);
|
||||
strcat(commandline, "\\bin\"\\");
|
||||
} else {
|
||||
commandline = (char*)malloc(len + 1);
|
||||
memset(commandline, 0, len + 1);
|
||||
}
|
||||
|
||||
|
||||
for(i=0;i<argc;i++) {
|
||||
|
||||
/* Append command args, and separate them with spaces. */
|
||||
strcat(commandline, argv[i]);
|
||||
|
||||
commandline[strlen(commandline)]=' ';
|
||||
}
|
||||
return commandline;
|
||||
}
|
||||
|
||||
|
||||
void plm_get_cluster_message(ICluster* pCluster)
|
||||
{
|
||||
HRESULT hr = S_OK;
|
||||
BSTR message = NULL;
|
||||
|
||||
hr = pCluster->get_ErrorMessage(&message);
|
||||
if (SUCCEEDED(hr)) {
|
||||
OPAL_OUTPUT_VERBOSE((1, orte_plm_globals.output,
|
||||
_com_util::ConvertBSTRToString(message)));
|
||||
SysFreeString(message);
|
||||
}
|
||||
else {
|
||||
OPAL_OUTPUT_VERBOSE((1, orte_plm_globals.output,
|
||||
"pCluster->get_ErrorMessage failed.\n"));
|
||||
}
|
||||
}
|
Загрузка…
Ссылка в новой задаче
Block a user