8f1b9b463e
This commit was SVN r32577.
430 строки
14 KiB
C
430 строки
14 KiB
C
/*
|
|
* Copyright (c) 2004-2005 The Trustees of Indiana University and Indiana
|
|
* University Research and Technology
|
|
* Corporation. All rights reserved.
|
|
* Copyright (c) 2004-2011 The University of Tennessee and The University
|
|
* of Tennessee Research Foundation. All rights
|
|
* reserved.
|
|
* Copyright (c) 2004-2005 High Performance Computing Center Stuttgart,
|
|
* University of Stuttgart. All rights reserved.
|
|
* Copyright (c) 2004-2005 The Regents of the University of California.
|
|
* All rights reserved.
|
|
* Copyright (c) 2008-2012 Cisco Systems, Inc. All rights reserved.
|
|
* Copyright (c) 2012-2013 Los Alamos National Security, LLC.
|
|
* All rights reserved.
|
|
* Copyright (c) 2013-2014 Intel, Inc. All rights reserved.
|
|
* $COPYRIGHT$
|
|
*
|
|
* Additional copyrights may follow
|
|
*
|
|
* $HEADER$
|
|
*
|
|
*/
|
|
|
|
#include "orte_config.h"
|
|
#include "orte/constants.h"
|
|
|
|
#ifdef HAVE_UNISTD_H
|
|
#include <unistd.h>
|
|
#endif /* HAVE_UNISTD_H */
|
|
#ifdef HAVE_STRING_H
|
|
#include <string.h>
|
|
#endif /* HAVE_STRING_H */
|
|
#include <ctype.h>
|
|
#ifdef HAVE_NETDB_H
|
|
#include <netdb.h>
|
|
#endif
|
|
#ifdef HAVE_IFADDRS_H
|
|
#include <ifaddrs.h>
|
|
#endif
|
|
|
|
#include "opal/util/opal_environ.h"
|
|
#include "opal/util/output.h"
|
|
#include "opal/util/argv.h"
|
|
#include "opal/class/opal_pointer_array.h"
|
|
#include "opal/mca/dstore/dstore.h"
|
|
#include "opal/mca/hwloc/base/base.h"
|
|
#include "opal/util/printf.h"
|
|
#include "opal/mca/pmix/pmix.h"
|
|
|
|
#include "orte/mca/errmgr/errmgr.h"
|
|
#include "orte/mca/grpcomm/grpcomm.h"
|
|
#include "orte/mca/rml/rml.h"
|
|
#include "orte/util/proc_info.h"
|
|
#include "orte/util/show_help.h"
|
|
#include "orte/util/name_fns.h"
|
|
#include "orte/util/pre_condition_transports.h"
|
|
#include "orte/util/regex.h"
|
|
#include "orte/runtime/orte_globals.h"
|
|
#include "orte/runtime/orte_wait.h"
|
|
|
|
#include "orte/mca/ess/ess.h"
|
|
#include "orte/mca/ess/base/base.h"
|
|
#include "orte/mca/ess/pmi/ess_pmi.h"
|
|
|
|
static int rte_init(void);
|
|
static int rte_finalize(void);
|
|
static void rte_abort(int error_code, bool report);
|
|
|
|
orte_ess_base_module_t orte_ess_pmi_module = {
|
|
rte_init,
|
|
rte_finalize,
|
|
rte_abort,
|
|
NULL /* ft_event */
|
|
};
|
|
|
|
static bool app_init_complete=false;
|
|
|
|
/**** MODULE FUNCTIONS ****/
|
|
|
|
static int rte_init(void)
|
|
{
|
|
int ret;
|
|
char *error = NULL;
|
|
char *envar, *ev1, *ev2;
|
|
uint64_t unique_key[2];
|
|
char *cs_env, *string_key;
|
|
orte_jobid_t jobid;
|
|
char *rmluri;
|
|
opal_value_t *kv, kvn;
|
|
opal_list_t vals;
|
|
|
|
/* run the prolog */
|
|
if (ORTE_SUCCESS != (ret = orte_ess_base_std_prolog())) {
|
|
error = "orte_ess_base_std_prolog";
|
|
goto error;
|
|
}
|
|
|
|
#if OPAL_HAVE_HWLOC
|
|
/* get the topology */
|
|
if (NULL == opal_hwloc_topology) {
|
|
if (OPAL_SUCCESS != opal_hwloc_base_get_topology()) {
|
|
error = "topology discovery";
|
|
goto error;
|
|
}
|
|
}
|
|
#endif
|
|
|
|
if (ORTE_PROC_IS_DAEMON) { /* I am a daemon, launched by mpirun */
|
|
/* ensure that we always exit with a non-zero status
|
|
* so that Slurm and other such RMs will terminate the
|
|
* job if any daemon exits, whether normal termination or not
|
|
*/
|
|
ORTE_UPDATE_EXIT_STATUS(ORTE_ERROR_DEFAULT_EXIT_CODE);
|
|
|
|
/* we had to be given a jobid */
|
|
if (NULL == orte_ess_base_jobid) {
|
|
error = "missing jobid";
|
|
ret = ORTE_ERR_FATAL;
|
|
goto error;
|
|
}
|
|
if (ORTE_SUCCESS != (ret = orte_util_convert_string_to_jobid(&jobid, orte_ess_base_jobid))) {
|
|
ORTE_ERROR_LOG(ret);
|
|
error = "convert jobid";
|
|
goto error;
|
|
}
|
|
ORTE_PROC_MY_NAME->jobid = jobid;
|
|
|
|
/* if we weren't given it, get our global rank from PMI */
|
|
if (NULL == orte_ess_base_vpid) {
|
|
if (!opal_pmix.get_attr(PMIX_RANK, &kv)) {
|
|
error = "getting rank";
|
|
ret = ORTE_ERR_NOT_FOUND;
|
|
goto error;
|
|
}
|
|
ORTE_PROC_MY_NAME->vpid = kv->data.uint32 + 1; // compensate for orterun
|
|
OBJ_RELEASE(kv);
|
|
}
|
|
|
|
/* if we weren't given it, get universe size */
|
|
if (orte_ess_base_num_procs < 0) {
|
|
if (!opal_pmix.get_attr(PMIX_UNIV_SIZE, &kv)) {
|
|
error = "getting univ size";
|
|
ret = ORTE_ERR_NOT_FOUND;
|
|
goto error;
|
|
}
|
|
orte_process_info.num_procs = kv->data.uint32 + 1; // compensate for orterun
|
|
OBJ_RELEASE(kv);
|
|
}
|
|
|
|
/* complete setup */
|
|
if (ORTE_SUCCESS != (ret = orte_ess_base_orted_setup(NULL))) {
|
|
ORTE_ERROR_LOG(ret);
|
|
error = "orte_ess_base_orted_setup";
|
|
goto error;
|
|
}
|
|
return ORTE_SUCCESS;
|
|
|
|
}
|
|
|
|
/**** THE FOLLOWING ARE REQUIRED VALUES ***/
|
|
/* get our jobid from PMI */
|
|
if (!opal_pmix.get_attr(PMIX_JOBID, &kv)) {
|
|
error = "getting jobid";
|
|
ret = ORTE_ERR_NOT_FOUND;
|
|
goto error;
|
|
}
|
|
ORTE_PROC_MY_NAME->jobid = kv->data.uint32;
|
|
OBJ_RELEASE(kv);
|
|
|
|
/* get our global rank from PMI */
|
|
if (!opal_pmix.get_attr(PMIX_RANK, &kv)) {
|
|
error = "getting rank";
|
|
ret = ORTE_ERR_NOT_FOUND;
|
|
goto error;
|
|
}
|
|
ORTE_PROC_MY_NAME->vpid = kv->data.uint32;
|
|
OBJ_RELEASE(kv);
|
|
|
|
/* get our local rank from PMI */
|
|
if (!opal_pmix.get_attr(PMIX_LOCAL_RANK, &kv)) {
|
|
error = "getting local rank";
|
|
ret = ORTE_ERR_NOT_FOUND;
|
|
goto error;
|
|
}
|
|
orte_process_info.my_local_rank = (orte_local_rank_t)kv->data.uint16;
|
|
OBJ_RELEASE(kv);
|
|
|
|
/* get our node rank from PMI */
|
|
if (!opal_pmix.get_attr(PMIX_NODE_RANK, &kv)) {
|
|
error = "getting node rank";
|
|
ret = ORTE_ERR_NOT_FOUND;
|
|
goto error;
|
|
}
|
|
orte_process_info.my_node_rank = (orte_local_rank_t)kv->data.uint16;
|
|
|
|
/* get universe size */
|
|
if (!opal_pmix.get_attr(PMIX_UNIV_SIZE, &kv)) {
|
|
error = "getting univ size";
|
|
ret = ORTE_ERR_NOT_FOUND;
|
|
goto error;
|
|
}
|
|
orte_process_info.num_procs = kv->data.uint32;
|
|
OBJ_RELEASE(kv);
|
|
/* push into the environ for pickup in MPI layer for
|
|
* MPI-3 required info key
|
|
*/
|
|
asprintf(&ev1, "OMPI_MCA_orte_ess_num_procs=%d", orte_process_info.num_procs);
|
|
putenv(ev1);
|
|
asprintf(&ev2, "OMPI_APP_CTX_NUM_PROCS=%d", orte_process_info.num_procs);
|
|
putenv(ev2);
|
|
|
|
|
|
/* get our app number from PMI - ok if not found */
|
|
if (opal_pmix.get_attr(PMIX_APPNUM, &kv)) {
|
|
orte_process_info.app_num = kv->data.uint32;
|
|
OBJ_RELEASE(kv);
|
|
} else {
|
|
orte_process_info.app_num = 0;
|
|
}
|
|
|
|
/* get the number of local peers - required for wireup of
|
|
* shared memory BTL */
|
|
if (opal_pmix.get_attr(PMIX_LOCAL_SIZE, &kv)) {
|
|
orte_process_info.num_local_peers = kv->data.uint32 - 1; // want number besides ourselves
|
|
OBJ_RELEASE(kv);
|
|
} else {
|
|
orte_process_info.num_local_peers = 0;
|
|
}
|
|
|
|
/* setup transport keys in case the MPI layer needs them -
|
|
* we can use the jobfam and stepid as unique keys
|
|
* because they are unique values assigned by the RM
|
|
*/
|
|
unique_key[0] = (uint64_t)(((ORTE_PROC_MY_NAME->jobid & 0xffff0000))) >> 32;
|
|
unique_key[1] = (uint64_t)(((ORTE_PROC_MY_NAME->jobid & 0x0000ffff))) >> 32;
|
|
if (NULL == (string_key = orte_pre_condition_transports_print(unique_key))) {
|
|
ORTE_ERROR_LOG(ORTE_ERR_OUT_OF_RESOURCE);
|
|
return ORTE_ERR_OUT_OF_RESOURCE;
|
|
}
|
|
if (OPAL_SUCCESS != mca_base_var_env_name ("orte_precondition_transports", &cs_env)) {
|
|
ORTE_ERROR_LOG(ORTE_ERR_OUT_OF_RESOURCE);
|
|
return ORTE_ERR_OUT_OF_RESOURCE;
|
|
}
|
|
asprintf(&envar, "%s=%s", cs_env, string_key);
|
|
putenv(envar);
|
|
/* cannot free the envar as that messes up our environ */
|
|
free(cs_env);
|
|
free(string_key);
|
|
|
|
/* we don't need to force the routed system to pick the
|
|
* "direct" component as that should happen automatically
|
|
* in those cases where we are direct launched (i.e., no
|
|
* HNP is defined in the environment */
|
|
|
|
/* now use the default procedure to finish my setup */
|
|
if (ORTE_SUCCESS != (ret = orte_ess_base_app_setup(false))) {
|
|
ORTE_ERROR_LOG(ret);
|
|
error = "orte_ess_base_app_setup";
|
|
goto error;
|
|
}
|
|
|
|
/* setup process binding */
|
|
if (ORTE_SUCCESS != (ret = orte_ess_base_proc_binding())) {
|
|
error = "proc_binding";
|
|
goto error;
|
|
}
|
|
|
|
/* this needs to be set to enable debugger use when direct launched */
|
|
if (NULL == orte_process_info.my_daemon_uri) {
|
|
orte_standalone_operation = true;
|
|
}
|
|
|
|
/* set max procs */
|
|
if (orte_process_info.max_procs < orte_process_info.num_procs) {
|
|
orte_process_info.max_procs = orte_process_info.num_procs;
|
|
}
|
|
|
|
/*** PUSH DATA FOR OTHERS TO FIND ***/
|
|
|
|
/* if our URI was not provided by the system, then
|
|
* push our URI so others can find us */
|
|
OBJ_CONSTRUCT(&vals, opal_list_t);
|
|
if (OPAL_SUCCESS != opal_dstore.fetch(opal_dstore_internal, &OPAL_PROC_MY_NAME,
|
|
OPAL_DSTORE_URI, &vals)) {
|
|
/* construct the RTE string */
|
|
rmluri = orte_rml.get_contact_info();
|
|
/* push it out for others to use */
|
|
OBJ_CONSTRUCT(&kvn, opal_value_t);
|
|
kvn.key = strdup(OPAL_DSTORE_URI);
|
|
kvn.type = OPAL_STRING;
|
|
kvn.data.string = strdup(rmluri);
|
|
if (ORTE_SUCCESS != (ret = opal_pmix.put(PMIX_GLOBAL, &kvn))) {
|
|
error = "db store uri";
|
|
OBJ_DESTRUCT(&kvn);
|
|
goto error;
|
|
}
|
|
OBJ_DESTRUCT(&kvn);
|
|
free(rmluri);
|
|
}
|
|
OPAL_LIST_DESTRUCT(&vals);
|
|
|
|
/* push our hostname so others can find us, if they need to */
|
|
OBJ_CONSTRUCT(&kvn, opal_value_t);
|
|
kvn.key = strdup(OPAL_DSTORE_HOSTNAME);
|
|
kvn.type = OPAL_STRING;
|
|
kvn.data.string = strdup(orte_process_info.nodename);
|
|
if (ORTE_SUCCESS != (ret = opal_pmix.put(PMIX_GLOBAL, &kvn))) {
|
|
error = "db store hostname";
|
|
OBJ_DESTRUCT(&kvn);
|
|
goto error;
|
|
}
|
|
OBJ_DESTRUCT(&kvn);
|
|
|
|
/* if our local rank was not provided by the system, then
|
|
* push our local rank so others can access it */
|
|
OBJ_CONSTRUCT(&vals, opal_list_t);
|
|
if (OPAL_SUCCESS != opal_dstore.fetch(opal_dstore_internal, &OPAL_PROC_MY_NAME,
|
|
OPAL_DSTORE_LOCALRANK, &vals)) {
|
|
OBJ_CONSTRUCT(&kvn, opal_value_t);
|
|
kvn.key = strdup(OPAL_DSTORE_LOCALRANK);
|
|
kvn.type = OPAL_UINT16;
|
|
kvn.data.uint16 = orte_process_info.my_local_rank;
|
|
if (ORTE_SUCCESS != (ret = opal_pmix.put(PMIX_GLOBAL, &kvn))) {
|
|
error = "db store local rank";
|
|
OBJ_DESTRUCT(&kvn);
|
|
goto error;
|
|
}
|
|
OBJ_DESTRUCT(&kvn);
|
|
}
|
|
OPAL_LIST_DESTRUCT(&vals);
|
|
|
|
/* if our node rank was not provided by the system, then
|
|
* push our node rank so others can access it */
|
|
OBJ_CONSTRUCT(&vals, opal_list_t);
|
|
if (OPAL_SUCCESS != opal_dstore.fetch(opal_dstore_internal, &OPAL_PROC_MY_NAME,
|
|
OPAL_DSTORE_NODERANK, &vals)) {
|
|
OBJ_CONSTRUCT(&kvn, opal_value_t);
|
|
kvn.key = strdup(OPAL_DSTORE_NODERANK);
|
|
kvn.type = OPAL_UINT16;
|
|
kvn.data.uint16 = orte_process_info.my_node_rank;
|
|
if (ORTE_SUCCESS != (ret = opal_pmix.put(PMIX_GLOBAL, &kvn))) {
|
|
error = "db store node rank";
|
|
OBJ_DESTRUCT(&kvn);
|
|
goto error;
|
|
}
|
|
OBJ_DESTRUCT(&kvn);
|
|
}
|
|
OPAL_LIST_DESTRUCT(&vals);
|
|
|
|
/* if we are an ORTE app - and not an MPI app - then
|
|
* we need to exchange our connection info here.
|
|
* MPI_Init has its own modex, so we don't need to do
|
|
* two of them. However, if we don't do a modex at all,
|
|
* then processes have no way to communicate
|
|
*
|
|
* NOTE: only do this when the process originally launches.
|
|
* Cannot do this on a restart as the rest of the processes
|
|
* in the job won't be executing this step, so we would hang
|
|
*/
|
|
if (ORTE_PROC_IS_NON_MPI && !orte_do_not_barrier) {
|
|
opal_pmix.fence(NULL, 0);
|
|
}
|
|
|
|
/* flag that we completed init */
|
|
app_init_complete = true;
|
|
|
|
return ORTE_SUCCESS;
|
|
|
|
error:
|
|
if (ORTE_ERR_SILENT != ret && !orte_report_silent_errors) {
|
|
orte_show_help("help-orte-runtime.txt",
|
|
"orte_init:startup:internal-failure",
|
|
true, error, ORTE_ERROR_NAME(ret), ret);
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
static int rte_finalize(void)
|
|
{
|
|
int ret;
|
|
|
|
if (app_init_complete) {
|
|
/* if I am a daemon, finalize using the default procedure */
|
|
if (ORTE_PROC_IS_DAEMON) {
|
|
if (ORTE_SUCCESS != (ret = orte_ess_base_orted_finalize())) {
|
|
ORTE_ERROR_LOG(ret);
|
|
return ret;
|
|
}
|
|
} else {
|
|
/* mark us as finalized */
|
|
opal_pmix.finalize();
|
|
|
|
/* remove the envars that we pushed into environ
|
|
* so we leave that structure intact
|
|
*/
|
|
unsetenv("OMPI_MCA_routed");
|
|
unsetenv("OMPI_MCA_orte_precondition_transports");
|
|
unsetenv("OMPI_MCA_orte_ess_num_procs");
|
|
unsetenv("OMPI_APP_CTX_NUM_PROCS");
|
|
/* use the default app procedure to finish */
|
|
if (ORTE_SUCCESS != (ret = orte_ess_base_app_finalize())) {
|
|
ORTE_ERROR_LOG(ret);
|
|
return ret;
|
|
}
|
|
}
|
|
}
|
|
|
|
return ORTE_SUCCESS;
|
|
}
|
|
|
|
static void rte_abort(int status, bool report)
|
|
{
|
|
OPAL_OUTPUT_VERBOSE((1, orte_ess_base_framework.framework_output,
|
|
"%s ess:pmi:abort: abort with status %d",
|
|
ORTE_NAME_PRINT(ORTE_PROC_MY_NAME),
|
|
status));
|
|
|
|
/* PMI doesn't like NULL messages, but our interface
|
|
* doesn't provide one - so rig one up here
|
|
*/
|
|
opal_pmix.abort(status, "N/A");
|
|
|
|
/* - Clean out the global structures
|
|
* (not really necessary, but good practice) */
|
|
orte_proc_info_finalize();
|
|
|
|
/* Now Exit */
|
|
exit(status);
|
|
}
|