1
1
openmpi/orte/mca/odls/default/odls_default_module.c
Jeff Squyres eaab8d0062 * Ensure to set paffinity_enabled in all cases
* Ensure to set the mask value before we use it

This commit was SVN r23861.
2010-10-07 15:48:49 +00:00

1593 строки
63 KiB
C

/*
* Copyright (c) 2004-2007 The Trustees of Indiana University and Indiana
* University Research and Technology
* Corporation. All rights reserved.
* Copyright (c) 2004-2008 The University of Tennessee and The University
* of Tennessee Research Foundation. All rights
* reserved.
* Copyright (c) 2004-2005 High Performance Computing Center Stuttgart,
* University of Stuttgart. All rights reserved.
* Copyright (c) 2004-2005 The Regents of the University of California.
* All rights reserved.
* Copyright (c) 2007-2010 Oracle and/or its affiliates. All rights reserved.
* Copyright (c) 2007 Evergrid, Inc. All rights reserved.
* Copyright (c) 2008-2010 Cisco Systems, Inc. All rights reserved.
* Copyright (c) 2010 IBM Corporation. All rights reserved.
*
* $COPYRIGHT$
*
* Additional copyrights may follow
*
* $HEADER$
*/
/*
* There is a complicated sequence of events that occurs when the
* parent forks a child process that is intended to launch the target
* executable.
*
* Before the child process exec's the target executable, it might tri
* to set the affinity of that new child process according to a
* complex series of rules. This binding may fail in a myriad of
* different ways. A lot of this code deals with reporting that error
* occurately to the end user. This is a complex task in itself
* because the child process is not "really" an ORTE process -- all
* error reporting must be proxied up to the parent who can use normal
* ORTE error reporting mechanisms.
*
* Here's a high-level description of what is occurring in this file:
*
* - parent opens a pipe
* - parent forks a child
* - parent blocks reading on the pipe: the pipe will either close
* (indicating that the child successfully exec'ed) or the child will
* write some proxied error data up the pipe
*
* - the child tries to set affinity and do other housekeeping in
* preparation of exec'ing the target executable
* - if the child fails anywhere along the way, it sends a message up
* the pipe to the parent indicating what happened -- including a
* rendered error message detailing the problem (i.e., human-readable).
* - it is important that the child renders the error message: there
* are so many errors that are possible that the child is really the
* only entity that has enough information to make an accuate error string
* to report back to the user.
* - the parent reads this message + rendered string in and uses ORTE
* reporting mechanisms to display it to the user
* - if the problem was only a warning, the child continues processing
* (potentially eventually exec'ing the target executable).
* - if the problem was an error, the child exits and the parent
* handles the death of the child as appropriate (i.e., this ODLS
* simply reports the error -- other things decide what to do).
*/
#include "orte_config.h"
#include "orte/constants.h"
#include "orte/types.h"
#ifdef HAVE_STRING_H
#include <string.h>
#endif
#include <stdlib.h>
#ifdef HAVE_UNISTD_H
#include <unistd.h>
#endif
#include <errno.h>
#ifdef HAVE_SYS_TYPES_H
#include <sys/types.h>
#endif
#ifdef HAVE_SYS_WAIT_H
#include <sys/wait.h>
#endif
#include <signal.h>
#ifdef HAVE_FCNTL_H
#include <fcntl.h>
#endif
#ifdef HAVE_SYS_TIME_H
#include <sys/time.h>
#endif
#ifdef HAVE_SYS_PARAM_H
#include <sys/param.h>
#endif
#ifdef HAVE_NETDB_H
#include <netdb.h>
#endif
#ifdef HAVE_STDLIB_H
#include <stdlib.h>
#endif
#ifdef HAVE_SYS_STAT_H
#include <sys/stat.h>
#endif /* HAVE_SYS_STAT_H */
#ifdef HAVE_STDARG_H
#include <stdarg.h>
#endif
#ifdef HAVE_SYS_SELECT_H
#include <sys/select.h>
#endif
#include "opal/mca/maffinity/base/base.h"
#include "opal/mca/paffinity/base/base.h"
#include "opal/class/opal_pointer_array.h"
#include "opal/util/opal_environ.h"
#include "opal/util/opal_sos.h"
#include "opal/util/show_help.h"
#include "opal/util/fd.h"
#include "orte/util/show_help.h"
#include "orte/runtime/orte_wait.h"
#include "orte/runtime/orte_globals.h"
#include "orte/mca/errmgr/errmgr.h"
#include "orte/mca/ess/ess.h"
#include "orte/mca/iof/base/iof_base_setup.h"
#include "orte/mca/plm/plm.h"
#include "orte/util/name_fns.h"
#include "orte/mca/odls/base/base.h"
#include "orte/mca/odls/base/odls_private.h"
#include "orte/mca/odls/default/odls_default.h"
/*
* Struct written up the pipe from the child to the parent.
*/
typedef struct {
/* True if the child has died; false if this is just a warning to
be printed. */
bool fatal;
/* Relevant only if fatal==true */
int exit_status;
/* Length of the strings that are written up the pipe after this
struct */
int file_str_len;
int topic_str_len;
int msg_str_len;
} pipe_err_msg_t;
/*
* Max length of strings from the pipe_err_msg_t
*/
#define MAX_FILE_LEN 511
#define MAX_TOPIC_LEN MAX_FILE_LEN
/*
* Module functions (function pointers used in a struct)
*/
static int orte_odls_default_launch_local_procs(opal_buffer_t *data);
static int orte_odls_default_kill_local_procs(opal_pointer_array_t *procs);
static int orte_odls_default_signal_local_procs(const orte_process_name_t *proc, int32_t signal);
static int orte_odls_default_restart_proc(orte_odls_child_t *child);
/*
* Explicitly declared functions so that we can get the noreturn
* attribute registered with the compiler.
*/
static void send_error_show_help(int fd, int exit_status,
const char *file, const char *topic, ...)
__opal_attribute_noreturn__;
static int do_child(orte_app_context_t* context,
orte_odls_child_t *child,
char **environ_copy,
orte_odls_job_t *jobdat, int write_fd,
orte_iof_base_io_conf_t opts)
__opal_attribute_noreturn__;
/*
* Module
*/
orte_odls_base_module_t orte_odls_default_module = {
orte_odls_base_default_get_add_procs_data,
orte_odls_default_launch_local_procs,
orte_odls_default_kill_local_procs,
orte_odls_default_signal_local_procs,
orte_odls_base_default_deliver_message,
orte_odls_base_default_require_sync,
orte_odls_default_restart_proc
};
static bool odls_default_child_died(orte_odls_child_t *child)
{
time_t end;
pid_t ret;
struct timeval t;
fd_set bogus;
end = time(NULL) + orte_odls_globals.timeout_before_sigkill;
do {
ret = waitpid(child->pid, &child->exit_code, WNOHANG);
if (child->pid == ret) {
OPAL_OUTPUT_VERBOSE((2, orte_odls_globals.output,
"%s odls:default:WAITPID INDICATES PROC %d IS DEAD",
ORTE_NAME_PRINT(ORTE_PROC_MY_NAME), (int)(child->pid)));
/* It died -- return success */
return true;
} else if (0 == ret) {
/* with NOHANG specified, if a process has already exited
* while waitpid was registered, then waitpid returns 0
* as there is no error - this is a race condition problem
* that occasionally causes us to incorrectly report a proc
* as refusing to die. Unfortunately, errno may not be reset
* by waitpid in this case, so we cannot check it - just assume
* the proc has indeed died
*/
OPAL_OUTPUT_VERBOSE((2, orte_odls_globals.output,
"%s odls:default:WAITPID INDICATES PROC %d HAS ALREADY EXITED",
ORTE_NAME_PRINT(ORTE_PROC_MY_NAME), (int)(child->pid)));
return true;
} else if (-1 == ret && ECHILD == errno) {
/* The pid no longer exists, so we'll call this "good
enough for government work" */
OPAL_OUTPUT_VERBOSE((2, orte_odls_globals.output,
"%s odls:default:WAITPID INDICATES PID %d NO LONGER EXISTS",
ORTE_NAME_PRINT(ORTE_PROC_MY_NAME), (int)(child->pid)));
return true;
}
/* Bogus delay for 1 usec (sched_yeild() -- even if we have it
-- changed behavior in 2.6.3x Linux flavors to be
undesirable. */
t.tv_sec = 0;
t.tv_usec = 1;
FD_ZERO(&bogus);
FD_SET(0, &bogus);
select(1, &bogus, NULL, NULL, &t);
} while (time(NULL) < end);
/* The child didn't die, so return false */
return false;
}
static int odls_default_kill_local(pid_t pid, int signum)
{
if (orte_forward_job_control) {
pid = -pid;
}
if (0 != kill(pid, signum)) {
if (ESRCH != errno) {
OPAL_OUTPUT_VERBOSE((2, orte_odls_globals.output,
"%s odls:default:SENT KILL %d TO PID %d GOT ERRNO %d",
ORTE_NAME_PRINT(ORTE_PROC_MY_NAME), signum, (int)pid, errno));
return errno;
}
}
OPAL_OUTPUT_VERBOSE((2, orte_odls_globals.output,
"%s odls:default:SENT KILL %d TO PID %d SUCCESS",
ORTE_NAME_PRINT(ORTE_PROC_MY_NAME), signum, (int)pid));
return 0;
}
int orte_odls_default_kill_local_procs(opal_pointer_array_t *procs)
{
int rc;
if (ORTE_SUCCESS != (rc = orte_odls_base_default_kill_local_procs(procs,
odls_default_kill_local, odls_default_child_died))) {
ORTE_ERROR_LOG(rc);
return rc;
}
return ORTE_SUCCESS;
}
static void set_handler_default(int sig)
{
struct sigaction act;
act.sa_handler = SIG_DFL;
act.sa_flags = 0;
sigemptyset(&act.sa_mask);
sigaction(sig, &act, (struct sigaction *)0);
}
/*
* Internal function to write a rendered show_help message back up the
* pipe to the waiting parent.
*/
static int write_help_msg(int fd, pipe_err_msg_t *msg, const char *file,
const char *topic, va_list ap)
{
int ret;
char *str;
if (NULL == file || NULL == topic) {
return OPAL_ERR_BAD_PARAM;
}
str = opal_show_help_vstring(file, topic, true, ap);
msg->file_str_len = (int) strlen(file);
if (msg->file_str_len > MAX_FILE_LEN) {
ORTE_ERROR_LOG(ORTE_ERR_BAD_PARAM);
return ORTE_ERR_BAD_PARAM;
}
msg->topic_str_len = (int) strlen(topic);
if (msg->topic_str_len > MAX_TOPIC_LEN) {
ORTE_ERROR_LOG(ORTE_ERR_BAD_PARAM);
return ORTE_ERR_BAD_PARAM;
}
msg->msg_str_len = (int) strlen(str);
/* Only keep writing if each write() succeeds */
if (OPAL_SUCCESS != (ret = opal_fd_write(fd, sizeof(*msg), msg))) {
goto out;
}
if (msg->file_str_len > 0 &&
OPAL_SUCCESS != (ret = opal_fd_write(fd, msg->file_str_len, file))) {
goto out;
}
if (msg->topic_str_len > 0 &&
OPAL_SUCCESS != (ret = opal_fd_write(fd, msg->topic_str_len, topic))) {
goto out;
}
if (msg->msg_str_len > 0 &&
OPAL_SUCCESS != (ret = opal_fd_write(fd, msg->msg_str_len, str))) {
goto out;
}
out:
free(str);
return ret;
}
/* Called from the child to send a warning show_help message up the
pipe to the waiting parent. */
static int send_warn_show_help(int fd, const char *file,
const char *topic, ...)
{
int ret;
va_list ap;
pipe_err_msg_t msg;
msg.fatal = false;
msg.exit_status = 0; /* ignored */
/* Send it */
va_start(ap, topic);
ret = write_help_msg(fd, &msg, file, topic, ap);
va_end(ap);
return ret;
}
/* Called from the child to send an error message up the pipe to the
waiting parent. */
static void send_error_show_help(int fd, int exit_status,
const char *file, const char *topic, ...)
{
int ret;
va_list ap;
pipe_err_msg_t msg;
msg.fatal = true;
msg.exit_status = exit_status;
/* Send it */
va_start(ap, topic);
ret = write_help_msg(fd, &msg, file, topic, ap);
va_end(ap);
exit(exit_status);
}
/*
* Bind the process to a specific slot list
*/
static int bind_to_slot_list(orte_app_context_t* context,
orte_odls_child_t *child,
orte_odls_job_t *jobdat,
bool *bound, int pipe_fd)
{
int rc;
opal_paffinity_base_cpu_set_t mask;
char *msg = NULL;
*bound = false;
OPAL_OUTPUT_VERBOSE((2, orte_odls_globals.output,
"%s odls:default:fork binding child %s to slot_list %s",
ORTE_NAME_PRINT(ORTE_PROC_MY_NAME),
ORTE_NAME_PRINT(child->name),
child->slot_list));
if (opal_paffinity_alone) {
send_error_show_help(pipe_fd, 1,
"help-orte-odls-default.txt",
"slot list and paffinity_alone",
orte_process_info.nodename, context->app);
/* Does not return */
}
if (orte_report_bindings) {
opal_output(0, "%s odls:default:fork binding child %s to slot_list %s",
ORTE_NAME_PRINT(ORTE_PROC_MY_NAME),
ORTE_NAME_PRINT(child->name), child->slot_list);
}
rc = opal_paffinity_base_slot_list_set((long)child->name->vpid,
child->slot_list, &mask);
if (ORTE_SUCCESS != rc) {
if (ORTE_ERR_NOT_SUPPORTED == OPAL_SOS_GET_ERROR_CODE(rc)) {
/* OS doesn't support providing topology information */
send_error_show_help(pipe_fd, 1, "help-orte-odls-default.txt",
"binding not supported",
orte_process_info.nodename, context->app);
/* Does not return */
}
asprintf(&msg, "opal_paffinity_base_slot_list_set() returned \"%s\"",
opal_strerror(OPAL_SOS_GET_ERROR_CODE(rc)));
if (NULL == msg) {
msg = "opal_paffinity_base_slot_list_set() returned failure";
}
send_error_show_help(pipe_fd, 1, "help-orte-odls-default.txt",
"binding generic error",
orte_process_info.nodename, context->app, msg,
__FILE__, __LINE__);
/* Does not return */
}
/* if we didn't wind up bound, then generate a warning unless
suppressed */
OPAL_PAFFINITY_PROCESS_IS_BOUND((mask), bound);
if (!bound && orte_odls_base.warn_if_not_bound) {
send_warn_show_help(pipe_fd, "help-orte-odls-base.txt",
"warn not bound", "slot list"
"Request resulted in binding to all available processors",
orte_process_info.nodename, context->app,
"bind to slot list", child->slot_list);
}
return ORTE_SUCCESS;
}
/*
* This function always prints a message: it may be a warning or an
* error.
*
* If binding is not required for this process, then print a simple
* warning message and return an error code. If binding *is*
* required, then send an error message up the pipe to the parent and
* exit.
*/
static int bind_failed_msg(const char *msg, orte_mapping_policy_t policy,
int return_code_if_warning,
int pipe_fd, const char *app_name,
const char *filename, int line_num)
{
/* If binding is not required, then send a warning up the pipe and
then return an error code. */
if (ORTE_BINDING_NOT_REQUIRED(policy)) {
send_warn_show_help(pipe_fd,
"help-orte-odls-default.txt", "not bound",
orte_process_info.nodename, app_name, msg,
filename, line_num);
return return_code_if_warning;
}
/* If binding is required, send an error up the pipe (which exits
-- it doesn't return). */
send_error_show_help(pipe_fd, 1, "help-orte-odls-default.txt",
"binding generic error",
orte_process_info.nodename, app_name, msg,
filename, line_num);
/* Does not return */
}
/*
* Similar to bind_failed_msg(), but if binding is not required, do
* not output a message (just return an error code). If binding is
* required, handling is the same as for bind_failed_msg().
*/
static int bind_failed(const char *msg, orte_mapping_policy_t policy,
int return_code_if_warning,
int pipe_fd, const char *app_name,
const char *filename, int line_num)
{
if (ORTE_BINDING_NOT_REQUIRED(policy)) {
return return_code_if_warning;
}
/* This won't return, but use "return" statement here so that the
compiler won't complain. */
return bind_failed_msg(msg, policy, 0, pipe_fd, app_name,
filename, line_num);
}
/*
* Bind the process to a core
*/
static int bind_to_core(orte_app_context_t* context,
orte_odls_child_t *child,
orte_odls_job_t *jobdat,
bool *bound, int pipe_fd)
{
bool flag;
int i, rc;
char *tmp, *msg;
int16_t n;
orte_node_rank_t nrank, lrank;
opal_paffinity_base_cpu_set_t mask;
int target_socket, npersocket, logical_skt;
int logical_cpu, phys_core, phys_cpu, ncpu;
*bound = false;
/* we want to bind this proc to a specific core, or multiple cores
if the cpus_per_rank is > 0 */
OPAL_OUTPUT_VERBOSE((5, orte_odls_globals.output,
"%s odls:default:fork binding child %s to core(s) cpus/rank %d stride %d",
ORTE_NAME_PRINT(ORTE_PROC_MY_NAME),
ORTE_NAME_PRINT(child->name),
(int)jobdat->cpus_per_rank, (int)jobdat->stride));
/* get the node rank */
if (ORTE_NODE_RANK_INVALID ==
(nrank = orte_ess.get_node_rank(child->name))) {
send_error_show_help(pipe_fd, 1, "help-orte-odls-default.txt",
"binding generic error",
orte_process_info.nodename, context->app,
"ess.get_node_rank returned NODE_RANK_INVALID",
__FILE__, __LINE__);
/* Does not return */
}
/* get the local rank */
if (ORTE_LOCAL_RANK_INVALID ==
(lrank = orte_ess.get_local_rank(child->name))) {
send_error_show_help(pipe_fd, 1, "help-orte-odls-default.txt",
"binding generic error",
orte_process_info.nodename, context->app,
"ess.get_local_rank returned LOCAL_RANK_INVALID",
__FILE__, __LINE__);
/* Does not return */
}
/* init the mask */
OPAL_PAFFINITY_CPU_ZERO(mask);
if (ORTE_MAPPING_NPERXXX & jobdat->policy) {
/* we need to balance the children from this job
across the available sockets */
npersocket = jobdat->num_local_procs / orte_odls_globals.num_sockets;
/* determine the socket to use based on those available */
if (npersocket < 2) {
/* if we only have 1/sock, or we have less procs than
sockets, then just put it on the lrank socket */
logical_skt = lrank;
} else if (ORTE_MAPPING_BYSOCKET & jobdat->policy) {
logical_skt = lrank % npersocket;
} else {
logical_skt = lrank / npersocket;
}
if (orte_odls_globals.bound) {
/* if we are already bound (by some other entity), use
this as an index into our available sockets */
for (n = target_socket = 0;
n < logical_skt &&
target_socket < opal_bitmap_size(&orte_odls_globals.sockets);
target_socket++) {
if (opal_bitmap_is_set_bit(&orte_odls_globals.sockets,
target_socket)) {
n++;
}
}
/* Did we have enough sockets? */
if (n < logical_skt) {
return bind_failed_msg("not enough processor sockets available",
jobdat->policy,
ORTE_ERR_NOT_FOUND,
pipe_fd, context->app,
__FILE__, __LINE__);
}
} else {
rc = opal_paffinity_base_get_physical_socket_id(logical_skt,
&target_socket);
if (ORTE_ERR_NOT_SUPPORTED == OPAL_SOS_GET_ERROR_CODE(rc)) {
return bind_failed_msg("OS does not provide processor topology info (physical socket ID)",
jobdat->policy,
ORTE_ERR_NOT_FOUND,
pipe_fd, context->app,
__FILE__, __LINE__);
}
}
OPAL_OUTPUT_VERBOSE((2, orte_odls_globals.output,
"%s odls:default:fork child %s local rank %d npersocket %d logical socket %d target socket %d",
ORTE_NAME_PRINT(ORTE_PROC_MY_NAME), ORTE_NAME_PRINT(child->name), lrank,
npersocket, logical_skt, target_socket));
/* set the starting point */
logical_cpu = (lrank % npersocket) * jobdat->cpus_per_rank;
/* bind to this socket */
goto bind_socket;
} else if (ORTE_MAPPING_BYSOCKET & jobdat->policy) {
/* this corresponds to a mapping policy where
* local rank 0 goes on socket 0, and local
* rank 1 goes on socket 1, etc. - round robin
* until all ranks are mapped
*
* NOTE: we already know our number of sockets
* from when we initialized
*/
rc = opal_paffinity_base_get_physical_socket_id(lrank % orte_odls_globals.num_sockets, &target_socket);
if (OPAL_SUCCESS != rc) {
/* This may be a small memory leak, but this child is
exiting soon anyway; keep the logic simple by not
worrying about the small leak. */
asprintf(&msg, "opal_paffinity_base_get_physical_socket_id(%d) returned \"%s\"",
(lrank % orte_odls_globals.num_sockets),
opal_strerror(OPAL_SOS_GET_ERROR_CODE(rc)));
if (NULL == msg) {
msg = "opal_paffinity_base_get_physical_socket_id() failed";
}
if (OPAL_ERR_NOT_SUPPORTED == OPAL_SOS_GET_ERROR_CODE(rc)) {
msg = "OS does not provide processor topology information (physical socket ID)";
}
return bind_failed(msg, jobdat->policy, ORTE_ERR_NOT_SUPPORTED,
pipe_fd, context->app, __FILE__, __LINE__);
}
OPAL_OUTPUT_VERBOSE((2, orte_odls_globals.output,
"bysocket lrank %d numsocks %d logical socket %d target socket %d", (int)lrank,
(int)orte_odls_globals.num_sockets,
(int)(lrank % orte_odls_globals.num_sockets),
target_socket));
/* my starting core within this socket has to be
offset by cpus_per_rank */
logical_cpu = (lrank / orte_odls_globals.num_sockets) * jobdat->cpus_per_rank;
bind_socket:
/* cycle across the cpus_per_rank */
for (n=0; n < jobdat->cpus_per_rank; n++) {
/* get the physical core within this target socket */
rc = opal_paffinity_base_get_physical_core_id(target_socket, logical_cpu, &phys_core);
if (OPAL_SUCCESS != rc) {
/* Seem comment above about "This may be a small
memory leak" */
asprintf(&msg, "opal_paffinity_base_get_physical_core_id(%d, %d) returned \"%s\"",
target_socket, logical_cpu,
opal_strerror(OPAL_SOS_GET_ERROR_CODE(rc)));
if (NULL == msg) {
msg = "opal_paffinity_base_get_physical_core_id() failed";
}
if (OPAL_ERR_NOT_SUPPORTED == OPAL_SOS_GET_ERROR_CODE(rc)) {
msg = "OS does not provide processor topology information (physical core ID)";
}
return bind_failed(msg, jobdat->policy, ORTE_ERR_NOT_SUPPORTED,
pipe_fd, context->app, __FILE__, __LINE__);
}
/* map this to a physical cpu on this node */
if (ORTE_SUCCESS != (rc = opal_paffinity_base_get_map_to_processor_id(target_socket, phys_core, &phys_cpu))) {
/* Seem comment above about "This may be a small
memory leak" */
asprintf(&msg, "opal_paffinity_base_get_map_to_processor_id(%d, %d) returned \"%s\"",
target_socket, phys_core,
opal_strerror(OPAL_SOS_GET_ERROR_CODE(rc)));
if (NULL == msg) {
msg = "opal_paffinity_base_get_map_to_processor_id() failed";
}
if (OPAL_ERR_NOT_SUPPORTED == OPAL_SOS_GET_ERROR_CODE(rc)) {
msg = "OS does not provide processor topology information (map socket,core->ID)";
}
return bind_failed(msg, jobdat->policy, ORTE_ERR_NOT_SUPPORTED,
pipe_fd, context->app, __FILE__, __LINE__);
}
/* are we bound? */
if (orte_odls_globals.bound) {
/* see if this physical cpu is available to us */
if (!OPAL_PAFFINITY_CPU_ISSET(phys_cpu, orte_odls_globals.my_cores)) {
/* no it isn't - skip it */
continue;
}
}
OPAL_OUTPUT_VERBOSE((2, orte_odls_globals.output,
"%s odls:default:fork mapping phys socket %d core %d to phys_cpu %d",
ORTE_NAME_PRINT(ORTE_PROC_MY_NAME),
target_socket, phys_core, phys_cpu));
OPAL_PAFFINITY_CPU_SET(phys_cpu, mask);
/* increment logical cpu */
logical_cpu += jobdat->stride;
}
if (orte_report_bindings) {
tmp = opal_paffinity_base_print_binding(mask);
opal_output(0, "%s odls:default:fork binding child %s to socket %d cpus %s",
ORTE_NAME_PRINT(ORTE_PROC_MY_NAME),
ORTE_NAME_PRINT(child->name), target_socket, tmp);
free(tmp);
}
} else {
/* my starting core has to be offset by cpus_per_rank */
logical_cpu = nrank * jobdat->cpus_per_rank;
for (n=0; n < jobdat->cpus_per_rank; n++) {
/* are we bound? */
if (orte_odls_globals.bound) {
/* if we are bound, then use the logical_cpu as an
index against our available cores */
ncpu = 0;
for (i = 0; i < OPAL_PAFFINITY_BITMASK_CPU_MAX &&
ncpu <= logical_cpu; i++) {
if (OPAL_PAFFINITY_CPU_ISSET(i,
orte_odls_globals.my_cores)) {
ncpu++;
phys_cpu = i;
}
}
/* if we don't have enough processors, that is an
error */
if (ncpu <= logical_cpu) {
if (ORTE_BINDING_NOT_REQUIRED(jobdat->policy)) {
return ORTE_ERR_NOT_SUPPORTED;
}
send_error_show_help(pipe_fd, 1,
"help-orte-odls-default.txt",
"binding generic error",
orte_process_info.nodename,
context->app,
"not enough logical processors",
__FILE__, __LINE__);
/* Does not return */
}
} else {
/* if we are not bound, then all processors are
available to us, so index into the node's array to
get the physical cpu */
rc = opal_paffinity_base_get_physical_processor_id(logical_cpu,
&phys_cpu);
if (OPAL_SUCCESS != rc) {
/* No processor to bind to */
/* Seem comment above about "This may be a small
memory leak" */
asprintf(&msg, "opal_paffinity_base_get_physical_processor_id(%d) returned \"%s\"",
logical_cpu,
opal_strerror(OPAL_SOS_GET_ERROR_CODE(rc)));
if (NULL == msg) {
msg = "opal_paffinity_base_get_physical_processor_id() failed";
}
if (OPAL_ERR_NOT_SUPPORTED == OPAL_SOS_GET_ERROR_CODE(rc)) {
msg = "OS does not provide processor topology information (physical processor ID)";
}
return bind_failed(msg, jobdat->policy,
ORTE_ERR_NOT_SUPPORTED,
pipe_fd, context->app,
__FILE__, __LINE__);
}
}
OPAL_PAFFINITY_CPU_SET(phys_cpu, mask);
/* increment logical cpu */
logical_cpu += jobdat->stride;
}
if (orte_report_bindings) {
tmp = opal_paffinity_base_print_binding(mask);
opal_output(0, "%s odls:default:fork binding child %s to cpus %s",
ORTE_NAME_PRINT(ORTE_PROC_MY_NAME),
ORTE_NAME_PRINT(child->name), tmp);
free(tmp);
}
}
/* Bind me! */
if (ORTE_SUCCESS != (rc = opal_paffinity_base_set(mask))) {
/* Seem comment above about "This may be a small memory
leak" */
asprintf(&msg, "opal_paffinity_base_set returned \"%s\"",
opal_strerror(OPAL_SOS_GET_ERROR_CODE(rc)));
if (NULL == msg) {
msg = "opal_paffinity_base_set() failed";
}
return bind_failed(msg,
jobdat->policy,
OPAL_SOS_GET_ERROR_CODE(rc),
pipe_fd, context->app, __FILE__, __LINE__);
}
*bound = true;
/* If the above work resulted in binding to everything (i.e.,
effectively not binding), warn -- unless the warning is
suppressed. */
OPAL_PAFFINITY_PROCESS_IS_BOUND(mask, &flag);
if (!flag && orte_odls_base.warn_if_not_bound) {
send_warn_show_help(pipe_fd,
"help-orte-odls-default.txt",
"bound to everything",
orte_process_info.nodename, context->app,
__FILE__, __LINE__);
}
return ORTE_SUCCESS;
}
static int bind_to_socket(orte_app_context_t* context,
orte_odls_child_t *child,
orte_odls_job_t *jobdat,
bool *bound, int pipe_fd)
{
bool flag;
int i, rc;
char *tmp, *msg;
int16_t n;
orte_node_rank_t lrank;
opal_paffinity_base_cpu_set_t mask;
int target_socket, npersocket, logical_skt;
int logical_cpu, phys_core, phys_cpu, ncpu;
*bound = false;
/* bind this proc to a socket */
OPAL_OUTPUT_VERBOSE((5, orte_odls_globals.output,
"%s odls:default:fork binding child %s to socket",
ORTE_NAME_PRINT(ORTE_PROC_MY_NAME),
ORTE_NAME_PRINT(child->name)));
/* layout this process across the sockets based on
* the provided mapping policy
*/
if (ORTE_LOCAL_RANK_INVALID == (lrank = orte_ess.get_local_rank(child->name))) {
send_error_show_help(pipe_fd, 1, "help-orte-odls-default.txt",
"binding generic error",
orte_process_info.nodename, context->app,
"ess.get_local_rank returned NODE_RANK_INVALID",
__FILE__, __LINE__);
/* Does not return */
}
if (ORTE_MAPPING_NPERXXX & jobdat->policy) {
/* we need to balance the children from this job
across the available sockets */
npersocket = jobdat->num_local_procs / orte_odls_globals.num_sockets;
/* determine the socket to use based on those available */
if (npersocket < 2) {
/* if we only have 1/sock, or we have less
* procs than sockets, then just put it on the
* lrank socket
*/
logical_skt = lrank;
} else if (ORTE_MAPPING_BYSOCKET & jobdat->policy) {
logical_skt = lrank % npersocket;
} else {
logical_skt = lrank / npersocket;
}
if (orte_odls_globals.bound) {
/* if we are bound, use this as an index into
our available sockets */
for (target_socket=0, n = 0; target_socket < opal_bitmap_size(&orte_odls_globals.sockets) && n < logical_skt; target_socket++) {
if (opal_bitmap_is_set_bit(&orte_odls_globals.sockets, target_socket)) {
n++;
}
}
/* if we don't have enough sockets, that is an error */
if (n < logical_skt) {
return bind_failed_msg("not enough processor sockets available",
jobdat->policy,
ORTE_ERR_NOT_FOUND,
pipe_fd, context->app,
__FILE__, __LINE__);
}
} else {
rc = opal_paffinity_base_get_physical_socket_id(logical_skt, &target_socket);
if (ORTE_ERR_NOT_SUPPORTED == OPAL_SOS_GET_ERROR_CODE(rc)) {
/* OS doesn't support providing topology
information */
return bind_failed_msg("OS does not provide processor topology info (physical socket ID)",
jobdat->policy,
ORTE_ERR_NOT_FOUND,
pipe_fd, context->app,
__FILE__, __LINE__);
}
}
OPAL_OUTPUT_VERBOSE((2, orte_odls_globals.output,
"%s odls:default:fork child %s local rank %d npersocket %d logical socket %d target socket %d",
ORTE_NAME_PRINT(ORTE_PROC_MY_NAME), ORTE_NAME_PRINT(child->name), lrank,
npersocket, logical_skt, target_socket));
} else if (ORTE_MAPPING_BYSOCKET & jobdat->policy) {
/* this corresponds to a mapping policy where
* local rank 0 goes on socket 0, and local
* rank 1 goes on socket 1, etc. - round robin
* until all ranks are mapped
*
* NOTE: we already know our number of sockets
* from when we initialized
*/
rc = opal_paffinity_base_get_physical_socket_id(lrank % orte_odls_globals.num_sockets, &target_socket);
if (ORTE_ERR_NOT_SUPPORTED == OPAL_SOS_GET_ERROR_CODE(rc)) {
/* OS does not support providing topology
information */
return bind_failed_msg("OS does not provide processor topology info(physical socket ID)",
jobdat->policy,
ORTE_ERR_NOT_FOUND,
pipe_fd, context->app,
__FILE__, __LINE__);
}
OPAL_OUTPUT_VERBOSE((2, orte_odls_globals.output,
"bysocket lrank %d numsocks %d logical socket %d target socket %d", (int)lrank,
(int)orte_odls_globals.num_sockets,
(int)(lrank % orte_odls_globals.num_sockets),
target_socket));
} else {
/* use a byslot-like policy where local rank 0 goes on
* socket 0, and local rank 1 goes on socket 0, etc.
* following round-robin until all ranks mapped
*/
if (orte_odls_globals.bound) {
/* if we are bound, then we compute the
* logical socket id based on the number of
* available cores in each socket so that each
* rank gets its own core, adjusting for the
* cpus_per_task
*/
/* Find the lrank available core, accounting
for cpus_per_task */
logical_cpu = lrank * jobdat->cpus_per_rank;
/* use the logical_cpu as an index against our
available cores */
ncpu = 0;
for (i=0; i < orte_odls_globals.num_processors && ncpu <= logical_cpu; i++) {
if (OPAL_PAFFINITY_CPU_ISSET(i, orte_odls_globals.my_cores)) {
ncpu++;
phys_cpu = i;
}
}
/* if we don't have enough processors, that is
an error */
if (ncpu < logical_cpu) {
send_error_show_help(pipe_fd, 1,
"help-orte-odls-default.txt",
"binding generic error",
orte_process_info.nodename,
context->app,
"not enough logical processors",
__FILE__, __LINE__);
/* Does not return */
}
/* get the physical socket of that cpu */
if (ORTE_SUCCESS != (rc = opal_paffinity_base_get_map_to_socket_core(phys_cpu, &target_socket, &phys_core))) {
/* Seem comment above about "This may be a small
memory leak" */
asprintf(&msg, "opal_paffinity_base_get_map_to_socket_core(%d) returned \"%s\"",
phys_cpu, opal_strerror(OPAL_SOS_GET_ERROR_CODE(rc)));
if (NULL == msg) {
msg = "opal_paffinity_base_get_map_to_socket_core() failed";
}
if (OPAL_ERR_NOT_SUPPORTED == OPAL_SOS_GET_ERROR_CODE(rc)) {
msg = "OS does not provide processor topology information (map socket,core->ID)";
}
return bind_failed(msg, jobdat->policy,
ORTE_ERR_NOT_SUPPORTED,
pipe_fd, context->app,
__FILE__, __LINE__);
}
} else {
/* if we are not bound, then just use all sockets */
if (1 == orte_odls_globals.num_sockets) {
/* if we only have one socket, then just
put it there */
rc = opal_paffinity_base_get_physical_socket_id(0, &target_socket);
if (ORTE_ERR_NOT_SUPPORTED == OPAL_SOS_GET_ERROR_CODE(rc)) {
/* OS doesn't support providing
topology information */
return bind_failed_msg("OS does not provide processor topology info (physical socket ID)",
jobdat->policy,
ORTE_ERR_NOT_FOUND,
pipe_fd, context->app,
__FILE__, __LINE__);
}
} else {
/* compute the logical socket,
compensating for the number of
cpus_per_rank */
logical_skt = lrank / (orte_default_num_cores_per_socket / jobdat->cpus_per_rank);
/* wrap that around the number of sockets
so we round-robin */
logical_skt = logical_skt % orte_odls_globals.num_sockets;
/* now get the target physical socket */
rc = opal_paffinity_base_get_physical_socket_id(logical_skt, &target_socket);
if (ORTE_ERR_NOT_SUPPORTED == OPAL_SOS_GET_ERROR_CODE(rc)) {
/* OS doesn't support providing
topology information */
return bind_failed_msg("OS does not provide processor topology info (physical socket ID)",
jobdat->policy,
ORTE_ERR_NOT_FOUND,
pipe_fd, context->app,
__FILE__, __LINE__);
}
}
OPAL_OUTPUT_VERBOSE((2, orte_odls_globals.output,
"byslot lrank %d socket %d", (int)lrank, target_socket));
}
}
OPAL_PAFFINITY_CPU_ZERO(mask);
for (n=0; n < orte_default_num_cores_per_socket; n++) {
/* get the physical core within this target socket */
rc = opal_paffinity_base_get_physical_core_id(target_socket, n, &phys_core);
if (OPAL_SUCCESS != rc) {
/* Seem comment above about "This may be a small memory
leak" */
asprintf(&msg, "opal_paffinity_base_get_physical_core_id(%d, %d) returned \"%s\"",
target_socket, n,
opal_strerror(OPAL_SOS_GET_ERROR_CODE(rc)));
if (NULL == msg) {
msg = "opal_paffinity_base_get_physical_core_id() failed";
}
if (OPAL_ERR_NOT_SUPPORTED == OPAL_SOS_GET_ERROR_CODE(rc)) {
msg = "OS does not provide processor topology information (physical core ID)";
}
return bind_failed(msg, jobdat->policy,
ORTE_ERR_NOT_SUPPORTED,
pipe_fd, context->app,
__FILE__, __LINE__);
}
/* map this to a physical cpu on this node */
if (ORTE_SUCCESS != (rc = opal_paffinity_base_get_map_to_processor_id(target_socket, phys_core, &phys_cpu))) {
/* Seem comment above about "This may be a small memory
leak" */
asprintf(&msg, "opal_paffinity_base_get_map_to_processor_id(%d, %d) returned \"%s\"",
target_socket, phys_core,
opal_strerror(OPAL_SOS_GET_ERROR_CODE(rc)));
if (NULL == msg) {
msg = "opal_paffinity_base_get_map_to_processor_id()";
}
if (OPAL_ERR_NOT_SUPPORTED == OPAL_SOS_GET_ERROR_CODE(rc)) {
msg = "OS does not provide processor topology information (map socket,core->ID)";
}
return bind_failed(msg, jobdat->policy,
ORTE_ERR_NOT_SUPPORTED,
pipe_fd, context->app,
__FILE__, __LINE__);
}
/* are we bound? */
if (orte_odls_globals.bound) {
/* see if this physical cpu is available to us */
if (!OPAL_PAFFINITY_CPU_ISSET(phys_cpu, orte_odls_globals.my_cores)) {
/* no it isn't - skip it */
continue;
}
}
OPAL_OUTPUT_VERBOSE((2, orte_odls_globals.output,
"%s odls:default:fork mapping phys socket %d core %d to phys_cpu %d",
ORTE_NAME_PRINT(ORTE_PROC_MY_NAME),
target_socket, phys_core, phys_cpu));
OPAL_PAFFINITY_CPU_SET(phys_cpu, mask);
}
/* Bind me! */
if (ORTE_SUCCESS != (rc = opal_paffinity_base_set(mask))) {
/* Seem comment above about "This may be a small memory
leak" */
asprintf(&msg, "opal_paffinity_base_set() returned \"%s\"",
opal_strerror(OPAL_SOS_GET_ERROR_CODE(rc)));
if (NULL == msg) {
msg = "opal_paffinity_base_set() failed";
}
return bind_failed(msg,
jobdat->policy,
OPAL_SOS_GET_ERROR_CODE(rc),
pipe_fd, context->app, __FILE__, __LINE__);
}
*bound = true;
/* If the above work resulted in binding to everything (i.e.,
effectively not binding), warn -- unless the warning is
suppressed. */
OPAL_PAFFINITY_PROCESS_IS_BOUND(mask, &flag);
if (!flag && orte_odls_base.warn_if_not_bound) {
send_warn_show_help(pipe_fd,
"help-orte-odls-default.txt",
"bound to everything",
orte_process_info.nodename, context->app,
__FILE__, __LINE__);
} else if (orte_report_bindings) {
tmp = opal_paffinity_base_print_binding(mask);
opal_output(0, "%s odls:default:fork binding child %s to socket %d cpus %s",
ORTE_NAME_PRINT(ORTE_PROC_MY_NAME),
ORTE_NAME_PRINT(child->name), target_socket, tmp);
free(tmp);
}
return ORTE_SUCCESS;
}
static int bind_to_board(orte_app_context_t* context,
orte_odls_child_t *child,
orte_odls_job_t *jobdat,
bool *bound, int pipe_fd)
{
/* Not currently supported until multi-board paffinity enabled.
But this is not an error -- for now. */
*bound = false;
if (orte_odls_base.warn_if_not_bound) {
send_warn_show_help(pipe_fd, "help-orte-odls-base.txt",
"warn not bound", "board",
"Not currently supported by Open MPI",
orte_process_info.nodename, context->app,
"Bind to board", "");
}
return ORTE_ERR_NOT_SUPPORTED;
}
static int do_child(orte_app_context_t* context,
orte_odls_child_t *child,
char **environ_copy,
orte_odls_job_t *jobdat, int write_fd,
orte_iof_base_io_conf_t opts)
{
int i;
sigset_t sigs;
long fd, fdmax = sysconf(_SC_OPEN_MAX);
bool paffinity_enabled = false;
char *param, *tmp;
opal_paffinity_base_cpu_set_t mask;
if (orte_forward_job_control) {
/* Set a new process group for this child, so that a
SIGSTOP can be sent to it without being sent to the
orted. */
setpgid(0, 0);
}
/* Setup the pipe to be close-on-exec */
fcntl(write_fd, F_SETFD, FD_CLOEXEC);
if (NULL != child) {
/* setup stdout/stderr so that any error messages that we
may print out will get displayed back at orterun.
NOTE: Definitely do this AFTER we check contexts so
that any error message from those two functions doesn't
come out to the user. IF we didn't do it in this order,
THEN a user who gives us a bad executable name or
working directory would get N error messages, where
N=num_procs. This would be very annoying for large
jobs, so instead we set things up so that orterun
always outputs a nice, single message indicating what
happened
*/
if (ORTE_SUCCESS != (i = orte_iof_base_setup_child(&opts,
&environ_copy))) {
ORTE_ERROR_LOG(i);
send_error_show_help(write_fd, 1,
"help-orte-odls-default.txt",
"iof setup failed",
orte_process_info.nodename, context->app);
/* Does not return */
}
/* Setup process affinity. Not for the meek. */
if (NULL != child->slot_list) {
bind_to_slot_list(context, child, jobdat,
&paffinity_enabled, write_fd);
} else if (ORTE_BIND_TO_CORE & jobdat->policy) {
bind_to_core(context, child, jobdat,
&paffinity_enabled, write_fd);
} else if (ORTE_BIND_TO_SOCKET & jobdat->policy) {
bind_to_socket(context, child, jobdat,
&paffinity_enabled, write_fd);
} else if (ORTE_BIND_TO_BOARD & jobdat->policy) {
bind_to_board(context, child, jobdat,
&paffinity_enabled, write_fd);
}
/* If we were able to set processor affinity, then also
setup memory affinity. */
if (paffinity_enabled) {
if (OPAL_SUCCESS == opal_maffinity_base_open() &&
OPAL_SUCCESS == opal_maffinity_base_select()) {
opal_maffinity_setup = true;
}
}
} else if (!(ORTE_JOB_CONTROL_FORWARD_OUTPUT & jobdat->controls)) {
/* tie stdin/out/err/internal to /dev/null */
int fdnull;
for (i=0; i < 3; i++) {
fdnull = open("/dev/null", O_RDONLY, 0);
if (fdnull > i && i != write_fd) {
dup2(fdnull, i);
}
close(fdnull);
}
fdnull = open("/dev/null", O_RDONLY, 0);
if (fdnull > opts.p_internal[1]) {
dup2(fdnull, opts.p_internal[1]);
}
close(fdnull);
}
/* If we are able to bind, then set an info MCA param that tells
the launched processes that it was bound by us (e.g., so that
MPI_INIT doesn't try to bind itself) */
if (paffinity_enabled) {
param = mca_base_param_environ_variable("paffinity","base","bound");
opal_setenv(param, "1", true, &environ_copy);
free(param);
/* ...and provide a nice string representation of what we
bound to */
if (OPAL_SUCCESS == opal_paffinity_base_get(&mask)) {
tmp = opal_paffinity_base_print_binding(mask);
if (NULL != tmp) {
param = mca_base_param_environ_variable("paffinity","base","applied_binding");
opal_setenv(param, tmp, true, &environ_copy);
free(tmp);
}
}
}
/* close all file descriptors w/ exception of stdin/stdout/stderr,
the pipe used for the IOF INTERNAL messages, and the pipe up to
the parent. */
for(fd=3; fd<fdmax; fd++) {
if (fd != opts.p_internal[1] && fd != write_fd) {
close(fd);
}
}
if (context->argv == NULL) {
context->argv = malloc(sizeof(char*)*2);
context->argv[0] = strdup(context->app);
context->argv[1] = NULL;
}
/* Set signal handlers back to the default. Do this close to
the exev() because the event library may (and likely will)
reset them. If we don't do this, the event library may
have left some set that, at least on some OS's, don't get
reset via fork() or exec(). Hence, the launched process
could be unkillable (for example). */
set_handler_default(SIGTERM);
set_handler_default(SIGINT);
set_handler_default(SIGHUP);
set_handler_default(SIGPIPE);
set_handler_default(SIGCHLD);
/* Unblock all signals, for many of the same reasons that we
set the default handlers, above. This is noticable on
Linux where the event library blocks SIGTERM, but we don't
want that blocked by the launched process. */
sigprocmask(0, 0, &sigs);
sigprocmask(SIG_UNBLOCK, &sigs, 0);
/* Exec the new executable */
if (10 < opal_output_get_verbosity(orte_odls_globals.output)) {
int jout;
opal_output(0, "%s STARTING %s", ORTE_NAME_PRINT(ORTE_PROC_MY_NAME), context->app);
for (jout=0; NULL != context->argv[jout]; jout++) {
opal_output(0, "%s\tARGV[%d]: %s", ORTE_NAME_PRINT(ORTE_PROC_MY_NAME), jout, context->argv[jout]);
}
for (jout=0; NULL != environ_copy[jout]; jout++) {
opal_output(0, "%s\tENVIRON[%d]: %s", ORTE_NAME_PRINT(ORTE_PROC_MY_NAME), jout, environ_copy[jout]);
}
}
execve(context->app, context->argv, environ_copy);
send_error_show_help(write_fd, 1,
"help-orte-odls-default.txt", "execve error",
context->app, strerror(errno));
/* Does not return */
}
static int do_parent(orte_app_context_t* context,
orte_odls_child_t *child,
char **environ_copy,
orte_odls_job_t *jobdat, int read_fd,
orte_iof_base_io_conf_t opts)
{
int rc;
pipe_err_msg_t msg;
char file[MAX_FILE_LEN + 1], topic[MAX_TOPIC_LEN + 1], *str = NULL;
if (NULL != child && (ORTE_JOB_CONTROL_FORWARD_OUTPUT & jobdat->controls)) {
/* connect endpoints IOF */
rc = orte_iof_base_setup_parent(child->name, &opts);
if (ORTE_SUCCESS != rc) {
ORTE_ERROR_LOG(rc);
close(read_fd);
if (NULL != child) {
child->state = ORTE_PROC_STATE_UNDEF;
}
return rc;
}
}
/* Block reading a message from the pipe */
while (1) {
rc = opal_fd_read(read_fd, sizeof(msg), &msg);
/* If the pipe closed, then the child successfully launched */
if (OPAL_ERR_TIMEOUT == OPAL_SOS_GET_ERROR_CODE(rc)) {
break;
}
/* If Something Bad happened in the read, error out */
if (OPAL_SUCCESS != rc) {
ORTE_ERROR_LOG(rc);
close(read_fd);
if (NULL != child) {
child->state = ORTE_PROC_STATE_UNDEF;
}
return rc;
}
/* Otherwise, we got a warning or error message from the child */
if (NULL != child) {
child->alive = msg.fatal ? 0 : 1;
}
/* Read in the strings; ensure to terminate them with \0 */
if (msg.file_str_len > 0) {
rc = opal_fd_read(read_fd, msg.file_str_len, file);
if (OPAL_SUCCESS != rc) {
orte_show_help("help-orte-odls-default.txt", "syscall fail",
true,
orte_process_info.nodename, context->app,
"opal_fd_read", __FILE__, __LINE__);
if (NULL != child) {
child->state = ORTE_PROC_STATE_UNDEF;
}
return rc;
}
file[msg.file_str_len] = '\0';
}
if (msg.topic_str_len > 0) {
rc = opal_fd_read(read_fd, msg.topic_str_len, topic);
if (OPAL_SUCCESS != rc) {
orte_show_help("help-orte-odls-default.txt", "syscall fail",
true,
orte_process_info.nodename, context->app,
"opal_fd_read", __FILE__, __LINE__);
if (NULL != child) {
child->state = ORTE_PROC_STATE_UNDEF;
}
return rc;
}
topic[msg.topic_str_len] = '\0';
}
if (msg.msg_str_len > 0) {
str = calloc(1, msg.msg_str_len + 1);
if (NULL == str) {
orte_show_help("help-orte-odls-default.txt", "syscall fail",
true,
orte_process_info.nodename, context->app,
"opal_fd_read", __FILE__, __LINE__);
if (NULL != child) {
child->state = ORTE_PROC_STATE_UNDEF;
}
return rc;
}
rc = opal_fd_read(read_fd, msg.msg_str_len, str);
}
/* Print out what we got. We already have a rendered string,
so use orte_show_help_norender(). */
if (msg.msg_str_len > 0) {
orte_show_help_norender(file, topic, false, str);
free(str);
str = NULL;
}
/* If msg.fatal is true, then the child exited with an error.
Otherwise, whatever we just printed was a warning, so loop
around and see what else is on the pipe (or if the pipe
closed, indicating that the child launched
successfully). */
if (msg.fatal) {
if (NULL != child) {
child->state = ORTE_PROC_STATE_FAILED_TO_START;
child->alive = false;
}
close(read_fd);
return ORTE_SUCCESS;
}
}
/* If we got here, it means that the pipe closed without
indication of a fatal error, meaning that the child process
launched successfully. */
if (NULL != child) {
child->state = ORTE_PROC_STATE_LAUNCHED;
child->alive = true;
}
close(read_fd);
return ORTE_SUCCESS;
}
/**
* Fork/exec the specified processes
*/
static int odls_default_fork_local_proc(orte_app_context_t* context,
orte_odls_child_t *child,
char **environ_copy,
orte_odls_job_t *jobdat)
{
orte_iof_base_io_conf_t opts;
int rc, p[2];
pid_t pid;
if (NULL != child) {
/* should pull this information from MPIRUN instead of going with
default */
opts.usepty = OPAL_ENABLE_PTY_SUPPORT;
/* do we want to setup stdin? */
if (NULL != child &&
(jobdat->stdin_target == ORTE_VPID_WILDCARD || child->name->vpid == jobdat->stdin_target)) {
opts.connect_stdin = true;
} else {
opts.connect_stdin = false;
}
if (ORTE_SUCCESS != (rc = orte_iof_base_setup_prefork(&opts))) {
ORTE_ERROR_LOG(rc);
if (NULL != child) {
child->state = ORTE_PROC_STATE_FAILED_TO_START;
child->exit_code = rc;
}
return rc;
}
}
/* A pipe is used to communicate between the parent and child to
indicate whether the exec ultimately succeeded or failed. The
child sets the pipe to be close-on-exec; the child only ever
writes anything to the pipe if there is an error (e.g.,
executable not found, exec() fails, etc.). The parent does a
blocking read on the pipe; if the pipe closed with no data,
then the exec() succeeded. If the parent reads something from
the pipe, then the child was letting us know why it failed. */
if (pipe(p) < 0) {
ORTE_ERROR_LOG(ORTE_ERR_SYS_LIMITS_PIPES);
if (NULL != child) {
child->state = ORTE_PROC_STATE_FAILED_TO_START;
child->exit_code = ORTE_ERR_SYS_LIMITS_PIPES;
}
return ORTE_ERR_SYS_LIMITS_PIPES;
}
/* Fork off the child */
pid = fork();
if (NULL != child) {
child->pid = pid;
}
if (pid < 0) {
ORTE_ERROR_LOG(ORTE_ERR_SYS_LIMITS_CHILDREN);
if (NULL != child) {
child->state = ORTE_PROC_STATE_FAILED_TO_START;
child->exit_code = ORTE_ERR_SYS_LIMITS_CHILDREN;
}
return ORTE_ERR_SYS_LIMITS_CHILDREN;
}
if (pid == 0) {
close(p[0]);
do_child(context, child, environ_copy, jobdat, p[1], opts);
/* Does not return */
}
close(p[1]);
return do_parent(context, child, environ_copy, jobdat, p[0], opts);
}
/**
* Launch all processes allocated to the current node.
*/
int orte_odls_default_launch_local_procs(opal_buffer_t *data)
{
int rc;
orte_jobid_t job;
orte_job_t *jdata;
/* construct the list of children we are to launch */
if (ORTE_SUCCESS != (rc = orte_odls_base_default_construct_child_list(data, &job))) {
OPAL_OUTPUT_VERBOSE((2, orte_odls_globals.output,
"%s odls:default:launch:local failed to construct child list on error %s",
ORTE_NAME_PRINT(ORTE_PROC_MY_NAME), ORTE_ERROR_NAME(rc)));
goto CLEANUP;
}
/* launch the local procs */
if (ORTE_SUCCESS != (rc = orte_odls_base_default_launch_local(job, odls_default_fork_local_proc))) {
OPAL_OUTPUT_VERBOSE((2, orte_odls_globals.output,
"%s odls:default:launch:local failed to launch on error %s",
ORTE_NAME_PRINT(ORTE_PROC_MY_NAME), ORTE_ERROR_NAME(rc)));
goto CLEANUP;
}
/* look up job data object */
if (NULL != (jdata = orte_get_job_data_object(job))) {
if (jdata->state & ORTE_JOB_STATE_SUSPENDED) {
if (ORTE_PROC_IS_HNP) {
/* Have the plm send the signal to all the nodes.
If the signal arrived before the orteds started,
then they won't know to suspend their procs.
The plm also arranges for any local procs to
be signaled.
*/
orte_plm.signal_job(jdata->jobid, SIGTSTP);
} else {
orte_odls_default_signal_local_procs(NULL, SIGTSTP);
}
}
}
CLEANUP:
return rc;
}
/**
* Send a sigal to a pid. Note that if we get an error, we set the
* return value and let the upper layer print out the message.
*/
static int send_signal(pid_t pid, int signal)
{
int rc = ORTE_SUCCESS;
OPAL_OUTPUT_VERBOSE((1, orte_odls_globals.output,
"%s sending signal %d to pid %ld",
ORTE_NAME_PRINT(ORTE_PROC_MY_NAME),
signal, (long)pid));
if (orte_forward_job_control) {
/* Send the signal to the process group rather than the
process. The child is the leader of its process group. */
pid = -pid;
}
if (kill(pid, signal) != 0) {
switch(errno) {
case EINVAL:
rc = ORTE_ERR_BAD_PARAM;
break;
case ESRCH:
/* This case can occur when we deliver a signal to a
process that is no longer there. This can happen if
we deliver a signal while the job is shutting down.
This does not indicate a real problem, so just
ignore the error. */
break;
case EPERM:
rc = ORTE_ERR_PERM;
break;
default:
rc = ORTE_ERROR;
}
}
return rc;
}
static int orte_odls_default_signal_local_procs(const orte_process_name_t *proc, int32_t signal)
{
int rc;
if (ORTE_SUCCESS != (rc = orte_odls_base_default_signal_local_procs(proc, signal, send_signal))) {
ORTE_ERROR_LOG(rc);
return rc;
}
return ORTE_SUCCESS;
}
static int orte_odls_default_restart_proc(orte_odls_child_t *child)
{
int rc;
/* restart the local proc */
if (ORTE_SUCCESS != (rc = orte_odls_base_default_restart_proc(child, odls_default_fork_local_proc))) {
OPAL_OUTPUT_VERBOSE((2, orte_odls_globals.output,
"%s odls:default:restart_proc failed to launch on error %s",
ORTE_NAME_PRINT(ORTE_PROC_MY_NAME), ORTE_ERROR_NAME(rc)));
}
return rc;
}