# -*- text -*- # # Copyright (c) 2004-2005 The Trustees of Indiana University and Indiana # University Research and Technology # Corporation. All rights reserved. # Copyright (c) 2004-2005 The University of Tennessee and The University # of Tennessee Research Foundation. All rights # reserved. # Copyright (c) 2004-2005 High Performance Computing Center Stuttgart, # University of Stuttgart. All rights reserved. # Copyright (c) 2004-2005 The Regents of the University of California. # All rights reserved. # $COPYRIGHT$ # # Additional copyrights may follow # # $HEADER$ # # This is the US/English general help file for Open RTE's orterun. # [bproc-vexecmove-launch] The bproc PLS component was not able to launch %s on node %d and therefore cannot continue. Errno was set to %d. [bproc-vexecmove-fork] The bproc PLS component was not able to fork and therefore cannot continue. Errno was set to %d. [no-orted] The bproc PLS component was not able to find the executable "%s" in the current directory, your PATH, or in the directory where Open MPI was initially installed, and therefore cannot continue. For reference, we looked for %s Your current PATH is: %s We also looked for orted in the following directory: %s You may need to set your PATH properly, or set the MCA parameter pls_bproc_orted to be the path to "orted". [daemon-launch-number] The bproc PLS component was not able to launch all the daemons on the remote nodes and therefore cannot continue. We attempted to launch %d daemons but only %d were actually launched. For reference, we tried to launch %s [daemon-launch-bad-pid] The bproc PLS component was not able to launch all the daemons on the remote nodes and therefore cannot continue. On node %d the daemon pid was %d and errno was set to %d. For reference, we tried to launch %s [daemon-died-no-signal] A daemon (pid %d) launched by the bproc PLS component on node %d died unexpectedly so we are aborting. This may be because the daemon was unable to find all the needed shared libraries on the remote node. You may set your LD_LIBRARY_PATH to have the location of the shared libraries on the remote nodes and this will automatically be forwarded to the remote nodes. [daemon-died-signal] A daemon (pid %d) launched by the bproc PLS component on node %d died unexpectedly on signal %d so we are aborting. This may be because the daemon was unable to find all the needed shared libraries on the remote node. You may set your LD_LIBRARY_PATH to have the location of the shared libraries on the remote nodes and this will automatically be forwarded to the remote nodes. [proc-launch-number] The bproc PLS component was not able to launch all the processes on the remote nodes and therefore cannot continue. We attempted to launch %d processes but only %d were actually launched. For reference, we tried to launch %s [proc-launch-bad-pid] The bproc PLS component was not able to launch all the processes on the remote nodes and therefore cannot continue. On node %d the process pid was %d and errno was set to %d. For reference, we tried to launch %s [mismatched-slots] The current bproc support requires that the number of available slots on each node be the same. Note that this does -not- mean that the number of processes you want to launch must be the same. It only requires that you have access to the same number of process slots on each node. This is not something inherent to Open MPI, but rather a reported characteristic of Bproc. We are in the process of confirming that this requirement remains in effect. If we find that it has been removed, then we will revise the system to support varying numbers of slots on the allocated nodes. Meantime, please revise your hostfile or other allocation so they report the same number of process slots on each node. If you want to force a particular mapping of numbers of processes to each node, please use any of the other Open MPI mechanisms for doing so.