Realy deleted old files now
This commit was SVN r18018.
Этот коммит содержится в:
родитель
f734ba51a4
Коммит
cb83a1287d
@ -1,105 +0,0 @@
|
||||
# Copyright (c) 2004-2005 The Regents of the University of California.
|
||||
# All rights reserved.
|
||||
# $COPYRIGHT$
|
||||
#
|
||||
# Additional copyrights may follow
|
||||
#
|
||||
# $HEADER$
|
||||
#
|
||||
# This is the US/English general help file for rankle utilities.
|
||||
#
|
||||
# Voltaire
|
||||
[no-rankfile]
|
||||
Open RTE was unable to open the rankfile:
|
||||
%s
|
||||
Check to make sure the path and filename are correct.
|
||||
|
||||
usage: mpirun mca rmaps_rankfile_path rankfile ./app
|
||||
|
||||
all unspecified by rankfile ranks are assigned using
|
||||
byslot or bynode policy.
|
||||
|
||||
example: cat hosfile
|
||||
host1
|
||||
host2
|
||||
host3
|
||||
host4
|
||||
cat rankfile
|
||||
rank 1=host1 slot=1:0,1
|
||||
rank 0=host2 slot=0:*
|
||||
rank 2=host4 slot=1-2
|
||||
rank 3=host3 slot=0:1,1:0-2
|
||||
|
||||
[parse_error_string]
|
||||
Open RTE detected a parse error in the rankfile:
|
||||
%s
|
||||
It occured on line number %d on token %d:
|
||||
%s
|
||||
|
||||
[parse_error_int]
|
||||
Open RTE detected a parse error in the rankfile:
|
||||
%s
|
||||
It occured on line number %d on token %d:
|
||||
%d
|
||||
|
||||
[parse_error]
|
||||
Open RTE detected a parse error in the rankfile:
|
||||
%s
|
||||
It occured on line number %d on token %d.
|
||||
|
||||
[not-all-mapped-alloc]
|
||||
Some of the requested ranks are not included in the current allocation.
|
||||
%s
|
||||
|
||||
Please verify that you have specified the allocated resources properly in
|
||||
the provided rankfile.
|
||||
|
||||
[bad-host]
|
||||
Rankfile claimed host %s that was not allocated or oversubscribed it's slots:
|
||||
|
||||
[orte-rmaps-rf:alloc-error]
|
||||
There are not enough slots available in the system to satisfy the %d slots
|
||||
that were requested by the application:
|
||||
%s
|
||||
|
||||
Either request fewer slots for your application, or make more slots available
|
||||
for use.
|
||||
|
||||
[bad-rankfile]
|
||||
Error, Invalid rank (%d) in the rankfile (%s)
|
||||
|
||||
[bad-assign]
|
||||
Error, rank %d is already assigned to %s, check %s
|
||||
|
||||
[orte-rmaps-rf:multi-apps-and-zero-np]
|
||||
RMAPS found multiple applications to be launched, with
|
||||
at least one that failed to specify the number of processes to execute.
|
||||
When specifying multiple applications, you must specify how many processes
|
||||
of each to launch via the -np argument.
|
||||
|
||||
[orte-rmaps-rf:per-node-and-too-many-procs]
|
||||
There are not enough nodes in your allocation to satisfy your request to
|
||||
launch
|
||||
%d processes on a per-node basis - only %d nodes were available.
|
||||
|
||||
Either request fewer processes, or obtain a larger allocation.
|
||||
[orte-rmaps-rf:n-per-node-and-too-many-procs]
|
||||
There are not enough nodes in your allocation to satisfy your request to
|
||||
launch
|
||||
%d processes on a %d per-node basis - only %d nodes with a total of %d slots
|
||||
%were available.
|
||||
|
||||
Either request fewer processes, or obtain a larger allocation.
|
||||
[orte-rmaps-rf:n-per-node-and-not-enough-slots]
|
||||
There are not enough slots on the nodes in your allocation to satisfy your
|
||||
request to launch on a %d process-per-node basis - only %d slots/node were
|
||||
available.
|
||||
|
||||
Either request fewer processes/node, or obtain a larger allocation.
|
||||
|
||||
[orte-rmaps-rf:no-np-and-user-map]
|
||||
You have specified a rank-to-node/slot mapping, but failed to provide
|
||||
the number of processes to be executed. This information is critical for Rank
|
||||
Mapping component.
|
||||
|
||||
|
@ -1,661 +0,0 @@
|
||||
/*
|
||||
* Copyright (c) 2004-2007 The Trustees of Indiana University and Indiana
|
||||
* University Research and Technology
|
||||
* Corporation. All rights reserved.
|
||||
* Copyright (c) 2004-2006 The University of Tennessee and The University
|
||||
* of Tennessee Research Foundation. All rights
|
||||
* reserved.
|
||||
* Copyright (c) 2004-2005 High Performance Computing Center Stuttgart,
|
||||
* University of Stuttgart. All rights reserved.
|
||||
* Copyright (c) 2004-2005 The Regents of the University of California.
|
||||
* All rights reserved.
|
||||
* Copyright (c) 2006 Cisco Systems, Inc. All rights reserved.
|
||||
*
|
||||
* Copyright (c) 2008 Voltaire. All rights reserved
|
||||
*
|
||||
* $COPYRIGHT$
|
||||
*
|
||||
* Additional copyrights may follow
|
||||
*
|
||||
* $HEADER$
|
||||
*/
|
||||
|
||||
#include "orte_config.h"
|
||||
#include "orte/constants.h"
|
||||
#include "orte/types.h"
|
||||
|
||||
#include <errno.h>
|
||||
#ifdef HAVE_UNISTD_H
|
||||
#include <unistd.h>
|
||||
#endif /* HAVE_UNISTD_H */
|
||||
#ifdef HAVE_STRING_H
|
||||
#include <string.h>
|
||||
#endif /* HAVE_STRING_H */
|
||||
|
||||
#include "opal/mca/base/mca_base_param.h"
|
||||
#include "opal/util/output.h"
|
||||
#include "opal/util/trace.h"
|
||||
#include "opal/util/show_help.h"
|
||||
#include "opal/util/argv.h"
|
||||
#include "opal/util/if.h"
|
||||
|
||||
#include "orte/mca/errmgr/errmgr.h"
|
||||
|
||||
#include "orte/mca/rmaps/base/rmaps_private.h"
|
||||
#include "orte/mca/rmaps/base/base.h"
|
||||
#include "rmaps_rf.h"
|
||||
#include "orte/mca/rmaps/rank_file/rankfile_lex.h"
|
||||
#include "orte/runtime/orte_globals.h"
|
||||
#include "orte/mca/ras/ras_types.h"
|
||||
|
||||
static int orte_rmaps_rank_file_parse(const char *, int);
|
||||
static char *rankfile_parse_string_or_int(void);
|
||||
char *rank_file_path = NULL;
|
||||
static const char *cur_rankfile_name = NULL;
|
||||
static opal_mutex_t rankfile_mutex;
|
||||
|
||||
/*
|
||||
* Local variable
|
||||
*/
|
||||
static opal_list_item_t *cur_node_item = NULL;
|
||||
orte_rmaps_rank_file_map_t *rankmap = NULL;
|
||||
|
||||
static int map_app_by_user_map(
|
||||
orte_app_context_t* app,
|
||||
orte_job_t* jdata,
|
||||
orte_vpid_t vpid_start,
|
||||
opal_list_t* nodes,
|
||||
opal_list_t* procs)
|
||||
{
|
||||
|
||||
int rc = ORTE_SUCCESS;
|
||||
opal_list_item_t *next;
|
||||
orte_node_t *node;
|
||||
orte_std_cntr_t num_alloc=0, round_cnt;
|
||||
|
||||
OPAL_TRACE(2);
|
||||
|
||||
while (num_alloc < app->num_procs) {
|
||||
/** see if any nodes remain unused and available. We need to do this check
|
||||
* each time since we may remove nodes from the list (as they become fully
|
||||
* used) as we cycle through the loop */
|
||||
if(0 >= opal_list_get_size(nodes) ) {
|
||||
/* No more nodes to allocate :( */
|
||||
opal_show_help("help-orte-rmaps-rf.txt", "orte-rmaps-rf:alloc-error",
|
||||
true, app->num_procs, app->app);
|
||||
return ORTE_ERR_SILENT;
|
||||
}
|
||||
|
||||
/* Save the next node we can use before claiming slots, since
|
||||
* we may need to prune the nodes list removing overused nodes.
|
||||
* Wrap around to beginning if we are at the end of the list */
|
||||
round_cnt=0;
|
||||
if ( -1 != rankmap[vpid_start + num_alloc].rank) {
|
||||
do {
|
||||
if (opal_list_get_end(nodes) == opal_list_get_next(cur_node_item)) {
|
||||
next = opal_list_get_first(nodes);
|
||||
round_cnt++;
|
||||
} else {
|
||||
next = opal_list_get_next(cur_node_item);
|
||||
}
|
||||
/* Allocate a slot on this node */
|
||||
node = (orte_node_t*) cur_node_item;
|
||||
cur_node_item = next;
|
||||
if ( round_cnt == 2 ) {
|
||||
opal_show_help("help-orte-rmaps-rf.txt","bad-host", true,rankmap[num_alloc+vpid_start].node_name);
|
||||
ORTE_ERROR_LOG(ORTE_ERR_BAD_PARAM);
|
||||
return ORTE_ERR_BAD_PARAM;
|
||||
}
|
||||
} while ( strcmp(node->name, rankmap[num_alloc + vpid_start].node_name));
|
||||
node->slot_list = strdup(rankmap[num_alloc+vpid_start].slot_list);
|
||||
if (mca_rmaps_rank_file_component.debug) {
|
||||
opal_output(0, "rank_file RMAPS component: [%s:%d]->slot_list=%s\n",
|
||||
rankmap[num_alloc + vpid_start].node_name,rankmap[num_alloc+vpid_start].rank, node->slot_list);
|
||||
}
|
||||
if (ORTE_SUCCESS != (rc = orte_rmaps_base_claim_slot(jdata, node, rankmap[num_alloc+vpid_start].rank, app->idx,
|
||||
nodes, jdata->map->oversubscribe))) {
|
||||
/** if the code is ORTE_ERR_NODE_FULLY_USED, then we know this
|
||||
* really isn't an error - we just need to break from the loop
|
||||
* since the node is fully used up. For now, just don't report
|
||||
* an error
|
||||
*/
|
||||
if (ORTE_ERR_NODE_FULLY_USED != rc) {
|
||||
ORTE_ERROR_LOG(rc);
|
||||
return rc;
|
||||
}
|
||||
}
|
||||
}
|
||||
++num_alloc;
|
||||
}
|
||||
return ORTE_SUCCESS;
|
||||
}
|
||||
|
||||
|
||||
|
||||
/*
|
||||
* Create a default mapping for the application, mapping rank by rank_file and
|
||||
* by node.
|
||||
*/
|
||||
static int map_app_by_node(
|
||||
orte_app_context_t* app,
|
||||
orte_job_t* jdata,
|
||||
orte_vpid_t vpid_start,
|
||||
opal_list_t* nodes )
|
||||
{
|
||||
int rc = ORTE_SUCCESS;
|
||||
opal_list_item_t *next;
|
||||
orte_node_t *node;
|
||||
orte_std_cntr_t num_alloc = 0;
|
||||
|
||||
OPAL_TRACE(2);
|
||||
|
||||
/* This loop continues until all procs have been mapped or we run
|
||||
out of resources. We determine that we have "run out of
|
||||
resources" when all nodes have slots_max processes mapped to them,
|
||||
thus there are no free slots for a process to be mapped, or we have
|
||||
hit the soft limit on all nodes and are in a "no oversubscribe" state.
|
||||
If we still have processes that haven't been mapped yet, then it's an
|
||||
"out of resources" error.
|
||||
|
||||
In this scenario, we rely on the claim_slot function to handle the
|
||||
oversubscribed case. The claim_slot function will leave a node on the
|
||||
list until it either reaches slots_max OR reaches the
|
||||
soft limit and the "no_oversubscribe" flag has been set - at which point,
|
||||
the node will be removed to prevent any more processes from being mapped to
|
||||
it. Since we are taking one slot from each node as we cycle through, the
|
||||
list, oversubscription is automatically taken care of via this logic.
|
||||
*/
|
||||
|
||||
while (num_alloc < app->num_procs) {
|
||||
if ( -1 != rankmap[num_alloc + vpid_start].rank) {
|
||||
++num_alloc;
|
||||
continue;
|
||||
}
|
||||
/** see if any nodes remain unused and available. We need to do this check
|
||||
* each time since we may remove nodes from the list (as they become fully
|
||||
* used) as we cycle through the loop */
|
||||
if(0 >= opal_list_get_size(nodes) ) {
|
||||
/* No more nodes to allocate :( */
|
||||
opal_show_help("help-orte-rmaps-rf.txt", "orte-rmaps-rf:alloc-error",
|
||||
true, app->num_procs, app->app);
|
||||
return ORTE_ERR_SILENT;
|
||||
}
|
||||
|
||||
/* Save the next node we can use before claiming slots, since
|
||||
* we may need to prune the nodes list removing overused nodes.
|
||||
* Wrap around to beginning if we are at the end of the list */
|
||||
if (opal_list_get_end(nodes) == opal_list_get_next(cur_node_item)) {
|
||||
next = opal_list_get_first(nodes);
|
||||
}
|
||||
else {
|
||||
next = opal_list_get_next(cur_node_item);
|
||||
}
|
||||
/* Allocate a slot on this node */
|
||||
node = (orte_node_t*) cur_node_item;
|
||||
if (ORTE_SUCCESS != (rc = orte_rmaps_base_claim_slot(jdata, node, vpid_start + num_alloc, app->idx,
|
||||
nodes, jdata->map->oversubscribe))) {
|
||||
/** if the code is ORTE_ERR_NODE_FULLY_USED, then we know this
|
||||
* really isn't an error - we just need to break from the loop
|
||||
* since the node is fully used up. For now, just don't report
|
||||
* an error
|
||||
*/
|
||||
if (ORTE_ERR_NODE_FULLY_USED != rc) {
|
||||
ORTE_ERROR_LOG(rc);
|
||||
return rc;
|
||||
}
|
||||
}
|
||||
++num_alloc;
|
||||
cur_node_item = next;
|
||||
}
|
||||
|
||||
return ORTE_SUCCESS;
|
||||
}
|
||||
|
||||
/*
|
||||
* Create a default mapping for the application, scheduling ranks byr rank_file
|
||||
* and by slot.
|
||||
*/
|
||||
static int map_app_by_slot(
|
||||
orte_app_context_t* app,
|
||||
orte_job_t* jdata,
|
||||
orte_vpid_t vpid_start,
|
||||
opal_list_t* nodes )
|
||||
{
|
||||
int rc = ORTE_SUCCESS;
|
||||
orte_std_cntr_t i, num_slots_to_take, num_alloc = 0;
|
||||
orte_node_t *node;
|
||||
opal_list_item_t *next;
|
||||
|
||||
OPAL_TRACE(2);
|
||||
/* This loop continues until all procs have been mapped or we run
|
||||
out of resources. We determine that we have "run out of
|
||||
resources" when either all nodes have slots_max processes mapped to them,
|
||||
(thus there are no free slots for a process to be mapped), OR all nodes
|
||||
have reached their soft limit and the user directed us to "no oversubscribe".
|
||||
If we still have processes that haven't been mapped yet, then it's an
|
||||
"out of resources" error. */
|
||||
|
||||
while ( num_alloc < app->num_procs) {
|
||||
/** see if any nodes remain unused and available. We need to do this check
|
||||
* each time since we may remove nodes from the list (as they become fully
|
||||
* used) as we cycle through the loop */
|
||||
if(0 >= opal_list_get_size(nodes) ) {
|
||||
/* Everything is at max usage! :( */
|
||||
opal_show_help("help-orte-rmaps-rf.txt", "orte-rmaps-rf:alloc-error",
|
||||
true, app->num_procs, app->app);
|
||||
return ORTE_ERR_SILENT;
|
||||
}
|
||||
|
||||
/* Save the next node we can use before claiming slots, since
|
||||
* we may need to prune the nodes list removing overused nodes.
|
||||
* Wrap around to beginning if we are at the end of the list */
|
||||
if (opal_list_get_end(nodes) == opal_list_get_next(cur_node_item)) {
|
||||
next = opal_list_get_first(nodes);
|
||||
} else {
|
||||
next = opal_list_get_next(cur_node_item);
|
||||
}
|
||||
|
||||
/** declare a shorter name for convenience in the code below */
|
||||
node = (orte_node_t*) cur_node_item;
|
||||
/* If we have available slots on this node, claim all of them
|
||||
* If node_slots == 0, assume 1 slot for that node.
|
||||
* JJH - is this assumption fully justified?
|
||||
*
|
||||
* If we are now oversubscribing the nodes, then we still take:
|
||||
* (a) if the node has not been used yet, we take a full node_slots
|
||||
* (b) if some of the slots are in-use, then we take the number of
|
||||
* remaining slots before hitting the soft limit (node_slots)
|
||||
* (c) if we are at or above the soft limit, we take a full node_slots
|
||||
*
|
||||
* Note: if node_slots is zero, then we always just take 1 slot
|
||||
*
|
||||
* We continue this process until either everything is done,
|
||||
* or all nodes have hit their hard limit. This algorithm ensures we
|
||||
* fully utilize each node before oversubscribing, and preserves the ratio
|
||||
* of processes between the nodes thereafter (e.g., if one node has twice as
|
||||
* many processes as another before oversubscribing, it will continue
|
||||
* to do so after oversubscribing).
|
||||
*/
|
||||
|
||||
|
||||
if (0 == node->slots_inuse ||
|
||||
node->slots_inuse >= node->slots) {
|
||||
num_slots_to_take = (node->slots == 0) ? 1 : node->slots;
|
||||
} else {
|
||||
num_slots_to_take = node->slots - node->slots_inuse;
|
||||
}
|
||||
|
||||
/* check if we are in npernode mode - if so, then set the num_slots_to_take
|
||||
* to the num_per_node
|
||||
*/
|
||||
if (jdata->map->pernode) {
|
||||
num_slots_to_take = jdata->map->npernode;
|
||||
}
|
||||
|
||||
for( i = 0; i < num_slots_to_take; ++i) {
|
||||
if ( -1 != rankmap[num_alloc + vpid_start].rank) {
|
||||
++num_alloc;
|
||||
continue;
|
||||
}
|
||||
if (ORTE_SUCCESS != (rc = orte_rmaps_base_claim_slot(jdata, node, vpid_start + num_alloc, app->idx,
|
||||
nodes, jdata->map->oversubscribe))) {
|
||||
/** if the code is ORTE_ERR_NODE_FULLY_USED, then we know this
|
||||
* really isn't an error - we just need to break from the loop
|
||||
* since the node is fully used up. For now, just don't report
|
||||
* an error
|
||||
*/
|
||||
if (ORTE_ERR_NODE_FULLY_USED != rc) {
|
||||
ORTE_ERROR_LOG(rc);
|
||||
return rc;
|
||||
}
|
||||
}
|
||||
/* Update the number of procs allocated */
|
||||
++num_alloc;
|
||||
|
||||
/** if all the procs have been mapped OR we have fully used up this node, then
|
||||
* break from the loop
|
||||
*/
|
||||
if(num_alloc == app->num_procs || ORTE_ERR_NODE_FULLY_USED == rc) {
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
/* we move on to the next node in all cases EXCEPT if we came
|
||||
* out of the loop without having taken a full bite AND the
|
||||
* node is NOT max'd out
|
||||
*
|
||||
*/
|
||||
if (i < (num_slots_to_take-1) && ORTE_ERR_NODE_FULLY_USED != rc) {
|
||||
continue;
|
||||
}
|
||||
cur_node_item = next;
|
||||
}
|
||||
|
||||
return ORTE_SUCCESS;
|
||||
}
|
||||
|
||||
|
||||
/*
|
||||
* Create a rank_file mapping for the job.
|
||||
*/
|
||||
static int orte_rmaps_rf_map(orte_job_t *jdata)
|
||||
{
|
||||
orte_job_map_t *map;
|
||||
orte_app_context_t *app, **apps;
|
||||
orte_std_cntr_t i, j;
|
||||
opal_list_t node_list, procs;
|
||||
opal_list_item_t *item;
|
||||
orte_node_t *node;
|
||||
orte_vpid_t vpid_start;
|
||||
orte_std_cntr_t num_nodes, num_slots;
|
||||
int rc;
|
||||
orte_std_cntr_t slots_per_node;
|
||||
|
||||
OPAL_TRACE(1);
|
||||
|
||||
/* conveniece def */
|
||||
map = jdata->map;
|
||||
apps = (orte_app_context_t**)jdata->apps->addr;
|
||||
|
||||
/* start at the beginning... */
|
||||
vpid_start = 0;
|
||||
|
||||
/* cycle through the app_contexts, mapping them sequentially */
|
||||
for(i=0; i < jdata->num_apps; i++) {
|
||||
app = apps[i];
|
||||
|
||||
/* if the number of processes wasn't specified, then we know there can be only
|
||||
* one app_context allowed in the launch, and that we are to launch it across
|
||||
* all available slots. We'll double-check the single app_context rule first
|
||||
*/
|
||||
if (0 == app->num_procs && 1 < jdata->num_apps) {
|
||||
opal_show_help("help-orte-rmaps-rf.txt", "orte-rmaps-rf:multi-apps-and-zero-np",
|
||||
true, jdata->num_apps, NULL);
|
||||
rc = ORTE_ERR_SILENT;
|
||||
goto error;
|
||||
}
|
||||
|
||||
/* for each app_context, we have to get the list of nodes that it can
|
||||
* use since that can now be modified with a hostfile and/or -host
|
||||
* option
|
||||
*/
|
||||
OBJ_CONSTRUCT(&node_list, opal_list_t);
|
||||
if(ORTE_SUCCESS != (rc = orte_rmaps_base_get_target_nodes(&node_list, &num_slots, app,
|
||||
map->no_use_local))) {
|
||||
ORTE_ERROR_LOG(rc);
|
||||
goto error;
|
||||
}
|
||||
num_nodes = (orte_std_cntr_t)opal_list_get_size(&node_list);
|
||||
|
||||
rankmap = (orte_rmaps_rank_file_map_t *) malloc ( app->num_procs * sizeof(orte_rmaps_rank_file_map_t));
|
||||
for ( j=0; j<app->num_procs; j++) {
|
||||
rankmap[j].rank = -1;
|
||||
rankmap[j].slot_list = (char *)malloc(64*sizeof(char));
|
||||
}
|
||||
if ( ORTE_SUCCESS != (rc = orte_rmaps_rank_file_parse(rank_file_path, app->num_procs))) {
|
||||
ORTE_ERROR_LOG(rc);
|
||||
goto error;
|
||||
}
|
||||
|
||||
/* if a bookmark exists from some prior mapping, set us to start there */
|
||||
if (NULL != jdata->bookmark) {
|
||||
cur_node_item = NULL;
|
||||
/* find this node on the list */
|
||||
for (item = opal_list_get_first(&node_list);
|
||||
item != opal_list_get_end(&node_list);
|
||||
item = opal_list_get_next(item)) {
|
||||
node = (orte_node_t*)item;
|
||||
|
||||
if (node->index == jdata->bookmark->index) {
|
||||
cur_node_item = item;
|
||||
break;
|
||||
}
|
||||
}
|
||||
/* see if we found it - if not, just start at the beginning */
|
||||
if (NULL == cur_node_item) {
|
||||
cur_node_item = opal_list_get_first(&node_list);
|
||||
}
|
||||
} else {
|
||||
/* if no bookmark, then just start at the beginning of the list */
|
||||
cur_node_item = opal_list_get_first(&node_list);
|
||||
}
|
||||
|
||||
if (map->pernode && map->npernode == 1) {
|
||||
/* there are three use-cases that we need to deal with:
|
||||
* (a) if -np was not provided, then we just use the number of nodes
|
||||
* (b) if -np was provided AND #procs > #nodes, then error out
|
||||
* (c) if -np was provided AND #procs <= #nodes, then launch
|
||||
* the specified #procs one/node. In this case, we just
|
||||
* leave app->num_procs alone
|
||||
*/
|
||||
if (0 == app->num_procs) {
|
||||
app->num_procs = num_nodes;
|
||||
} else if (app->num_procs > num_nodes) {
|
||||
opal_show_help("help-orte-rmaps-rf.txt", "orte-rmaps-rf:per-node-and-too-many-procs",
|
||||
true, app->num_procs, num_nodes, NULL);
|
||||
rc = ORTE_ERR_SILENT;
|
||||
goto error;
|
||||
}
|
||||
} else if (map->pernode && map->npernode > 1) {
|
||||
/* first, let's check to see if there are enough slots/node to
|
||||
* meet the request - error out if not
|
||||
*/
|
||||
slots_per_node = num_slots / num_nodes;
|
||||
if (map->npernode > slots_per_node) {
|
||||
opal_show_help("help-orte-rmaps-rf.txt", "orte-rmaps-rf:n-per-node-and-not-enough-slots",
|
||||
true, map->npernode, slots_per_node, NULL);
|
||||
rc = ORTE_ERR_SILENT;
|
||||
goto error;
|
||||
}
|
||||
/* there are three use-cases that we need to deal with:
|
||||
* (a) if -np was not provided, then we just use the n/node * #nodes
|
||||
* (b) if -np was provided AND #procs > (n/node * #nodes), then error out
|
||||
* (c) if -np was provided AND #procs <= (n/node * #nodes), then launch
|
||||
* the specified #procs n/node. In this case, we just
|
||||
* leave app->num_procs alone
|
||||
*/
|
||||
if (0 == app->num_procs) {
|
||||
/* set the num_procs to equal the specified num/node * the number of nodes */
|
||||
app->num_procs = map->npernode * num_nodes;
|
||||
} else if (app->num_procs > (map->npernode * num_nodes)) {
|
||||
opal_show_help("help-orte-rmaps-rf.txt", "orte-rmaps-rf:n-per-node-and-too-many-procs",
|
||||
true, app->num_procs, map->npernode, num_nodes, num_slots, NULL);
|
||||
rc = ORTE_ERR_SILENT;
|
||||
goto error;
|
||||
}
|
||||
} else if (0 == app->num_procs) {
|
||||
/* we can't handle this - it should have been set when we got
|
||||
* the map info. If it wasn't, then we can only error out
|
||||
*/
|
||||
opal_show_help("help-orte-rmaps-rf.txt", "orte-rmaps-rf:no-np-and-user-map",
|
||||
true, app->num_procs, map->npernode, num_nodes, num_slots, NULL);
|
||||
rc = ORTE_ERR_SILENT;
|
||||
goto error;
|
||||
}
|
||||
/** track the total number of processes we mapped */
|
||||
jdata->num_procs += app->num_procs;
|
||||
|
||||
/* Make assignments */
|
||||
rc = map_app_by_user_map(app, jdata, vpid_start, &node_list, &procs);
|
||||
|
||||
/* assign unassigned ranks by map policy */
|
||||
if (map->policy == ORTE_RMAPS_BYNODE) {
|
||||
rc = map_app_by_node(app, jdata, vpid_start, &node_list);
|
||||
} else {
|
||||
rc = map_app_by_slot(app, jdata, vpid_start, &node_list);
|
||||
}
|
||||
|
||||
if (ORTE_SUCCESS != rc) {
|
||||
ORTE_ERROR_LOG(rc);
|
||||
goto error;
|
||||
}
|
||||
|
||||
/* save the bookmark */
|
||||
jdata->bookmark = (orte_node_t*)cur_node_item;
|
||||
|
||||
/* cleanup the node list - it can differ from one app_context
|
||||
* to another, so we have to get it every time
|
||||
*/
|
||||
while(NULL != (item = opal_list_remove_first(&node_list))) {
|
||||
OBJ_RELEASE(item);
|
||||
}
|
||||
OBJ_DESTRUCT(&node_list);
|
||||
}
|
||||
|
||||
/* compute and save convenience values */
|
||||
if (ORTE_SUCCESS != (rc = orte_rmaps_base_compute_usage(jdata))) {
|
||||
ORTE_ERROR_LOG(rc);
|
||||
return rc;
|
||||
}
|
||||
|
||||
/* define the daemons that we will use for this job */
|
||||
if (ORTE_SUCCESS != (rc = orte_rmaps_base_define_daemons(map))) {
|
||||
ORTE_ERROR_LOG(rc);
|
||||
return rc;
|
||||
}
|
||||
for (j=0; j<app->num_procs; j++) {
|
||||
if (NULL != rankmap[j].slot_list) {
|
||||
free (rankmap[j].slot_list );
|
||||
}
|
||||
}
|
||||
if (NULL != rankmap) {
|
||||
free(rankmap);
|
||||
}
|
||||
return ORTE_SUCCESS;
|
||||
|
||||
error:
|
||||
while(NULL != (item = opal_list_remove_first(&node_list))) {
|
||||
OBJ_RELEASE(item);
|
||||
}
|
||||
OBJ_DESTRUCT(&node_list);
|
||||
for (j=0; j<app->num_procs; j++) {
|
||||
if (NULL != rankmap[j].slot_list) {
|
||||
free (rankmap[j].slot_list );
|
||||
}
|
||||
}
|
||||
if (NULL != rankmap) {
|
||||
free(rankmap);
|
||||
}
|
||||
return rc;
|
||||
}
|
||||
|
||||
orte_rmaps_base_module_t orte_rmaps_rank_file_module = {
|
||||
orte_rmaps_rf_map
|
||||
};
|
||||
|
||||
|
||||
static int orte_rmaps_rank_file_parse(const char *rankfile, int np)
|
||||
{
|
||||
int token;
|
||||
int rc = ORTE_SUCCESS;
|
||||
int line_number = 1;
|
||||
int cnt;
|
||||
char* node_name = NULL;
|
||||
char* username = NULL;
|
||||
char** argv;
|
||||
char buff[64];
|
||||
char* value;
|
||||
int ival;
|
||||
|
||||
OPAL_THREAD_LOCK(&rankfile_mutex);
|
||||
|
||||
cur_rankfile_name = rankfile;
|
||||
rank_file_done = false;
|
||||
rank_file_in = fopen(rankfile, "r");
|
||||
|
||||
if ( NULL == rank_file_in) {
|
||||
opal_show_help("help-orte-rmaps-rf.txt", "no-rankfile", true, rankfile, np);
|
||||
rc = OPAL_ERR_NOT_FOUND;
|
||||
goto unlock;
|
||||
}
|
||||
|
||||
while (!rank_file_done) {
|
||||
token = rank_file_lex();
|
||||
switch (token) {
|
||||
case ORTE_RANKFILE_DONE:
|
||||
rank_file_done = true;
|
||||
break;
|
||||
case ORTE_RANKFILE_NEWLINE:
|
||||
line_number++;
|
||||
break;
|
||||
case ORTE_RANKFILE_RANK:
|
||||
break;
|
||||
case ORTE_RANKFILE_EQUAL:
|
||||
ival = rank_file_value.ival;
|
||||
if ( ival > (np-1) ) {
|
||||
opal_show_help("help-orte-rmaps-rf.txt", "bad-rankfile", true, ival, rankfile);
|
||||
rc = ORTE_ERR_BAD_PARAM;
|
||||
goto unlock;
|
||||
}
|
||||
token = rank_file_lex();
|
||||
switch (token) {
|
||||
case ORTE_RANKFILE_HOSTNAME:
|
||||
case ORTE_RANKFILE_IPV4:
|
||||
case ORTE_RANKFILE_IPV6:
|
||||
case ORTE_RANKFILE_STRING:
|
||||
case ORTE_RANKFILE_INT:
|
||||
if(ORTE_RANKFILE_INT == token) {
|
||||
sprintf(buff,"%d", rank_file_value.ival);
|
||||
value = buff;
|
||||
} else {
|
||||
value = rank_file_value.sval;
|
||||
}
|
||||
argv = opal_argv_split (value, '@');
|
||||
cnt = opal_argv_count (argv);
|
||||
if (1 == cnt) {
|
||||
node_name = strdup(argv[0]);
|
||||
} else if (2 == cnt) {
|
||||
username = strdup(argv[0]);
|
||||
node_name = strdup(argv[1]);
|
||||
}
|
||||
else {
|
||||
opal_output(0, "WARNING: Unhandled user@host-combination\n"); /* XXX */
|
||||
}
|
||||
opal_argv_free (argv);
|
||||
rankmap[ival].rank = ival;
|
||||
rankmap[ival].node_name = strdup(node_name);
|
||||
/* convert this into something globally unique */
|
||||
if (strcmp(node_name, "localhost") == 0 || opal_ifislocal(node_name)) {
|
||||
/* Nodename has been allocated, that is for sure */
|
||||
free (node_name);
|
||||
}
|
||||
}
|
||||
break;
|
||||
case ORTE_RANKFILE_SLOT:
|
||||
rankmap[ival].slot_list = strdup(rankfile_parse_string_or_int());
|
||||
break;
|
||||
}
|
||||
}
|
||||
fclose(rank_file_in);
|
||||
rank_file_in = NULL;
|
||||
|
||||
unlock:
|
||||
cur_rankfile_name = NULL;
|
||||
OPAL_THREAD_UNLOCK(&rankfile_mutex);
|
||||
return rc;
|
||||
}
|
||||
|
||||
|
||||
static char *rankfile_parse_string_or_int(void)
|
||||
{
|
||||
int rc;
|
||||
char tmp_str[64];
|
||||
|
||||
if (ORTE_RANKFILE_EQUAL != rank_file_lex()){
|
||||
return NULL;
|
||||
}
|
||||
|
||||
rc = rank_file_lex();
|
||||
switch (rc) {
|
||||
case ORTE_RANKFILE_STRING:
|
||||
return strdup(rank_file_value.sval);
|
||||
case ORTE_RANKFILE_INT:
|
||||
sprintf(tmp_str,"%d",rank_file_value.ival);
|
||||
return strdup(tmp_str);
|
||||
default:
|
||||
return NULL;
|
||||
|
||||
}
|
||||
|
||||
}
|
||||
|
@ -1,71 +0,0 @@
|
||||
/*
|
||||
* Copyright (c) 2004-2005 The Trustees of Indiana University and Indiana
|
||||
* University Research and Technology
|
||||
* Corporation. All rights reserved.
|
||||
* Copyright (c) 2004-2006 The University of Tennessee and The University
|
||||
* of Tennessee Research Foundation. All rights
|
||||
* reserved.
|
||||
* Copyright (c) 2004-2005 High Performance Computing Center Stuttgart,
|
||||
* University of Stuttgart. All rights reserved.
|
||||
* Copyright (c) 2004-2005 The Regents of the University of California.
|
||||
* All rights reserved.
|
||||
* Copyright (c) 2008 Voltaire. All rights reserved
|
||||
*
|
||||
* $COPYRIGHT$
|
||||
*
|
||||
* Additional copyrights may follow
|
||||
*
|
||||
* $HEADER$
|
||||
*/
|
||||
/**
|
||||
* @file
|
||||
*
|
||||
* Resource Mapping
|
||||
*/
|
||||
|
||||
|
||||
#include "opal_config.h"
|
||||
|
||||
#include "opal/mca/paffinity/paffinity.h"
|
||||
|
||||
#ifndef ORTE_RMAPS_RF_H
|
||||
#define ORTE_RMAPS_RF_H
|
||||
|
||||
#include "orte/mca/rmaps/rmaps.h"
|
||||
#if defined(c_plusplus) || defined(__cplusplus)
|
||||
extern "C" {
|
||||
#endif
|
||||
|
||||
/**
|
||||
* RMGR Component
|
||||
*/
|
||||
struct orte_rmaps_rank_file_component_t {
|
||||
orte_rmaps_base_component_t super;
|
||||
int debug;
|
||||
int priority;
|
||||
};
|
||||
typedef struct orte_rmaps_rank_file_component_t orte_rmaps_rank_file_component_t;
|
||||
|
||||
ORTE_MODULE_DECLSPEC extern orte_rmaps_rank_file_component_t mca_rmaps_rank_file_component;
|
||||
extern orte_rmaps_base_module_t orte_rmaps_rank_file_module;
|
||||
|
||||
|
||||
extern char *rank_file_path;
|
||||
|
||||
typedef struct cpu_socket_t cpu_socket_t;
|
||||
|
||||
struct orte_rmaps_rank_file_map_t {
|
||||
int rank;
|
||||
char* node_name;
|
||||
char* slot_list;
|
||||
};
|
||||
typedef struct orte_rmaps_rank_file_map_t orte_rmaps_rank_file_map_t;
|
||||
|
||||
|
||||
ORTE_DECLSPEC OBJ_CLASS_DECLARATION(orte_rmaps_rank_file_map_t);
|
||||
|
||||
#if defined(c_plusplus) || defined(__cplusplus)
|
||||
}
|
||||
#endif
|
||||
|
||||
#endif
|
@ -1,128 +0,0 @@
|
||||
/*
|
||||
* Copyright (c) 2004-2007 The Trustees of Indiana University and Indiana
|
||||
* University Research and Technology
|
||||
* Corporation. All rights reserved.
|
||||
* Copyright (c) 2004-2005 The University of Tennessee and The University
|
||||
* of Tennessee Research Foundation. All rights
|
||||
* reserved.
|
||||
* Copyright (c) 2004-2005 High Performance Computing Center Stuttgart,
|
||||
* University of Stuttgart. All rights reserved.
|
||||
* Copyright (c) 2004-2005 The Regents of the University of California.
|
||||
* All rights reserved.
|
||||
* Copyright (c) 2008 Voltaire. All rights reserved
|
||||
*
|
||||
* $COPYRIGHT$
|
||||
*
|
||||
* Additional copyrights may follow
|
||||
*
|
||||
* $HEADER$
|
||||
*/
|
||||
|
||||
#include "orte_config.h"
|
||||
#include "orte/constants.h"
|
||||
|
||||
#include "orte/runtime/orte_globals.h"
|
||||
#include "orte/mca/ras/ras_types.h"
|
||||
|
||||
#include "opal/mca/base/base.h"
|
||||
#include "opal/mca/base/mca_base_param.h"
|
||||
#include "opal/util/output.h"
|
||||
|
||||
#include "orte/mca/rmaps/base/rmaps_private.h"
|
||||
#include "rmaps_rf.h"
|
||||
#include "orte/mca/rmaps/rank_file/rankfile_lex.h"
|
||||
|
||||
/*
|
||||
* Local functions
|
||||
*/
|
||||
|
||||
static int orte_rmaps_rank_file_open(void);
|
||||
static int orte_rmaps_rank_file_close(void);
|
||||
static orte_rmaps_base_module_t* orte_rmaps_rank_file_init(int* priority);
|
||||
|
||||
orte_rmaps_rank_file_component_t mca_rmaps_rank_file_component = {
|
||||
{
|
||||
/* First, the mca_base_component_t struct containing meta
|
||||
information about the component itself */
|
||||
|
||||
{
|
||||
/* Indicate that we are a rmaps v1.3.0 component (which also
|
||||
implies a specific MCA version) */
|
||||
|
||||
ORTE_RMAPS_BASE_VERSION_1_3_0,
|
||||
|
||||
"rank_file", /* MCA component name */
|
||||
ORTE_MAJOR_VERSION, /* MCA component major version */
|
||||
ORTE_MINOR_VERSION, /* MCA component minor version */
|
||||
ORTE_RELEASE_VERSION, /* MCA component release version */
|
||||
orte_rmaps_rank_file_open, /* component open */
|
||||
orte_rmaps_rank_file_close /* component close */
|
||||
},
|
||||
|
||||
/* Next the MCA v1.0.0 component meta data */
|
||||
{
|
||||
/* The component is checkpoint ready */
|
||||
MCA_BASE_METADATA_PARAM_CHECKPOINT
|
||||
},
|
||||
|
||||
orte_rmaps_rank_file_init
|
||||
}
|
||||
};
|
||||
|
||||
|
||||
/**
|
||||
* component open/close/init function
|
||||
*/
|
||||
static int orte_rmaps_rank_file_open(void)
|
||||
{
|
||||
int index;
|
||||
|
||||
mca_base_param_reg_int(&mca_rmaps_rank_file_component.super.rmaps_version, "debug",
|
||||
"Toggle debug output for Rank File RMAPS component",
|
||||
false, false, 0,
|
||||
&mca_rmaps_rank_file_component.debug);
|
||||
|
||||
mca_base_param_reg_int(&mca_rmaps_rank_file_component.super.rmaps_version, "priority",
|
||||
"Selection priority for Rank File RMAPS component",
|
||||
false, false, 1,
|
||||
&mca_rmaps_rank_file_component.priority);
|
||||
|
||||
mca_base_param_reg_string(&mca_rmaps_rank_file_component.super.rmaps_version,
|
||||
"path",
|
||||
"The path to the rank mapping file",
|
||||
false, false, NULL, &rank_file_path);
|
||||
if (NULL != rank_file_path) {
|
||||
index = mca_base_param_find("rank_file",NULL,NULL);
|
||||
if ( OPAL_ERROR != index) {
|
||||
mca_base_param_set_string(index,"rank_file");
|
||||
}
|
||||
mca_base_param_lookup_string(index, &rank_file_path);
|
||||
/* if rankfile path is present than set higher priority to this component */
|
||||
mca_rmaps_rank_file_component.priority = 1000000;
|
||||
} else {
|
||||
mca_rmaps_rank_file_component.priority = 0;
|
||||
}
|
||||
return ORTE_SUCCESS;
|
||||
}
|
||||
|
||||
static orte_rmaps_base_module_t*
|
||||
orte_rmaps_rank_file_init(int *priority)
|
||||
{
|
||||
/* the RMAPS framework is -only- opened on HNP's,
|
||||
* so no need to check for that here
|
||||
*/
|
||||
|
||||
*priority = mca_rmaps_rank_file_component.priority;
|
||||
|
||||
return &orte_rmaps_rank_file_module;
|
||||
}
|
||||
|
||||
/**
|
||||
* Close all subsystems.
|
||||
*/
|
||||
|
||||
static int orte_rmaps_rank_file_close(void)
|
||||
{
|
||||
return ORTE_SUCCESS;
|
||||
}
|
||||
|
Загрузка…
Ссылка в новой задаче
Block a user