2015-02-19 23:41:41 +03:00
|
|
|
/* -*- Mode: C; c-basic-offset:4 ; indent-tabs-mode:nil -*- */
|
George did the work and deserves all the credit for it. Ralph did the merge, and deserves whatever blame results from errors in it :-)
WHAT: Open our low-level communication infrastructure by moving all necessary components (btl/rcache/allocator/mpool) down in OPAL
All the components required for inter-process communications are currently deeply integrated in the OMPI layer. Several groups/institutions have express interest in having a more generic communication infrastructure, without all the OMPI layer dependencies. This communication layer should be made available at a different software level, available to all layers in the Open MPI software stack. As an example, our ORTE layer could replace the current OOB and instead use the BTL directly, gaining access to more reactive network interfaces than TCP. Similarly, external software libraries could take advantage of our highly optimized AM (active message) communication layer for their own purpose. UTK with support from Sandia, developped a version of Open MPI where the entire communication infrastucture has been moved down to OPAL (btl/rcache/allocator/mpool). Most of the moved components have been updated to match the new schema, with few exceptions (mainly BTLs where I have no way of compiling/testing them). Thus, the completion of this RFC is tied to being able to completing this move for all BTLs. For this we need help from the rest of the Open MPI community, especially those supporting some of the BTLs. A non-exhaustive list of BTLs that qualify here is: mx, portals4, scif, udapl, ugni, usnic.
This commit was SVN r32317.
2014-07-26 04:47:28 +04:00
|
|
|
/*
|
|
|
|
* Copyright (c) 2004-2005 The Trustees of Indiana University and Indiana
|
|
|
|
* University Research and Technology
|
|
|
|
* Corporation. All rights reserved.
|
|
|
|
* Copyright (c) 2004-2005 The University of Tennessee and The University
|
|
|
|
* of Tennessee Research Foundation. All rights
|
|
|
|
* reserved.
|
2015-06-24 06:59:57 +03:00
|
|
|
* Copyright (c) 2004-2005 High Performance Computing Center Stuttgart,
|
George did the work and deserves all the credit for it. Ralph did the merge, and deserves whatever blame results from errors in it :-)
WHAT: Open our low-level communication infrastructure by moving all necessary components (btl/rcache/allocator/mpool) down in OPAL
All the components required for inter-process communications are currently deeply integrated in the OMPI layer. Several groups/institutions have express interest in having a more generic communication infrastructure, without all the OMPI layer dependencies. This communication layer should be made available at a different software level, available to all layers in the Open MPI software stack. As an example, our ORTE layer could replace the current OOB and instead use the BTL directly, gaining access to more reactive network interfaces than TCP. Similarly, external software libraries could take advantage of our highly optimized AM (active message) communication layer for their own purpose. UTK with support from Sandia, developped a version of Open MPI where the entire communication infrastucture has been moved down to OPAL (btl/rcache/allocator/mpool). Most of the moved components have been updated to match the new schema, with few exceptions (mainly BTLs where I have no way of compiling/testing them). Thus, the completion of this RFC is tied to being able to completing this move for all BTLs. For this we need help from the rest of the Open MPI community, especially those supporting some of the BTLs. A non-exhaustive list of BTLs that qualify here is: mx, portals4, scif, udapl, ugni, usnic.
This commit was SVN r32317.
2014-07-26 04:47:28 +04:00
|
|
|
* University of Stuttgart. All rights reserved.
|
|
|
|
* Copyright (c) 2004-2005 The Regents of the University of California.
|
|
|
|
* All rights reserved.
|
2015-02-19 23:41:41 +03:00
|
|
|
* Copyright (c) 2015 Los Alamos National Security, LLC. All rights
|
|
|
|
* reserved.
|
George did the work and deserves all the credit for it. Ralph did the merge, and deserves whatever blame results from errors in it :-)
WHAT: Open our low-level communication infrastructure by moving all necessary components (btl/rcache/allocator/mpool) down in OPAL
All the components required for inter-process communications are currently deeply integrated in the OMPI layer. Several groups/institutions have express interest in having a more generic communication infrastructure, without all the OMPI layer dependencies. This communication layer should be made available at a different software level, available to all layers in the Open MPI software stack. As an example, our ORTE layer could replace the current OOB and instead use the BTL directly, gaining access to more reactive network interfaces than TCP. Similarly, external software libraries could take advantage of our highly optimized AM (active message) communication layer for their own purpose. UTK with support from Sandia, developped a version of Open MPI where the entire communication infrastucture has been moved down to OPAL (btl/rcache/allocator/mpool). Most of the moved components have been updated to match the new schema, with few exceptions (mainly BTLs where I have no way of compiling/testing them). Thus, the completion of this RFC is tied to being able to completing this move for all BTLs. For this we need help from the rest of the Open MPI community, especially those supporting some of the BTLs. A non-exhaustive list of BTLs that qualify here is: mx, portals4, scif, udapl, ugni, usnic.
This commit was SVN r32317.
2014-07-26 04:47:28 +04:00
|
|
|
* $COPYRIGHT$
|
2015-06-24 06:59:57 +03:00
|
|
|
*
|
George did the work and deserves all the credit for it. Ralph did the merge, and deserves whatever blame results from errors in it :-)
WHAT: Open our low-level communication infrastructure by moving all necessary components (btl/rcache/allocator/mpool) down in OPAL
All the components required for inter-process communications are currently deeply integrated in the OMPI layer. Several groups/institutions have express interest in having a more generic communication infrastructure, without all the OMPI layer dependencies. This communication layer should be made available at a different software level, available to all layers in the Open MPI software stack. As an example, our ORTE layer could replace the current OOB and instead use the BTL directly, gaining access to more reactive network interfaces than TCP. Similarly, external software libraries could take advantage of our highly optimized AM (active message) communication layer for their own purpose. UTK with support from Sandia, developped a version of Open MPI where the entire communication infrastucture has been moved down to OPAL (btl/rcache/allocator/mpool). Most of the moved components have been updated to match the new schema, with few exceptions (mainly BTLs where I have no way of compiling/testing them). Thus, the completion of this RFC is tied to being able to completing this move for all BTLs. For this we need help from the rest of the Open MPI community, especially those supporting some of the BTLs. A non-exhaustive list of BTLs that qualify here is: mx, portals4, scif, udapl, ugni, usnic.
This commit was SVN r32317.
2014-07-26 04:47:28 +04:00
|
|
|
* Additional copyrights may follow
|
2015-06-24 06:59:57 +03:00
|
|
|
*
|
George did the work and deserves all the credit for it. Ralph did the merge, and deserves whatever blame results from errors in it :-)
WHAT: Open our low-level communication infrastructure by moving all necessary components (btl/rcache/allocator/mpool) down in OPAL
All the components required for inter-process communications are currently deeply integrated in the OMPI layer. Several groups/institutions have express interest in having a more generic communication infrastructure, without all the OMPI layer dependencies. This communication layer should be made available at a different software level, available to all layers in the Open MPI software stack. As an example, our ORTE layer could replace the current OOB and instead use the BTL directly, gaining access to more reactive network interfaces than TCP. Similarly, external software libraries could take advantage of our highly optimized AM (active message) communication layer for their own purpose. UTK with support from Sandia, developped a version of Open MPI where the entire communication infrastucture has been moved down to OPAL (btl/rcache/allocator/mpool). Most of the moved components have been updated to match the new schema, with few exceptions (mainly BTLs where I have no way of compiling/testing them). Thus, the completion of this RFC is tied to being able to completing this move for all BTLs. For this we need help from the rest of the Open MPI community, especially those supporting some of the BTLs. A non-exhaustive list of BTLs that qualify here is: mx, portals4, scif, udapl, ugni, usnic.
This commit was SVN r32317.
2014-07-26 04:47:28 +04:00
|
|
|
* $HEADER$
|
|
|
|
*
|
|
|
|
*/
|
|
|
|
|
|
|
|
/** @file
|
|
|
|
*
|
|
|
|
* A red black tree
|
|
|
|
*/
|
|
|
|
|
|
|
|
#ifndef OPAL_RB_TREE_H
|
|
|
|
#define OPAL_RB_TREE_H
|
|
|
|
|
|
|
|
#include "opal_config.h"
|
|
|
|
#include <stdlib.h>
|
|
|
|
#include "opal/constants.h"
|
|
|
|
#include "opal/class/opal_object.h"
|
2015-02-19 23:41:41 +03:00
|
|
|
#include "opal/class/opal_free_list.h"
|
George did the work and deserves all the credit for it. Ralph did the merge, and deserves whatever blame results from errors in it :-)
WHAT: Open our low-level communication infrastructure by moving all necessary components (btl/rcache/allocator/mpool) down in OPAL
All the components required for inter-process communications are currently deeply integrated in the OMPI layer. Several groups/institutions have express interest in having a more generic communication infrastructure, without all the OMPI layer dependencies. This communication layer should be made available at a different software level, available to all layers in the Open MPI software stack. As an example, our ORTE layer could replace the current OOB and instead use the BTL directly, gaining access to more reactive network interfaces than TCP. Similarly, external software libraries could take advantage of our highly optimized AM (active message) communication layer for their own purpose. UTK with support from Sandia, developped a version of Open MPI where the entire communication infrastucture has been moved down to OPAL (btl/rcache/allocator/mpool). Most of the moved components have been updated to match the new schema, with few exceptions (mainly BTLs where I have no way of compiling/testing them). Thus, the completion of this RFC is tied to being able to completing this move for all BTLs. For this we need help from the rest of the Open MPI community, especially those supporting some of the BTLs. A non-exhaustive list of BTLs that qualify here is: mx, portals4, scif, udapl, ugni, usnic.
This commit was SVN r32317.
2014-07-26 04:47:28 +04:00
|
|
|
|
|
|
|
BEGIN_C_DECLS
|
|
|
|
/*
|
|
|
|
* Data structures and datatypes
|
|
|
|
*/
|
|
|
|
|
|
|
|
/**
|
|
|
|
* red and black enum
|
|
|
|
*/
|
|
|
|
typedef enum {RED, BLACK} opal_rb_tree_nodecolor_t;
|
|
|
|
|
|
|
|
/**
|
|
|
|
* node data structure
|
|
|
|
*/
|
|
|
|
struct opal_rb_tree_node_t
|
|
|
|
{
|
2015-02-19 23:41:41 +03:00
|
|
|
opal_free_list_item_t super; /**< the parent class */
|
George did the work and deserves all the credit for it. Ralph did the merge, and deserves whatever blame results from errors in it :-)
WHAT: Open our low-level communication infrastructure by moving all necessary components (btl/rcache/allocator/mpool) down in OPAL
All the components required for inter-process communications are currently deeply integrated in the OMPI layer. Several groups/institutions have express interest in having a more generic communication infrastructure, without all the OMPI layer dependencies. This communication layer should be made available at a different software level, available to all layers in the Open MPI software stack. As an example, our ORTE layer could replace the current OOB and instead use the BTL directly, gaining access to more reactive network interfaces than TCP. Similarly, external software libraries could take advantage of our highly optimized AM (active message) communication layer for their own purpose. UTK with support from Sandia, developped a version of Open MPI where the entire communication infrastucture has been moved down to OPAL (btl/rcache/allocator/mpool). Most of the moved components have been updated to match the new schema, with few exceptions (mainly BTLs where I have no way of compiling/testing them). Thus, the completion of this RFC is tied to being able to completing this move for all BTLs. For this we need help from the rest of the Open MPI community, especially those supporting some of the BTLs. A non-exhaustive list of BTLs that qualify here is: mx, portals4, scif, udapl, ugni, usnic.
This commit was SVN r32317.
2014-07-26 04:47:28 +04:00
|
|
|
opal_rb_tree_nodecolor_t color; /**< the node color */
|
|
|
|
struct opal_rb_tree_node_t * parent;/**< the parent node, can be NULL */
|
|
|
|
struct opal_rb_tree_node_t * left; /**< the left child - can be nill */
|
|
|
|
struct opal_rb_tree_node_t * right; /**< the right child - can be nill */
|
|
|
|
void *key; /**< a pointer to the key */
|
|
|
|
void *value; /**< a pointer to the value */
|
|
|
|
};
|
|
|
|
typedef struct opal_rb_tree_node_t opal_rb_tree_node_t;
|
|
|
|
|
|
|
|
/**
|
|
|
|
* the compare function typedef. This function is used to compare 2 nodes.
|
|
|
|
*/
|
|
|
|
typedef int (*opal_rb_tree_comp_fn_t)(void *key1, void *key2);
|
|
|
|
|
|
|
|
/**
|
|
|
|
* the data structure that holds all the needed information about the tree.
|
|
|
|
*/
|
|
|
|
struct opal_rb_tree_t {
|
|
|
|
opal_object_t parent; /**< the parent class */
|
|
|
|
/* this root pointer doesn't actually point to the root of the tree.
|
|
|
|
* rather, it points to a sentinal node who's left branch is the real
|
|
|
|
* root of the tree. This is done to eliminate special cases */
|
|
|
|
opal_rb_tree_node_t * root_ptr; /**< a pointer to the root of the tree */
|
|
|
|
opal_rb_tree_node_t * nill; /**< the nill sentinal node */
|
|
|
|
opal_rb_tree_comp_fn_t comp; /**< the compare function */
|
2015-02-19 23:41:41 +03:00
|
|
|
opal_free_list_t free_list; /**< the free list to get the memory from */
|
George did the work and deserves all the credit for it. Ralph did the merge, and deserves whatever blame results from errors in it :-)
WHAT: Open our low-level communication infrastructure by moving all necessary components (btl/rcache/allocator/mpool) down in OPAL
All the components required for inter-process communications are currently deeply integrated in the OMPI layer. Several groups/institutions have express interest in having a more generic communication infrastructure, without all the OMPI layer dependencies. This communication layer should be made available at a different software level, available to all layers in the Open MPI software stack. As an example, our ORTE layer could replace the current OOB and instead use the BTL directly, gaining access to more reactive network interfaces than TCP. Similarly, external software libraries could take advantage of our highly optimized AM (active message) communication layer for their own purpose. UTK with support from Sandia, developped a version of Open MPI where the entire communication infrastucture has been moved down to OPAL (btl/rcache/allocator/mpool). Most of the moved components have been updated to match the new schema, with few exceptions (mainly BTLs where I have no way of compiling/testing them). Thus, the completion of this RFC is tied to being able to completing this move for all BTLs. For this we need help from the rest of the Open MPI community, especially those supporting some of the BTLs. A non-exhaustive list of BTLs that qualify here is: mx, portals4, scif, udapl, ugni, usnic.
This commit was SVN r32317.
2014-07-26 04:47:28 +04:00
|
|
|
size_t tree_size; /**< the size of the tree */
|
|
|
|
};
|
|
|
|
typedef struct opal_rb_tree_t opal_rb_tree_t;
|
|
|
|
|
|
|
|
/** declare the tree node as a class */
|
|
|
|
OPAL_DECLSPEC OBJ_CLASS_DECLARATION(opal_rb_tree_node_t);
|
|
|
|
/** declare the tree as a class */
|
|
|
|
OPAL_DECLSPEC OBJ_CLASS_DECLARATION(opal_rb_tree_t);
|
|
|
|
|
|
|
|
/* Function pointers for map traversal function */
|
|
|
|
/**
|
|
|
|
* this function is used for the opal_rb_tree_traverse function.
|
|
|
|
* it is passed a pointer to the value for each node and, if it returns
|
|
|
|
* a one, the action function is called on that node. Otherwise, the node is ignored.
|
|
|
|
*/
|
|
|
|
typedef int (*opal_rb_tree_condition_fn_t)(void *);
|
|
|
|
/**
|
|
|
|
* this function is used for the user to perform any action on the passed
|
|
|
|
* values. The first argument is the key and the second is the value.
|
|
|
|
* note that this function SHOULD NOT modify the keys, as that would
|
|
|
|
* mess up the tree.
|
|
|
|
*/
|
|
|
|
typedef void (*opal_rb_tree_action_fn_t)(void *, void *);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Public function protoypes
|
|
|
|
*/
|
|
|
|
|
|
|
|
/**
|
|
|
|
* the function creates a new tree
|
|
|
|
*
|
|
|
|
* @param tree a pointer to an allocated area of memory for the main
|
|
|
|
* tree data structure.
|
|
|
|
* @param comp a pointer to the function to use for comaparing 2 nodes
|
|
|
|
*
|
|
|
|
* @retval OPAL_SUCCESS if it is successful
|
|
|
|
* @retval OPAL_ERR_TEMP_OUT_OF_RESOURCE if unsuccessful
|
|
|
|
*/
|
|
|
|
OPAL_DECLSPEC int opal_rb_tree_init(opal_rb_tree_t * tree, opal_rb_tree_comp_fn_t comp);
|
|
|
|
|
|
|
|
|
|
|
|
/**
|
|
|
|
* inserts a node into the tree
|
|
|
|
*
|
|
|
|
* @param tree a pointer to the tree data structure
|
|
|
|
* @param key the key for the node
|
|
|
|
* @param value the value for the node
|
|
|
|
*
|
|
|
|
* @retval OPAL_SUCCESS
|
|
|
|
* @retval OPAL_ERR_TEMP_OUT_OF_RESOURCE if unsuccessful
|
|
|
|
*/
|
|
|
|
OPAL_DECLSPEC int opal_rb_tree_insert(opal_rb_tree_t *tree, void * key, void * value);
|
|
|
|
|
|
|
|
/**
|
|
|
|
* finds a value in the tree based on the passed key using passed
|
|
|
|
* compare function
|
|
|
|
*
|
|
|
|
* @param tree a pointer to the tree data structure
|
|
|
|
* @param key a pointer to the key
|
|
|
|
* @param compare function
|
|
|
|
*
|
|
|
|
* @retval pointer to the value if found
|
|
|
|
* @retval NULL if not found
|
|
|
|
*/
|
|
|
|
OPAL_DECLSPEC void * opal_rb_tree_find_with(opal_rb_tree_t *tree, void *key, opal_rb_tree_comp_fn_t compfn);
|
|
|
|
|
|
|
|
/**
|
|
|
|
* finds a value in the tree based on the passed key
|
|
|
|
*
|
|
|
|
* @param tree a pointer to the tree data structure
|
|
|
|
* @param key a pointer to the key
|
|
|
|
*
|
|
|
|
* @retval pointer to the value if found
|
|
|
|
* @retval NULL if not found
|
|
|
|
*/
|
|
|
|
static inline void * opal_rb_tree_find(opal_rb_tree_t *tree, void *key)
|
|
|
|
{
|
|
|
|
return opal_rb_tree_find_with(tree, key, tree->comp);
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* deletes a node based on its key
|
|
|
|
*
|
|
|
|
* @param tree a pointer to the tree data structure
|
|
|
|
* @param key a pointer to the key
|
|
|
|
*
|
|
|
|
* @retval OPAL_SUCCESS if the node is found and deleted
|
|
|
|
* @retval OPAL_ERR_NOT_FOUND if the node is not found
|
|
|
|
*/
|
|
|
|
OPAL_DECLSPEC int opal_rb_tree_delete(opal_rb_tree_t *tree, void *key);
|
|
|
|
|
|
|
|
/**
|
|
|
|
* frees all the nodes on the tree
|
|
|
|
*
|
|
|
|
* @param tree a pointer to the tree data structure
|
|
|
|
*
|
|
|
|
* @retval OPAL_SUCCESS
|
|
|
|
*/
|
|
|
|
OPAL_DECLSPEC int opal_rb_tree_destroy(opal_rb_tree_t *tree);
|
|
|
|
|
|
|
|
/**
|
|
|
|
* traverses the entire tree, performing the cond function on each of the
|
|
|
|
* values and if it returns one it performs the action function on the values
|
|
|
|
*
|
|
|
|
* @param tree a pointer to the tree
|
|
|
|
* @param cond a pointer to the condition function
|
|
|
|
* @param action a pointer to the action function
|
|
|
|
*
|
|
|
|
* @retval OPAL_SUCCESS
|
|
|
|
* @retval OPAL_ERROR if there is an error
|
|
|
|
*/
|
|
|
|
OPAL_DECLSPEC int opal_rb_tree_traverse(opal_rb_tree_t *tree,
|
|
|
|
opal_rb_tree_condition_fn_t cond,
|
|
|
|
opal_rb_tree_action_fn_t action);
|
|
|
|
|
|
|
|
/**
|
|
|
|
* returns the size of the tree
|
|
|
|
*
|
|
|
|
* @param tree a pointer to the tree data structure
|
|
|
|
*
|
|
|
|
* @retval int the nuber of items on the tree
|
|
|
|
*/
|
|
|
|
OPAL_DECLSPEC int opal_rb_tree_size(opal_rb_tree_t *tree);
|
|
|
|
|
|
|
|
END_C_DECLS
|
|
|
|
#endif /* OPAL_RB_TREE_H */
|
|
|
|
|