Для этого сайта требуется поддержка JavaScript.
Обзор
Помощь
Вход
ports
/
openmpi
Следить
1
В избранное
1
Форкнуть
0
Вы уже форкнули openmpi
Код
Релизы
Активность
openmpi
/
orte
/
mca
/
oob
/
tcp
История
Ralph Castain
3cee4152fc
Fix the intercommunictor issue reported by Gilles. Instead of directly checking the reachability bitmap, ask the component if the proc is reachable when doing a send as the component is the final arbiter in such cases. Recirculate any messages that a daemon is trying to send to void race conditions. Cleanup listener sockets so we don't leak them
2015-05-11 09:16:25 -07:00
..
configure.m4
Remove the old configure option for disabling full rte support - we now use the OMPI rte framework for such purposes
2013-02-28 01:35:55 +00:00
help-oob-tcp.txt
ckpt
2015-03-28 07:59:20 -07:00
Makefile.am
Use the correct abstraction layer name for the data dirs
2014-05-08 14:32:24 +00:00
oob_tcp_common.c
We don't need to know about non-fatal errors from setting socket options
2015-03-20 07:16:31 -07:00
oob_tcp_common.h
Now that the BTLs are moving down to OPAL and becoming available to ORTE, there no longer is a need/desire to push performance in the OOB/TCP component. So we don't need multiple modules driving NICs in parallel, and can drop all the complicated distribution logic. Fall back to the simplified single module model, but retain the ability to run that module in its own progress thread if so directed.
2014-06-06 02:24:17 +00:00
oob_tcp_component.c
Rework the OOB selection logic to allow a component (e.g., usock) to direct that it be the sole active component. Remove prior disqualifying code in the oob/tcp component as it was too restrictive - if usock wasn't able to run, it left apps with no way to communicate to their daemon. Have the local daemon check the global modex for the RML URI info of the local procs so it can route messages between them when tcp is the primary channel.
2015-05-08 11:15:21 -07:00
oob_tcp_component.h
Support attempts to connect async processes by allowing the oob/tcp connection to retry the attempt to connect to a peer. Off by default, operates if someone specifies how long to wait between retry attempts.
2015-04-01 20:21:23 -07:00
oob_tcp_connection.c
Support attempts to connect async processes by allowing the oob/tcp connection to retry the attempt to connect to a peer. Off by default, operates if someone specifies how long to wait between retry attempts.
2015-04-01 20:21:23 -07:00
oob_tcp_connection.h
Now that the BTLs are moving down to OPAL and becoming available to ORTE, there no longer is a need/desire to push performance in the OOB/TCP component. So we don't need multiple modules driving NICs in parallel, and can drop all the complicated distribution logic. Fall back to the simplified single module model, but retain the ability to run that module in its own progress thread if so directed.
2014-06-06 02:24:17 +00:00
oob_tcp_hdr.h
Consolidate all the QOS changes into one clean commit
2015-05-06 19:48:42 -07:00
oob_tcp_listener.c
Fix the intercommunictor issue reported by Gilles. Instead of directly checking the reachability bitmap, ask the component if the proc is reachable when doing a send as the component is the final arbiter in such cases. Recirculate any messages that a daemon is trying to send to void race conditions. Cleanup listener sockets so we don't leak them
2015-05-11 09:16:25 -07:00
oob_tcp_listener.h
As per the RFC, bring in the ORTE async progress code and the rewrite of OOB:
2013-08-22 16:37:40 +00:00
oob_tcp_peer.h
Support attempts to connect async processes by allowing the oob/tcp connection to retry the attempt to connect to a peer. Off by default, operates if someone specifies how long to wait between retry attempts.
2015-04-01 20:21:23 -07:00
oob_tcp_ping.h
Now that the BTLs are moving down to OPAL and becoming available to ORTE, there no longer is a need/desire to push performance in the OOB/TCP component. So we don't need multiple modules driving NICs in parallel, and can drop all the complicated distribution logic. Fall back to the simplified single module model, but retain the ability to run that module in its own progress thread if so directed.
2014-06-06 02:24:17 +00:00
oob_tcp_sendrecv.c
Consolidate all the QOS changes into one clean commit
2015-05-06 19:48:42 -07:00
oob_tcp_sendrecv.h
Consolidate all the QOS changes into one clean commit
2015-05-06 19:48:42 -07:00
oob_tcp.c
Fix typo in oob_tcp.c when IPV6 enabled. Cleanup a few other warnings, including a type in coll_sm that prevented that component from registering its MCA params!
2015-05-07 21:05:08 -07:00
oob_tcp.h
Now that the BTLs are moving down to OPAL and becoming available to ORTE, there no longer is a need/desire to push performance in the OOB/TCP component. So we don't need multiple modules driving NICs in parallel, and can drop all the complicated distribution logic. Fall back to the simplified single module model, but retain the ability to run that module in its own progress thread if so directed.
2014-06-06 02:24:17 +00:00
owner.txt
add owner files to opa/ompi/orte mca directories
2015-02-22 15:10:23 -07:00