Move the openmpi-nightly-coverity.pl script into the directory where
all the other build server scripts live.
Leave the coverity-model.c file in the coverity dir, because it's
specific to coverity. Other items can be added into the
coverity-model.c, if helpful (and then we can re-submit that model
file).
Per feedback from rhc, manually set the base_ptr member
of the opal_buffer_t variable to NULL prior to calling
OBJ_RELEASE. A similar feature of opal_dss.load also
exists so likewise reset the base_ptr to NULL prior to
invoking it.
Hopefully the opal_buffer_t struct does not change
frequently.
Minor cleanups to reduce output when pmix_base_verbose
mca paramater is set.
Ensure to build support for:
* usnic
* PSM
* MXM
* OSHMEM
* Fortran (MPI and OSHMEM)
* Java (MPI and OSHMEM)
So that the OMPI code for all of these networks can be analyzed by
Coverity.
build moar stuff -- squashme
This is a first cut at scripts to run the Coverity Scan tool and
upload the results (Coverity used to do this for us automatically; now
we need to do it ourselves).
usnic_fls() can actually return 0, leading us to incorrectly free() a
buffer instead of OMPI_FREE_LIST_RETURN_MT'ing it.
So add an explicit bool in the struct that tracks whether the buffer
came from malloc or a freelist.
This was CID 1269660.
Coverity alerted us to the fact that there are places where
the synonym_for param is hard-coded to -1 when calling
register_variable(). It would be a coding error if synonym_for==-1
and (flags & MCA_BASE_VAR_FLAG_SYNONYM)>0, so let's add that to the
debug-only check at the top of the function.
This was CID 993717.
Coverity identified that we treated the possibility that one of the
message buffers could be NULL in some places (because strdup() could
fail), but not in others.
So just use stack buffers that will never be NULL.
This was CID 1269914.
Remove use of the Cray PMI KVS - which is designed for a lighweight
MPI that exchanges only a minimimal amount of connection info
(about 128 bytes per rank) - within cray/pmix. Use Cray PMI
collective extensions instead.
This is the first of several steps to accelerate launch of
Open MPI on Cray systems using either native aprun or nativized
slurm.
in this code and it ended up changing the logic that is used to set up eager RDMA.
Rather than setting up eager RDMA with a high priority message, it did it the other
way around. For some reason, CUDA-aware support did not like this. So, basically,
restore the logic to the way it was prior to the refactoring. The refactoring did not
intend to change this. Lightly reviewed by hjelmn.
This is a bit of overkill, but I'm cleaning out a bunch of other
libltdl-support assumptions, so I might as well do this one, too. The
test isn't built if we don't have libltdl support, but it had
half-hearted #if protection in it to make it safe to build even if we
didn't have libltdl support. This commit finishes that half-hearted
support.