Skip to content

Conversation

@bosilca
Copy link
Member

@bosilca bosilca commented Mar 5, 2015

This PR contains two commits that are combined into a single PR as they are related. First, it adds a new topo module allowing for process reordering based on user-provided weights. Second, it adds a new overlay PML (one that works in combination with another PML), allowing to accurately track the amount of data send between processes.

@mellanox-github
Copy link

Refer to this link for build results (access rights to CI server needed):
http://bgate.mellanox.com/jenkins/job/gh-ompi-master-pr/324/
Test PASSed.

@mike-dubman
Copy link
Member

this is great addon!
is that possible to add communicator ID and size to the output report?

bosilca added 3 commits March 27, 2015 01:47
at Inria Bordeaux. This allows us to take advantage of the remap
capability of MPI to rearrange the ranks beased on the weights
povided by the application.
counting or not the collective traffic as a separate entity. The need
for such a PML is simply because the PMPI interface doesn't allow us to
identify the collective generated traffic.
@bosilca
Copy link
Member Author

bosilca commented Mar 27, 2015

Once pushed in the master we can add all the features we want, including additional output information and per rank output files (as long as we update the scripts to be able to concatenate these files).

Regarding extending the same mechanism to the coll framework, there is no need as the coll framework has an automatic fallback to the next module, allowing us to implement the same functionality without the need for any tricks.

@mellanox-github
Copy link

Refer to this link for build results (access rights to CI server needed):
http://bgate.mellanox.com/jenkins/job/gh-ompi-master-pr/383/

Build Log
last 50 lines

[...truncated 22104 lines...]
make[4]: Entering directory `/scrap/jenkins/jenkins/jobs/gh-ompi-master-pr/workspace/test/monitoring'
FAIL: monitoring_test
============================================================================
Testsuite summary for Open MPI gitclone
============================================================================
# TOTAL: 1
# PASS:  0
# SKIP:  0
# XFAIL: 0
# FAIL:  1
# XPASS: 0
# ERROR: 0
============================================================================
See test/monitoring/test-suite.log
Please report to http://www.open-mpi.org/community/help/
============================================================================
make[4]: *** [test-suite.log] Error 1
make[4]: Leaving directory `/scrap/jenkins/jenkins/jobs/gh-ompi-master-pr/workspace/test/monitoring'
make[3]: *** [check-TESTS] Error 2
make[3]: Leaving directory `/scrap/jenkins/jenkins/jobs/gh-ompi-master-pr/workspace/test/monitoring'
make[2]: *** [check-am] Error 2
make[2]: Leaving directory `/scrap/jenkins/jenkins/jobs/gh-ompi-master-pr/workspace/test/monitoring'
make[1]: *** [check-recursive] Error 1
make[1]: Leaving directory `/scrap/jenkins/jenkins/jobs/gh-ompi-master-pr/workspace/test'
make: *** [check-recursive] Error 1
+ exit 12
Build step 'Execute shell' marked build as failure
TAP Reports Processing: START
Looking for TAP results report in workspace using pattern: **/*.tap
Saving reports...
Processing '/var/lib/jenkins/jobs/gh-ompi-master-pr/builds/383/tap-master-files/cov_stat.tap'
Parsing TAP test result [/var/lib/jenkins/jobs/gh-ompi-master-pr/builds/383/tap-master-files/cov_stat.tap].
not ok - coverity detected 996 failures in all_383 # SKIP http://bgate.mellanox.com:8888/jenkins/job/gh-ompi-master-pr//ws/cov_build/all_383/output/errors/index.html
not ok - coverity detected 5 failures in oshmem_383 # TODO http://bgate.mellanox.com:8888/jenkins/job/gh-ompi-master-pr//ws/cov_build/oshmem_383/output/errors/index.html
ok - coverity found no issues for yalla_383
ok - coverity found no issues for mxm_383
not ok - coverity detected 2 failures in fca_383 # TODO http://bgate.mellanox.com:8888/jenkins/job/gh-ompi-master-pr//ws/cov_build/fca_383/output/errors/index.html
ok - coverity found no issues for hcoll_383

TAP Reports Processing: FINISH
coverity_for_all    http://bgate.mellanox.com:8888/jenkins/job/gh-ompi-master-pr//ws/cov_build/all_383/output/errors/index.html
coverity_for_oshmem http://bgate.mellanox.com:8888/jenkins/job/gh-ompi-master-pr//ws/cov_build/oshmem_383/output/errors/index.html
coverity_for_fca    http://bgate.mellanox.com:8888/jenkins/job/gh-ompi-master-pr//ws/cov_build/fca_383/output/errors/index.html
[copy-to-slave] The build is taking place on the master node, no copy back to the master will take place.
Setting commit status on GitHub for https://github.com/open-mpi/ompi/commit/dea83edf12eb66fc8d0653500e9eda37673f30e1
[BFA] Scanning build for known causes...

[BFA] Done. 0s
Setting status of 946f7a0718290bad2123fc263aceec9a9d8241f4 to FAILURE with url http://bgate.mellanox.com:8888/jenkins/job/gh-ompi-master-pr/383/ and message: Merged build finished.

Test FAILed.

@mellanox-github
Copy link

Refer to this link for build results (access rights to CI server needed):
http://bgate.mellanox.com/jenkins/job/gh-ompi-master-pr/384/

Build Log
last 50 lines

[...truncated 22104 lines...]
make[4]: Entering directory `/scrap/jenkins/jenkins/jobs/gh-ompi-master-pr/workspace-2/test/monitoring'
FAIL: monitoring_test
============================================================================
Testsuite summary for Open MPI gitclone
============================================================================
# TOTAL: 1
# PASS:  0
# SKIP:  0
# XFAIL: 0
# FAIL:  1
# XPASS: 0
# ERROR: 0
============================================================================
See test/monitoring/test-suite.log
Please report to http://www.open-mpi.org/community/help/
============================================================================
make[4]: *** [test-suite.log] Error 1
make[4]: Leaving directory `/scrap/jenkins/jenkins/jobs/gh-ompi-master-pr/workspace-2/test/monitoring'
make[3]: *** [check-TESTS] Error 2
make[3]: Leaving directory `/scrap/jenkins/jenkins/jobs/gh-ompi-master-pr/workspace-2/test/monitoring'
make[2]: *** [check-am] Error 2
make[2]: Leaving directory `/scrap/jenkins/jenkins/jobs/gh-ompi-master-pr/workspace-2/test/monitoring'
make[1]: *** [check-recursive] Error 1
make[1]: Leaving directory `/scrap/jenkins/jenkins/jobs/gh-ompi-master-pr/workspace-2/test'
make: *** [check-recursive] Error 1
+ exit 12
Build step 'Execute shell' marked build as failure
TAP Reports Processing: START
Looking for TAP results report in workspace using pattern: **/*.tap
Saving reports...
Processing '/var/lib/jenkins/jobs/gh-ompi-master-pr/builds/384/tap-master-files/cov_stat.tap'
Parsing TAP test result [/var/lib/jenkins/jobs/gh-ompi-master-pr/builds/384/tap-master-files/cov_stat.tap].
not ok - coverity detected 996 failures in all_384 # SKIP http://bgate.mellanox.com:8888/jenkins/job/gh-ompi-master-pr//ws/cov_build/all_384/output/errors/index.html
not ok - coverity detected 5 failures in oshmem_384 # TODO http://bgate.mellanox.com:8888/jenkins/job/gh-ompi-master-pr//ws/cov_build/oshmem_384/output/errors/index.html
ok - coverity found no issues for yalla_384
ok - coverity found no issues for mxm_384
not ok - coverity detected 2 failures in fca_384 # TODO http://bgate.mellanox.com:8888/jenkins/job/gh-ompi-master-pr//ws/cov_build/fca_384/output/errors/index.html
ok - coverity found no issues for hcoll_384

TAP Reports Processing: FINISH
coverity_for_all    http://bgate.mellanox.com:8888/jenkins/job/gh-ompi-master-pr//ws/cov_build/all_384/output/errors/index.html
coverity_for_oshmem http://bgate.mellanox.com:8888/jenkins/job/gh-ompi-master-pr//ws/cov_build/oshmem_384/output/errors/index.html
coverity_for_fca    http://bgate.mellanox.com:8888/jenkins/job/gh-ompi-master-pr//ws/cov_build/fca_384/output/errors/index.html
[copy-to-slave] The build is taking place on the master node, no copy back to the master will take place.
Setting commit status on GitHub for https://github.com/open-mpi/ompi/commit/dea83edf12eb66fc8d0653500e9eda37673f30e1
[BFA] Scanning build for known causes...

[BFA] Done. 0s
Setting status of 2cb27fca22b38cd406f3d7e803f12eaa273355aa to FAILURE with url http://bgate.mellanox.com:8888/jenkins/job/gh-ompi-master-pr/384/ and message: Merged build finished.

Test FAILed.

@jsquyres
Copy link
Member

@bosilca Any idea why the monitoring test is failing at Jenkins?

@jsquyres
Copy link
Member

@bosilca This monitoring PML is an interesting concept. Is it conceptually different than a PMPI-based tool?

Also, what is the mechanism to select the underlying ("real") PML? E.g., if I want to use the use of cm or ob1, can I do that?

@bosilca
Copy link
Member Author

bosilca commented Mar 27, 2015

On Fri, Mar 27, 2015 at 8:59 AM, Jeff Squyres notifications@github.com
wrote:

@bosilca https://github.com/bosilca This monitoring PML is an
interesting concept. Is it conceptually different than a PMPI-based tool?

The major difference with the PMPI interface is that this tool allow to
gather more precise information about all messages exchanged between
processes (including some of the collective traffic). It is somewhere
between the PERUSE and the PMPI concept, as it open a part of the internal
MPI traffic to the tools, but does not allow to identify the rate of
transfer for each message.

Also, what is the mechanism to select the underlying ("real") PML? E.g.,
if I want to use the use of cm or ob1, can I do that?

By default this "fake" PML discard itself from the selection process,
except if a particular MCA parameter is set. In this case, after the
default OMPI selection logic takes place the monitoring PML alters the
global ompi_pml structure, and gracefully replace the PML visible to the
upper layers.

@mellanox-github
Copy link

Refer to this link for build results (access rights to CI server needed):
http://bgate.mellanox.com/jenkins/job/gh-ompi-master-pr/391/

Build Log
last 50 lines

[...truncated 22064 lines...]
make[4]: Entering directory `/scrap/jenkins/jenkins/jobs/gh-ompi-master-pr/workspace/test/monitoring'
FAIL: monitoring_test
============================================================================
Testsuite summary for Open MPI gitclone
============================================================================
# TOTAL: 1
# PASS:  0
# SKIP:  0
# XFAIL: 0
# FAIL:  1
# XPASS: 0
# ERROR: 0
============================================================================
See test/monitoring/test-suite.log
Please report to http://www.open-mpi.org/community/help/
============================================================================
make[4]: *** [test-suite.log] Error 1
make[4]: Leaving directory `/scrap/jenkins/jenkins/jobs/gh-ompi-master-pr/workspace/test/monitoring'
make[3]: *** [check-TESTS] Error 2
make[3]: Leaving directory `/scrap/jenkins/jenkins/jobs/gh-ompi-master-pr/workspace/test/monitoring'
make[2]: *** [check-am] Error 2
make[2]: Leaving directory `/scrap/jenkins/jenkins/jobs/gh-ompi-master-pr/workspace/test/monitoring'
make[1]: *** [check-recursive] Error 1
make[1]: Leaving directory `/scrap/jenkins/jenkins/jobs/gh-ompi-master-pr/workspace/test'
make: *** [check-recursive] Error 1
+ exit 12
Build step 'Execute shell' marked build as failure
TAP Reports Processing: START
Looking for TAP results report in workspace using pattern: **/*.tap
Saving reports...
Processing '/var/lib/jenkins/jobs/gh-ompi-master-pr/builds/391/tap-master-files/cov_stat.tap'
Parsing TAP test result [/var/lib/jenkins/jobs/gh-ompi-master-pr/builds/391/tap-master-files/cov_stat.tap].
not ok - coverity detected 987 failures in all_391 # SKIP http://bgate.mellanox.com:8888/jenkins/job/gh-ompi-master-pr//ws/cov_build/all_391/output/errors/index.html
not ok - coverity detected 5 failures in oshmem_391 # TODO http://bgate.mellanox.com:8888/jenkins/job/gh-ompi-master-pr//ws/cov_build/oshmem_391/output/errors/index.html
ok - coverity found no issues for yalla_391
ok - coverity found no issues for mxm_391
not ok - coverity detected 2 failures in fca_391 # TODO http://bgate.mellanox.com:8888/jenkins/job/gh-ompi-master-pr//ws/cov_build/fca_391/output/errors/index.html
ok - coverity found no issues for hcoll_391

TAP Reports Processing: FINISH
coverity_for_all    http://bgate.mellanox.com:8888/jenkins/job/gh-ompi-master-pr//ws/cov_build/all_391/output/errors/index.html
coverity_for_oshmem http://bgate.mellanox.com:8888/jenkins/job/gh-ompi-master-pr//ws/cov_build/oshmem_391/output/errors/index.html
coverity_for_fca    http://bgate.mellanox.com:8888/jenkins/job/gh-ompi-master-pr//ws/cov_build/fca_391/output/errors/index.html
[copy-to-slave] The build is taking place on the master node, no copy back to the master will take place.
Setting commit status on GitHub for https://api.github.com/repos/open-mpi/ompi/commit/925f1593b8fef72c3615119a6852cefe65ed5b99
[BFA] Scanning build for known causes...

[BFA] Done. 0s
Setting status of 1b8b998909f42005dd9e92624f8087ac80ab05ee to FAILURE with url http://bgate.mellanox.com:8888/jenkins/job/gh-ompi-master-pr/391/ and message: Merged build finished.

Test FAILed.

Ralph Castain and others added 3 commits March 28, 2015 07:59
…g all available credentials, and letting the receiver use the highest priority one they have in common.
sm dstore: set pmix segment size to proper value
@hppritcha
Copy link
Member

I'm getting a failure in the monitoring_test when doing make check. Should this test really be run as part of make check?

Ralph Castain added 2 commits March 30, 2015 09:56
@jsquyres
Copy link
Member

@hppritcha +1 -- I'm not sure we do.

@bosilca thoughts?

@bosilca
Copy link
Member Author

bosilca commented Mar 30, 2015

We are using a hidden trick to do some ugly things, and it appears that the extremely picky compile flags expose it out. The long story is that we need (wanted) to expose a flush function, to allow the application to dump files at desired intervals (allowing us to identify the stages in a parallel application). Because there is no "legal" way to interact with a component, we tried to expose this internal function pointer as an read-only MCA parameter, but apparently the compiler disagrees.

@jsquyres
Copy link
Member

It looks like treematch/treematch is a wholesale import of a separate library. As such, my comment about no copyrights in the individual treematch/treematch/*.[ch] files is moot -- but we should have some kind of top-level license/copyright for that directory somewhere. And we do need to make sure that the license is compatible with OMPI's overall license.

@hppritcha
Copy link
Member

George are you talking about the monitor_test?

@bosilca
Copy link
Member Author

bosilca commented Mar 30, 2015

Let me take this one back. The compiler expose our trick but the compilation step succeed. What fails is that we need at least 2 processes to run this test, and that "make check" doesn't run parallel tests. So I think that @hppritcha is right and we should not run time test as part of "make check".

@bosilca
Copy link
Member Author

bosilca commented Mar 30, 2015

This library is developed by researchers that signed the OMPI collaboration agreement (Inria Bordeaux). Moreover the Cecill-B license is another derivative of the BSD and 100% compatible with it. The main difference is the requirement for citation (http://www.cecill.info/faq.en.html#bsd).

@bosilca
Copy link
Member Author

bosilca commented Apr 3, 2015

They seem to be used in the PriorityQueue which is then used in the k-partitioning.c.

@bgoglin
Copy link
Contributor

bgoglin commented Apr 3, 2015

k-partitioning.c isn't actually used in tm_kpartitioning.c because USE_KL_KPART is 0.
This patch makes things compile without fibo.* k-partitioning.* and PriorityQueue.*

--- tm_kpartitioning.c  (révision 266)
+++ tm_kpartitioning.c  (copie de travail)
@@ -1,11 +1,14 @@
 #include "tm_mapping.h"
 #include "tm_mt.h"
 #include "tm_kpartitioning.h"
-#include "k-partitioning.h"
 #include <stdlib.h>
 #include <stdio.h>

 #define USE_KL_KPART 0
+#if USE_KL_KPART
+#include "k-partitioning.h"
+#endif
+
 #define KL_KPART_GREEDY_TRIALS 0

 static int verbose_level = ERROR;

We'll check things more in depth after the week-end and confirm everything.

@bosilca
Copy link
Member Author

bosilca commented Apr 3, 2015

@bgoglin thanks for the tip. Things seems to work correctly after removing the fibo.[ch], PriorityQueue.[ch] and k-partitioning.[ch].

@mellanox-github
Copy link

Refer to this link for build results (access rights to CI server needed):
http://bgate.mellanox.com/jenkins/job/gh-ompi-master-pr/412/
Test PASSed.

@jsquyres
Copy link
Member

@bgoglin Any progress on the license issue?

@bgoglin
Copy link
Contributor

bgoglin commented Apr 13, 2015

Sorry for the delay, they just confirmed that putting everything in BSD3 (the one in the upstream repo COPYING https://gforge.inria.fr/scm/viewvc.php/COPYING?view=markup&root=treematch) is OK after removing those files.

@mellanox-github
Copy link

Refer to this link for build results (access rights to CI server needed):
http://bgate.mellanox.com/jenkins/job/gh-ompi-master-pr/430/
Test PASSed.

@hppritcha
Copy link
Member

Can this PR be merged now that legal issues have been resolved and problematic code has been removed?

@hppritcha
Copy link
Member

Can one of the admins verify this patch?

@bgoglin
Copy link
Contributor

bgoglin commented Jun 5, 2015

Any news about merging this PR ?

@lanl-ompi
Copy link
Contributor

Refer to this link for build results (access rights to CI server needed):
http://jenkins.open-mpi.org/job/ompi_master_pr_distcheck/10/

Build Log
last 20 lines

[...truncated 5438 lines...]
     am__remove_distdir=: am__skip_length_check=: am__skip_mode_fix=: distdir)
make[3]: Entering directory `/var/lib/jenkins/workspace/ompi_master_pr_distcheck/ompi/mca/osc/pt2pt'
make[3]: Leaving directory `/var/lib/jenkins/workspace/ompi_master_pr_distcheck/ompi/mca/osc/pt2pt'
 (cd mca/pml/cm && make  top_distdir=../../../../openmpi-gitclone distdir=../../../../openmpi-gitclone/ompi/mca/pml/cm \
     am__remove_distdir=: am__skip_length_check=: am__skip_mode_fix=: distdir)
make[3]: Entering directory `/var/lib/jenkins/workspace/ompi_master_pr_distcheck/ompi/mca/pml/cm'
make[3]: Leaving directory `/var/lib/jenkins/workspace/ompi_master_pr_distcheck/ompi/mca/pml/cm'
 (cd mca/pml/monitoring && make  top_distdir=../../../../openmpi-gitclone distdir=../../../../openmpi-gitclone/ompi/mca/pml/monitoring \
     am__remove_distdir=: am__skip_length_check=: am__skip_mode_fix=: distdir)
make[3]: Entering directory `/var/lib/jenkins/workspace/ompi_master_pr_distcheck/ompi/mca/pml/monitoring'
make[3]: *** No rule to make target `pml_monitoring_comm.h', needed by `distdir'.  Stop.
make[3]: Leaving directory `/var/lib/jenkins/workspace/ompi_master_pr_distcheck/ompi/mca/pml/monitoring'
make[2]: *** [distdir] Error 1
make[2]: Leaving directory `/var/lib/jenkins/workspace/ompi_master_pr_distcheck/ompi'
make[1]: *** [distdir] Error 1
make[1]: Leaving directory `/var/lib/jenkins/workspace/ompi_master_pr_distcheck'
make: *** [dist] Error 2
Build step 'Execute shell' marked build as failure
Setting status of 090061ff6ebb5586a4e4c887da1d4669b21a8263 to FAILURE with url http://jenkins.open-mpi.org/job/ompi_master_pr_distcheck/10/ and message: Build finished.

Test FAILed.

@lanl-ompi
Copy link
Contributor

Refer to this link for build results (access rights to CI server needed):
http://jenkins.open-mpi.org/job/ompi_master_pr_cle5.2up02/99/

Build Log
last 20 lines

[...truncated 8641 lines...]
pml_monitoring.c:185:17: warning: too many arguments for format [-Wformat-extra-args]
                 fprintf(stderr, "E\t%d\t%d\t" PRIu64 " bytes\t" PRIu64 " msgs sent\n", my_rank, i, all_filtered_sent_data[i], all_filtered_messages_count[i]); fflush(stderr);
                 ^
pml_monitoring.c: In function 'ompi_mca_pml_monitoring_flush':
pml_monitoring.c:211:7: warning: too many arguments for format [-Wformat-extra-args]
       fprintf(pf, "I\t%d\t%d\t" PRIu64 " bytes\t" PRIu64 " msgs sent\n", my_rank, i, sent_data[i], messages_count[i]); fflush(pf);
       ^
pml_monitoring.c:224:5: warning: too many arguments for format [-Wformat-extra-args]
     fprintf(pf, "E\t%d\t%d\t" PRIu64 " bytes\t" PRIu64 " msgs sent\n", my_rank, i, filtered_sent_data[i], filtered_messages_count[i]); fflush(pf);
     ^
make[2]: *** [pml_monitoring_component.lo] Error 1
make[2]: *** Waiting for unfinished jobs....
make[2]: Leaving directory `/global/u2/h/hpp/jenkins_edison/workspace/ompi_master_pr_cle5.2up02/ompi/mca/pml/monitoring'
make[1]: *** [all-recursive] Error 1
make[1]: Leaving directory `/global/u2/h/hpp/jenkins_edison/workspace/ompi_master_pr_cle5.2up02/ompi'
make: *** [all-recursive] Error 1
Build step 'Execute shell' marked build as failure
GCM: Sending notification to: hpp
Setting status of 090061ff6ebb5586a4e4c887da1d4669b21a8263 to FAILURE with url http://jenkins.open-mpi.org/job/ompi_master_pr_cle5.2up02/99/ and message: Build finished.

Test FAILed.

@hppritcha
Copy link
Member

looks like this PR needs work. Let's see if the MLNX jenkins doesn't like it either.
bot:retest

@lanl-ompi
Copy link
Contributor

Refer to this link for build results (access rights to CI server needed):
http://jenkins.open-mpi.org/job/ompi_master_pr_distcheck/12/

Build Log
last 20 lines

[...truncated 5438 lines...]
     am__remove_distdir=: am__skip_length_check=: am__skip_mode_fix=: distdir)
make[3]: Entering directory `/var/lib/jenkins/workspace/ompi_master_pr_distcheck/ompi/mca/osc/pt2pt'
make[3]: Leaving directory `/var/lib/jenkins/workspace/ompi_master_pr_distcheck/ompi/mca/osc/pt2pt'
 (cd mca/pml/cm && make  top_distdir=../../../../openmpi-gitclone distdir=../../../../openmpi-gitclone/ompi/mca/pml/cm \
     am__remove_distdir=: am__skip_length_check=: am__skip_mode_fix=: distdir)
make[3]: Entering directory `/var/lib/jenkins/workspace/ompi_master_pr_distcheck/ompi/mca/pml/cm'
make[3]: Leaving directory `/var/lib/jenkins/workspace/ompi_master_pr_distcheck/ompi/mca/pml/cm'
 (cd mca/pml/monitoring && make  top_distdir=../../../../openmpi-gitclone distdir=../../../../openmpi-gitclone/ompi/mca/pml/monitoring \
     am__remove_distdir=: am__skip_length_check=: am__skip_mode_fix=: distdir)
make[3]: Entering directory `/var/lib/jenkins/workspace/ompi_master_pr_distcheck/ompi/mca/pml/monitoring'
make[3]: *** No rule to make target `pml_monitoring_comm.h', needed by `distdir'.  Stop.
make[3]: Leaving directory `/var/lib/jenkins/workspace/ompi_master_pr_distcheck/ompi/mca/pml/monitoring'
make[2]: *** [distdir] Error 1
make[2]: Leaving directory `/var/lib/jenkins/workspace/ompi_master_pr_distcheck/ompi'
make[1]: *** [distdir] Error 1
make[1]: Leaving directory `/var/lib/jenkins/workspace/ompi_master_pr_distcheck'
make: *** [dist] Error 2
Build step 'Execute shell' marked build as failure
Setting status of 090061ff6ebb5586a4e4c887da1d4669b21a8263 to FAILURE with url http://jenkins.open-mpi.org/job/ompi_master_pr_distcheck/12/ and message: Build finished.

Test FAILed.

@lanl-ompi
Copy link
Contributor

Refer to this link for build results (access rights to CI server needed):
http://jenkins.open-mpi.org/job/ompi_master_pr_cle5.2up02/101/

Build Log
last 20 lines

[...truncated 8643 lines...]
  CC       pml_monitoring_irecv.lo
pml_monitoring_component.c:40:51: warning: ISO C forbids conversion of function pointer to object pointer type [-Wpedantic]
 static unsigned long hidden_fct = (unsigned long)((void*)ompi_mca_pml_monitoring_flush);
                                                   ^
pml_monitoring_component.c: In function 'mca_pml_monitoring_component_close':
pml_monitoring_component.c:77:13: error: implicit declaration of function 'mca_base_component_repository_retain_component' [-Werror=implicit-function-declaration]
             mca_base_component_repository_retain_component(mca_pml_monitoring_component.pmlm_version.mca_type_name,
             ^
  CC       pml_monitoring_isend.lo
cc1: some warnings being treated as errors
make[2]: *** [pml_monitoring_component.lo] Error 1
make[2]: *** Waiting for unfinished jobs....
make[2]: Leaving directory `/global/u2/h/hpp/jenkins_edison/workspace/ompi_master_pr_cle5.2up02/ompi/mca/pml/monitoring'
make[1]: *** [all-recursive] Error 1
make[1]: Leaving directory `/global/u2/h/hpp/jenkins_edison/workspace/ompi_master_pr_cle5.2up02/ompi'
make: *** [all-recursive] Error 1
Build step 'Execute shell' marked build as failure
GCM: Sending notification to: hpp
Setting status of 090061ff6ebb5586a4e4c887da1d4669b21a8263 to FAILURE with url http://jenkins.open-mpi.org/job/ompi_master_pr_cle5.2up02/101/ and message: Build finished.

Test FAILed.

@mike-dubman
Copy link
Member

bot:retest

@mike-dubman
Copy link
Member

will check why it ignored prev command from @hppritcha

@lanl-ompi
Copy link
Contributor

Refer to this link for build results (access rights to CI server needed):
http://jenkins.open-mpi.org/job/ompi_master_pr_distcheck/13/

Build Log
last 20 lines

[...truncated 5439 lines...]
     am__remove_distdir=: am__skip_length_check=: am__skip_mode_fix=: distdir)
make[3]: Entering directory `/var/lib/jenkins/workspace/ompi_master_pr_distcheck/ompi/mca/osc/pt2pt'
make[3]: Leaving directory `/var/lib/jenkins/workspace/ompi_master_pr_distcheck/ompi/mca/osc/pt2pt'
 (cd mca/pml/cm && make  top_distdir=../../../../openmpi-gitclone distdir=../../../../openmpi-gitclone/ompi/mca/pml/cm \
     am__remove_distdir=: am__skip_length_check=: am__skip_mode_fix=: distdir)
make[3]: Entering directory `/var/lib/jenkins/workspace/ompi_master_pr_distcheck/ompi/mca/pml/cm'
make[3]: Leaving directory `/var/lib/jenkins/workspace/ompi_master_pr_distcheck/ompi/mca/pml/cm'
 (cd mca/pml/monitoring && make  top_distdir=../../../../openmpi-gitclone distdir=../../../../openmpi-gitclone/ompi/mca/pml/monitoring \
     am__remove_distdir=: am__skip_length_check=: am__skip_mode_fix=: distdir)
make[3]: Entering directory `/var/lib/jenkins/workspace/ompi_master_pr_distcheck/ompi/mca/pml/monitoring'
make[3]: *** No rule to make target `pml_monitoring_comm.h', needed by `distdir'.  Stop.
make[3]: Leaving directory `/var/lib/jenkins/workspace/ompi_master_pr_distcheck/ompi/mca/pml/monitoring'
make[2]: *** [distdir] Error 1
make[2]: Leaving directory `/var/lib/jenkins/workspace/ompi_master_pr_distcheck/ompi'
make[1]: *** [distdir] Error 1
make[1]: Leaving directory `/var/lib/jenkins/workspace/ompi_master_pr_distcheck'
make: *** [dist] Error 2
Build step 'Execute shell' marked build as failure
Setting status of 090061ff6ebb5586a4e4c887da1d4669b21a8263 to FAILURE with url http://jenkins.open-mpi.org/job/ompi_master_pr_distcheck/13/ and message: Build finished.

Test FAILed.

@lanl-ompi
Copy link
Contributor

Refer to this link for build results (access rights to CI server needed):
http://jenkins.open-mpi.org/job/ompi_master_pr_cle5.2up02/102/

Build Log
last 20 lines

[...truncated 6282 lines...]
make[2]: Entering directory `/global/u2/h/hpp/jenkins_edison/workspace/ompi_master_pr_cle5.2up02/ompi/mca/pml/monitoring'
  CC       pml_monitoring_start.lo
  CC       pml_monitoring_component.lo
pml_monitoring_component.c:40:51: warning: ISO C forbids conversion of function pointer to object pointer type [-Wpedantic]
 static unsigned long hidden_fct = (unsigned long)((void*)ompi_mca_pml_monitoring_flush);
                                                   ^
pml_monitoring_component.c: In function 'mca_pml_monitoring_component_close':
pml_monitoring_component.c:77:13: error: implicit declaration of function 'mca_base_component_repository_retain_component' [-Werror=implicit-function-declaration]
             mca_base_component_repository_retain_component(mca_pml_monitoring_component.pmlm_version.mca_type_name,
             ^
cc1: some warnings being treated as errors
make[2]: *** [pml_monitoring_component.lo] Error 1
make[2]: *** Waiting for unfinished jobs....
make[2]: Leaving directory `/global/u2/h/hpp/jenkins_edison/workspace/ompi_master_pr_cle5.2up02/ompi/mca/pml/monitoring'
make[1]: *** [all-recursive] Error 1
make[1]: Leaving directory `/global/u2/h/hpp/jenkins_edison/workspace/ompi_master_pr_cle5.2up02/ompi'
make: *** [all-recursive] Error 1
Build step 'Execute shell' marked build as failure
Setting status of 090061ff6ebb5586a4e4c887da1d4669b21a8263 to FAILURE with url http://jenkins.open-mpi.org/job/ompi_master_pr_cle5.2up02/102/ and message: Build finished.

Test FAILed.

@jsquyres
Copy link
Member

jsquyres commented Jun 8, 2015

@bosilca Can you have a look at all these Jenkins failures? They look legit.

@jsquyres
Copy link
Member

jsquyres commented Jul 7, 2015

@bosilca ping -- these jenkins errors look legit. Can you have a look?

@bosilca
Copy link
Member Author

bosilca commented Jul 16, 2015

This PR has been replaced by #724 and #725 .

@bosilca bosilca deleted the treematch branch July 16, 2015 16:15
jsquyres pushed a commit to jsquyres/ompi that referenced this pull request Sep 19, 2016
Bring over the hwloc assembly updates
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

7 participants