Enhance the runing time info in the debug model #439
Conversation
[APPROVALNOTIFIER] This PR is NOT APPROVED This pull-request has been approved by: The full list of commands accepted by this bot can be found here.
Needs approval from an approver in each of these files:
Approvers can indicate their approval by writing |
/cc @jinchihe , Could you review this enhance, many thanks. |
Thanks @xauthulei for the contribution. I'm thinking if need to feature that's set the LOGLEVEL from env variable, and we can define the default value in the constants, as below, so user can set the log level from env if need debug.
What do you think? |
Seems the cluster is removed,
Checked, no kubeflow-ci-fairing cluster. I will take a look this, thanks.
|
@xauthulei the cluster has been changed to |
Thanks @jinchihe , as your suggestion, I have moved the |
Great, thanks @xauthulei |
The current error log output is confuse to me, Is there some wrong from my part. Thanks @jinchihe , I am sorry for this. |
@xauthulei I think that should be cuased by test env problem. |
The CI test hangs again... |
/test kubeflow-fairing-presubmit |
1 similar comment
/test kubeflow-fairing-presubmit |
Still working on these tests. They run successfully when I launch the tests
from my desktop onto the cluster. But fail when running via prow.
…On Thu, Jan 9, 2020 at 5:22 PM Kubernetes Prow Robot < ***@***.***> wrote:
@xauthulei <https://github.com/xauthulei>: The following test *failed*,
say /retest to rerun all failed tests:
Test name Commit Details Rerun command
kubeflow-fairing-presubmit 9850457
<9850457>
link
<https://prow.k8s.io/view/gcs/kubernetes-jenkins/pr-logs/pull/kubeflow_fairing/439/kubeflow-fairing-presubmit/1215398305979174912/> /test
kubeflow-fairing-presubmit
Full PR test history
<https://prow.k8s.io/pr-history?org=kubeflow&repo=fairing&pr=439>. Your
PR dashboard <https://gubernator.k8s.io/pr/xauthulei>. Please help us cut
down on flakes by linking to
<https://git.k8s.io/community/contributors/devel/sig-testing/flaky-tests.md#filing-issues-for-flaky-tests>
an open issue
<https://github.com/kubeflow/fairing/issues?q=is:issue+is:open> when you
hit one in your PR.
Instructions for interacting with me using PR comments are available here
<https://git.k8s.io/community/contributors/guide/pull-requests.md>. If
you have questions or suggestions related to my behavior, please file an
issue against the kubernetes/test-infra
<https://github.com/kubernetes/test-infra/issues/new?title=Prow%20issue:>
repository. I understand the commands that are listed here
<https://go.k8s.io/bot-commands>.
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
<#439?email_source=notifications&email_token=ACZ2UZSIWOEKBTL2VGOW7EDQ47EW5A5CNFSM4KEDTAF2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGOEISLLTI#issuecomment-572831181>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/ACZ2UZRIZ7ZKOMOBVU7TUKTQ47EW5ANCNFSM4KEDTAFQ>
.
|
/retest |
@xauthulei please merge upstream/master into your branch to pick up CI test fixes |
/retest |
@abhi-g , Thanks for your great efforts, I have refreshed my local, but it seems fail on other parts, could you check it again. |
/retest |
For some reason your tests are failing with errors. Such as when running pytorch jobs, the master fails with: |
Also, there are failure errors in the unittests which you should be able to run locally and verify the errors as well. |
For eg. 1 of the failures.. ____________________ test_overwrite_file_for_multiple_runs _____________________
preprocessors/test_converted_notebook_preprocessor.py:27: ../../kubeflow/fairing/preprocessors/converted_notebook.py:121: in preprocess self = <nbconvert.exporters.python.PythonExporter object at 0x7fce2cc330b8>
E ValueError: No template sub-directory with name 'python' found in the following paths: |
I'd suggest that you try running the unittests locally on your deskop/laptop and that might help debugging. I see some failures in the unittests which shouldn;'t be happening. See logs of unittests at http://testing-argo.kubeflow.org/workflows/kubeflow-test-infra/kubeflow-fairing-presubmit-e2e-439-0ef38e1-9312-8e09?tab=workflow for the currently running test. |
you should be able to run the unittests locally by cding to the kubeflow/fairing/tests dir and running the command pytest -vv --durations=10 unit/. In my case on the master branch all these tests run locally and pass without errors. |
/test kubeflow-fairing-presubmit |
format=constants.FAIRING_LOG_FORMAT, | ||
datefmt=constants.FAIRING_LOG_DATEFMT, | ||
) | ||
logging.getLogger().setLevel(constants.FAIRING_LOG_LEVEL) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
One of the things I noticed that setLevel generally takes a number value defined in log.py such as
DEBUG = 1
INFO = 2
WARN = 3
ERROR = 4
FATAL = 5
Whereas with this change, it seems like it becomes a string value. Could that be causing issues?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
@xauthulei Is that possiable to set level in basicConfig
? such as below?
logging.basicConfig(level=kubeflow.fairing.constants.FAIRING_LOG_LEVEL)
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
@abhi-g , That works in my local, even if the level is string value
, :
llhu@huleis-mbp fairing % python3
Python 3.7.3 (default, Nov 15 2019, 04:04:52)
[Clang 11.0.0 (clang-1100.0.33.16)] on darwin
Type "help", "copyright", "credits" or "license" for more information.
>>> import logging
>>> logging.getLogger().setLevel(logging.INFO)
>>> logging.info("HI")
INFO:root:HI
>>> logging.getLogger().setLevel('ERROR')
>>> logging.info("HI")
>>> logging.critical("HI")
CRITICAL:root:HI
>>> logging.error("HI")
ERROR:root:HI
>>> logging.info("HI")
>>>
@xauthulei That's strange, I executed unit tests using your branch, that's passed.
|
@xauthulei: The following test failed, say
Full PR test history. Your PR dashboard. Please help us cut down on flakes by linking to an open issue when you hit one in your PR. Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. I understand the commands that are listed here. |
@abhi-g , I fond same error when doing nbconvert , in the testing env which we used , |
Closed this PR, I have used another PR #447( it have been merged), Thanks every one here for the review efforts. |
@xauthulei: Closed this PR. In response to this:
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. |
Our current debug info is so simple,especially, the confuse time stamp, also don't show the target debug file's location ,
[I 200108 13:36:53 config:131] Using preprocessor: <kubeflow.fairing.preprocessors.base.BasePreProcessor object at 0x1017af9b0>
after changed, it will be more meaningful, for example :
INFO|2020-01-08 13:18:33|/Users/llhu/Library/Python/3.7/lib/python/site-packages/werkzeug/_internal.py|122| * Running on http://127.0.0.1:8080/
This change is