Skip to content

e2e-upi template: gcp enable 02_lb_int.py workflow#7571

Merged
openshift-merge-robot merged 2 commits intoopenshift:masterfrom
jstuever:cors1394
Mar 12, 2020
Merged

e2e-upi template: gcp enable 02_lb_int.py workflow#7571
openshift-merge-robot merged 2 commits intoopenshift:masterfrom
jstuever:cors1394

Conversation

@jstuever
Copy link
Contributor

@jstuever jstuever commented Mar 11, 2020

This change is required to assist with the modified GCP UPI workflow
defined by openshift/installer#3270. It also fixes a typo to enable openshift/installer#2574.

Note: if [ -f 02_lb_int.py ] is used throughout to determine if the internal load balancer code is in play. This allows the various version to continue to work depending on what features they have available.

@jstuever
Copy link
Contributor Author

/cc @patrickdillon
/assign @abhinavdahiya

@jstuever
Copy link
Contributor Author

/retest

1 similar comment
@jstuever
Copy link
Contributor Author

/retest

@jstuever
Copy link
Contributor Author

/hold for e2e-gcp-upi

@openshift-ci-robot openshift-ci-robot added the do-not-merge/hold Indicates that a PR should not merge because someone has issued a /hold command. label Mar 11, 2020
@openshift-ci-robot openshift-ci-robot added the size/L Denotes a PR that changes 100-499 lines, ignoring generated files. label Mar 11, 2020
@jstuever jstuever force-pushed the cors1394 branch 2 times, most recently from f9f3dea to 1c06884 Compare March 11, 2020 16:35
@jstuever
Copy link
Contributor Author

/test pj-rehearse
There were concurrent policy changes. Please retry the whole read-modify-write with exponential backoff.

@jstuever
Copy link
Contributor Author

/retest

@jstuever jstuever force-pushed the cors1394 branch 2 times, most recently from 19d703f to 1bc6105 Compare March 12, 2020 00:47
@jstuever
Copy link
Contributor Author

/test pj-rehearse
There were concurrent policy changes. Please retry the whole read-modify-write with exponential backoff.

@jstuever
Copy link
Contributor Author

/test pj-rehearse

@jstuever
Copy link
Contributor Author

/test pj-rehearse
Pod pending timeout.

@patrickdillon
Copy link
Contributor

/test pj-rehearse

@patrickdillon
Copy link
Contributor

Generally LGTM. Let's see if the tests turn anything up, or ping me for LGTM.

@realnedsanders
Copy link

realnedsanders commented Mar 12, 2020

Why is azure trying to pull /vhd/rhcos.vhd from ci-op-liltk7zy-cb9d7sa.blob.core.windows.net?
This matches none of the other resources that were successfully created.
InfraID: ci-op-liltk7zy-cb9d7-cdfpx
ResourceGroup: ci-op-liltk7zy-cb9d7-cdfpx-rg
Identity: ci-op-liltk7zy-cb9d7-cdfpx-identity
Where can I find the source for this CI job?
@patrickdillon
@abhinavdahiya

@jstuever
Copy link
Contributor Author

@patrickdillon
Copy link
Contributor

/lgtm

@openshift-ci-robot openshift-ci-robot added the lgtm Indicates that a PR is ready to be merged. label Mar 12, 2020
@patrickdillon
Copy link
Contributor

Why is azure trying to pull /vhd/rhcos.vhd from ci-op-liltk7zy-cb9d7sa.blob.core.windows.net?
This matches none of the other resources that were successfully created.
InfraID: ci-op-liltk7zy-cb9d7-cdfpx
ResourceGroup: ci-op-liltk7zy-cb9d7-cdfpx-rg
Identity: ci-op-liltk7zy-cb9d7-cdfpx-identity

I am guessing that the vhd is created just using the cluster name and the rest of the resources use the infra id which includes a random string.

Where can I find the source for this CI job?

azure upi source is in this very template.

@jstuever
Copy link
Contributor Author

Why is azure trying to pull /vhd/rhcos.vhd from ci-op-liltk7zy-cb9d7sa.blob.core.windows.net?
This matches none of the other resources that were successfully created.
InfraID: ci-op-liltk7zy-cb9d7-cdfpx
ResourceGroup: ci-op-liltk7zy-cb9d7-cdfpx-rg
Identity: ci-op-liltk7zy-cb9d7-cdfpx-identity
Where can I find the source for this CI job?
@patrickdillon
@abhinavdahiya

This appears to be originating from az storage account create[1]. It has the same identifier ci-op-liltk7zy-cb9d7. I would assume the sa stands for storage account. The rhcos.vhd was copied there from the location identified in rhcos.json [2][3].

[1]

az storage account create -g $RESOURCE_GROUP --location $AZURE_REGION --name $ACCOUNT_NAME --kind Storage --sku Standard_LRS

[2]
VHD_URL="$(cat /var/lib/openshift-install/rhcos.json | jq -r .azure.url)"

[3]
az storage blob copy start --account-name $ACCOUNT_NAME --account-key $ACCOUNT_KEY --destination-blob "rhcos.vhd" --destination-container vhd --source-uri "$VHD_URL"

@patrickdillon
Copy link
Contributor

@angelofdeauth I would recommend reaching out to @fabianofranz with azure upi questions in slack forum-installer

@jstuever
Copy link
Contributor Author

/hold cancel

@openshift-ci-robot openshift-ci-robot removed the do-not-merge/hold Indicates that a PR should not merge because someone has issued a /hold command. label Mar 12, 2020
@abhinavdahiya
Copy link
Contributor

No change required for destroy ?

@jstuever
Copy link
Contributor Author

No change required for destroy ?

Destroy changes are here:
https://github.com/openshift/release/pull/7571/files#diff-2b1b845b92f8062711789a2bfdb27290L2002

@abhinavdahiya
Copy link
Contributor

No change required for destroy ?

Destroy changes are here:
https://github.com/openshift/release/pull/7571/files#diff-2b1b845b92f8062711789a2bfdb27290L2002

that's bootstrap destroy, part of setup, i was more concerned about the teardown.

@abhinavdahiya
Copy link
Contributor

/approve

@openshift-ci-robot
Copy link
Contributor

[APPROVALNOTIFIER] This PR is APPROVED

This pull-request has been approved by: abhinavdahiya, jstuever, patrickdillon

The full list of commands accepted by this bot can be found here.

The pull request process is described here

Details Needs approval from an approver in each of these files:

Approvers can indicate their approval by writing /approve in a comment
Approvers can cancel approval by writing /approve cancel in a comment

@openshift-ci-robot openshift-ci-robot added the approved Indicates a PR has been approved by an approver from all required OWNERS files. label Mar 12, 2020
@jstuever
Copy link
Contributor Author

jstuever commented Mar 12, 2020

No change required for destroy ?

Destroy changes are here:
https://github.com/openshift/release/pull/7571/files#diff-2b1b845b92f8062711789a2bfdb27290L2002

that's bootstrap destroy, part of setup, i was more concerned about the teardown.

Nothing special in destroy... it uses the openshift-install destroy cluster and then calls gcloud deployment-manager deployments delete on each of the templates. Because everything is prefixed with ${INFRA_ID}, and we are deploying the same bits as IPI, destroy cluster will get everything that's not in a template.

@openshift-bot
Copy link
Contributor

/retest

Please review the full test history for this PR and help us cut down flakes.

@openshift-merge-robot openshift-merge-robot merged commit ccaba53 into openshift:master Mar 12, 2020
@openshift-ci-robot
Copy link
Contributor

@jstuever: Updated the following 3 configmaps:

  • prow-job-cluster-launch-installer-upi-e2e configmap in namespace ci at cluster ci/api-build01-ci-devcluster-openshift-com:6443 using the following files:
    • key cluster-launch-installer-upi-e2e.yaml using file ci-operator/templates/openshift/installer/cluster-launch-installer-upi-e2e.yaml
  • prow-job-cluster-launch-installer-upi-e2e configmap in namespace ci at cluster default using the following files:
    • key cluster-launch-installer-upi-e2e.yaml using file ci-operator/templates/openshift/installer/cluster-launch-installer-upi-e2e.yaml
  • prow-job-cluster-launch-installer-upi-e2e configmap in namespace ci-stg at cluster default using the following files:
    • key cluster-launch-installer-upi-e2e.yaml using file ci-operator/templates/openshift/installer/cluster-launch-installer-upi-e2e.yaml
Details

In response to this:

This change is required to assist with the modified GCP UPI workflow
defined by openshift/installer#3270. It also fixes a typo to enable openshift/installer#2574.

Note: if [ -f 02_lb_int.py ] is used throughout to determine if the internal load balancer code is in play. This allows the various version to continue to work depending on what features they have available.

Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository.

@openshift-ci-robot
Copy link
Contributor

@jstuever: The following tests failed, say /retest to rerun all failed tests:

Test name Commit Details Rerun command
ci/rehearse/openshift/installer/master/e2e-gcp-upi 00a2fb5 link /test pj-rehearse
ci/rehearse/openshift/installer/master/e2e-azure-upi 00a2fb5 link /test pj-rehearse
ci/rehearse/openshift/cluster-image-registry-operator/master/e2e-vsphere 00a2fb5 link /test pj-rehearse
ci/prow/pj-rehearse 00a2fb5 link /test pj-rehearse

Full PR test history. Your PR dashboard. Please help us cut down on flakes by linking to an open issue when you hit one in your PR.

Details

Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. I understand the commands that are listed here.

@realnedsanders
Copy link

realnedsanders commented Mar 12, 2020

I am guessing that the vhd is created just using the cluster name and the rest of the resources use the infra id which includes a random string.

Where can I find the source for this CI job?

azure upi source is in this very template.

@patrickdillon Awesome, thanks. Still getting my bearings on the installer but I'm hoping to help with development in the near future. I'm not currently on an engagement utilizing Azure, so I most likely won't be doing much with that as of right now, but I noticed the test was failing and that seemed to be the rabbit hole where the issue started. My current client is one of the largest GCP + OCP customers -- I will be doing custom work with them, hopefully we're able to utilize it upstream.

This appears to be originating from az storage account create[1]. It has the same identifier ci-op-liltk7zy-cb9d7. I would assume the sa stands for storage account. The rhcos.vhd was copied there from the location identified in rhcos.json [2][3].

[1]

az storage account create -g $RESOURCE_GROUP --location $AZURE_REGION --name $ACCOUNT_NAME --kind Storage --sku Standard_LRS

[2]

VHD_URL="$(cat /var/lib/openshift-install/rhcos.json | jq -r .azure.url)"

[3]

az storage blob copy start --account-name $ACCOUNT_NAME --account-key $ACCOUNT_KEY --destination-blob "rhcos.vhd" --destination-container vhd --source-uri "$VHD_URL"

@jstuever Thanks for the explanation, as well as links to the relevant lines. Aside from openshift/release and openshift/installer, what repos are relevant to the installer? Are there architecture/design/requirements/planning documents I can take a look at? Chat chans I can join? BlueJeans meetings? I've wanted to add features before, however I'm hesitant to make changes without looking at the contract and existing/planned obligations for the install-config and the existing roadmap for the installer.

@patrickdillon
Copy link
Contributor

@patrickdillon Awesome, thanks. Still getting my bearings on the installer but I'm hoping to help with development in the near future.

If you need intro on these CI templates ping me on slack. I think it would save you a lot of time rather than trying to figure it out yourself. These CI templates will be moving to a new system which should make them easier to digest.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

approved Indicates a PR has been approved by an approver from all required OWNERS files. lgtm Indicates that a PR is ready to be merged. size/L Denotes a PR that changes 100-499 lines, ignoring generated files.

Projects

None yet

Development

Successfully merging this pull request may close these issues.

7 participants