Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

followed "without a git repo checkout" steps, failed to launch qemu #65

Closed
juzhao opened this issue Feb 23, 2021 · 5 comments
Closed

followed "without a git repo checkout" steps, failed to launch qemu #65

juzhao opened this issue Feb 23, 2021 · 5 comments
Labels
lifecycle/rotten Denotes an issue or PR that has aged beyond stale and will be auto-closed.

Comments

@juzhao
Copy link

juzhao commented Feb 23, 2021

Description

# oc version
Client Version: 4.7.0-rc.3
Server Version: 4.7.0-0.nightly-2021-02-18-110409
Kubernetes Version: v1.20.0+bd9e442

followed https://github.com/openshift/kata-operator
without a git repo checkout

1. Make sure that oc is configured to talk to the cluster

2. To deploy the operator and create a custom resource (which installs Kata on all worker nodes), run curl https://raw.githubusercontent.com/openshift/kata-operator/master/deploy/install.sh | bash

result we got

# oc describe kataconfig example-kataconfig
...
Status:
  Installation Status:
    Completed:
      Completed Nodes Count:  3
      Completed Nodes List:
        ip-10-0-207-246
        ip-10-0-149-128
        ip-10-0-176-142
    Failed:
    In Progress:
...
# oc get node -o wide | grep worker
ip-10-0-149-128.us-east-2.compute.internal   Ready    worker   108m   v1.20.0+ba45583   10.0.149.128   <none>        Red Hat Enterprise Linux CoreOS 47.83.202102090044-0 (Ootpa)   4.18.0-240.10.1.el8_3.x86_64   cri-o://1.20.0-0.rhaos4.7.git8921e00.el8.51
ip-10-0-176-142.us-east-2.compute.internal   Ready    worker   108m   v1.20.0+ba45583   10.0.176.142   <none>        Red Hat Enterprise Linux CoreOS 47.83.202102090044-0 (Ootpa)   4.18.0-240.10.1.el8_3.x86_64   cri-o://1.20.0-0.rhaos4.7.git8921e00.el8.51
ip-10-0-207-246.us-east-2.compute.internal   Ready    worker   108m   v1.20.0+ba45583   10.0.207.246   <none>        Red Hat Enterprise Linux CoreOS 47.83.202102090044-0 (Ootpa)   4.18.0-240.10.1.el8_3.x86_64   cri-o://1.20.0-0.rhaos4.7.git8921e00.el8.51

Apply example-fedora.yaml from kata-operator repo

# oc -n default get po
NAME             READY   STATUS              RESTARTS   AGE
example-fedora   0/1     ContainerCreating   0          113s
# oc -n default describe pod example-fedora
...
Events:
  Type     Reason                  Age                             From               Message
  ----     ------                  ----                            ----               -------
  Normal   Scheduled               <invalid>                       default-scheduler  Successfully assigned default/example-fedora to ip-10-0-176-142.us-east-2.compute.internal
  Normal   AddedInterface          <invalid>                       multus             Add eth0 [10.131.0.212/23]
  Normal   AddedInterface          <invalid>                       multus             Add eth0 [10.131.0.213/23]
  Normal   AddedInterface          <invalid>                       multus             Add eth0 [10.131.0.214/23]
  Normal   AddedInterface          <invalid>                       multus             Add eth0 [10.131.0.215/23]
  Normal   AddedInterface          <invalid>                       multus             Add eth0 [10.131.0.216/23]
  Normal   AddedInterface          <invalid>                       multus             Add eth0 [10.131.0.217/23]
  Normal   AddedInterface          <invalid>                       multus             Add eth0 [10.131.0.218/23]
  Normal   AddedInterface          <invalid>                       multus             Add eth0 [10.131.0.219/23]
  Normal   AddedInterface          <invalid>                       multus             Add eth0 [10.131.0.220/23]
  Normal   AddedInterface          <invalid>                       multus             Add eth0 [10.131.0.221/23]
  Normal   AddedInterface          <invalid>                       multus             Add eth0 [10.131.0.222/23]
  Normal   AddedInterface          <invalid>                       multus             Add eth0 [10.131.0.223/23]
  Normal   AddedInterface          <invalid>                       multus             Add eth0 [10.131.0.225/23]
  Normal   AddedInterface          <invalid>                       multus             Add eth0 [10.131.0.226/23]
  Normal   AddedInterface          <invalid>                       multus             Add eth0 [10.131.0.227/23]
  Normal   AddedInterface          <invalid>                       multus             Add eth0 [10.131.0.228/23]
  Normal   AddedInterface          <invalid>                       multus             Add eth0 [10.131.0.229/23]
  Normal   AddedInterface          <invalid>                       multus             Add eth0 [10.131.0.230/23]
  Normal   AddedInterface          <invalid>                       multus             Add eth0 [10.131.0.231/23]
  Normal   AddedInterface          <invalid>                       multus             Add eth0 [10.131.0.232/23]
  Normal   AddedInterface          <invalid>                       multus             Add eth0 [10.131.0.233/23]
  Normal   AddedInterface          <invalid>                       multus             Add eth0 [10.131.0.234/23]
  Normal   AddedInterface          <invalid>                       multus             Add eth0 [10.131.0.235/23]
  Normal   AddedInterface          <invalid>                       multus             Add eth0 [10.131.0.236/23]
  Normal   AddedInterface          <invalid>                       multus             Add eth0 [10.131.0.237/23]
  Warning  FailedCreatePodSandBox  <invalid> (x25 over <invalid>)  kubelet            Failed to create pod sandbox: rpc error: code = Unknown desc = CreateContainer failed: failed to launch qemu: exit status 1, error messages from qemu log: Could not access KVM kernel module: No such file or directory
qemu-kvm: failed to initialize kvm: No such file or directory
: unknown

Additional environment details (platform, options, etc.):
4.7 aws cluster

# TYPE virt_platform gauge
virt_platform{type="kvm"} 1
virt_platform{type="aws"} 1
@jensfr
Copy link
Contributor

jensfr commented May 18, 2021

Hi @juzhao , I apologise for never having replied to your issues. Very sorry, I just noticed it now!

I believe you saw this problem because your cluster in AWS runs on machines that don't support nested virtualization. If you're still interested I recommend you try the operator on Openshift 4.8 (https://github.com/openshift/sandboxed-containers-operator/tree/release-4.8 )in either GCP or Azure (make sure you use machines with support for nested virt).

@openshift-bot
Copy link

Issues go stale after 90d of inactivity.

Mark the issue as fresh by commenting /remove-lifecycle stale.
Stale issues rot after an additional 30d of inactivity and eventually close.
Exclude this issue from closing by commenting /lifecycle frozen.

If this issue is safe to close now please do so with /close.

/lifecycle stale

@openshift-ci openshift-ci bot added the lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale. label Aug 16, 2021
@openshift-bot
Copy link

Stale issues rot after 30d of inactivity.

Mark the issue as fresh by commenting /remove-lifecycle rotten.
Rotten issues close after an additional 30d of inactivity.
Exclude this issue from closing by commenting /lifecycle frozen.

If this issue is safe to close now please do so with /close.

/lifecycle rotten
/remove-lifecycle stale

@openshift-ci openshift-ci bot added lifecycle/rotten Denotes an issue or PR that has aged beyond stale and will be auto-closed. and removed lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale. labels Sep 16, 2021
@openshift-bot
Copy link

Rotten issues close after 30d of inactivity.

Reopen the issue by commenting /reopen.
Mark the issue as fresh by commenting /remove-lifecycle rotten.
Exclude this issue from closing again by commenting /lifecycle frozen.

/close

@openshift-ci
Copy link

openshift-ci bot commented Oct 16, 2021

@openshift-bot: Closing this issue.

In response to this:

Rotten issues close after 30d of inactivity.

Reopen the issue by commenting /reopen.
Mark the issue as fresh by commenting /remove-lifecycle rotten.
Exclude this issue from closing again by commenting /lifecycle frozen.

/close

Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository.

@openshift-ci openshift-ci bot closed this as completed Oct 16, 2021
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
lifecycle/rotten Denotes an issue or PR that has aged beyond stale and will be auto-closed.
Projects
None yet
Development

No branches or pull requests

3 participants