Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Create the custom hosts file, there is no minikube with IP added. #11612

Open
charleech opened this issue Jun 9, 2021 · 3 comments
Open

Create the custom hosts file, there is no minikube with IP added. #11612

charleech opened this issue Jun 9, 2021 · 3 comments
Labels
co/docker-driver Issues related to kubernetes in container kind/bug Categorizes issue or PR as related to a bug. lifecycle/frozen Indicates that an issue or PR should not be auto-closed due to staleness. priority/backlog Higher priority than priority/awaiting-more-evidence.

Comments

@charleech
Copy link

  • minikube v1.20.0 on Centos 7.7.1908
    • MINIKUBE_HOME=/opt/minikube.home
  • Using the docker driver based on user configuration
  • Starting control plane node minikube in cluster minikube
  • Pulling base image ...
  • Creating docker container (CPUs=16, Memory=8000MB) ...
  • Preparing Kubernetes v1.20.2 on Docker 20.10.6 ...
    • opt bip=172.18.0.1/16
    • Generating certificates and keys ...
    • Booting up control plane ...
    • Configuring RBAC rules ...
  • Verifying Kubernetes components...
    • Using image gcr.io/k8s-minikube/storage-provisioner:v5
  • Enabled addons: storage-provisioner, default-storageclass
  • Done! kubectl is now configured to use "minikube" cluster and "default" namespace by default

Steps to reproduce the issue:

  1. Create the $MINIKUBE_HOME/.minikube/files/etc/hosts with the following content: -
127.0.0.1       localhost
::1     localhost ip6-localhost ip6-loopback
fe00::0 ip6-localnet
ff00::0 ip6-mcastprefix
ff02::1 ip6-allnodes
ff02::2 ip6-allrouters

172.17.x.x my.some.host1
172.17.x.y my.some.host2
  1. Start the minikube.
  2. Access the shell via minikube ssh.
  3. Inside the shell execute the cat /etc/hosts.
::1     localhost ip6-localhost ip6-loopback
fe00::0 ip6-localnet
ff00::0 ip6-mcastprefix
ff02::1 ip6-allnodes
ff02::2 ip6-allrouters

172.17.x.x my.some.host1
172.17.x.y my.some.host2
#
#<--- There is no `192.168.49.2 minikube` here.
#
192.168.49.1    host.minikube.internal
192.168.49.2    control-plane.minikube.internal

Full output of minikube logs command:
N/A

Full output of failed command:
N/A

@afbjorklund
Copy link
Collaborator

afbjorklund commented Jun 13, 2021

Related issue to #11220

Works here, though:

192.168.49.2	minikube
192.168.49.1	host.minikube.internal
192.168.49.2	control-plane.minikube.internal

When not providing your own file, without it...

Looks like the hostname is set before files are copied.

@afbjorklund afbjorklund added triage/duplicate Indicates an issue is a duplicate of other open issue. co/docker-driver Issues related to kubernetes in container kind/bug Categorizes issue or PR as related to a bug. and removed triage/duplicate Indicates an issue is a duplicate of other open issue. labels Jun 13, 2021
@spowelljr spowelljr added the priority/important-longterm Important over the long term, but may not be staffed and/or may need multiple releases to complete. label Jun 15, 2021
@k8s-triage-robot
Copy link

The Kubernetes project currently lacks enough contributors to adequately respond to all issues and PRs.

This bot triages issues and PRs according to the following rules:

  • After 90d of inactivity, lifecycle/stale is applied
  • After 30d of inactivity since lifecycle/stale was applied, lifecycle/rotten is applied
  • After 30d of inactivity since lifecycle/rotten was applied, the issue is closed

You can:

  • Mark this issue or PR as fresh with /remove-lifecycle stale
  • Mark this issue or PR as rotten with /lifecycle rotten
  • Close this issue or PR with /close
  • Offer to help out with Issue Triage

Please send feedback to sig-contributor-experience at kubernetes/community.

/lifecycle stale

@k8s-ci-robot k8s-ci-robot added the lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale. label Sep 13, 2021
@k8s-triage-robot
Copy link

The Kubernetes project currently lacks enough active contributors to adequately respond to all issues and PRs.

This bot triages issues and PRs according to the following rules:

  • After 90d of inactivity, lifecycle/stale is applied
  • After 30d of inactivity since lifecycle/stale was applied, lifecycle/rotten is applied
  • After 30d of inactivity since lifecycle/rotten was applied, the issue is closed

You can:

  • Mark this issue or PR as fresh with /remove-lifecycle rotten
  • Close this issue or PR with /close
  • Offer to help out with Issue Triage

Please send feedback to sig-contributor-experience at kubernetes/community.

/lifecycle rotten

@k8s-ci-robot k8s-ci-robot added lifecycle/rotten Denotes an issue or PR that has aged beyond stale and will be auto-closed. and removed lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale. labels Oct 13, 2021
@spowelljr spowelljr added lifecycle/frozen Indicates that an issue or PR should not be auto-closed due to staleness. and removed lifecycle/rotten Denotes an issue or PR that has aged beyond stale and will be auto-closed. labels Nov 3, 2021
@spowelljr spowelljr added priority/backlog Higher priority than priority/awaiting-more-evidence. and removed priority/important-longterm Important over the long term, but may not be staffed and/or may need multiple releases to complete. labels Feb 23, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
co/docker-driver Issues related to kubernetes in container kind/bug Categorizes issue or PR as related to a bug. lifecycle/frozen Indicates that an issue or PR should not be auto-closed due to staleness. priority/backlog Higher priority than priority/awaiting-more-evidence.
Projects
None yet
Development

No branches or pull requests

5 participants