==> Audit <== |---------|--------------------------------|----------|-----------------------|---------|---------------------|---------------------| | Command | Args | Profile | User | Version | Start Time | End Time | |---------|--------------------------------|----------|-----------------------|---------|---------------------|---------------------| | start | --driver=hyperv | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 22 Apr 24 17:39 IST | | | start | | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 22 Apr 24 17:52 IST | | | start | --driver=hyperv | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 22 Apr 24 21:03 IST | | | start | --driver=docker | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 22 Apr 24 21:11 IST | | | delete | | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 22 Apr 24 21:12 IST | 22 Apr 24 21:12 IST | | start | --driver=docker | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 22 Apr 24 21:12 IST | | | start | --driver=docker | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 22 Apr 24 21:14 IST | | | start | --driver=docker | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 22 Apr 24 21:29 IST | | | start | --driver=docker | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 22 Apr 24 21:32 IST | | | start | --alsologtostderr -v=1 | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 22 Apr 24 21:35 IST | | | start | --driver=docker | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 22 Apr 24 21:38 IST | 22 Apr 24 21:39 IST | | start | --driver qemu --network | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 22 Apr 24 21:57 IST | | | | socket_vmnet | | | | | | | delete | | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 22 Apr 24 21:57 IST | 22 Apr 24 21:57 IST | | start | --driver qemu --network | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 22 Apr 24 21:57 IST | | | | socket_vmnet | | | | | | | start | --driver qemu --network | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 22 Apr 24 21:58 IST | | | | socket_vmnet | | | | | | | start | --driver qemu --network | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 22 Apr 24 22:01 IST | | | | socket_vmnet | | | | | | | start | --driver qemu --network | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 22 Apr 24 22:10 IST | | | | socket_vmnet | | | | | | | start | | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 22 Apr 24 22:13 IST | | | start | --driver=docker | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 22 Apr 24 22:15 IST | | | delete | | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 22 Apr 24 22:15 IST | 22 Apr 24 22:15 IST | | start | --driver=docker | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 22 Apr 24 22:15 IST | | | start | --driver=docker | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 22 Apr 24 22:16 IST | | | start | --driver=docker | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 22 Apr 24 22:17 IST | 22 Apr 24 22:18 IST | | start | --driver=hyperv | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 22 Apr 24 22:33 IST | | | delete | | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 22 Apr 24 22:36 IST | 22 Apr 24 22:36 IST | | start | --driver=hyperv | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 22 Apr 24 22:36 IST | | | start | --driver=hyperv | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 22 Apr 24 22:41 IST | | | start | --driver=hyperv | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 22 Apr 24 22:52 IST | | | start | --vm-driver hyperv | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 22 Apr 24 22:58 IST | | | | --hyperv-virtual-switch | | | | | | | | Minikube | | | | | | | start | --vm-driver=hyperv | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 22 Apr 24 22:59 IST | | | | --alsologtostderr -v=8 | | | | | | | start | --vm-driver hyperv | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 22 Apr 24 23:00 IST | | | | --hyperv-virtual-switch | | | | | | | | Default Switch | | | | | | | | --alsologtostderr -v=8 | | | | | | | delete | --all | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 22 Apr 24 23:05 IST | 22 Apr 24 23:05 IST | | start | --vm-driver hyperv | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 22 Apr 24 23:07 IST | | | | --cpus 4 --memory=4096 | | | | | | | | --hyperv-virtual-switch HV | | | | | | | | Virtual Switch | | | | | | | start | --vm-driver hyperv --cpus 4 | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 22 Apr 24 23:07 IST | | | | --memory=4096 | | | | | | | start | --vm-driver hyperv | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 22 Apr 24 23:09 IST | | | | --hyperv-virtual-switch | | | | | | | | Default Switch | | | | | | | | --alsologtostderr -v=8 | | | | | | | start | --vm-driver hyperv | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 22 Apr 24 23:10 IST | | | | --cpus 4 --memory=4096 | | | | | | | | --hyperv-virtual-switch | | | | | | | | Default Switch | | | | | | | start | | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 23 Apr 24 10:02 IST | | | delete | --all | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 23 Apr 24 10:05 IST | 23 Apr 24 10:05 IST | | start | | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 23 Apr 24 10:06 IST | 23 Apr 24 10:07 IST | | start | --driver=hyperv | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 23 Apr 24 10:10 IST | | | delete | --all | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 23 Apr 24 10:10 IST | 23 Apr 24 10:10 IST | | start | --driver=hyperv | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 23 Apr 24 10:10 IST | | | start | --driver=hyperv | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 23 Apr 24 10:53 IST | | | start | --driver=virtualbox | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 23 Apr 24 17:46 IST | | | delete | --all | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 23 Apr 24 17:46 IST | 23 Apr 24 17:46 IST | | start | --driver=virtualbox | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 23 Apr 24 17:47 IST | | | | --no-vtx-check | | | | | | | start | | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 23 Apr 24 17:51 IST | | | start | | minikube | LAPTOP-9JMN7DNV\Admin | v1.33.0 | 23 Apr 24 17:59 IST | | |---------|--------------------------------|----------|-----------------------|---------|---------------------|---------------------| ==> Last Start <== Log file created at: 2024/04/23 17:59:45 Running on machine: LAPTOP-9JMN7DNV Binary: Built with gc go1.22.1 for windows/amd64 Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg I0423 17:59:45.372197 8992 out.go:291] Setting OutFile to fd 96 ... I0423 17:59:45.372720 8992 out.go:338] TERM=,COLORTERM=, which probably does not support color I0423 17:59:45.372720 8992 out.go:304] Setting ErrFile to fd 100... I0423 17:59:45.372720 8992 out.go:338] TERM=,COLORTERM=, which probably does not support color W0423 17:59:45.383493 8992 root.go:314] Error reading config file at C:\Users\Admin\.minikube\config\config.json: open C:\Users\Admin\.minikube\config\config.json: The system cannot find the file specified. I0423 17:59:45.396309 8992 out.go:298] Setting JSON to false I0423 17:59:45.399875 8992 start.go:129] hostinfo: {"hostname":"LAPTOP-9JMN7DNV","uptime":561,"bootTime":1713874824,"procs":315,"os":"windows","platform":"Microsoft Windows 11 Pro","platformFamily":"Standalone Workstation","platformVersion":"10.0.22631.3447 Build 22631.3447","kernelVersion":"10.0.22631.3447 Build 22631.3447","kernelArch":"x86_64","virtualizationSystem":"","virtualizationRole":"","hostId":"10b71527-c726-4aab-a8e8-b931aa355178"} W0423 17:59:45.399875 8992 start.go:137] gopshost.Virtualization returned error: not implemented yet I0423 17:59:45.402906 8992 out.go:177] * minikube v1.33.0 on Microsoft Windows 11 Pro 10.0.22631.3447 Build 22631.3447 I0423 17:59:45.421942 8992 notify.go:220] Checking for updates... I0423 17:59:45.423020 8992 config.go:182] Loaded profile config "minikube": Driver=virtualbox, ContainerRuntime=docker, KubernetesVersion=v1.30.0 I0423 17:59:45.423020 8992 driver.go:392] Setting default libvirt URI to qemu:///system I0423 17:59:45.459767 8992 virtualbox.go:136] virtual box version: 7.0.16r162802 I0423 17:59:45.462295 8992 out.go:177] * Using the virtualbox driver based on existing profile I0423 17:59:45.466073 8992 start.go:297] selected driver: virtualbox I0423 17:59:45.466073 8992 start.go:901] validating driver "virtualbox" against &{Name:minikube KeepContext:false EmbedCerts:false MinikubeISO:https://storage.googleapis.com/minikube/iso/minikube-v1.33.0-amd64.iso KicBaseImage:gcr.io/k8s-minikube/kicbase:v0.0.43@sha256:7ff490df401cc0fbf19a4521544ae8f4a00cc163e92a95017a8d8bfdb1422737 Memory:4000 CPUs:2 DiskSize:20000 Driver:virtualbox HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:true DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.30.0 ClusterName:minikube Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.59.101 Port:8443 KubernetesVersion:v1.30.0 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop: ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:C:\Users\Admin:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} I0423 17:59:45.466356 8992 start.go:912] status for virtualbox: {Installed:true Healthy:true Running:false NeedsImprovement:false Error: Reason: Fix: Doc: Version:7.0.16r162802 } I0423 17:59:45.491304 8992 cni.go:84] Creating CNI manager for "" I0423 17:59:45.491304 8992 cni.go:158] "virtualbox" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge I0423 17:59:45.491304 8992 start.go:340] cluster config: {Name:minikube KeepContext:false EmbedCerts:false MinikubeISO:https://storage.googleapis.com/minikube/iso/minikube-v1.33.0-amd64.iso KicBaseImage:gcr.io/k8s-minikube/kicbase:v0.0.43@sha256:7ff490df401cc0fbf19a4521544ae8f4a00cc163e92a95017a8d8bfdb1422737 Memory:4000 CPUs:2 DiskSize:20000 Driver:virtualbox HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:true DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.30.0 ClusterName:minikube Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.59.101 Port:8443 KubernetesVersion:v1.30.0 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop: ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:C:\Users\Admin:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} I0423 17:59:45.491304 8992 iso.go:125] acquiring lock: {Name:mk4205d387919714760d9448c9c700c5eff9b606 Clock:{} Delay:500ms Timeout:10m0s Cancel:} I0423 17:59:45.496270 8992 out.go:177] * Starting "minikube" primary control-plane node in "minikube" cluster I0423 17:59:45.498574 8992 preload.go:132] Checking if preload exists for k8s version v1.30.0 and runtime docker I0423 17:59:45.498574 8992 preload.go:147] Found local preload: C:\Users\Admin\.minikube\cache\preloaded-tarball\preloaded-images-k8s-v18-v1.30.0-docker-overlay2-amd64.tar.lz4 I0423 17:59:45.498574 8992 cache.go:56] Caching tarball of preloaded images I0423 17:59:45.498574 8992 preload.go:173] Found C:\Users\Admin\.minikube\cache\preloaded-tarball\preloaded-images-k8s-v18-v1.30.0-docker-overlay2-amd64.tar.lz4 in cache, skipping download I0423 17:59:45.498574 8992 cache.go:59] Finished verifying existence of preloaded tar for v1.30.0 on docker I0423 17:59:45.499078 8992 profile.go:143] Saving config to C:\Users\Admin\.minikube\profiles\minikube\config.json ... I0423 17:59:45.500123 8992 start.go:360] acquireMachinesLock for minikube: {Name:mk754ef25f30054fc66f6ac44b71c78369c41dbf Clock:{} Delay:500ms Timeout:13m0s Cancel:} I0423 17:59:45.500123 8992 start.go:364] duration metric: took 0s to acquireMachinesLock for "minikube" I0423 17:59:45.500123 8992 start.go:96] Skipping create...Using existing machine configuration I0423 17:59:45.500123 8992 fix.go:54] fixHost starting: I0423 17:59:45.500123 8992 main.go:141] libmachine: COMMAND: C:\Program Files\Oracle\VirtualBox\VBoxManage.exe showvminfo minikube --machinereadable I0423 17:59:45.626150 8992 main.go:141] libmachine: STDOUT: { name="minikube" encryption="disabled" groups="/" ostype="Linux 2.6 / 3.x / 4.x / 5.x (64-bit)" UUID="9b5e3e57-d91e-432d-add7-21d97242bc13" CfgFile="C:\\Users\\Admin\\.minikube\\machines\\minikube\\minikube\\minikube.vbox" SnapFldr="C:\\Users\\Admin\\.minikube\\machines\\minikube\\minikube\\Snapshots" LogFldr="C:\\Users\\Admin\\.minikube\\machines\\minikube\\minikube\\Logs" hardwareuuid="9b5e3e57-d91e-432d-add7-21d97242bc13" memory=4000 pagefusion="off" vram=8 cpuexecutioncap=100 hpet="on" cpu-profile="host" chipset="piix3" firmware="BIOS" cpus=2 pae="on" longmode="on" triplefaultreset="off" apic="on" x2apic="off" nested-hw-virt="off" cpuid-portability-level=0 bootmenu="disabled" boot1="dvd" boot2="dvd" boot3="disk" boot4="none" acpi="on" ioapic="on" biosapic="apic" biossystemtimeoffset=0 NvramFile="C:\\Users\\Admin\\.minikube\\machines\\minikube\\minikube\\minikube.nvram" rtcuseutc="on" hwvirtex="on" nestedpaging="on" largepages="on" vtxvpid="on" vtxux="on" virtvmsavevmload="on" iommu="none" paravirtprovider="default" effparavirtprovider="kvm" VMState="running" VMStateChangeTime="2024-04-23T12:21:55.061000000" graphicscontroller="vboxvga" monitorcount=1 accelerate3d="off" accelerate2dvideo="off" teleporterenabled="off" teleporterport=0 teleporteraddress="" teleporterpassword="" tracing-enabled="off" tracing-allow-vm-access="off" tracing-config="" autostart-enabled="off" autostart-delay=0 defaultfrontend="" vmprocpriority="default" storagecontrollername0="SATA" storagecontrollertype0="IntelAhci" storagecontrollerinstance0="0" storagecontrollermaxportcount0="30" storagecontrollerportcount0="30" storagecontrollerbootable0="on" "SATA-0-0"="C:\\Users\\Admin\\.minikube\\machines\\minikube\\boot2docker.iso" "SATA-ImageUUID-0-0"="e0f2fbc8-be91-4bbd-a49a-129556e63202" "SATA-tempeject-0-0"="off" "SATA-IsEjected-0-0"="off" "SATA-hot-pluggable-0-0"="off" "SATA-nonrotational-0-0"="off" "SATA-discard-0-0"="off" "SATA-1-0"="C:\\Users\\Admin\\.minikube\\machines\\minikube\\disk.vmdk" "SATA-ImageUUID-1-0"="7d9f9db6-e8c6-40f8-a240-8d3dacae006d" "SATA-hot-pluggable-1-0"="off" "SATA-nonrotational-1-0"="off" "SATA-discard-1-0"="off" "SATA-2-0"="none" "SATA-3-0"="none" "SATA-4-0"="none" "SATA-5-0"="none" "SATA-6-0"="none" "SATA-7-0"="none" "SATA-8-0"="none" "SATA-9-0"="none" "SATA-10-0"="none" "SATA-11-0"="none" "SATA-12-0"="none" "SATA-13-0"="none" "SATA-14-0"="none" "SATA-15-0"="none" "SATA-16-0"="none" "SATA-17-0"="none" "SATA-18-0"="none" "SATA-19-0"="none" "SATA-20-0"="none" "SATA-21-0"="none" "SATA-22-0"="none" "SATA-23-0"="none" "SATA-24-0"="none" "SATA-25-0"="none" "SATA-26-0"="none" "SATA-27-0"="none" "SATA-28-0"="none" "SATA-29-0"="none" natnet1="nat" macaddress1="080027C72164" cableconnected1="on" nic1="nat" nictype1="virtio" nicspeed1="0" mtu="0" sockSnd="64" sockRcv="64" tcpWndSnd="64" tcpWndRcv="64" Forwarding(0)="ssh,tcp,127.0.0.1,50096,,22" hostonlyadapter2="VirtualBox Host-Only Ethernet Adapter #2" macaddress2="0800270FF676" cableconnected2="on" nic2="hostonly" nictype2="virtio" nicspeed2="0" nic3="none" nic4="none" nic5="none" nic6="none" nic7="none" nic8="none" hidpointing="ps2mouse" hidkeyboard="ps2kbd" uart1="off" uart2="off" uart3="off" uart4="off" lpt1="off" lpt2="off" audio="default" audio_out="off" audio_in="off" clipboard="disabled" draganddrop="disabled" SessionName="headless" VideoMode="720,400,0"@0,0 1 vrde="off" usb="off" ehci="off" xhci="off" SharedFolderNameMachineMapping1="c/Users" SharedFolderPathMachineMapping1="\\\\?\\c:\\Users" VRDEActiveConnection="off" VRDEClients==0 recording_enabled="off" recording_screens=1 rec_screen0 rec_screen_enabled="on" rec_screen_id=0 rec_screen_video_enabled="on" rec_screen_audio_enabled="off" rec_screen_dest="File" rec_screen_dest_filename="C:\\Users\\Admin\\.minikube\\machines\\minikube\\minikube\\minikube-screen0.webm" rec_screen_opts="vc_enabled=true,ac_enabled=false,ac_profile=med" rec_screen_video_res_xy="1024x768" rec_screen_video_rate_kbps=512 rec_screen_video_fps=25 GuestMemoryBalloon=0 GuestOSType="Linux26_64" GuestAdditionsRunLevel=2 GuestAdditionsVersion="6.0.0 r127566" GuestAdditionsFacility_VirtualBox Base Driver=50,1713874957680 GuestAdditionsFacility_VirtualBox System Service=50,1713874958183 GuestAdditionsFacility_Seamless Mode=0,1713874957669 GuestAdditionsFacility_Graphics Mode=0,1713874957669 } I0423 17:59:45.626150 8992 main.go:141] libmachine: STDERR: { } I0423 17:59:45.626150 8992 fix.go:112] recreateIfNeeded on minikube: state=Running err= W0423 17:59:45.626150 8992 fix.go:138] unexpected machine state, will restart: I0423 17:59:45.628743 8992 out.go:177] * Updating the running virtualbox "minikube" VM ... I0423 17:59:45.630847 8992 machine.go:94] provisionDockerMachine start ... I0423 17:59:45.632947 8992 main.go:141] libmachine: Using SSH client type: native I0423 17:59:45.633535 8992 main.go:141] libmachine: &{{{ 0 [] [] []} docker [0x137a1c0] 0x137cda0 [] 0s} 127.0.0.1 50096 } I0423 17:59:45.633535 8992 main.go:141] libmachine: About to run SSH command: hostname I0423 17:59:45.756693 8992 main.go:141] libmachine: SSH cmd err, output: : minikube I0423 17:59:45.756693 8992 buildroot.go:166] provisioning hostname "minikube" I0423 17:59:45.759014 8992 main.go:141] libmachine: Using SSH client type: native I0423 17:59:45.759524 8992 main.go:141] libmachine: &{{{ 0 [] [] []} docker [0x137a1c0] 0x137cda0 [] 0s} 127.0.0.1 50096 } I0423 17:59:45.759524 8992 main.go:141] libmachine: About to run SSH command: sudo hostname minikube && echo "minikube" | sudo tee /etc/hostname I0423 17:59:48.380766 8992 main.go:141] libmachine: SSH cmd err, output: : minikube I0423 17:59:48.383130 8992 main.go:141] libmachine: Using SSH client type: native I0423 17:59:48.383209 8992 main.go:141] libmachine: &{{{ 0 [] [] []} docker [0x137a1c0] 0x137cda0 [] 0s} 127.0.0.1 50096 } I0423 17:59:48.383209 8992 main.go:141] libmachine: About to run SSH command: if ! grep -xq '.*\sminikube' /etc/hosts; then if grep -xq '127.0.1.1\s.*' /etc/hosts; then sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 minikube/g' /etc/hosts; else echo '127.0.1.1 minikube' | sudo tee -a /etc/hosts; fi fi I0423 17:59:48.569613 8992 main.go:141] libmachine: SSH cmd err, output: : I0423 17:59:48.569613 8992 buildroot.go:172] set auth options {CertDir:C:\Users\Admin\.minikube CaCertPath:C:\Users\Admin\.minikube\certs\ca.pem CaPrivateKeyPath:C:\Users\Admin\.minikube\certs\ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:C:\Users\Admin\.minikube\machines\server.pem ServerKeyPath:C:\Users\Admin\.minikube\machines\server-key.pem ClientKeyPath:C:\Users\Admin\.minikube\certs\key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:C:\Users\Admin\.minikube\certs\cert.pem ServerCertSANs:[] StorePath:C:\Users\Admin\.minikube} I0423 17:59:48.569613 8992 buildroot.go:174] setting up certificates I0423 17:59:48.569613 8992 provision.go:84] configureAuth start I0423 17:59:48.569613 8992 main.go:141] libmachine: COMMAND: C:\Program Files\Oracle\VirtualBox\VBoxManage.exe showvminfo minikube --machinereadable I0423 17:59:48.632774 8992 main.go:141] libmachine: STDOUT: { name="minikube" encryption="disabled" groups="/" ostype="Linux 2.6 / 3.x / 4.x / 5.x (64-bit)" UUID="9b5e3e57-d91e-432d-add7-21d97242bc13" CfgFile="C:\\Users\\Admin\\.minikube\\machines\\minikube\\minikube\\minikube.vbox" SnapFldr="C:\\Users\\Admin\\.minikube\\machines\\minikube\\minikube\\Snapshots" LogFldr="C:\\Users\\Admin\\.minikube\\machines\\minikube\\minikube\\Logs" hardwareuuid="9b5e3e57-d91e-432d-add7-21d97242bc13" memory=4000 pagefusion="off" vram=8 cpuexecutioncap=100 hpet="on" cpu-profile="host" chipset="piix3" firmware="BIOS" cpus=2 pae="on" longmode="on" triplefaultreset="off" apic="on" x2apic="off" nested-hw-virt="off" cpuid-portability-level=0 bootmenu="disabled" boot1="dvd" boot2="dvd" boot3="disk" boot4="none" acpi="on" ioapic="on" biosapic="apic" biossystemtimeoffset=0 NvramFile="C:\\Users\\Admin\\.minikube\\machines\\minikube\\minikube\\minikube.nvram" rtcuseutc="on" hwvirtex="on" nestedpaging="on" largepages="on" vtxvpid="on" vtxux="on" virtvmsavevmload="on" iommu="none" paravirtprovider="default" effparavirtprovider="kvm" VMState="running" VMStateChangeTime="2024-04-23T12:21:55.061000000" graphicscontroller="vboxvga" monitorcount=1 accelerate3d="off" accelerate2dvideo="off" teleporterenabled="off" teleporterport=0 teleporteraddress="" teleporterpassword="" tracing-enabled="off" tracing-allow-vm-access="off" tracing-config="" autostart-enabled="off" autostart-delay=0 defaultfrontend="" vmprocpriority="default" storagecontrollername0="SATA" storagecontrollertype0="IntelAhci" storagecontrollerinstance0="0" storagecontrollermaxportcount0="30" storagecontrollerportcount0="30" storagecontrollerbootable0="on" "SATA-0-0"="C:\\Users\\Admin\\.minikube\\machines\\minikube\\boot2docker.iso" "SATA-ImageUUID-0-0"="e0f2fbc8-be91-4bbd-a49a-129556e63202" "SATA-tempeject-0-0"="off" "SATA-IsEjected-0-0"="off" "SATA-hot-pluggable-0-0"="off" "SATA-nonrotational-0-0"="off" "SATA-discard-0-0"="off" "SATA-1-0"="C:\\Users\\Admin\\.minikube\\machines\\minikube\\disk.vmdk" "SATA-ImageUUID-1-0"="7d9f9db6-e8c6-40f8-a240-8d3dacae006d" "SATA-hot-pluggable-1-0"="off" "SATA-nonrotational-1-0"="off" "SATA-discard-1-0"="off" "SATA-2-0"="none" "SATA-3-0"="none" "SATA-4-0"="none" "SATA-5-0"="none" "SATA-6-0"="none" "SATA-7-0"="none" "SATA-8-0"="none" "SATA-9-0"="none" "SATA-10-0"="none" "SATA-11-0"="none" "SATA-12-0"="none" "SATA-13-0"="none" "SATA-14-0"="none" "SATA-15-0"="none" "SATA-16-0"="none" "SATA-17-0"="none" "SATA-18-0"="none" "SATA-19-0"="none" "SATA-20-0"="none" "SATA-21-0"="none" "SATA-22-0"="none" "SATA-23-0"="none" "SATA-24-0"="none" "SATA-25-0"="none" "SATA-26-0"="none" "SATA-27-0"="none" "SATA-28-0"="none" "SATA-29-0"="none" natnet1="nat" macaddress1="080027C72164" cableconnected1="on" nic1="nat" nictype1="virtio" nicspeed1="0" mtu="0" sockSnd="64" sockRcv="64" tcpWndSnd="64" tcpWndRcv="64" Forwarding(0)="ssh,tcp,127.0.0.1,50096,,22" hostonlyadapter2="VirtualBox Host-Only Ethernet Adapter #2" macaddress2="0800270FF676" cableconnected2="on" nic2="hostonly" nictype2="virtio" nicspeed2="0" nic3="none" nic4="none" nic5="none" nic6="none" nic7="none" nic8="none" hidpointing="ps2mouse" hidkeyboard="ps2kbd" uart1="off" uart2="off" uart3="off" uart4="off" lpt1="off" lpt2="off" audio="default" audio_out="off" audio_in="off" clipboard="disabled" draganddrop="disabled" SessionName="headless" VideoMode="720,400,0"@0,0 1 vrde="off" usb="off" ehci="off" xhci="off" SharedFolderNameMachineMapping1="c/Users" SharedFolderPathMachineMapping1="\\\\?\\c:\\Users" VRDEActiveConnection="off" VRDEClients==0 recording_enabled="off" recording_screens=1 rec_screen0 rec_screen_enabled="on" rec_screen_id=0 rec_screen_video_enabled="on" rec_screen_audio_enabled="off" rec_screen_dest="File" rec_screen_dest_filename="C:\\Users\\Admin\\.minikube\\machines\\minikube\\minikube\\minikube-screen0.webm" rec_screen_opts="vc_enabled=true,ac_enabled=false,ac_profile=med" rec_screen_video_res_xy="1024x768" rec_screen_video_rate_kbps=512 rec_screen_video_fps=25 GuestMemoryBalloon=0 GuestOSType="Linux26_64" GuestAdditionsRunLevel=2 GuestAdditionsVersion="6.0.0 r127566" GuestAdditionsFacility_VirtualBox Base Driver=50,1713874957680 GuestAdditionsFacility_VirtualBox System Service=50,1713874958183 GuestAdditionsFacility_Seamless Mode=0,1713874957669 GuestAdditionsFacility_Graphics Mode=0,1713874957669 } I0423 17:59:48.632774 8992 main.go:141] libmachine: STDERR: { } I0423 17:59:48.632774 8992 main.go:141] libmachine: COMMAND: C:\Program Files\Oracle\VirtualBox\VBoxManage.exe showvminfo minikube --machinereadable I0423 17:59:48.692501 8992 main.go:141] libmachine: STDOUT: { name="minikube" encryption="disabled" groups="/" ostype="Linux 2.6 / 3.x / 4.x / 5.x (64-bit)" UUID="9b5e3e57-d91e-432d-add7-21d97242bc13" CfgFile="C:\\Users\\Admin\\.minikube\\machines\\minikube\\minikube\\minikube.vbox" SnapFldr="C:\\Users\\Admin\\.minikube\\machines\\minikube\\minikube\\Snapshots" LogFldr="C:\\Users\\Admin\\.minikube\\machines\\minikube\\minikube\\Logs" hardwareuuid="9b5e3e57-d91e-432d-add7-21d97242bc13" memory=4000 pagefusion="off" vram=8 cpuexecutioncap=100 hpet="on" cpu-profile="host" chipset="piix3" firmware="BIOS" cpus=2 pae="on" longmode="on" triplefaultreset="off" apic="on" x2apic="off" nested-hw-virt="off" cpuid-portability-level=0 bootmenu="disabled" boot1="dvd" boot2="dvd" boot3="disk" boot4="none" acpi="on" ioapic="on" biosapic="apic" biossystemtimeoffset=0 NvramFile="C:\\Users\\Admin\\.minikube\\machines\\minikube\\minikube\\minikube.nvram" rtcuseutc="on" hwvirtex="on" nestedpaging="on" largepages="on" vtxvpid="on" vtxux="on" virtvmsavevmload="on" iommu="none" paravirtprovider="default" effparavirtprovider="kvm" VMState="running" VMStateChangeTime="2024-04-23T12:21:55.061000000" graphicscontroller="vboxvga" monitorcount=1 accelerate3d="off" accelerate2dvideo="off" teleporterenabled="off" teleporterport=0 teleporteraddress="" teleporterpassword="" tracing-enabled="off" tracing-allow-vm-access="off" tracing-config="" autostart-enabled="off" autostart-delay=0 defaultfrontend="" vmprocpriority="default" storagecontrollername0="SATA" storagecontrollertype0="IntelAhci" storagecontrollerinstance0="0" storagecontrollermaxportcount0="30" storagecontrollerportcount0="30" storagecontrollerbootable0="on" "SATA-0-0"="C:\\Users\\Admin\\.minikube\\machines\\minikube\\boot2docker.iso" "SATA-ImageUUID-0-0"="e0f2fbc8-be91-4bbd-a49a-129556e63202" "SATA-tempeject-0-0"="off" "SATA-IsEjected-0-0"="off" "SATA-hot-pluggable-0-0"="off" "SATA-nonrotational-0-0"="off" "SATA-discard-0-0"="off" "SATA-1-0"="C:\\Users\\Admin\\.minikube\\machines\\minikube\\disk.vmdk" "SATA-ImageUUID-1-0"="7d9f9db6-e8c6-40f8-a240-8d3dacae006d" "SATA-hot-pluggable-1-0"="off" "SATA-nonrotational-1-0"="off" "SATA-discard-1-0"="off" "SATA-2-0"="none" "SATA-3-0"="none" "SATA-4-0"="none" "SATA-5-0"="none" "SATA-6-0"="none" "SATA-7-0"="none" "SATA-8-0"="none" "SATA-9-0"="none" "SATA-10-0"="none" "SATA-11-0"="none" "SATA-12-0"="none" "SATA-13-0"="none" "SATA-14-0"="none" "SATA-15-0"="none" "SATA-16-0"="none" "SATA-17-0"="none" "SATA-18-0"="none" "SATA-19-0"="none" "SATA-20-0"="none" "SATA-21-0"="none" "SATA-22-0"="none" "SATA-23-0"="none" "SATA-24-0"="none" "SATA-25-0"="none" "SATA-26-0"="none" "SATA-27-0"="none" "SATA-28-0"="none" "SATA-29-0"="none" natnet1="nat" macaddress1="080027C72164" cableconnected1="on" nic1="nat" nictype1="virtio" nicspeed1="0" mtu="0" sockSnd="64" sockRcv="64" tcpWndSnd="64" tcpWndRcv="64" Forwarding(0)="ssh,tcp,127.0.0.1,50096,,22" hostonlyadapter2="VirtualBox Host-Only Ethernet Adapter #2" macaddress2="0800270FF676" cableconnected2="on" nic2="hostonly" nictype2="virtio" nicspeed2="0" nic3="none" nic4="none" nic5="none" nic6="none" nic7="none" nic8="none" hidpointing="ps2mouse" hidkeyboard="ps2kbd" uart1="off" uart2="off" uart3="off" uart4="off" lpt1="off" lpt2="off" audio="default" audio_out="off" audio_in="off" clipboard="disabled" draganddrop="disabled" SessionName="headless" VideoMode="720,400,0"@0,0 1 vrde="off" usb="off" ehci="off" xhci="off" SharedFolderNameMachineMapping1="c/Users" SharedFolderPathMachineMapping1="\\\\?\\c:\\Users" VRDEActiveConnection="off" VRDEClients==0 recording_enabled="off" recording_screens=1 rec_screen0 rec_screen_enabled="on" rec_screen_id=0 rec_screen_video_enabled="on" rec_screen_audio_enabled="off" rec_screen_dest="File" rec_screen_dest_filename="C:\\Users\\Admin\\.minikube\\machines\\minikube\\minikube\\minikube-screen0.webm" rec_screen_opts="vc_enabled=true,ac_enabled=false,ac_profile=med" rec_screen_video_res_xy="1024x768" rec_screen_video_rate_kbps=512 rec_screen_video_fps=25 GuestMemoryBalloon=0 GuestOSType="Linux26_64" GuestAdditionsRunLevel=2 GuestAdditionsVersion="6.0.0 r127566" GuestAdditionsFacility_VirtualBox Base Driver=50,1713874957680 GuestAdditionsFacility_VirtualBox System Service=50,1713874958183 GuestAdditionsFacility_Seamless Mode=0,1713874957669 GuestAdditionsFacility_Graphics Mode=0,1713874957669 } I0423 17:59:48.692501 8992 main.go:141] libmachine: STDERR: { } I0423 17:59:48.692501 8992 main.go:141] libmachine: Host-only MAC: 0800270ff676 I0423 17:59:48.694595 8992 main.go:141] libmachine: Using SSH client type: native I0423 17:59:48.694595 8992 main.go:141] libmachine: &{{{ 0 [] [] []} docker [0x137a1c0] 0x137cda0 [] 0s} 127.0.0.1 50096 } I0423 17:59:48.694595 8992 main.go:141] libmachine: About to run SSH command: ip addr show I0423 17:59:48.854249 8992 main.go:141] libmachine: SSH cmd err, output: : 1: lo: mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000 link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 inet 127.0.0.1/8 scope host lo valid_lft forever preferred_lft forever 2: eth0: mtu 1500 qdisc pfifo_fast state UP group default qlen 1000 link/ether 08:00:27:c7:21:64 brd ff:ff:ff:ff:ff:ff inet 10.0.2.15/24 metric 1024 brd 10.0.2.255 scope global dynamic eth0 valid_lft 85969sec preferred_lft 85969sec 3: eth1: mtu 1500 qdisc pfifo_fast state UP group default qlen 1000 link/ether 08:00:27:0f:f6:76 brd ff:ff:ff:ff:ff:ff inet 192.168.59.101/24 metric 1024 brd 192.168.59.255 scope global dynamic eth1 valid_lft 469sec preferred_lft 469sec 4: sit0@NONE: mtu 1480 qdisc noop state DOWN group default qlen 1000 link/sit 0.0.0.0 brd 0.0.0.0 5: docker0: mtu 1500 qdisc noqueue state DOWN group default link/ether 02:42:48:91:f7:26 brd ff:ff:ff:ff:ff:ff inet 172.17.0.1/16 brd 172.17.255.255 scope global docker0 valid_lft forever preferred_lft forever 6: bridge: mtu 1500 qdisc noqueue state UP group default qlen 1000 link/ether 6a:c1:68:84:40:54 brd ff:ff:ff:ff:ff:ff inet 10.244.0.1/16 brd 10.244.255.255 scope global bridge valid_lft forever preferred_lft forever 7: veth8445026a@if3: mtu 1500 qdisc noqueue master bridge state UP group default link/ether ca:15:76:15:5b:db brd ff:ff:ff:ff:ff:ff link-netnsid 0 8: veth7a242195@if3: mtu 1500 qdisc noqueue master bridge state UP group default link/ether 8a:35:8a:56:8a:40 brd ff:ff:ff:ff:ff:ff link-netnsid 1 I0423 17:59:48.854249 8992 main.go:141] libmachine: SSH returned: 1: lo: mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000 link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 inet 127.0.0.1/8 scope host lo valid_lft forever preferred_lft forever 2: eth0: mtu 1500 qdisc pfifo_fast state UP group default qlen 1000 link/ether 08:00:27:c7:21:64 brd ff:ff:ff:ff:ff:ff inet 10.0.2.15/24 metric 1024 brd 10.0.2.255 scope global dynamic eth0 valid_lft 85969sec preferred_lft 85969sec 3: eth1: mtu 1500 qdisc pfifo_fast state UP group default qlen 1000 link/ether 08:00:27:0f:f6:76 brd ff:ff:ff:ff:ff:ff inet 192.168.59.101/24 metric 1024 brd 192.168.59.255 scope global dynamic eth1 valid_lft 469sec preferred_lft 469sec 4: sit0@NONE: mtu 1480 qdisc noop state DOWN group default qlen 1000 link/sit 0.0.0.0 brd 0.0.0.0 5: docker0: mtu 1500 qdisc noqueue state DOWN group default link/ether 02:42:48:91:f7:26 brd ff:ff:ff:ff:ff:ff inet 172.17.0.1/16 brd 172.17.255.255 scope global docker0 valid_lft forever preferred_lft forever 6: bridge: mtu 1500 qdisc noqueue state UP group default qlen 1000 link/ether 6a:c1:68:84:40:54 brd ff:ff:ff:ff:ff:ff inet 10.244.0.1/16 brd 10.244.255.255 scope global bridge valid_lft forever preferred_lft forever 7: veth8445026a@if3: mtu 1500 qdisc noqueue master bridge state UP group default link/ether ca:15:76:15:5b:db brd ff:ff:ff:ff:ff:ff link-netnsid 0 8: veth7a242195@if3: mtu 1500 qdisc noqueue master bridge state UP group default link/ether 8a:35:8a:56:8a:40 brd ff:ff:ff:ff:ff:ff link-netnsid 1 END SSH I0423 17:59:48.854249 8992 provision.go:143] copyHostCerts I0423 17:59:48.854249 8992 exec_runner.go:144] found C:\Users\Admin\.minikube/ca.pem, removing ... I0423 17:59:48.854249 8992 exec_runner.go:203] rm: C:\Users\Admin\.minikube\ca.pem I0423 17:59:48.854801 8992 exec_runner.go:151] cp: C:\Users\Admin\.minikube\certs\ca.pem --> C:\Users\Admin\.minikube/ca.pem (1074 bytes) I0423 17:59:48.855329 8992 exec_runner.go:144] found C:\Users\Admin\.minikube/cert.pem, removing ... I0423 17:59:48.855329 8992 exec_runner.go:203] rm: C:\Users\Admin\.minikube\cert.pem I0423 17:59:48.855866 8992 exec_runner.go:151] cp: C:\Users\Admin\.minikube\certs\cert.pem --> C:\Users\Admin\.minikube/cert.pem (1119 bytes) I0423 17:59:48.856399 8992 exec_runner.go:144] found C:\Users\Admin\.minikube/key.pem, removing ... I0423 17:59:48.856399 8992 exec_runner.go:203] rm: C:\Users\Admin\.minikube\key.pem I0423 17:59:48.856399 8992 exec_runner.go:151] cp: C:\Users\Admin\.minikube\certs\key.pem --> C:\Users\Admin\.minikube/key.pem (1675 bytes) I0423 17:59:48.869315 8992 provision.go:117] generating server cert: C:\Users\Admin\.minikube\machines\server.pem ca-key=C:\Users\Admin\.minikube\certs\ca.pem private-key=C:\Users\Admin\.minikube\certs\ca-key.pem org=Admin.minikube san=[127.0.0.1 192.168.59.101 localhost minikube] I0423 17:59:49.011896 8992 provision.go:177] copyRemoteCerts I0423 17:59:49.019399 8992 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker I0423 17:59:49.019399 8992 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:50096 SSHKeyPath:C:\Users\Admin\.minikube\machines\minikube\id_rsa Username:docker} I0423 17:59:49.100101 8992 ssh_runner.go:362] scp C:\Users\Admin\.minikube\certs\ca.pem --> /etc/docker/ca.pem (1074 bytes) I0423 17:59:49.157863 8992 ssh_runner.go:362] scp C:\Users\Admin\.minikube\machines\server.pem --> /etc/docker/server.pem (1176 bytes) I0423 17:59:49.222654 8992 ssh_runner.go:362] scp C:\Users\Admin\.minikube\machines\server-key.pem --> /etc/docker/server-key.pem (1679 bytes) I0423 17:59:49.283066 8992 provision.go:87] duration metric: took 713.4528ms to configureAuth I0423 17:59:49.283066 8992 buildroot.go:189] setting minikube options for container-runtime I0423 17:59:49.283066 8992 config.go:182] Loaded profile config "minikube": Driver=virtualbox, ContainerRuntime=docker, KubernetesVersion=v1.30.0 I0423 17:59:49.285356 8992 main.go:141] libmachine: Using SSH client type: native I0423 17:59:49.285972 8992 main.go:141] libmachine: &{{{ 0 [] [] []} docker [0x137a1c0] 0x137cda0 [] 0s} 127.0.0.1 50096 } I0423 17:59:49.285972 8992 main.go:141] libmachine: About to run SSH command: df --output=fstype / | tail -n 1 I0423 17:59:49.436707 8992 main.go:141] libmachine: SSH cmd err, output: : tmpfs I0423 17:59:49.436707 8992 buildroot.go:70] root file system type: tmpfs I0423 17:59:49.436707 8992 provision.go:314] Updating docker unit: /lib/systemd/system/docker.service ... I0423 17:59:49.438953 8992 main.go:141] libmachine: Using SSH client type: native I0423 17:59:49.439529 8992 main.go:141] libmachine: &{{{ 0 [] [] []} docker [0x137a1c0] 0x137cda0 [] 0s} 127.0.0.1 50096 } I0423 17:59:49.439529 8992 main.go:141] libmachine: About to run SSH command: sudo mkdir -p /lib/systemd/system && printf %!s(MISSING) "[Unit] Description=Docker Application Container Engine Documentation=https://docs.docker.com After=network.target minikube-automount.service docker.socket Requires= minikube-automount.service docker.socket StartLimitBurst=3 StartLimitIntervalSec=60 [Service] Type=notify Restart=on-failure # This file is a systemd drop-in unit that inherits from the base dockerd configuration. # The base configuration already specifies an 'ExecStart=...' command. The first directive # here is to clear out that command inherited from the base configuration. Without this, # the command from the base configuration and the command specified here are treated as # a sequence of commands, which is not the desired behavior, nor is it valid -- systemd # will catch this invalid input and refuse to start the service with an error like: # Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services. # NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other # container runtimes. If left unlimited, it may result in OOM issues with MySQL. ExecStart= ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=virtualbox --insecure-registry 10.96.0.0/12 ExecReload=/bin/kill -s HUP \$MAINPID # Having non-zero Limit*s causes performance problems due to accounting overhead # in the kernel. We recommend using cgroups to do container-local accounting. LimitNOFILE=infinity LimitNPROC=infinity LimitCORE=infinity # Uncomment TasksMax if your systemd version supports it. # Only systemd 226 and above support this version. TasksMax=infinity TimeoutStartSec=0 # set delegate yes so that systemd does not reset the cgroups of docker containers Delegate=yes # kill only the docker process, not all processes in the cgroup KillMode=process [Install] WantedBy=multi-user.target " | sudo tee /lib/systemd/system/docker.service.new I0423 17:59:49.568283 8992 main.go:141] libmachine: SSH cmd err, output: : [Unit] Description=Docker Application Container Engine Documentation=https://docs.docker.com After=network.target minikube-automount.service docker.socket Requires= minikube-automount.service docker.socket StartLimitBurst=3 StartLimitIntervalSec=60 [Service] Type=notify Restart=on-failure # This file is a systemd drop-in unit that inherits from the base dockerd configuration. # The base configuration already specifies an 'ExecStart=...' command. The first directive # here is to clear out that command inherited from the base configuration. Without this, # the command from the base configuration and the command specified here are treated as # a sequence of commands, which is not the desired behavior, nor is it valid -- systemd # will catch this invalid input and refuse to start the service with an error like: # Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services. # NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other # container runtimes. If left unlimited, it may result in OOM issues with MySQL. ExecStart= ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=virtualbox --insecure-registry 10.96.0.0/12 ExecReload=/bin/kill -s HUP $MAINPID # Having non-zero Limit*s causes performance problems due to accounting overhead # in the kernel. We recommend using cgroups to do container-local accounting. LimitNOFILE=infinity LimitNPROC=infinity LimitCORE=infinity # Uncomment TasksMax if your systemd version supports it. # Only systemd 226 and above support this version. TasksMax=infinity TimeoutStartSec=0 # set delegate yes so that systemd does not reset the cgroups of docker containers Delegate=yes # kill only the docker process, not all processes in the cgroup KillMode=process [Install] WantedBy=multi-user.target I0423 17:59:49.571081 8992 main.go:141] libmachine: Using SSH client type: native I0423 17:59:49.572279 8992 main.go:141] libmachine: &{{{ 0 [] [] []} docker [0x137a1c0] 0x137cda0 [] 0s} 127.0.0.1 50096 } I0423 17:59:49.572279 8992 main.go:141] libmachine: About to run SSH command: sudo diff -u /lib/systemd/system/docker.service /lib/systemd/system/docker.service.new || { sudo mv /lib/systemd/system/docker.service.new /lib/systemd/system/docker.service; sudo systemctl -f daemon-reload && sudo systemctl -f enable docker && sudo systemctl -f restart docker; } I0423 17:59:49.695743 8992 main.go:141] libmachine: SSH cmd err, output: : I0423 17:59:49.695743 8992 machine.go:97] duration metric: took 4.0648953s to provisionDockerMachine I0423 17:59:49.695743 8992 start.go:293] postStartSetup for "minikube" (driver="virtualbox") I0423 17:59:49.695743 8992 start.go:322] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs] I0423 17:59:49.702678 8992 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs I0423 17:59:49.703187 8992 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:50096 SSHKeyPath:C:\Users\Admin\.minikube\machines\minikube\id_rsa Username:docker} I0423 17:59:49.778906 8992 ssh_runner.go:195] Run: cat /etc/os-release I0423 17:59:49.791999 8992 info.go:137] Remote host: Buildroot 2023.02.9 I0423 17:59:49.791999 8992 filesync.go:126] Scanning C:\Users\Admin\.minikube\addons for local assets ... I0423 17:59:49.791999 8992 filesync.go:126] Scanning C:\Users\Admin\.minikube\files for local assets ... I0423 17:59:49.791999 8992 start.go:296] duration metric: took 96.256ms for postStartSetup I0423 17:59:49.791999 8992 fix.go:56] duration metric: took 4.2918756s for fixHost I0423 17:59:49.794345 8992 main.go:141] libmachine: Using SSH client type: native I0423 17:59:49.794853 8992 main.go:141] libmachine: &{{{ 0 [] [] []} docker [0x137a1c0] 0x137cda0 [] 0s} 127.0.0.1 50096 } I0423 17:59:49.794853 8992 main.go:141] libmachine: About to run SSH command: date +%!s(MISSING).%!N(MISSING) I0423 17:59:49.921521 8992 main.go:141] libmachine: SSH cmd err, output: : 1713875389.807189568 I0423 17:59:49.921521 8992 fix.go:216] guest clock: 1713875389.807189568 I0423 17:59:49.921521 8992 fix.go:229] Guest: 2024-04-23 17:59:49.807189568 +0530 IST Remote: 2024-04-23 17:59:49.7919991 +0530 IST m=+4.465887101 (delta=15.190468ms) I0423 17:59:49.921521 8992 fix.go:200] guest clock delta is within tolerance: 15.190468ms I0423 17:59:49.921521 8992 start.go:83] releasing machines lock for "minikube", held for 4.4213981s I0423 17:59:49.921521 8992 main.go:141] libmachine: COMMAND: C:\Program Files\Oracle\VirtualBox\VBoxManage.exe showvminfo minikube --machinereadable I0423 17:59:49.982993 8992 main.go:141] libmachine: STDOUT: { name="minikube" encryption="disabled" groups="/" ostype="Linux 2.6 / 3.x / 4.x / 5.x (64-bit)" UUID="9b5e3e57-d91e-432d-add7-21d97242bc13" CfgFile="C:\\Users\\Admin\\.minikube\\machines\\minikube\\minikube\\minikube.vbox" SnapFldr="C:\\Users\\Admin\\.minikube\\machines\\minikube\\minikube\\Snapshots" LogFldr="C:\\Users\\Admin\\.minikube\\machines\\minikube\\minikube\\Logs" hardwareuuid="9b5e3e57-d91e-432d-add7-21d97242bc13" memory=4000 pagefusion="off" vram=8 cpuexecutioncap=100 hpet="on" cpu-profile="host" chipset="piix3" firmware="BIOS" cpus=2 pae="on" longmode="on" triplefaultreset="off" apic="on" x2apic="off" nested-hw-virt="off" cpuid-portability-level=0 bootmenu="disabled" boot1="dvd" boot2="dvd" boot3="disk" boot4="none" acpi="on" ioapic="on" biosapic="apic" biossystemtimeoffset=0 NvramFile="C:\\Users\\Admin\\.minikube\\machines\\minikube\\minikube\\minikube.nvram" rtcuseutc="on" hwvirtex="on" nestedpaging="on" largepages="on" vtxvpid="on" vtxux="on" virtvmsavevmload="on" iommu="none" paravirtprovider="default" effparavirtprovider="kvm" VMState="running" VMStateChangeTime="2024-04-23T12:21:55.061000000" graphicscontroller="vboxvga" monitorcount=1 accelerate3d="off" accelerate2dvideo="off" teleporterenabled="off" teleporterport=0 teleporteraddress="" teleporterpassword="" tracing-enabled="off" tracing-allow-vm-access="off" tracing-config="" autostart-enabled="off" autostart-delay=0 defaultfrontend="" vmprocpriority="default" storagecontrollername0="SATA" storagecontrollertype0="IntelAhci" storagecontrollerinstance0="0" storagecontrollermaxportcount0="30" storagecontrollerportcount0="30" storagecontrollerbootable0="on" "SATA-0-0"="C:\\Users\\Admin\\.minikube\\machines\\minikube\\boot2docker.iso" "SATA-ImageUUID-0-0"="e0f2fbc8-be91-4bbd-a49a-129556e63202" "SATA-tempeject-0-0"="off" "SATA-IsEjected-0-0"="off" "SATA-hot-pluggable-0-0"="off" "SATA-nonrotational-0-0"="off" "SATA-discard-0-0"="off" "SATA-1-0"="C:\\Users\\Admin\\.minikube\\machines\\minikube\\disk.vmdk" "SATA-ImageUUID-1-0"="7d9f9db6-e8c6-40f8-a240-8d3dacae006d" "SATA-hot-pluggable-1-0"="off" "SATA-nonrotational-1-0"="off" "SATA-discard-1-0"="off" "SATA-2-0"="none" "SATA-3-0"="none" "SATA-4-0"="none" "SATA-5-0"="none" "SATA-6-0"="none" "SATA-7-0"="none" "SATA-8-0"="none" "SATA-9-0"="none" "SATA-10-0"="none" "SATA-11-0"="none" "SATA-12-0"="none" "SATA-13-0"="none" "SATA-14-0"="none" "SATA-15-0"="none" "SATA-16-0"="none" "SATA-17-0"="none" "SATA-18-0"="none" "SATA-19-0"="none" "SATA-20-0"="none" "SATA-21-0"="none" "SATA-22-0"="none" "SATA-23-0"="none" "SATA-24-0"="none" "SATA-25-0"="none" "SATA-26-0"="none" "SATA-27-0"="none" "SATA-28-0"="none" "SATA-29-0"="none" natnet1="nat" macaddress1="080027C72164" cableconnected1="on" nic1="nat" nictype1="virtio" nicspeed1="0" mtu="0" sockSnd="64" sockRcv="64" tcpWndSnd="64" tcpWndRcv="64" Forwarding(0)="ssh,tcp,127.0.0.1,50096,,22" hostonlyadapter2="VirtualBox Host-Only Ethernet Adapter #2" macaddress2="0800270FF676" cableconnected2="on" nic2="hostonly" nictype2="virtio" nicspeed2="0" nic3="none" nic4="none" nic5="none" nic6="none" nic7="none" nic8="none" hidpointing="ps2mouse" hidkeyboard="ps2kbd" uart1="off" uart2="off" uart3="off" uart4="off" lpt1="off" lpt2="off" audio="default" audio_out="off" audio_in="off" clipboard="disabled" draganddrop="disabled" SessionName="headless" VideoMode="720,400,0"@0,0 1 vrde="off" usb="off" ehci="off" xhci="off" SharedFolderNameMachineMapping1="c/Users" SharedFolderPathMachineMapping1="\\\\?\\c:\\Users" VRDEActiveConnection="off" VRDEClients==0 recording_enabled="off" recording_screens=1 rec_screen0 rec_screen_enabled="on" rec_screen_id=0 rec_screen_video_enabled="on" rec_screen_audio_enabled="off" rec_screen_dest="File" rec_screen_dest_filename="C:\\Users\\Admin\\.minikube\\machines\\minikube\\minikube\\minikube-screen0.webm" rec_screen_opts="vc_enabled=true,ac_enabled=false,ac_profile=med" rec_screen_video_res_xy="1024x768" rec_screen_video_rate_kbps=512 rec_screen_video_fps=25 GuestMemoryBalloon=0 GuestOSType="Linux26_64" GuestAdditionsRunLevel=2 GuestAdditionsVersion="6.0.0 r127566" GuestAdditionsFacility_VirtualBox Base Driver=50,1713874957680 GuestAdditionsFacility_VirtualBox System Service=50,1713874958183 GuestAdditionsFacility_Seamless Mode=0,1713874957669 GuestAdditionsFacility_Graphics Mode=0,1713874957669 } I0423 17:59:49.982993 8992 main.go:141] libmachine: STDERR: { } I0423 17:59:49.982993 8992 main.go:141] libmachine: COMMAND: C:\Program Files\Oracle\VirtualBox\VBoxManage.exe showvminfo minikube --machinereadable I0423 17:59:50.042741 8992 main.go:141] libmachine: STDOUT: { name="minikube" encryption="disabled" groups="/" ostype="Linux 2.6 / 3.x / 4.x / 5.x (64-bit)" UUID="9b5e3e57-d91e-432d-add7-21d97242bc13" CfgFile="C:\\Users\\Admin\\.minikube\\machines\\minikube\\minikube\\minikube.vbox" SnapFldr="C:\\Users\\Admin\\.minikube\\machines\\minikube\\minikube\\Snapshots" LogFldr="C:\\Users\\Admin\\.minikube\\machines\\minikube\\minikube\\Logs" hardwareuuid="9b5e3e57-d91e-432d-add7-21d97242bc13" memory=4000 pagefusion="off" vram=8 cpuexecutioncap=100 hpet="on" cpu-profile="host" chipset="piix3" firmware="BIOS" cpus=2 pae="on" longmode="on" triplefaultreset="off" apic="on" x2apic="off" nested-hw-virt="off" cpuid-portability-level=0 bootmenu="disabled" boot1="dvd" boot2="dvd" boot3="disk" boot4="none" acpi="on" ioapic="on" biosapic="apic" biossystemtimeoffset=0 NvramFile="C:\\Users\\Admin\\.minikube\\machines\\minikube\\minikube\\minikube.nvram" rtcuseutc="on" hwvirtex="on" nestedpaging="on" largepages="on" vtxvpid="on" vtxux="on" virtvmsavevmload="on" iommu="none" paravirtprovider="default" effparavirtprovider="kvm" VMState="running" VMStateChangeTime="2024-04-23T12:21:55.061000000" graphicscontroller="vboxvga" monitorcount=1 accelerate3d="off" accelerate2dvideo="off" teleporterenabled="off" teleporterport=0 teleporteraddress="" teleporterpassword="" tracing-enabled="off" tracing-allow-vm-access="off" tracing-config="" autostart-enabled="off" autostart-delay=0 defaultfrontend="" vmprocpriority="default" storagecontrollername0="SATA" storagecontrollertype0="IntelAhci" storagecontrollerinstance0="0" storagecontrollermaxportcount0="30" storagecontrollerportcount0="30" storagecontrollerbootable0="on" "SATA-0-0"="C:\\Users\\Admin\\.minikube\\machines\\minikube\\boot2docker.iso" "SATA-ImageUUID-0-0"="e0f2fbc8-be91-4bbd-a49a-129556e63202" "SATA-tempeject-0-0"="off" "SATA-IsEjected-0-0"="off" "SATA-hot-pluggable-0-0"="off" "SATA-nonrotational-0-0"="off" "SATA-discard-0-0"="off" "SATA-1-0"="C:\\Users\\Admin\\.minikube\\machines\\minikube\\disk.vmdk" "SATA-ImageUUID-1-0"="7d9f9db6-e8c6-40f8-a240-8d3dacae006d" "SATA-hot-pluggable-1-0"="off" "SATA-nonrotational-1-0"="off" "SATA-discard-1-0"="off" "SATA-2-0"="none" "SATA-3-0"="none" "SATA-4-0"="none" "SATA-5-0"="none" "SATA-6-0"="none" "SATA-7-0"="none" "SATA-8-0"="none" "SATA-9-0"="none" "SATA-10-0"="none" "SATA-11-0"="none" "SATA-12-0"="none" "SATA-13-0"="none" "SATA-14-0"="none" "SATA-15-0"="none" "SATA-16-0"="none" "SATA-17-0"="none" "SATA-18-0"="none" "SATA-19-0"="none" "SATA-20-0"="none" "SATA-21-0"="none" "SATA-22-0"="none" "SATA-23-0"="none" "SATA-24-0"="none" "SATA-25-0"="none" "SATA-26-0"="none" "SATA-27-0"="none" "SATA-28-0"="none" "SATA-29-0"="none" natnet1="nat" macaddress1="080027C72164" cableconnected1="on" nic1="nat" nictype1="virtio" nicspeed1="0" mtu="0" sockSnd="64" sockRcv="64" tcpWndSnd="64" tcpWndRcv="64" Forwarding(0)="ssh,tcp,127.0.0.1,50096,,22" hostonlyadapter2="VirtualBox Host-Only Ethernet Adapter #2" macaddress2="0800270FF676" cableconnected2="on" nic2="hostonly" nictype2="virtio" nicspeed2="0" nic3="none" nic4="none" nic5="none" nic6="none" nic7="none" nic8="none" hidpointing="ps2mouse" hidkeyboard="ps2kbd" uart1="off" uart2="off" uart3="off" uart4="off" lpt1="off" lpt2="off" audio="default" audio_out="off" audio_in="off" clipboard="disabled" draganddrop="disabled" SessionName="headless" VideoMode="720,400,0"@0,0 1 vrde="off" usb="off" ehci="off" xhci="off" SharedFolderNameMachineMapping1="c/Users" SharedFolderPathMachineMapping1="\\\\?\\c:\\Users" VRDEActiveConnection="off" VRDEClients==0 recording_enabled="off" recording_screens=1 rec_screen0 rec_screen_enabled="on" rec_screen_id=0 rec_screen_video_enabled="on" rec_screen_audio_enabled="off" rec_screen_dest="File" rec_screen_dest_filename="C:\\Users\\Admin\\.minikube\\machines\\minikube\\minikube\\minikube-screen0.webm" rec_screen_opts="vc_enabled=true,ac_enabled=false,ac_profile=med" rec_screen_video_res_xy="1024x768" rec_screen_video_rate_kbps=512 rec_screen_video_fps=25 GuestMemoryBalloon=0 GuestOSType="Linux26_64" GuestAdditionsRunLevel=2 GuestAdditionsVersion="6.0.0 r127566" GuestAdditionsFacility_VirtualBox Base Driver=50,1713874957680 GuestAdditionsFacility_VirtualBox System Service=50,1713874958183 GuestAdditionsFacility_Seamless Mode=0,1713874957669 GuestAdditionsFacility_Graphics Mode=0,1713874957669 } I0423 17:59:50.042741 8992 main.go:141] libmachine: STDERR: { } I0423 17:59:50.042741 8992 main.go:141] libmachine: Host-only MAC: 0800270ff676 I0423 17:59:50.044809 8992 main.go:141] libmachine: Using SSH client type: native I0423 17:59:50.044809 8992 main.go:141] libmachine: &{{{ 0 [] [] []} docker [0x137a1c0] 0x137cda0 [] 0s} 127.0.0.1 50096 } I0423 17:59:50.044809 8992 main.go:141] libmachine: About to run SSH command: ip addr show I0423 17:59:50.142866 8992 main.go:141] libmachine: SSH cmd err, output: : 1: lo: mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000 link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 inet 127.0.0.1/8 scope host lo valid_lft forever preferred_lft forever 2: eth0: mtu 1500 qdisc pfifo_fast state UP group default qlen 1000 link/ether 08:00:27:c7:21:64 brd ff:ff:ff:ff:ff:ff inet 10.0.2.15/24 metric 1024 brd 10.0.2.255 scope global dynamic eth0 valid_lft 85968sec preferred_lft 85968sec 3: eth1: mtu 1500 qdisc pfifo_fast state UP group default qlen 1000 link/ether 08:00:27:0f:f6:76 brd ff:ff:ff:ff:ff:ff inet 192.168.59.101/24 metric 1024 brd 192.168.59.255 scope global dynamic eth1 valid_lft 467sec preferred_lft 467sec 4: sit0@NONE: mtu 1480 qdisc noop state DOWN group default qlen 1000 link/sit 0.0.0.0 brd 0.0.0.0 5: docker0: mtu 1500 qdisc noqueue state DOWN group default link/ether 02:42:48:91:f7:26 brd ff:ff:ff:ff:ff:ff inet 172.17.0.1/16 brd 172.17.255.255 scope global docker0 valid_lft forever preferred_lft forever 6: bridge: mtu 1500 qdisc noqueue state UP group default qlen 1000 link/ether 6a:c1:68:84:40:54 brd ff:ff:ff:ff:ff:ff inet 10.244.0.1/16 brd 10.244.255.255 scope global bridge valid_lft forever preferred_lft forever 7: veth8445026a@if3: mtu 1500 qdisc noqueue master bridge state UP group default link/ether ca:15:76:15:5b:db brd ff:ff:ff:ff:ff:ff link-netnsid 0 8: veth7a242195@if3: mtu 1500 qdisc noqueue master bridge state UP group default link/ether 8a:35:8a:56:8a:40 brd ff:ff:ff:ff:ff:ff link-netnsid 1 I0423 17:59:50.142866 8992 main.go:141] libmachine: SSH returned: 1: lo: mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000 link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 inet 127.0.0.1/8 scope host lo valid_lft forever preferred_lft forever 2: eth0: mtu 1500 qdisc pfifo_fast state UP group default qlen 1000 link/ether 08:00:27:c7:21:64 brd ff:ff:ff:ff:ff:ff inet 10.0.2.15/24 metric 1024 brd 10.0.2.255 scope global dynamic eth0 valid_lft 85968sec preferred_lft 85968sec 3: eth1: mtu 1500 qdisc pfifo_fast state UP group default qlen 1000 link/ether 08:00:27:0f:f6:76 brd ff:ff:ff:ff:ff:ff inet 192.168.59.101/24 metric 1024 brd 192.168.59.255 scope global dynamic eth1 valid_lft 467sec preferred_lft 467sec 4: sit0@NONE: mtu 1480 qdisc noop state DOWN group default qlen 1000 link/sit 0.0.0.0 brd 0.0.0.0 5: docker0: mtu 1500 qdisc noqueue state DOWN group default link/ether 02:42:48:91:f7:26 brd ff:ff:ff:ff:ff:ff inet 172.17.0.1/16 brd 172.17.255.255 scope global docker0 valid_lft forever preferred_lft forever 6: bridge: mtu 1500 qdisc noqueue state UP group default qlen 1000 link/ether 6a:c1:68:84:40:54 brd ff:ff:ff:ff:ff:ff inet 10.244.0.1/16 brd 10.244.255.255 scope global bridge valid_lft forever preferred_lft forever 7: veth8445026a@if3: mtu 1500 qdisc noqueue master bridge state UP group default link/ether ca:15:76:15:5b:db brd ff:ff:ff:ff:ff:ff link-netnsid 0 8: veth7a242195@if3: mtu 1500 qdisc noqueue master bridge state UP group default link/ether 8a:35:8a:56:8a:40 brd ff:ff:ff:ff:ff:ff link-netnsid 1 END SSH I0423 17:59:50.144951 8992 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/ I0423 17:59:50.145493 8992 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:50096 SSHKeyPath:C:\Users\Admin\.minikube\machines\minikube\id_rsa Username:docker} I0423 17:59:50.150260 8992 ssh_runner.go:195] Run: cat /version.json I0423 17:59:50.150784 8992 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:50096 SSHKeyPath:C:\Users\Admin\.minikube\machines\minikube\id_rsa Username:docker} I0423 17:59:52.223948 8992 ssh_runner.go:235] Completed: cat /version.json: (2.0736877s) I0423 17:59:52.223948 8992 ssh_runner.go:235] Completed: curl -sS -m 2 https://registry.k8s.io/: (2.0789973s) W0423 17:59:52.223948 8992 start.go:860] [curl -sS -m 2 https://registry.k8s.io/] failed: curl -sS -m 2 https://registry.k8s.io/: Process exited with status 28 stdout: stderr: curl: (28) Resolving timed out after 2001 milliseconds W0423 17:59:52.223948 8992 out.go:239] ! This VM is having trouble accessing https://registry.k8s.io W0423 17:59:52.224457 8992 out.go:239] * To pull new external images, you may need to configure a proxy: https://minikube.sigs.k8s.io/docs/reference/networking/proxy/ I0423 17:59:52.237066 8992 ssh_runner.go:195] Run: systemctl --version I0423 17:59:52.260689 8992 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*" W0423 17:59:52.279828 8992 cni.go:209] loopback cni configuration skipped: "/etc/cni/net.d/*loopback.conf*" not found I0423 17:59:52.286799 8992 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%!p(MISSING), " -exec sh -c "sudo mv {} {}.mk_disabled" ; I0423 17:59:52.314108 8992 cni.go:259] no active bridge cni configs found in "/etc/cni/net.d" - nothing to disable I0423 17:59:52.314108 8992 start.go:494] detecting cgroup driver to use... I0423 17:59:52.314108 8992 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %!s(MISSING) "runtime-endpoint: unix:///run/containerd/containerd.sock " | sudo tee /etc/crictl.yaml" I0423 17:59:52.383205 8992 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)sandbox_image = .*$|\1sandbox_image = "registry.k8s.io/pause:3.9"|' /etc/containerd/config.toml" I0423 17:59:52.428831 8992 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)restrict_oom_score_adj = .*$|\1restrict_oom_score_adj = false|' /etc/containerd/config.toml" I0423 17:59:52.453859 8992 containerd.go:146] configuring containerd to use "cgroupfs" as cgroup driver... I0423 17:59:52.460605 8992 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)SystemdCgroup = .*$|\1SystemdCgroup = false|g' /etc/containerd/config.toml" I0423 17:59:52.500451 8992 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runtime.v1.linux"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml" I0423 17:59:52.533376 8992 ssh_runner.go:195] Run: sh -c "sudo sed -i '/systemd_cgroup/d' /etc/containerd/config.toml" I0423 17:59:52.574780 8992 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runc.v1"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml" I0423 17:59:52.611817 8992 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk" I0423 17:59:52.647985 8992 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)conf_dir = .*$|\1conf_dir = "/etc/cni/net.d"|g' /etc/containerd/config.toml" I0423 17:59:52.694528 8992 ssh_runner.go:195] Run: sh -c "sudo sed -i '/^ *enable_unprivileged_ports = .*/d' /etc/containerd/config.toml" I0423 17:59:52.731230 8992 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)\[plugins."io.containerd.grpc.v1.cri"\]|&\n\1 enable_unprivileged_ports = true|' /etc/containerd/config.toml" I0423 17:59:52.779047 8992 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables I0423 17:59:52.818142 8992 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward" I0423 17:59:52.870882 8992 ssh_runner.go:195] Run: sudo systemctl daemon-reload I0423 17:59:53.208748 8992 ssh_runner.go:195] Run: sudo systemctl restart containerd I0423 17:59:53.266053 8992 start.go:494] detecting cgroup driver to use... I0423 17:59:53.275699 8992 ssh_runner.go:195] Run: sudo systemctl cat docker.service I0423 17:59:53.332903 8992 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service containerd I0423 17:59:53.405104 8992 ssh_runner.go:195] Run: sudo systemctl stop -f containerd I0423 17:59:53.472192 8992 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service containerd I0423 17:59:53.512681 8992 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio I0423 17:59:53.549643 8992 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %!s(MISSING) "runtime-endpoint: unix:///var/run/cri-dockerd.sock " | sudo tee /etc/crictl.yaml" I0423 17:59:53.606603 8992 ssh_runner.go:195] Run: which cri-dockerd I0423 17:59:53.626940 8992 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/cri-docker.service.d I0423 17:59:53.654112 8992 ssh_runner.go:362] scp memory --> /etc/systemd/system/cri-docker.service.d/10-cni.conf (189 bytes) I0423 17:59:53.704467 8992 ssh_runner.go:195] Run: sudo systemctl unmask docker.service I0423 17:59:54.014307 8992 ssh_runner.go:195] Run: sudo systemctl enable docker.socket I0423 17:59:54.294404 8992 docker.go:574] configuring docker to use "cgroupfs" as cgroup driver... I0423 17:59:54.294404 8992 ssh_runner.go:362] scp memory --> /etc/docker/daemon.json (130 bytes) I0423 17:59:54.350775 8992 ssh_runner.go:195] Run: sudo systemctl daemon-reload I0423 17:59:54.724177 8992 ssh_runner.go:195] Run: sudo systemctl restart docker I0423 18:00:08.554874 8992 ssh_runner.go:235] Completed: sudo systemctl restart docker: (13.8306973s) I0423 18:00:08.560985 8992 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.socket I0423 18:00:08.612450 8992 ssh_runner.go:195] Run: sudo systemctl stop cri-docker.socket I0423 18:00:08.781465 8992 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service I0423 18:00:08.882069 8992 ssh_runner.go:195] Run: sudo systemctl unmask cri-docker.socket I0423 18:00:09.104860 8992 ssh_runner.go:195] Run: sudo systemctl enable cri-docker.socket I0423 18:00:09.332879 8992 ssh_runner.go:195] Run: sudo systemctl daemon-reload I0423 18:00:09.551285 8992 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.socket I0423 18:00:09.598528 8992 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service I0423 18:00:09.656107 8992 ssh_runner.go:195] Run: sudo systemctl daemon-reload I0423 18:00:09.945634 8992 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.service I0423 18:00:10.194781 8992 start.go:541] Will wait 60s for socket path /var/run/cri-dockerd.sock I0423 18:00:10.201632 8992 ssh_runner.go:195] Run: stat /var/run/cri-dockerd.sock I0423 18:00:10.221353 8992 start.go:562] Will wait 60s for crictl version I0423 18:00:10.228632 8992 ssh_runner.go:195] Run: which crictl I0423 18:00:10.245742 8992 ssh_runner.go:195] Run: sudo /usr/bin/crictl version I0423 18:00:10.312075 8992 start.go:578] Version: 0.1.0 RuntimeName: docker RuntimeVersion: 26.0.1 RuntimeApiVersion: v1 I0423 18:00:10.317098 8992 ssh_runner.go:195] Run: docker version --format {{.Server.Version}} I0423 18:00:10.374138 8992 ssh_runner.go:195] Run: docker version --format {{.Server.Version}} I0423 18:00:10.450151 8992 out.go:204] * Preparing Kubernetes v1.30.0 on Docker 26.0.1 ... I0423 18:00:10.713495 8992 ssh_runner.go:195] Run: grep 192.168.59.1 host.minikube.internal$ /etc/hosts I0423 18:00:10.725184 8992 kubeadm.go:877] updating cluster {Name:minikube KeepContext:false EmbedCerts:false MinikubeISO:https://storage.googleapis.com/minikube/iso/minikube-v1.33.0-amd64.iso KicBaseImage:gcr.io/k8s-minikube/kicbase:v0.0.43@sha256:7ff490df401cc0fbf19a4521544ae8f4a00cc163e92a95017a8d8bfdb1422737 Memory:4000 CPUs:2 DiskSize:20000 Driver:virtualbox HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:true DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.30.0 ClusterName:minikube Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.59.101 Port:8443 KubernetesVersion:v1.30.0 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop: ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:C:\Users\Admin:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} ... I0423 18:00:10.725184 8992 preload.go:132] Checking if preload exists for k8s version v1.30.0 and runtime docker I0423 18:00:10.730027 8992 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}} I0423 18:00:10.777583 8992 docker.go:685] Got preloaded images: -- stdout -- registry.k8s.io/kube-apiserver:v1.30.0 registry.k8s.io/kube-controller-manager:v1.30.0 registry.k8s.io/kube-scheduler:v1.30.0 registry.k8s.io/kube-proxy:v1.30.0 registry.k8s.io/etcd:3.5.12-0 registry.k8s.io/coredns/coredns:v1.11.1 registry.k8s.io/pause:3.9 gcr.io/k8s-minikube/storage-provisioner:v5 -- /stdout -- I0423 18:00:10.777583 8992 docker.go:615] Images already preloaded, skipping extraction I0423 18:00:10.782586 8992 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}} I0423 18:00:10.928835 8992 docker.go:685] Got preloaded images: -- stdout -- registry.k8s.io/kube-apiserver:v1.30.0 registry.k8s.io/kube-controller-manager:v1.30.0 registry.k8s.io/kube-scheduler:v1.30.0 registry.k8s.io/kube-proxy:v1.30.0 registry.k8s.io/etcd:3.5.12-0 registry.k8s.io/coredns/coredns:v1.11.1 registry.k8s.io/pause:3.9 gcr.io/k8s-minikube/storage-provisioner:v5 -- /stdout -- I0423 18:00:10.928835 8992 cache_images.go:84] Images are preloaded, skipping loading I0423 18:00:10.928835 8992 kubeadm.go:928] updating node { 192.168.59.101 8443 v1.30.0 docker true true} ... I0423 18:00:10.928835 8992 kubeadm.go:940] kubelet [Unit] Wants=docker.socket [Service] ExecStart= ExecStart=/var/lib/minikube/binaries/v1.30.0/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --config=/var/lib/kubelet/config.yaml --hostname-override=minikube --kubeconfig=/etc/kubernetes/kubelet.conf --node-ip=192.168.59.101 [Install] config: {KubernetesVersion:v1.30.0 ClusterName:minikube Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} I0423 18:00:10.933776 8992 ssh_runner.go:195] Run: docker info --format {{.CgroupDriver}} I0423 18:00:11.389150 8992 cni.go:84] Creating CNI manager for "" I0423 18:00:11.389150 8992 cni.go:158] "virtualbox" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge I0423 18:00:11.389150 8992 kubeadm.go:84] Using pod CIDR: 10.244.0.0/16 I0423 18:00:11.389150 8992 kubeadm.go:181] kubeadm options: {CertDir:/var/lib/minikube/certs ServiceCIDR:10.96.0.0/12 PodSubnet:10.244.0.0/16 AdvertiseAddress:192.168.59.101 APIServerPort:8443 KubernetesVersion:v1.30.0 EtcdDataDir:/var/lib/minikube/etcd EtcdExtraArgs:map[] ClusterName:minikube NodeName:minikube DNSDomain:cluster.local CRISocket:/var/run/cri-dockerd.sock ImageRepository: ComponentOptions:[{Component:apiServer ExtraArgs:map[enable-admission-plugins:NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota] Pairs:map[certSANs:["127.0.0.1", "localhost", "192.168.59.101"]]} {Component:controllerManager ExtraArgs:map[allocate-node-cidrs:true leader-elect:false] Pairs:map[]} {Component:scheduler ExtraArgs:map[leader-elect:false] Pairs:map[]}] FeatureArgs:map[] NodeIP:192.168.59.101 CgroupDriver:cgroupfs ClientCAFile:/var/lib/minikube/certs/ca.crt StaticPodPath:/etc/kubernetes/manifests ControlPlaneAddress:control-plane.minikube.internal KubeProxyOptions:map[] ResolvConfSearchRegression:false KubeletConfigOpts:map[containerRuntimeEndpoint:unix:///var/run/cri-dockerd.sock hairpinMode:hairpin-veth runtimeRequestTimeout:15m] PrependCriSocketUnix:true} I0423 18:00:11.389150 8992 kubeadm.go:187] kubeadm config: apiVersion: kubeadm.k8s.io/v1beta3 kind: InitConfiguration localAPIEndpoint: advertiseAddress: 192.168.59.101 bindPort: 8443 bootstrapTokens: - groups: - system:bootstrappers:kubeadm:default-node-token ttl: 24h0m0s usages: - signing - authentication nodeRegistration: criSocket: unix:///var/run/cri-dockerd.sock name: "minikube" kubeletExtraArgs: node-ip: 192.168.59.101 taints: [] --- apiVersion: kubeadm.k8s.io/v1beta3 kind: ClusterConfiguration apiServer: certSANs: ["127.0.0.1", "localhost", "192.168.59.101"] extraArgs: enable-admission-plugins: "NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota" controllerManager: extraArgs: allocate-node-cidrs: "true" leader-elect: "false" scheduler: extraArgs: leader-elect: "false" certificatesDir: /var/lib/minikube/certs clusterName: mk controlPlaneEndpoint: control-plane.minikube.internal:8443 etcd: local: dataDir: /var/lib/minikube/etcd extraArgs: proxy-refresh-interval: "70000" kubernetesVersion: v1.30.0 networking: dnsDomain: cluster.local podSubnet: "10.244.0.0/16" serviceSubnet: 10.96.0.0/12 --- apiVersion: kubelet.config.k8s.io/v1beta1 kind: KubeletConfiguration authentication: x509: clientCAFile: /var/lib/minikube/certs/ca.crt cgroupDriver: cgroupfs containerRuntimeEndpoint: unix:///var/run/cri-dockerd.sock hairpinMode: hairpin-veth runtimeRequestTimeout: 15m clusterDomain: "cluster.local" # disable disk resource management by default imageGCHighThresholdPercent: 100 evictionHard: nodefs.available: "0%!"(MISSING) nodefs.inodesFree: "0%!"(MISSING) imagefs.available: "0%!"(MISSING) failSwapOn: false staticPodPath: /etc/kubernetes/manifests --- apiVersion: kubeproxy.config.k8s.io/v1alpha1 kind: KubeProxyConfiguration clusterCIDR: "10.244.0.0/16" metricsBindAddress: 0.0.0.0:10249 conntrack: maxPerCore: 0 # Skip setting "net.netfilter.nf_conntrack_tcp_timeout_established" tcpEstablishedTimeout: 0s # Skip setting "net.netfilter.nf_conntrack_tcp_timeout_close" tcpCloseWaitTimeout: 0s I0423 18:00:11.396905 8992 ssh_runner.go:195] Run: sudo ls /var/lib/minikube/binaries/v1.30.0 I0423 18:00:11.459079 8992 binaries.go:44] Found k8s binaries, skipping transfer I0423 18:00:11.468952 8992 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/kubelet.service.d /lib/systemd/system /var/tmp/minikube I0423 18:00:11.526493 8992 ssh_runner.go:362] scp memory --> /etc/systemd/system/kubelet.service.d/10-kubeadm.conf (309 bytes) I0423 18:00:11.637280 8992 ssh_runner.go:362] scp memory --> /lib/systemd/system/kubelet.service (352 bytes) I0423 18:00:11.726881 8992 ssh_runner.go:362] scp memory --> /var/tmp/minikube/kubeadm.yaml.new (2156 bytes) I0423 18:00:11.856099 8992 ssh_runner.go:195] Run: grep 192.168.59.101 control-plane.minikube.internal$ /etc/hosts I0423 18:00:11.878565 8992 ssh_runner.go:195] Run: sudo systemctl daemon-reload I0423 18:00:12.698033 8992 ssh_runner.go:195] Run: sudo systemctl start kubelet I0423 18:00:12.788050 8992 certs.go:68] Setting up C:\Users\Admin\.minikube\profiles\minikube for IP: 192.168.59.101 I0423 18:00:12.788050 8992 certs.go:194] generating shared ca certs ... I0423 18:00:12.788050 8992 certs.go:226] acquiring lock for ca certs: {Name:mkf28197c9a65e99074f3ba820d4a88215c788d1 Clock:{} Delay:500ms Timeout:1m0s Cancel:} I0423 18:00:12.788557 8992 certs.go:235] skipping valid "minikubeCA" ca cert: C:\Users\Admin\.minikube\ca.key I0423 18:00:12.788557 8992 certs.go:235] skipping valid "proxyClientCA" ca cert: C:\Users\Admin\.minikube\proxy-client-ca.key I0423 18:00:12.788557 8992 certs.go:256] generating profile certs ... I0423 18:00:12.789093 8992 certs.go:359] skipping valid signed profile cert regeneration for "minikube-user": C:\Users\Admin\.minikube\profiles\minikube\client.key I0423 18:00:12.789093 8992 certs.go:359] skipping valid signed profile cert regeneration for "minikube": C:\Users\Admin\.minikube\profiles\minikube\apiserver.key.88b027f6 I0423 18:00:12.789093 8992 certs.go:359] skipping valid signed profile cert regeneration for "aggregator": C:\Users\Admin\.minikube\profiles\minikube\proxy-client.key I0423 18:00:12.789615 8992 certs.go:484] found cert: C:\Users\Admin\.minikube\certs\ca-key.pem (1675 bytes) I0423 18:00:12.790157 8992 certs.go:484] found cert: C:\Users\Admin\.minikube\certs\ca.pem (1074 bytes) I0423 18:00:12.790157 8992 certs.go:484] found cert: C:\Users\Admin\.minikube\certs\cert.pem (1119 bytes) I0423 18:00:12.790157 8992 certs.go:484] found cert: C:\Users\Admin\.minikube\certs\key.pem (1675 bytes) I0423 18:00:12.791207 8992 ssh_runner.go:362] scp C:\Users\Admin\.minikube\ca.crt --> /var/lib/minikube/certs/ca.crt (1111 bytes) I0423 18:00:13.017431 8992 ssh_runner.go:362] scp C:\Users\Admin\.minikube\ca.key --> /var/lib/minikube/certs/ca.key (1675 bytes) I0423 18:00:13.094629 8992 ssh_runner.go:362] scp C:\Users\Admin\.minikube\proxy-client-ca.crt --> /var/lib/minikube/certs/proxy-client-ca.crt (1119 bytes) I0423 18:00:13.361772 8992 ssh_runner.go:362] scp C:\Users\Admin\.minikube\proxy-client-ca.key --> /var/lib/minikube/certs/proxy-client-ca.key (1675 bytes) I0423 18:00:13.533529 8992 ssh_runner.go:362] scp C:\Users\Admin\.minikube\profiles\minikube\apiserver.crt --> /var/lib/minikube/certs/apiserver.crt (1411 bytes) I0423 18:00:13.770294 8992 ssh_runner.go:362] scp C:\Users\Admin\.minikube\profiles\minikube\apiserver.key --> /var/lib/minikube/certs/apiserver.key (1675 bytes) I0423 18:00:13.889263 8992 ssh_runner.go:362] scp C:\Users\Admin\.minikube\profiles\minikube\proxy-client.crt --> /var/lib/minikube/certs/proxy-client.crt (1147 bytes) I0423 18:00:13.981263 8992 ssh_runner.go:362] scp C:\Users\Admin\.minikube\profiles\minikube\proxy-client.key --> /var/lib/minikube/certs/proxy-client.key (1675 bytes) I0423 18:00:14.064670 8992 ssh_runner.go:362] scp C:\Users\Admin\.minikube\ca.crt --> /usr/share/ca-certificates/minikubeCA.pem (1111 bytes) I0423 18:00:14.228296 8992 ssh_runner.go:362] scp memory --> /var/lib/minikube/kubeconfig (738 bytes) I0423 18:00:14.331205 8992 ssh_runner.go:195] Run: openssl version I0423 18:00:14.349912 8992 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/minikubeCA.pem && ln -fs /usr/share/ca-certificates/minikubeCA.pem /etc/ssl/certs/minikubeCA.pem" I0423 18:00:14.429477 8992 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/minikubeCA.pem I0423 18:00:14.441671 8992 certs.go:528] hashing: -rw-r--r-- 1 root root 1111 Apr 22 15:58 /usr/share/ca-certificates/minikubeCA.pem I0423 18:00:14.448307 8992 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/minikubeCA.pem I0423 18:00:14.488427 8992 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/b5213941.0 || ln -fs /etc/ssl/certs/minikubeCA.pem /etc/ssl/certs/b5213941.0" I0423 18:00:14.531095 8992 ssh_runner.go:195] Run: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt I0423 18:00:14.554325 8992 ssh_runner.go:195] Run: openssl x509 -noout -in /var/lib/minikube/certs/apiserver-etcd-client.crt -checkend 86400 I0423 18:00:14.584540 8992 ssh_runner.go:195] Run: openssl x509 -noout -in /var/lib/minikube/certs/apiserver-kubelet-client.crt -checkend 86400 I0423 18:00:14.612587 8992 ssh_runner.go:195] Run: openssl x509 -noout -in /var/lib/minikube/certs/etcd/server.crt -checkend 86400 I0423 18:00:14.637751 8992 ssh_runner.go:195] Run: openssl x509 -noout -in /var/lib/minikube/certs/etcd/healthcheck-client.crt -checkend 86400 I0423 18:00:14.660499 8992 ssh_runner.go:195] Run: openssl x509 -noout -in /var/lib/minikube/certs/etcd/peer.crt -checkend 86400 I0423 18:00:14.693495 8992 ssh_runner.go:195] Run: openssl x509 -noout -in /var/lib/minikube/certs/front-proxy-client.crt -checkend 86400 I0423 18:00:14.744467 8992 kubeadm.go:391] StartCluster: {Name:minikube KeepContext:false EmbedCerts:false MinikubeISO:https://storage.googleapis.com/minikube/iso/minikube-v1.33.0-amd64.iso KicBaseImage:gcr.io/k8s-minikube/kicbase:v0.0.43@sha256:7ff490df401cc0fbf19a4521544ae8f4a00cc163e92a95017a8d8bfdb1422737 Memory:4000 CPUs:2 DiskSize:20000 Driver:virtualbox HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:true DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.30.0 ClusterName:minikube Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.59.101 Port:8443 KubernetesVersion:v1.30.0 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop: ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:C:\Users\Admin:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} I0423 18:00:14.749844 8992 ssh_runner.go:195] Run: docker ps --filter status=paused --filter=name=k8s_.*_(kube-system)_ --format={{.ID}} I0423 18:00:14.813570 8992 ssh_runner.go:195] Run: sudo ls /var/lib/kubelet/kubeadm-flags.env /var/lib/kubelet/config.yaml /var/lib/minikube/etcd W0423 18:00:14.850561 8992 kubeadm.go:404] apiserver tunnel failed: apiserver port not set I0423 18:00:14.850561 8992 kubeadm.go:407] found existing configuration files, will attempt cluster restart I0423 18:00:14.863302 8992 kubeadm.go:587] restartPrimaryControlPlane start ... I0423 18:00:14.869629 8992 ssh_runner.go:195] Run: sudo test -d /data/minikube I0423 18:00:15.025935 8992 kubeadm.go:129] /data/minikube skipping compat symlinks: sudo test -d /data/minikube: Process exited with status 1 stdout: stderr: I0423 18:00:15.027042 8992 kubeconfig.go:47] verify endpoint returned: get endpoint: read kubeconfig: decode kubeconfig from "C:\\Users\\Admin\\.kube\\config": decode data: : yaml: control characters are not allowed W0423 18:00:15.027561 8992 kubeadm.go:610] unable to update kubeconfig (cluster will likely require a reset): get kubeconfig: decode kubeconfig from "C:\\Users\\Admin\\.kube\\config": decode data: : yaml: control characters are not allowed I0423 18:00:15.029393 8992 kubeadm.go:591] duration metric: took 166.0916ms to restartPrimaryControlPlane W0423 18:00:15.029393 8992 out.go:239] ! Unable to restart control-plane node(s), will reset cluster: I0423 18:00:15.029899 8992 ssh_runner.go:195] Run: /bin/bash -c "sudo env PATH="/var/lib/minikube/binaries/v1.30.0:$PATH" kubeadm reset --cri-socket /var/run/cri-dockerd.sock --force" I0423 18:00:37.752993 8992 ssh_runner.go:235] Completed: /bin/bash -c "sudo env PATH="/var/lib/minikube/binaries/v1.30.0:$PATH" kubeadm reset --cri-socket /var/run/cri-dockerd.sock --force": (22.7225698s) I0423 18:00:37.772238 8992 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service kubelet I0423 18:00:37.811474 8992 ssh_runner.go:195] Run: sudo cp /var/tmp/minikube/kubeadm.yaml.new /var/tmp/minikube/kubeadm.yaml I0423 18:00:37.835527 8992 ssh_runner.go:195] Run: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf I0423 18:00:37.857714 8992 kubeadm.go:154] config check failed, skipping stale config cleanup: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf: Process exited with status 2 stdout: stderr: ls: cannot access '/etc/kubernetes/admin.conf': No such file or directory ls: cannot access '/etc/kubernetes/kubelet.conf': No such file or directory ls: cannot access '/etc/kubernetes/controller-manager.conf': No such file or directory ls: cannot access '/etc/kubernetes/scheduler.conf': No such file or directory I0423 18:00:37.857714 8992 kubeadm.go:156] found existing configuration files: I0423 18:00:37.864122 8992 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf I0423 18:00:37.880869 8992 kubeadm.go:162] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/admin.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf: Process exited with status 2 stdout: stderr: grep: /etc/kubernetes/admin.conf: No such file or directory I0423 18:00:37.887920 8992 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/admin.conf I0423 18:00:37.913270 8992 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf I0423 18:00:37.932643 8992 kubeadm.go:162] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/kubelet.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf: Process exited with status 2 stdout: stderr: grep: /etc/kubernetes/kubelet.conf: No such file or directory I0423 18:00:37.938895 8992 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/kubelet.conf I0423 18:00:37.964034 8992 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf I0423 18:00:37.990244 8992 kubeadm.go:162] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/controller-manager.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf: Process exited with status 2 stdout: stderr: grep: /etc/kubernetes/controller-manager.conf: No such file or directory I0423 18:00:37.997202 8992 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/controller-manager.conf I0423 18:00:38.031087 8992 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf I0423 18:00:38.056774 8992 kubeadm.go:162] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/scheduler.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf: Process exited with status 2 stdout: stderr: grep: /etc/kubernetes/scheduler.conf: No such file or directory I0423 18:00:38.064183 8992 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/scheduler.conf I0423 18:00:38.089248 8992 ssh_runner.go:286] Start: /bin/bash -c "sudo env PATH="/var/lib/minikube/binaries/v1.30.0:$PATH" kubeadm init --config /var/tmp/minikube/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests,DirAvailable--var-lib-minikube,DirAvailable--var-lib-minikube-etcd,FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml,FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml,FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml,FileAvailable--etc-kubernetes-manifests-etcd.yaml,Port-10250,Swap,NumCPU,Mem" I0423 18:00:38.500153 8992 kubeadm.go:309] [WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service' I0423 18:00:53.213290 8992 kubeadm.go:309] [init] Using Kubernetes version: v1.30.0 I0423 18:00:53.213290 8992 kubeadm.go:309] [preflight] Running pre-flight checks I0423 18:00:53.213290 8992 kubeadm.go:309] [preflight] Pulling images required for setting up a Kubernetes cluster I0423 18:00:53.213290 8992 kubeadm.go:309] [preflight] This might take a minute or two, depending on the speed of your internet connection I0423 18:00:53.213290 8992 kubeadm.go:309] [preflight] You can also perform this action in beforehand using 'kubeadm config images pull' I0423 18:00:53.213795 8992 kubeadm.go:309] [certs] Using certificateDir folder "/var/lib/minikube/certs" I0423 18:00:53.216473 8992 out.go:204] - Generating certificates and keys ... I0423 18:00:53.217001 8992 kubeadm.go:309] [certs] Using existing ca certificate authority I0423 18:00:53.217001 8992 kubeadm.go:309] [certs] Using existing apiserver certificate and key on disk I0423 18:00:53.217001 8992 kubeadm.go:309] [certs] Using existing apiserver-kubelet-client certificate and key on disk I0423 18:00:53.217001 8992 kubeadm.go:309] [certs] Using existing front-proxy-ca certificate authority I0423 18:00:53.217001 8992 kubeadm.go:309] [certs] Using existing front-proxy-client certificate and key on disk I0423 18:00:53.217695 8992 kubeadm.go:309] [certs] Using existing etcd/ca certificate authority I0423 18:00:53.217695 8992 kubeadm.go:309] [certs] Using existing etcd/server certificate and key on disk I0423 18:00:53.217695 8992 kubeadm.go:309] [certs] Using existing etcd/peer certificate and key on disk I0423 18:00:53.217695 8992 kubeadm.go:309] [certs] Using existing etcd/healthcheck-client certificate and key on disk I0423 18:00:53.217695 8992 kubeadm.go:309] [certs] Using existing apiserver-etcd-client certificate and key on disk I0423 18:00:53.217695 8992 kubeadm.go:309] [certs] Using the existing "sa" key I0423 18:00:53.217695 8992 kubeadm.go:309] [kubeconfig] Using kubeconfig folder "/etc/kubernetes" I0423 18:00:53.218220 8992 kubeadm.go:309] [kubeconfig] Writing "admin.conf" kubeconfig file I0423 18:00:53.218220 8992 kubeadm.go:309] [kubeconfig] Writing "super-admin.conf" kubeconfig file I0423 18:00:53.218220 8992 kubeadm.go:309] [kubeconfig] Writing "kubelet.conf" kubeconfig file I0423 18:00:53.218220 8992 kubeadm.go:309] [kubeconfig] Writing "controller-manager.conf" kubeconfig file I0423 18:00:53.218220 8992 kubeadm.go:309] [kubeconfig] Writing "scheduler.conf" kubeconfig file I0423 18:00:53.218220 8992 kubeadm.go:309] [etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests" I0423 18:00:53.218220 8992 kubeadm.go:309] [control-plane] Using manifest folder "/etc/kubernetes/manifests" I0423 18:00:53.233305 8992 out.go:204] - Booting up control plane ... I0423 18:00:53.234113 8992 kubeadm.go:309] [control-plane] Creating static Pod manifest for "kube-apiserver" I0423 18:00:53.234818 8992 kubeadm.go:309] [control-plane] Creating static Pod manifest for "kube-controller-manager" I0423 18:00:53.234818 8992 kubeadm.go:309] [control-plane] Creating static Pod manifest for "kube-scheduler" I0423 18:00:53.234818 8992 kubeadm.go:309] [kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env" I0423 18:00:53.234818 8992 kubeadm.go:309] [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml" I0423 18:00:53.234818 8992 kubeadm.go:309] [kubelet-start] Starting the kubelet I0423 18:00:53.234818 8992 kubeadm.go:309] [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests" I0423 18:00:53.235353 8992 kubeadm.go:309] [kubelet-check] Waiting for a healthy kubelet. This can take up to 4m0s I0423 18:00:53.235353 8992 kubeadm.go:309] [kubelet-check] The kubelet is healthy after 1.029407804s I0423 18:00:53.235353 8992 kubeadm.go:309] [api-check] Waiting for a healthy API server. This can take up to 4m0s I0423 18:00:53.235353 8992 kubeadm.go:309] [api-check] The API server is healthy after 10.505584555s I0423 18:00:53.235353 8992 kubeadm.go:309] [upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace I0423 18:00:53.235882 8992 kubeadm.go:309] [kubelet] Creating a ConfigMap "kubelet-config" in namespace kube-system with the configuration for the kubelets in the cluster I0423 18:00:53.235882 8992 kubeadm.go:309] [upload-certs] Skipping phase. Please see --upload-certs I0423 18:00:53.235882 8992 kubeadm.go:309] [mark-control-plane] Marking the node minikube as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers] I0423 18:00:53.235882 8992 kubeadm.go:309] [bootstrap-token] Using token: tgawmp.q02cspkajc0ae1zm I0423 18:00:53.238525 8992 out.go:204] - Configuring RBAC rules ... I0423 18:00:53.239044 8992 kubeadm.go:309] [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles I0423 18:00:53.239577 8992 kubeadm.go:309] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes I0423 18:00:53.239577 8992 kubeadm.go:309] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials I0423 18:00:53.239577 8992 kubeadm.go:309] [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token I0423 18:00:53.240095 8992 kubeadm.go:309] [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster I0423 18:00:53.240095 8992 kubeadm.go:309] [bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace I0423 18:00:53.240095 8992 kubeadm.go:309] [kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key I0423 18:00:53.240095 8992 kubeadm.go:309] [addons] Applied essential addon: CoreDNS I0423 18:00:53.240095 8992 kubeadm.go:309] [addons] Applied essential addon: kube-proxy I0423 18:00:53.240095 8992 kubeadm.go:309] I0423 18:00:53.240095 8992 kubeadm.go:309] Your Kubernetes control-plane has initialized successfully! I0423 18:00:53.240095 8992 kubeadm.go:309] I0423 18:00:53.240622 8992 kubeadm.go:309] To start using your cluster, you need to run the following as a regular user: I0423 18:00:53.240622 8992 kubeadm.go:309] I0423 18:00:53.240622 8992 kubeadm.go:309] mkdir -p $HOME/.kube I0423 18:00:53.240622 8992 kubeadm.go:309] sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config I0423 18:00:53.240622 8992 kubeadm.go:309] sudo chown $(id -u):$(id -g) $HOME/.kube/config I0423 18:00:53.240622 8992 kubeadm.go:309] I0423 18:00:53.240622 8992 kubeadm.go:309] Alternatively, if you are the root user, you can run: I0423 18:00:53.240622 8992 kubeadm.go:309] I0423 18:00:53.240622 8992 kubeadm.go:309] export KUBECONFIG=/etc/kubernetes/admin.conf I0423 18:00:53.240622 8992 kubeadm.go:309] I0423 18:00:53.240622 8992 kubeadm.go:309] You should now deploy a pod network to the cluster. I0423 18:00:53.241144 8992 kubeadm.go:309] Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at: I0423 18:00:53.241144 8992 kubeadm.go:309] https://kubernetes.io/docs/concepts/cluster-administration/addons/ I0423 18:00:53.241144 8992 kubeadm.go:309] I0423 18:00:53.241144 8992 kubeadm.go:309] You can now join any number of control-plane nodes by copying certificate authorities I0423 18:00:53.241144 8992 kubeadm.go:309] and service account keys on each node and then running the following as root: I0423 18:00:53.241144 8992 kubeadm.go:309] I0423 18:00:53.241144 8992 kubeadm.go:309] kubeadm join control-plane.minikube.internal:8443 --token tgawmp.q02cspkajc0ae1zm \ I0423 18:00:53.241666 8992 kubeadm.go:309] --discovery-token-ca-cert-hash sha256:14763fe1585f37c76fb4d66e35947f73abf958764776cee50fee3d964200e2d6 \ I0423 18:00:53.241666 8992 kubeadm.go:309] --control-plane I0423 18:00:53.241666 8992 kubeadm.go:309] I0423 18:00:53.241666 8992 kubeadm.go:309] Then you can join any number of worker nodes by running the following on each as root: I0423 18:00:53.241666 8992 kubeadm.go:309] I0423 18:00:53.241666 8992 kubeadm.go:309] kubeadm join control-plane.minikube.internal:8443 --token tgawmp.q02cspkajc0ae1zm \ I0423 18:00:53.241666 8992 kubeadm.go:309] --discovery-token-ca-cert-hash sha256:14763fe1585f37c76fb4d66e35947f73abf958764776cee50fee3d964200e2d6 I0423 18:00:53.241666 8992 cni.go:84] Creating CNI manager for "" I0423 18:00:53.241666 8992 cni.go:158] "virtualbox" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge I0423 18:00:53.244297 8992 out.go:177] * Configuring bridge CNI (Container Networking Interface) ... I0423 18:00:53.256174 8992 ssh_runner.go:195] Run: sudo mkdir -p /etc/cni/net.d I0423 18:00:53.378105 8992 ssh_runner.go:362] scp memory --> /etc/cni/net.d/1-k8s.conflist (496 bytes) I0423 18:00:53.507229 8992 ssh_runner.go:195] Run: /bin/bash -c "cat /proc/$(pgrep kube-apiserver)/oom_adj" I0423 18:00:53.517085 8992 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig label --overwrite nodes minikube minikube.k8s.io/updated_at=2024_04_23T18_00_53_0700 minikube.k8s.io/version=v1.33.0 minikube.k8s.io/commit=86fc9d54fca63f295d8737c8eacdbb7987e89c67 minikube.k8s.io/name=minikube minikube.k8s.io/primary=true I0423 18:00:53.518172 8992 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.0/kubectl create clusterrolebinding minikube-rbac --clusterrole=cluster-admin --serviceaccount=kube-system:default --kubeconfig=/var/lib/minikube/kubeconfig I0423 18:00:53.618469 8992 ops.go:34] apiserver oom_adj: -16 I0423 18:00:53.938890 8992 kubeadm.go:1107] duration metric: took 431.6605ms to wait for elevateKubeSystemPrivileges W0423 18:00:53.938890 8992 kubeadm.go:286] apiserver tunnel failed: apiserver port not set I0423 18:00:53.938890 8992 kubeadm.go:393] duration metric: took 39.1944226s to StartCluster I0423 18:00:53.938890 8992 settings.go:142] acquiring lock: {Name:mk98c2b42cbe328c01bb9ae373473b3510453e13 Clock:{} Delay:500ms Timeout:1m0s Cancel:} I0423 18:00:53.938890 8992 settings.go:150] Updating kubeconfig: C:\Users\Admin\.kube\config I0423 18:00:53.942678 8992 out.go:177] W0423 18:00:53.945494 8992 out.go:239] X Exiting due to GUEST_START: failed to start node: Failed kubeconfig update: decode kubeconfig from "C:\\Users\\Admin\\.kube\\config": decode data: : yaml: control characters are not allowed W0423 18:00:53.945494 8992 out.go:239] * W0423 18:00:53.950966 8992 out.go:239] ╭─────────────────────────────────────────────────────────────────────────────────────────────╮ │ │ │ * If the above advice does not help, please let us know: │ │ https://github.com/kubernetes/minikube/issues/new/choose │ │ │ │ * Please run `minikube logs --file=logs.txt` and attach logs.txt to the GitHub issue. │ │ │ ╰─────────────────────────────────────────────────────────────────────────────────────────────╯ I0423 18:00:53.955162 8992 out.go:177] ==> Docker <== Apr 23 12:30:41 minikube dockerd[4239]: time="2024-04-23T12:30:41.919512486Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 23 12:30:41 minikube dockerd[4239]: time="2024-04-23T12:30:41.947047947Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 23 12:30:41 minikube dockerd[4239]: time="2024-04-23T12:30:41.947127880Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 23 12:30:41 minikube dockerd[4239]: time="2024-04-23T12:30:41.947149071Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 23 12:30:41 minikube dockerd[4239]: time="2024-04-23T12:30:41.947526490Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 23 12:30:42 minikube cri-dockerd[4547]: time="2024-04-23T12:30:42Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/e28e5c40f426dee3243db0de6e63edc90175875504174d9c2420ad7497e64a0a/resolv.conf as [nameserver 10.0.2.3]" Apr 23 12:30:42 minikube cri-dockerd[4547]: time="2024-04-23T12:30:42Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/c905f8147990a8d040e97863a972b207a454f9078a28cf5b302afc8900573aec/resolv.conf as [nameserver 10.0.2.3]" Apr 23 12:30:42 minikube cri-dockerd[4547]: time="2024-04-23T12:30:42Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/3b34159df3e24755860c568f15e8a533562d3647b866e0411116878e78c49fe1/resolv.conf as [nameserver 10.0.2.3]" Apr 23 12:30:42 minikube cri-dockerd[4547]: time="2024-04-23T12:30:42Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/4841ef16bce8222ab9f627d9eaf7110d474b31822c21ab33b4b04e09b103913d/resolv.conf as [nameserver 10.0.2.3]" Apr 23 12:30:42 minikube dockerd[4239]: time="2024-04-23T12:30:42.703513849Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 23 12:30:42 minikube dockerd[4239]: time="2024-04-23T12:30:42.703830885Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 23 12:30:42 minikube dockerd[4239]: time="2024-04-23T12:30:42.703849813Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 23 12:30:42 minikube dockerd[4239]: time="2024-04-23T12:30:42.704177321Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 23 12:30:42 minikube dockerd[4239]: time="2024-04-23T12:30:42.717216314Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 23 12:30:42 minikube dockerd[4239]: time="2024-04-23T12:30:42.718444566Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 23 12:30:42 minikube dockerd[4239]: time="2024-04-23T12:30:42.719262251Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 23 12:30:42 minikube dockerd[4239]: time="2024-04-23T12:30:42.720770920Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 23 12:30:42 minikube dockerd[4239]: time="2024-04-23T12:30:42.746923595Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 23 12:30:42 minikube dockerd[4239]: time="2024-04-23T12:30:42.748014169Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 23 12:30:42 minikube dockerd[4239]: time="2024-04-23T12:30:42.750284722Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 23 12:30:42 minikube dockerd[4239]: time="2024-04-23T12:30:42.750392853Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 23 12:30:42 minikube dockerd[4239]: time="2024-04-23T12:30:42.755635512Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 23 12:30:42 minikube dockerd[4239]: time="2024-04-23T12:30:42.756541020Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 23 12:30:42 minikube dockerd[4239]: time="2024-04-23T12:30:42.756808899Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 23 12:30:42 minikube dockerd[4239]: time="2024-04-23T12:30:42.770925383Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 23 12:31:07 minikube dockerd[4239]: time="2024-04-23T12:31:07.412703753Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 23 12:31:07 minikube dockerd[4239]: time="2024-04-23T12:31:07.412875307Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 23 12:31:07 minikube dockerd[4239]: time="2024-04-23T12:31:07.412900739Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 23 12:31:07 minikube dockerd[4239]: time="2024-04-23T12:31:07.412990402Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 23 12:31:07 minikube dockerd[4239]: time="2024-04-23T12:31:07.758579395Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 23 12:31:07 minikube dockerd[4239]: time="2024-04-23T12:31:07.760983721Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 23 12:31:07 minikube dockerd[4239]: time="2024-04-23T12:31:07.764188172Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 23 12:31:07 minikube dockerd[4239]: time="2024-04-23T12:31:07.766241793Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 23 12:31:07 minikube cri-dockerd[4547]: time="2024-04-23T12:31:07Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/59fb619499875a442a0a01554e0e668570a0960d887fd3ed48476534e6828c2a/resolv.conf as [nameserver 10.0.2.3]" Apr 23 12:31:08 minikube cri-dockerd[4547]: time="2024-04-23T12:31:08Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/64c37cca31f058fcec9a30d9358a7f0d32f7b4e35111b96605369b21bfd1a996/resolv.conf as [nameserver 10.0.2.3]" Apr 23 12:31:08 minikube dockerd[4239]: time="2024-04-23T12:31:08.478608360Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 23 12:31:08 minikube dockerd[4239]: time="2024-04-23T12:31:08.478738839Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 23 12:31:08 minikube dockerd[4239]: time="2024-04-23T12:31:08.479603070Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 23 12:31:08 minikube dockerd[4239]: time="2024-04-23T12:31:08.479916804Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 23 12:31:08 minikube dockerd[4239]: time="2024-04-23T12:31:08.960741468Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 23 12:31:08 minikube dockerd[4239]: time="2024-04-23T12:31:08.960826445Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 23 12:31:08 minikube dockerd[4239]: time="2024-04-23T12:31:08.960841180Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 23 12:31:08 minikube dockerd[4239]: time="2024-04-23T12:31:08.960918942Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 23 12:31:08 minikube dockerd[4239]: time="2024-04-23T12:31:08.979757212Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 23 12:31:08 minikube dockerd[4239]: time="2024-04-23T12:31:08.979840959Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 23 12:31:08 minikube dockerd[4239]: time="2024-04-23T12:31:08.982615541Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 23 12:31:08 minikube dockerd[4239]: time="2024-04-23T12:31:08.985033332Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 23 12:31:09 minikube cri-dockerd[4547]: time="2024-04-23T12:31:09Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/2f28330b2bc431f808a2fda7a4da12065569f482ce66b5bb16c9b2bceafa0a72/resolv.conf as [nameserver 10.0.2.3]" Apr 23 12:31:09 minikube dockerd[4239]: time="2024-04-23T12:31:09.580232362Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 23 12:31:09 minikube dockerd[4239]: time="2024-04-23T12:31:09.580763425Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 23 12:31:09 minikube dockerd[4239]: time="2024-04-23T12:31:09.580804638Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 23 12:31:09 minikube dockerd[4239]: time="2024-04-23T12:31:09.582999185Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 23 12:31:14 minikube cri-dockerd[4547]: time="2024-04-23T12:31:14Z" level=info msg="Docker cri received runtime config &RuntimeConfig{NetworkConfig:&NetworkConfig{PodCidr:10.244.0.0/24,},}" Apr 23 12:32:50 minikube dockerd[4233]: 2024/04/23 12:32:50 http: superfluous response.WriteHeader call from go.opentelemetry.io/contrib/instrumentation/net/http/otelhttp.(*respWriterWrapper).WriteHeader (wrap.go:98) Apr 23 12:32:50 minikube dockerd[4233]: 2024/04/23 12:32:50 http: superfluous response.WriteHeader call from go.opentelemetry.io/contrib/instrumentation/net/http/otelhttp.(*respWriterWrapper).WriteHeader (wrap.go:98) Apr 23 12:32:51 minikube dockerd[4233]: 2024/04/23 12:32:51 http: superfluous response.WriteHeader call from go.opentelemetry.io/contrib/instrumentation/net/http/otelhttp.(*respWriterWrapper).WriteHeader (wrap.go:98) Apr 23 12:32:51 minikube dockerd[4233]: 2024/04/23 12:32:51 http: superfluous response.WriteHeader call from go.opentelemetry.io/contrib/instrumentation/net/http/otelhttp.(*respWriterWrapper).WriteHeader (wrap.go:98) Apr 23 12:32:51 minikube dockerd[4233]: 2024/04/23 12:32:51 http: superfluous response.WriteHeader call from go.opentelemetry.io/contrib/instrumentation/net/http/otelhttp.(*respWriterWrapper).WriteHeader (wrap.go:98) Apr 23 12:32:51 minikube dockerd[4233]: 2024/04/23 12:32:51 http: superfluous response.WriteHeader call from go.opentelemetry.io/contrib/instrumentation/net/http/otelhttp.(*respWriterWrapper).WriteHeader (wrap.go:98) Apr 23 12:32:51 minikube dockerd[4233]: 2024/04/23 12:32:51 http: superfluous response.WriteHeader call from go.opentelemetry.io/contrib/instrumentation/net/http/otelhttp.(*respWriterWrapper).WriteHeader (wrap.go:98) ==> container status <== CONTAINER IMAGE CREATED STATE NAME ATTEMPT POD ID POD e7194fc4f8e4c a0bf559e280cf 2 minutes ago Running kube-proxy 0 2f28330b2bc43 kube-proxy-jx599 3ad2c1ad293f8 cbb01a7bd410d 2 minutes ago Running coredns 0 64c37cca31f05 coredns-7db6d8ff4d-fqg4r 0ba7a52ba945a cbb01a7bd410d 2 minutes ago Running coredns 0 59fb619499875 coredns-7db6d8ff4d-z7ds7 96d000bf1f734 c42f13656d0b2 2 minutes ago Running kube-apiserver 0 3b34159df3e24 kube-apiserver-minikube 877fba6696c66 c7aad43836fa5 2 minutes ago Running kube-controller-manager 0 4841ef16bce82 kube-controller-manager-minikube 4fcdb023cded4 259c8277fcbbc 2 minutes ago Running kube-scheduler 0 c905f8147990a kube-scheduler-minikube f717fd3558f51 3861cfcd7c04c 2 minutes ago Running etcd 0 e28e5c40f426d etcd-minikube ==> coredns [0ba7a52ba945] <== .:53 [INFO] plugin/reload: Running configuration SHA512 = 591cf328cccc12bc490481273e738df59329c62c0b729d94e8b61db9961c2fa5f046dd37f1cf888b953814040d180f52594972691cd6ff41be96639138a43908 CoreDNS-1.11.1 linux/amd64, go1.20.7, ae2bbc2 ==> coredns [3ad2c1ad293f] <== .:53 [INFO] plugin/reload: Running configuration SHA512 = 591cf328cccc12bc490481273e738df59329c62c0b729d94e8b61db9961c2fa5f046dd37f1cf888b953814040d180f52594972691cd6ff41be96639138a43908 CoreDNS-1.11.1 linux/amd64, go1.20.7, ae2bbc2 ==> describe nodes <== Name: minikube Roles: control-plane Labels: beta.kubernetes.io/arch=amd64 beta.kubernetes.io/os=linux kubernetes.io/arch=amd64 kubernetes.io/hostname=minikube kubernetes.io/os=linux minikube.k8s.io/commit=86fc9d54fca63f295d8737c8eacdbb7987e89c67 minikube.k8s.io/name=minikube minikube.k8s.io/primary=true minikube.k8s.io/updated_at=2024_04_23T18_00_53_0700 minikube.k8s.io/version=v1.33.0 node-role.kubernetes.io/control-plane= node.kubernetes.io/exclude-from-external-load-balancers= Annotations: kubeadm.alpha.kubernetes.io/cri-socket: unix:///var/run/cri-dockerd.sock node.alpha.kubernetes.io/ttl: 0 volumes.kubernetes.io/controller-managed-attach-detach: true CreationTimestamp: Tue, 23 Apr 2024 12:30:48 +0000 Taints: Unschedulable: false Lease: HolderIdentity: minikube AcquireTime: RenewTime: Tue, 23 Apr 2024 12:33:18 +0000 Conditions: Type Status LastHeartbeatTime LastTransitionTime Reason Message ---- ------ ----------------- ------------------ ------ ------- MemoryPressure False Tue, 23 Apr 2024 12:31:14 +0000 Tue, 23 Apr 2024 12:30:44 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available DiskPressure False Tue, 23 Apr 2024 12:31:14 +0000 Tue, 23 Apr 2024 12:30:44 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure PIDPressure False Tue, 23 Apr 2024 12:31:14 +0000 Tue, 23 Apr 2024 12:30:44 +0000 KubeletHasSufficientPID kubelet has sufficient PID available Ready True Tue, 23 Apr 2024 12:31:14 +0000 Tue, 23 Apr 2024 12:31:03 +0000 KubeletReady kubelet is posting ready status Addresses: InternalIP: 192.168.59.101 Hostname: minikube Capacity: cpu: 2 ephemeral-storage: 17734596Ki hugepages-2Mi: 0 memory: 3912864Ki pods: 110 Allocatable: cpu: 2 ephemeral-storage: 17734596Ki hugepages-2Mi: 0 memory: 3912864Ki pods: 110 System Info: Machine ID: e9c91f47bceb4ef69aa4b9d98e236b8e System UUID: 573e5e9b-1ed9-2d43-add7-21d97242bc13 Boot ID: 35abea12-85c3-4a28-8f7a-3a23a36d45b5 Kernel Version: 5.10.207 OS Image: Buildroot 2023.02.9 Operating System: linux Architecture: amd64 Container Runtime Version: docker://26.0.1 Kubelet Version: v1.30.0 Kube-Proxy Version: v1.30.0 PodCIDR: 10.244.0.0/24 PodCIDRs: 10.244.0.0/24 Non-terminated Pods: (7 in total) Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits Age --------- ---- ------------ ---------- --------------- ------------- --- kube-system coredns-7db6d8ff4d-fqg4r 100m (5%!)(MISSING) 0 (0%!)(MISSING) 70Mi (1%!)(MISSING) 170Mi (4%!)(MISSING) 2m16s kube-system coredns-7db6d8ff4d-z7ds7 100m (5%!)(MISSING) 0 (0%!)(MISSING) 70Mi (1%!)(MISSING) 170Mi (4%!)(MISSING) 2m16s kube-system etcd-minikube 100m (5%!)(MISSING) 0 (0%!)(MISSING) 100Mi (2%!)(MISSING) 0 (0%!)(MISSING) 2m29s kube-system kube-apiserver-minikube 250m (12%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 2m31s kube-system kube-controller-manager-minikube 200m (10%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 2m29s kube-system kube-proxy-jx599 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 2m16s kube-system kube-scheduler-minikube 100m (5%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 2m32s Allocated resources: (Total limits may be over 100 percent, i.e., overcommitted.) Resource Requests Limits -------- -------- ------ cpu 850m (42%!)(MISSING) 0 (0%!)(MISSING) memory 240Mi (6%!)(MISSING) 340Mi (8%!)(MISSING) ephemeral-storage 0 (0%!)(MISSING) 0 (0%!)(MISSING) hugepages-2Mi 0 (0%!)(MISSING) 0 (0%!)(MISSING) Events: Type Reason Age From Message ---- ------ ---- ---- ------- Normal Starting 2m12s kube-proxy Normal Starting 2m42s kubelet Starting kubelet. Normal NodeHasSufficientMemory 2m42s (x8 over 2m42s) kubelet Node minikube status is now: NodeHasSufficientMemory Normal NodeHasNoDiskPressure 2m42s (x8 over 2m42s) kubelet Node minikube status is now: NodeHasNoDiskPressure Normal NodeHasSufficientPID 2m42s (x7 over 2m42s) kubelet Node minikube status is now: NodeHasSufficientPID Normal NodeAllocatableEnforced 2m42s kubelet Updated Node Allocatable limit across pods Normal Starting 2m30s kubelet Starting kubelet. Normal NodeHasSufficientMemory 2m29s kubelet Node minikube status is now: NodeHasSufficientMemory Normal NodeHasNoDiskPressure 2m29s kubelet Node minikube status is now: NodeHasNoDiskPressure Normal NodeHasSufficientPID 2m29s kubelet Node minikube status is now: NodeHasSufficientPID Normal NodeAllocatableEnforced 2m29s kubelet Updated Node Allocatable limit across pods Normal NodeNotReady 2m29s kubelet Node minikube status is now: NodeNotReady Normal NodeReady 2m19s kubelet Node minikube status is now: NodeReady Normal RegisteredNode 2m17s node-controller Node minikube event: Registered Node minikube in Controller ==> dmesg <== [Apr23 12:22] You have booted with nomodeset. This means your GPU drivers are DISABLED [ +0.000000] Any video related functionality will be severely degraded, and you may not even be able to suspend the system properly [ +0.000001] Unless you actually understand what nomodeset does, you should reboot without enabling it [ +0.386681] Spectre V2 : WARNING: Unprivileged eBPF is enabled with eIBRS on, data leaks possible via Spectre v2 BHB attacks! [ +0.130870] APIC calibration not consistent with PM-Timer: 104ms instead of 100ms [ +0.010995] TSC synchronization [CPU#0 -> CPU#1]: [ +0.000000] Measured 71646 cycles TSC warp between CPUs, turning off TSC clock. [ +0.048806] acpi PNP0A03:00: fail to add MMCONFIG information, can't access extended PCI configuration space under this bridge. [ +8.124002] platform regulatory.0: Direct firmware load for regulatory.db failed with error -2 [ +22.123425] systemd-fstab-generator[199]: Ignoring "noauto" option for root device [ +2.121639] NFSD: Using /var/lib/nfs/v4recovery as the NFSv4 state recovery directory [ +0.000007] NFSD: unable to find recovery directory /var/lib/nfs/v4recovery [ +0.000002] NFSD: Unable to initialize client recovery tracking! (-2) [ +13.110826] systemd-fstab-generator[613]: Ignoring "noauto" option for root device [ +0.204700] systemd-fstab-generator[626]: Ignoring "noauto" option for root device [ +5.792826] systemd-fstab-generator[860]: Ignoring "noauto" option for root device [ +0.069958] kauditd_printk_skb: 69 callbacks suppressed [ +0.603641] systemd-fstab-generator[897]: Ignoring "noauto" option for root device [ +0.226514] systemd-fstab-generator[909]: Ignoring "noauto" option for root device [ +0.256374] systemd-fstab-generator[924]: Ignoring "noauto" option for root device [Apr23 12:23] systemd-fstab-generator[1095]: Ignoring "noauto" option for root device [ +0.193187] systemd-fstab-generator[1107]: Ignoring "noauto" option for root device [ +0.207221] systemd-fstab-generator[1119]: Ignoring "noauto" option for root device [ +0.067352] kauditd_printk_skb: 149 callbacks suppressed [ +0.217267] systemd-fstab-generator[1134]: Ignoring "noauto" option for root device [ +44.935549] systemd-fstab-generator[1232]: Ignoring "noauto" option for root device [ +0.302076] kauditd_printk_skb: 30 callbacks suppressed [ +5.108387] kauditd_printk_skb: 25 callbacks suppressed [ +4.455383] systemd-fstab-generator[1429]: Ignoring "noauto" option for root device [Apr23 12:24] systemd-fstab-generator[1642]: Ignoring "noauto" option for root device [ +0.127980] kauditd_printk_skb: 34 callbacks suppressed [ +13.699267] systemd-fstab-generator[2053]: Ignoring "noauto" option for root device [ +0.296680] kauditd_printk_skb: 60 callbacks suppressed [ +15.392269] kauditd_printk_skb: 10 callbacks suppressed [ +6.424127] kauditd_printk_skb: 64 callbacks suppressed [Apr23 12:29] systemd-fstab-generator[3742]: Ignoring "noauto" option for root device [ +0.815992] systemd-fstab-generator[3793]: Ignoring "noauto" option for root device [ +0.306438] systemd-fstab-generator[3805]: Ignoring "noauto" option for root device [ +0.390569] systemd-fstab-generator[3819]: Ignoring "noauto" option for root device [ +5.432152] kauditd_printk_skb: 79 callbacks suppressed [Apr23 12:30] systemd-fstab-generator[4425]: Ignoring "noauto" option for root device [ +0.198844] systemd-fstab-generator[4437]: Ignoring "noauto" option for root device [ +0.252282] systemd-fstab-generator[4449]: Ignoring "noauto" option for root device [ +0.355145] systemd-fstab-generator[4468]: Ignoring "noauto" option for root device [ +1.721320] kauditd_printk_skb: 112 callbacks suppressed [ +0.762959] systemd-fstab-generator[4843]: Ignoring "noauto" option for root device [ +6.001541] kauditd_printk_skb: 86 callbacks suppressed [ +21.798392] systemd-fstab-generator[7206]: Ignoring "noauto" option for root device [ +0.104915] kauditd_printk_skb: 11 callbacks suppressed [ +12.290748] systemd-fstab-generator[7644]: Ignoring "noauto" option for root device [ +0.136541] kauditd_printk_skb: 50 callbacks suppressed [Apr23 12:31] kauditd_printk_skb: 10 callbacks suppressed [Apr23 12:32] kauditd_printk_skb: 36 callbacks suppressed ==> etcd [f717fd3558f5] <== {"level":"warn","ts":"2024-04-23T12:30:43.329776Z","caller":"embed/config.go:679","msg":"Running http and grpc server on single port. This is not recommended for production."} {"level":"info","ts":"2024-04-23T12:30:43.347715Z","caller":"etcdmain/etcd.go:73","msg":"Running: ","args":["etcd","--advertise-client-urls=https://192.168.59.101:2379","--cert-file=/var/lib/minikube/certs/etcd/server.crt","--client-cert-auth=true","--data-dir=/var/lib/minikube/etcd","--experimental-initial-corrupt-check=true","--experimental-watch-progress-notify-interval=5s","--initial-advertise-peer-urls=https://192.168.59.101:2380","--initial-cluster=minikube=https://192.168.59.101:2380","--key-file=/var/lib/minikube/certs/etcd/server.key","--listen-client-urls=https://127.0.0.1:2379,https://192.168.59.101:2379","--listen-metrics-urls=http://127.0.0.1:2381","--listen-peer-urls=https://192.168.59.101:2380","--name=minikube","--peer-cert-file=/var/lib/minikube/certs/etcd/peer.crt","--peer-client-cert-auth=true","--peer-key-file=/var/lib/minikube/certs/etcd/peer.key","--peer-trusted-ca-file=/var/lib/minikube/certs/etcd/ca.crt","--proxy-refresh-interval=70000","--snapshot-count=10000","--trusted-ca-file=/var/lib/minikube/certs/etcd/ca.crt"]} {"level":"warn","ts":"2024-04-23T12:30:43.356133Z","caller":"embed/config.go:679","msg":"Running http and grpc server on single port. This is not recommended for production."} {"level":"info","ts":"2024-04-23T12:30:43.356488Z","caller":"embed/etcd.go:127","msg":"configuring peer listeners","listen-peer-urls":["https://192.168.59.101:2380"]} {"level":"info","ts":"2024-04-23T12:30:43.356542Z","caller":"embed/etcd.go:494","msg":"starting with peer TLS","tls-info":"cert = /var/lib/minikube/certs/etcd/peer.crt, key = /var/lib/minikube/certs/etcd/peer.key, client-cert=, client-key=, trusted-ca = /var/lib/minikube/certs/etcd/ca.crt, client-cert-auth = true, crl-file = ","cipher-suites":[]} {"level":"info","ts":"2024-04-23T12:30:43.365973Z","caller":"embed/etcd.go:135","msg":"configuring client listeners","listen-client-urls":["https://127.0.0.1:2379","https://192.168.59.101:2379"]} {"level":"info","ts":"2024-04-23T12:30:43.366141Z","caller":"embed/etcd.go:308","msg":"starting an etcd server","etcd-version":"3.5.12","git-sha":"e7b3bb6cc","go-version":"go1.20.13","go-os":"linux","go-arch":"amd64","max-cpu-set":2,"max-cpu-available":2,"member-initialized":false,"name":"minikube","data-dir":"/var/lib/minikube/etcd","wal-dir":"","wal-dir-dedicated":"","member-dir":"/var/lib/minikube/etcd/member","force-new-cluster":false,"heartbeat-interval":"100ms","election-timeout":"1s","initial-election-tick-advance":true,"snapshot-count":10000,"max-wals":5,"max-snapshots":5,"snapshot-catchup-entries":5000,"initial-advertise-peer-urls":["https://192.168.59.101:2380"],"listen-peer-urls":["https://192.168.59.101:2380"],"advertise-client-urls":["https://192.168.59.101:2379"],"listen-client-urls":["https://127.0.0.1:2379","https://192.168.59.101:2379"],"listen-metrics-urls":["http://127.0.0.1:2381"],"cors":["*"],"host-whitelist":["*"],"initial-cluster":"minikube=https://192.168.59.101:2380","initial-cluster-state":"new","initial-cluster-token":"etcd-cluster","quota-backend-bytes":2147483648,"max-request-bytes":1572864,"max-concurrent-streams":4294967295,"pre-vote":true,"initial-corrupt-check":true,"corrupt-check-time-interval":"0s","compact-check-time-enabled":false,"compact-check-time-interval":"1m0s","auto-compaction-mode":"periodic","auto-compaction-retention":"0s","auto-compaction-interval":"0s","discovery-url":"","discovery-proxy":"","downgrade-check-interval":"5s"} {"level":"info","ts":"2024-04-23T12:30:43.389879Z","caller":"etcdserver/backend.go:81","msg":"opened backend db","path":"/var/lib/minikube/etcd/member/snap/db","took":"9.232546ms"} {"level":"info","ts":"2024-04-23T12:30:43.441215Z","caller":"etcdserver/raft.go:495","msg":"starting local member","local-member-id":"a2cd0164016c046c","cluster-id":"dc9cc21855b32f2e"} {"level":"info","ts":"2024-04-23T12:30:43.442984Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"a2cd0164016c046c switched to configuration voters=()"} {"level":"info","ts":"2024-04-23T12:30:43.457497Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"a2cd0164016c046c became follower at term 0"} {"level":"info","ts":"2024-04-23T12:30:43.458101Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"newRaft a2cd0164016c046c [peers: [], term: 0, commit: 0, applied: 0, lastindex: 0, lastterm: 0]"} {"level":"info","ts":"2024-04-23T12:30:43.45815Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"a2cd0164016c046c became follower at term 1"} {"level":"info","ts":"2024-04-23T12:30:43.458212Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"a2cd0164016c046c switched to configuration voters=(11731034133402223724)"} {"level":"warn","ts":"2024-04-23T12:30:43.488438Z","caller":"auth/store.go:1241","msg":"simple token is not cryptographically signed"} {"level":"info","ts":"2024-04-23T12:30:43.515028Z","caller":"mvcc/kvstore.go:407","msg":"kvstore restored","current-rev":1} {"level":"info","ts":"2024-04-23T12:30:43.518528Z","caller":"etcdserver/quota.go:94","msg":"enabled backend quota with default value","quota-name":"v3-applier","quota-size-bytes":2147483648,"quota-size":"2.1 GB"} {"level":"info","ts":"2024-04-23T12:30:43.524784Z","caller":"etcdserver/server.go:860","msg":"starting etcd server","local-member-id":"a2cd0164016c046c","local-server-version":"3.5.12","cluster-version":"to_be_decided"} {"level":"info","ts":"2024-04-23T12:30:43.525616Z","caller":"fileutil/purge.go:50","msg":"started to purge file","dir":"/var/lib/minikube/etcd/member/snap","suffix":"snap.db","max":5,"interval":"30s"} {"level":"info","ts":"2024-04-23T12:30:43.526913Z","caller":"fileutil/purge.go:50","msg":"started to purge file","dir":"/var/lib/minikube/etcd/member/snap","suffix":"snap","max":5,"interval":"30s"} {"level":"info","ts":"2024-04-23T12:30:43.52767Z","caller":"fileutil/purge.go:50","msg":"started to purge file","dir":"/var/lib/minikube/etcd/member/wal","suffix":"wal","max":5,"interval":"30s"} {"level":"info","ts":"2024-04-23T12:30:43.532585Z","caller":"etcdserver/server.go:744","msg":"started as single-node; fast-forwarding election ticks","local-member-id":"a2cd0164016c046c","forward-ticks":9,"forward-duration":"900ms","election-ticks":10,"election-timeout":"1s"} {"level":"info","ts":"2024-04-23T12:30:43.644412Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"a2cd0164016c046c switched to configuration voters=(11731034133402223724)"} {"level":"info","ts":"2024-04-23T12:30:43.64481Z","caller":"membership/cluster.go:421","msg":"added member","cluster-id":"dc9cc21855b32f2e","local-member-id":"a2cd0164016c046c","added-peer-id":"a2cd0164016c046c","added-peer-peer-urls":["https://192.168.59.101:2380"]} {"level":"info","ts":"2024-04-23T12:30:43.652831Z","caller":"embed/etcd.go:726","msg":"starting with client TLS","tls-info":"cert = /var/lib/minikube/certs/etcd/server.crt, key = /var/lib/minikube/certs/etcd/server.key, client-cert=, client-key=, trusted-ca = /var/lib/minikube/certs/etcd/ca.crt, client-cert-auth = true, crl-file = ","cipher-suites":[]} {"level":"info","ts":"2024-04-23T12:30:43.653113Z","caller":"embed/etcd.go:277","msg":"now serving peer/client/metrics","local-member-id":"a2cd0164016c046c","initial-advertise-peer-urls":["https://192.168.59.101:2380"],"listen-peer-urls":["https://192.168.59.101:2380"],"advertise-client-urls":["https://192.168.59.101:2379"],"listen-client-urls":["https://127.0.0.1:2379","https://192.168.59.101:2379"],"listen-metrics-urls":["http://127.0.0.1:2381"]} {"level":"info","ts":"2024-04-23T12:30:43.667403Z","caller":"embed/etcd.go:857","msg":"serving metrics","address":"http://127.0.0.1:2381"} {"level":"info","ts":"2024-04-23T12:30:43.667513Z","caller":"embed/etcd.go:597","msg":"serving peer traffic","address":"192.168.59.101:2380"} {"level":"info","ts":"2024-04-23T12:30:43.66752Z","caller":"embed/etcd.go:569","msg":"cmux::serve","address":"192.168.59.101:2380"} {"level":"info","ts":"2024-04-23T12:30:45.077845Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"a2cd0164016c046c is starting a new election at term 1"} {"level":"info","ts":"2024-04-23T12:30:45.077891Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"a2cd0164016c046c became pre-candidate at term 1"} {"level":"info","ts":"2024-04-23T12:30:45.077929Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"a2cd0164016c046c received MsgPreVoteResp from a2cd0164016c046c at term 1"} {"level":"info","ts":"2024-04-23T12:30:45.077949Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"a2cd0164016c046c became candidate at term 2"} {"level":"info","ts":"2024-04-23T12:30:45.077959Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"a2cd0164016c046c received MsgVoteResp from a2cd0164016c046c at term 2"} {"level":"info","ts":"2024-04-23T12:30:45.07797Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"a2cd0164016c046c became leader at term 2"} {"level":"info","ts":"2024-04-23T12:30:45.07798Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"raft.node: a2cd0164016c046c elected leader a2cd0164016c046c at term 2"} {"level":"info","ts":"2024-04-23T12:30:45.248578Z","caller":"etcdserver/server.go:2578","msg":"setting up initial cluster version using v2 API","cluster-version":"3.5"} {"level":"info","ts":"2024-04-23T12:30:45.254015Z","caller":"membership/cluster.go:584","msg":"set initial cluster version","cluster-id":"dc9cc21855b32f2e","local-member-id":"a2cd0164016c046c","cluster-version":"3.5"} {"level":"info","ts":"2024-04-23T12:30:45.254426Z","caller":"api/capability.go:75","msg":"enabled capabilities for version","cluster-version":"3.5"} {"level":"info","ts":"2024-04-23T12:30:45.255617Z","caller":"etcdserver/server.go:2602","msg":"cluster version is updated","cluster-version":"3.5"} {"level":"info","ts":"2024-04-23T12:30:45.255706Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"} {"level":"info","ts":"2024-04-23T12:30:45.259582Z","caller":"etcdserver/server.go:2068","msg":"published local member to cluster through raft","local-member-id":"a2cd0164016c046c","local-member-attributes":"{Name:minikube ClientURLs:[https://192.168.59.101:2379]}","request-path":"/0/members/a2cd0164016c046c/attributes","cluster-id":"dc9cc21855b32f2e","publish-timeout":"7s"} {"level":"info","ts":"2024-04-23T12:30:45.261961Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"} {"level":"info","ts":"2024-04-23T12:30:45.266263Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"127.0.0.1:2379"} {"level":"info","ts":"2024-04-23T12:30:45.269883Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"192.168.59.101:2379"} {"level":"info","ts":"2024-04-23T12:30:45.65178Z","caller":"etcdmain/main.go:44","msg":"notifying init daemon"} {"level":"info","ts":"2024-04-23T12:30:45.652025Z","caller":"etcdmain/main.go:50","msg":"successfully notified init daemon"} ==> kernel <== 12:33:22 up 11 min, 0 users, load average: 2.22, 2.75, 1.77 Linux minikube 5.10.207 #1 SMP Thu Apr 18 22:28:35 UTC 2024 x86_64 GNU/Linux PRETTY_NAME="Buildroot 2023.02.9" ==> kube-apiserver [96d000bf1f73] <== I0423 12:30:48.495900 1 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/var/lib/minikube/certs/ca.crt" I0423 12:30:48.496286 1 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/var/lib/minikube/certs/ca.crt" I0423 12:30:48.496398 1 dynamic_cafile_content.go:157] "Starting controller" name="request-header::/var/lib/minikube/certs/front-proxy-ca.crt" I0423 12:30:48.494683 1 tlsconfig.go:240] "Starting DynamicServingCertificateController" I0423 12:30:48.496787 1 controller.go:139] Starting OpenAPI controller I0423 12:30:48.496858 1 controller.go:87] Starting OpenAPI V3 controller I0423 12:30:48.496906 1 naming_controller.go:291] Starting NamingConditionController I0423 12:30:48.496952 1 establishing_controller.go:76] Starting EstablishingController I0423 12:30:48.496999 1 nonstructuralschema_controller.go:192] Starting NonStructuralSchemaConditionController I0423 12:30:48.497049 1 apiapproval_controller.go:186] Starting KubernetesAPIApprovalPolicyConformantConditionController I0423 12:30:48.497190 1 crd_finalizer.go:266] Starting CRDFinalizer I0423 12:30:48.506950 1 gc_controller.go:78] Starting apiserver lease garbage collector I0423 12:30:48.507509 1 dynamic_serving_content.go:132] "Starting controller" name="aggregator-proxy-cert::/var/lib/minikube/certs/front-proxy-client.crt::/var/lib/minikube/certs/front-proxy-client.key" I0423 12:30:48.510792 1 available_controller.go:423] Starting AvailableConditionController I0423 12:30:48.510882 1 cache.go:32] Waiting for caches to sync for AvailableConditionController controller I0423 12:30:48.510904 1 controller.go:78] Starting OpenAPI AggregationController I0423 12:30:48.510984 1 controller.go:116] Starting legacy_token_tracking_controller I0423 12:30:48.510990 1 shared_informer.go:313] Waiting for caches to sync for configmaps I0423 12:30:48.511841 1 aggregator.go:163] waiting for initial CRD sync... I0423 12:30:48.512150 1 system_namespaces_controller.go:67] Starting system namespaces controller I0423 12:30:48.584954 1 crdregistration_controller.go:111] Starting crd-autoregister controller I0423 12:30:48.585174 1 shared_informer.go:313] Waiting for caches to sync for crd-autoregister I0423 12:30:48.659251 1 shared_informer.go:320] Caches are synced for *generic.policySource[*k8s.io/api/admissionregistration/v1.ValidatingAdmissionPolicy,*k8s.io/api/admissionregistration/v1.ValidatingAdmissionPolicyBinding,k8s.io/apiserver/pkg/admission/plugin/policy/validating.Validator] I0423 12:30:48.659758 1 policy_source.go:224] refreshing policies I0423 12:30:48.660246 1 shared_informer.go:320] Caches are synced for node_authorizer I0423 12:30:48.686686 1 shared_informer.go:320] Caches are synced for crd-autoregister I0423 12:30:48.686756 1 aggregator.go:165] initial CRD sync complete... I0423 12:30:48.686781 1 autoregister_controller.go:141] Starting autoregister controller I0423 12:30:48.686802 1 cache.go:32] Waiting for caches to sync for autoregister controller I0423 12:30:48.686821 1 cache.go:39] Caches are synced for autoregister controller I0423 12:30:48.694927 1 shared_informer.go:320] Caches are synced for cluster_authentication_trust_controller I0423 12:30:48.708057 1 cache.go:39] Caches are synced for APIServiceRegistrationController controller I0423 12:30:48.708230 1 apf_controller.go:379] Running API Priority and Fairness config worker I0423 12:30:48.708258 1 apf_controller.go:382] Running API Priority and Fairness periodic rebalancing process I0423 12:30:48.708969 1 handler_discovery.go:447] Starting ResourceDiscoveryManager I0423 12:30:48.714344 1 cache.go:39] Caches are synced for AvailableConditionController controller I0423 12:30:48.723553 1 shared_informer.go:320] Caches are synced for configmaps I0423 12:30:48.724733 1 controller.go:615] quota admission added evaluator for: namespaces E0423 12:30:48.785097 1 controller.go:145] "Failed to ensure lease exists, will retry" err="namespaces \"kube-system\" not found" interval="200ms" I0423 12:30:48.998129 1 controller.go:615] quota admission added evaluator for: leases.coordination.k8s.io I0423 12:30:49.524112 1 storage_scheduling.go:95] created PriorityClass system-node-critical with value 2000001000 I0423 12:30:49.587440 1 storage_scheduling.go:95] created PriorityClass system-cluster-critical with value 2000000000 I0423 12:30:49.589777 1 storage_scheduling.go:111] all system priority classes are created successfully or already exist. I0423 12:30:51.358415 1 controller.go:615] quota admission added evaluator for: roles.rbac.authorization.k8s.io I0423 12:30:51.438824 1 controller.go:615] quota admission added evaluator for: rolebindings.rbac.authorization.k8s.io I0423 12:30:51.664003 1 alloc.go:330] "allocated clusterIPs" service="default/kubernetes" clusterIPs={"IPv4":"10.96.0.1"} I0423 12:30:51.682635 1 controller.go:615] quota admission added evaluator for: serviceaccounts W0423 12:30:51.729269 1 lease.go:265] Resetting endpoints for master service "kubernetes" to [192.168.59.101] I0423 12:30:51.735057 1 controller.go:615] quota admission added evaluator for: endpoints I0423 12:30:51.761610 1 controller.go:615] quota admission added evaluator for: endpointslices.discovery.k8s.io I0423 12:30:52.716353 1 controller.go:615] quota admission added evaluator for: deployments.apps I0423 12:30:52.773403 1 alloc.go:330] "allocated clusterIPs" service="kube-system/kube-dns" clusterIPs={"IPv4":"10.96.0.10"} I0423 12:30:52.806321 1 controller.go:615] quota admission added evaluator for: daemonsets.apps E0423 12:31:05.594125 1 queueset.go:474] "Overflow" err="queueset::currentR overflow" QS="workload-high" when="2024-04-23 12:31:05.594103113" prevR="0.68517043ss" incrR="184467440737.09551547ss" currentR="0.68516974ss" E0423 12:31:05.660516 1 queueset.go:474] "Overflow" err="queueset::currentR overflow" QS="workload-high" when="2024-04-23 12:31:05.660500615" prevR="0.72518864ss" incrR="184467440737.09549487ss" currentR="0.72516735ss" I0423 12:31:06.107006 1 controller.go:615] quota admission added evaluator for: replicasets.apps I0423 12:31:06.717835 1 controller.go:615] quota admission added evaluator for: controllerrevisions.apps E0423 12:31:09.384683 1 queueset.go:474] "Overflow" err="queueset::currentR overflow" QS="workload-high" when="2024-04-23 12:31:09.384666915" prevR="1.44885733ss" incrR="184467440737.09549439ss" currentR="1.44883556ss" E0423 12:31:10.032831 1 queueset.go:474] "Overflow" err="queueset::currentR overflow" QS="workload-high" when="2024-04-23 12:31:10.032790329" prevR="1.65256622ss" incrR="184467440737.09551274ss" currentR="1.65256280ss" E0423 12:31:10.035133 1 queueset.go:474] "Overflow" err="queueset::currentR overflow" QS="workload-high" when="2024-04-23 12:31:10.035119348" prevR="1.65489482ss" incrR="184467440737.09551318ss" currentR="1.65489184ss" ==> kube-controller-manager [877fba6696c6] <== I0423 12:31:05.624589 1 shared_informer.go:320] Caches are synced for validatingadmissionpolicy-status I0423 12:31:05.625805 1 shared_informer.go:320] Caches are synced for TTL after finished I0423 12:31:05.630414 1 shared_informer.go:320] Caches are synced for PV protection I0423 12:31:05.630521 1 shared_informer.go:320] Caches are synced for certificate-csrsigning-kubelet-serving I0423 12:31:05.630662 1 shared_informer.go:320] Caches are synced for endpoint I0423 12:31:05.632792 1 shared_informer.go:320] Caches are synced for ReplicationController I0423 12:31:05.632466 1 shared_informer.go:320] Caches are synced for cronjob I0423 12:31:05.634455 1 shared_informer.go:320] Caches are synced for ephemeral I0423 12:31:05.632502 1 shared_informer.go:320] Caches are synced for HPA I0423 12:31:05.634533 1 shared_informer.go:320] Caches are synced for deployment I0423 12:31:05.632508 1 shared_informer.go:320] Caches are synced for PVC protection I0423 12:31:05.636415 1 shared_informer.go:320] Caches are synced for ReplicaSet I0423 12:31:05.636456 1 shared_informer.go:320] Caches are synced for endpoint_slice I0423 12:31:05.638385 1 shared_informer.go:313] Waiting for caches to sync for garbage collector I0423 12:31:05.640244 1 shared_informer.go:320] Caches are synced for GC I0423 12:31:05.638831 1 shared_informer.go:320] Caches are synced for certificate-csrapproving I0423 12:31:05.640492 1 shared_informer.go:320] Caches are synced for certificate-csrsigning-legacy-unknown I0423 12:31:05.649567 1 shared_informer.go:320] Caches are synced for bootstrap_signer I0423 12:31:05.649876 1 shared_informer.go:320] Caches are synced for crt configmap I0423 12:31:05.650279 1 shared_informer.go:320] Caches are synced for TTL I0423 12:31:05.652916 1 shared_informer.go:320] Caches are synced for ClusterRoleAggregator I0423 12:31:05.651796 1 shared_informer.go:320] Caches are synced for disruption I0423 12:31:05.657033 1 shared_informer.go:320] Caches are synced for service account I0423 12:31:05.657719 1 shared_informer.go:320] Caches are synced for namespace I0423 12:31:05.658269 1 shared_informer.go:320] Caches are synced for taint-eviction-controller I0423 12:31:05.652675 1 shared_informer.go:320] Caches are synced for persistent volume I0423 12:31:05.658633 1 shared_informer.go:320] Caches are synced for job I0423 12:31:05.656699 1 shared_informer.go:320] Caches are synced for legacy-service-account-token-cleaner I0423 12:31:05.665134 1 shared_informer.go:320] Caches are synced for node I0423 12:31:05.665715 1 range_allocator.go:175] "Sending events to api server" logger="node-ipam-controller" I0423 12:31:05.666174 1 range_allocator.go:179] "Starting range CIDR allocator" logger="node-ipam-controller" I0423 12:31:05.666490 1 shared_informer.go:313] Waiting for caches to sync for cidrallocator I0423 12:31:05.667057 1 shared_informer.go:320] Caches are synced for cidrallocator I0423 12:31:05.674655 1 shared_informer.go:320] Caches are synced for expand I0423 12:31:05.675685 1 shared_informer.go:320] Caches are synced for stateful set I0423 12:31:05.686646 1 range_allocator.go:381] "Set node PodCIDR" logger="node-ipam-controller" node="minikube" podCIDRs=["10.244.0.0/24"] I0423 12:31:05.699059 1 shared_informer.go:320] Caches are synced for endpoint_slice_mirroring I0423 12:31:05.699229 1 shared_informer.go:320] Caches are synced for taint I0423 12:31:05.700373 1 node_lifecycle_controller.go:1227] "Initializing eviction metric for zone" logger="node-lifecycle-controller" zone="" I0423 12:31:05.701012 1 node_lifecycle_controller.go:879] "Missing timestamp for Node. Assuming now as a timestamp" logger="node-lifecycle-controller" node="minikube" I0423 12:31:05.701380 1 node_lifecycle_controller.go:1073] "Controller detected that zone is now in new state" logger="node-lifecycle-controller" zone="" newState="Normal" I0423 12:31:05.709006 1 shared_informer.go:320] Caches are synced for certificate-csrsigning-kubelet-client I0423 12:31:05.751762 1 shared_informer.go:320] Caches are synced for daemon sets I0423 12:31:05.765777 1 shared_informer.go:320] Caches are synced for resource quota I0423 12:31:05.814693 1 shared_informer.go:320] Caches are synced for resource quota I0423 12:31:05.851041 1 shared_informer.go:320] Caches are synced for attach detach I0423 12:31:06.347167 1 shared_informer.go:320] Caches are synced for garbage collector I0423 12:31:06.355589 1 shared_informer.go:320] Caches are synced for garbage collector I0423 12:31:06.355643 1 garbagecollector.go:157] "All resource monitors have synced. Proceeding to collect garbage" logger="garbage-collector-controller" I0423 12:31:06.925920 1 replica_set.go:676] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/coredns-7db6d8ff4d" duration="803.14151ms" I0423 12:31:06.958940 1 replica_set.go:676] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/coredns-7db6d8ff4d" duration="30.443186ms" I0423 12:31:06.959251 1 replica_set.go:676] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/coredns-7db6d8ff4d" duration="76.948µs" I0423 12:31:06.983702 1 replica_set.go:676] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/coredns-7db6d8ff4d" duration="107.222µs" I0423 12:31:07.023276 1 replica_set.go:676] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/coredns-7db6d8ff4d" duration="104.152µs" I0423 12:31:09.381400 1 replica_set.go:676] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/coredns-7db6d8ff4d" duration="56.027µs" I0423 12:31:09.480980 1 replica_set.go:676] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/coredns-7db6d8ff4d" duration="25.202357ms" I0423 12:31:09.489151 1 replica_set.go:676] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/coredns-7db6d8ff4d" duration="61.475µs" I0423 12:31:09.533081 1 replica_set.go:676] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/coredns-7db6d8ff4d" duration="43.258µs" I0423 12:31:09.608958 1 replica_set.go:676] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/coredns-7db6d8ff4d" duration="32.426377ms" I0423 12:31:09.609649 1 replica_set.go:676] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/coredns-7db6d8ff4d" duration="65.747µs" ==> kube-proxy [e7194fc4f8e4] <== I0423 12:31:09.827498 1 server_linux.go:69] "Using iptables proxy" I0423 12:31:09.845663 1 server.go:1062] "Successfully retrieved node IP(s)" IPs=["192.168.59.101"] I0423 12:31:09.998538 1 server_linux.go:143] "No iptables support for family" ipFamily="IPv6" I0423 12:31:09.998567 1 server.go:661] "kube-proxy running in single-stack mode" ipFamily="IPv4" I0423 12:31:09.998583 1 server_linux.go:165] "Using iptables Proxier" I0423 12:31:10.019274 1 proxier.go:243] "Setting route_localnet=1 to allow node-ports on localhost; to change this either disable iptables.localhostNodePorts (--iptables-localhost-nodeports) or set nodePortAddresses (--nodeport-addresses) to filter loopback addresses" I0423 12:31:10.025489 1 server.go:872] "Version info" version="v1.30.0" I0423 12:31:10.027138 1 server.go:874] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" I0423 12:31:10.029496 1 config.go:192] "Starting service config controller" I0423 12:31:10.029667 1 shared_informer.go:313] Waiting for caches to sync for service config I0423 12:31:10.029697 1 config.go:101] "Starting endpoint slice config controller" I0423 12:31:10.029704 1 shared_informer.go:313] Waiting for caches to sync for endpoint slice config I0423 12:31:10.031246 1 config.go:319] "Starting node config controller" I0423 12:31:10.031516 1 shared_informer.go:313] Waiting for caches to sync for node config I0423 12:31:10.129887 1 shared_informer.go:320] Caches are synced for endpoint slice config I0423 12:31:10.130227 1 shared_informer.go:320] Caches are synced for service config I0423 12:31:10.132018 1 shared_informer.go:320] Caches are synced for node config ==> kube-scheduler [4fcdb023cded] <== E0423 12:30:48.671802 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.ReplicationController: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User "system:kube-scheduler" cannot list resource "replicationcontrollers" in API group "" at the cluster scope W0423 12:30:48.657539 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csistoragecapacities" in API group "storage.k8s.io" at the cluster scope W0423 12:30:48.673524 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes is forbidden: User "system:kube-scheduler" cannot list resource "nodes" in API group "" at the cluster scope E0423 12:30:48.673643 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes is forbidden: User "system:kube-scheduler" cannot list resource "nodes" in API group "" at the cluster scope W0423 12:30:48.674988 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csinodes" in API group "storage.k8s.io" at the cluster scope E0423 12:30:48.676469 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSINode: failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csinodes" in API group "storage.k8s.io" at the cluster scope W0423 12:30:48.679494 1 reflector.go:547] runtime/asm_amd64.s:1695: failed to list *v1.ConfigMap: configmaps "extension-apiserver-authentication" is forbidden: User "system:kube-scheduler" cannot list resource "configmaps" in API group "" in the namespace "kube-system" E0423 12:30:48.679964 1 reflector.go:150] runtime/asm_amd64.s:1695: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps "extension-apiserver-authentication" is forbidden: User "system:kube-scheduler" cannot list resource "configmaps" in API group "" in the namespace "kube-system" E0423 12:30:48.681726 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIStorageCapacity: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csistoragecapacities" in API group "storage.k8s.io" at the cluster scope W0423 12:30:48.657610 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Pod: pods is forbidden: User "system:kube-scheduler" cannot list resource "pods" in API group "" at the cluster scope W0423 12:30:48.689116 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Namespace: namespaces is forbidden: User "system:kube-scheduler" cannot list resource "namespaces" in API group "" at the cluster scope E0423 12:30:48.695107 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Namespace: failed to list *v1.Namespace: namespaces is forbidden: User "system:kube-scheduler" cannot list resource "namespaces" in API group "" at the cluster scope E0423 12:30:48.691758 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Pod: failed to list *v1.Pod: pods is forbidden: User "system:kube-scheduler" cannot list resource "pods" in API group "" at the cluster scope W0423 12:30:48.658187 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumeclaims" in API group "" at the cluster scope E0423 12:30:48.698634 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolumeClaim: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumeclaims" in API group "" at the cluster scope W0423 12:30:48.658670 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:kube-scheduler" cannot list resource "services" in API group "" at the cluster scope E0423 12:30:48.698712 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User "system:kube-scheduler" cannot list resource "services" in API group "" at the cluster scope W0423 12:30:48.658787 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User "system:kube-scheduler" cannot list resource "replicasets" in API group "apps" at the cluster scope E0423 12:30:48.698768 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.ReplicaSet: failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User "system:kube-scheduler" cannot list resource "replicasets" in API group "apps" at the cluster scope W0423 12:30:48.662776 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope E0423 12:30:48.699723 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope W0423 12:30:48.667892 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "storageclasses" in API group "storage.k8s.io" at the cluster scope W0423 12:30:48.671607 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumes" in API group "" at the cluster scope W0423 12:30:48.671711 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User "system:kube-scheduler" cannot list resource "poddisruptionbudgets" in API group "policy" at the cluster scope W0423 12:30:48.658135 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User "system:kube-scheduler" cannot list resource "statefulsets" in API group "apps" at the cluster scope E0423 12:30:48.714633 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.StatefulSet: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User "system:kube-scheduler" cannot list resource "statefulsets" in API group "apps" at the cluster scope E0423 12:30:48.714732 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.StorageClass: failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "storageclasses" in API group "storage.k8s.io" at the cluster scope E0423 12:30:48.714775 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolume: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumes" in API group "" at the cluster scope E0423 12:30:48.714803 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PodDisruptionBudget: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User "system:kube-scheduler" cannot list resource "poddisruptionbudgets" in API group "policy" at the cluster scope W0423 12:30:49.585143 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User "system:kube-scheduler" cannot list resource "replicasets" in API group "apps" at the cluster scope E0423 12:30:49.588910 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.ReplicaSet: failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User "system:kube-scheduler" cannot list resource "replicasets" in API group "apps" at the cluster scope W0423 12:30:49.717814 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumeclaims" in API group "" at the cluster scope E0423 12:30:49.718181 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolumeClaim: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumeclaims" in API group "" at the cluster scope W0423 12:30:49.737974 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:kube-scheduler" cannot list resource "services" in API group "" at the cluster scope E0423 12:30:49.738294 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User "system:kube-scheduler" cannot list resource "services" in API group "" at the cluster scope W0423 12:30:49.868862 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope E0423 12:30:49.869160 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope W0423 12:30:49.973620 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes is forbidden: User "system:kube-scheduler" cannot list resource "nodes" in API group "" at the cluster scope E0423 12:30:49.973715 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes is forbidden: User "system:kube-scheduler" cannot list resource "nodes" in API group "" at the cluster scope W0423 12:30:49.992727 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Namespace: namespaces is forbidden: User "system:kube-scheduler" cannot list resource "namespaces" in API group "" at the cluster scope E0423 12:30:49.992899 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Namespace: failed to list *v1.Namespace: namespaces is forbidden: User "system:kube-scheduler" cannot list resource "namespaces" in API group "" at the cluster scope W0423 12:30:50.020008 1 reflector.go:547] runtime/asm_amd64.s:1695: failed to list *v1.ConfigMap: configmaps "extension-apiserver-authentication" is forbidden: User "system:kube-scheduler" cannot list resource "configmaps" in API group "" in the namespace "kube-system" E0423 12:30:50.020041 1 reflector.go:150] runtime/asm_amd64.s:1695: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps "extension-apiserver-authentication" is forbidden: User "system:kube-scheduler" cannot list resource "configmaps" in API group "" in the namespace "kube-system" W0423 12:30:50.029873 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csinodes" in API group "storage.k8s.io" at the cluster scope E0423 12:30:50.030516 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSINode: failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csinodes" in API group "storage.k8s.io" at the cluster scope W0423 12:30:50.056040 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Pod: pods is forbidden: User "system:kube-scheduler" cannot list resource "pods" in API group "" at the cluster scope E0423 12:30:50.058959 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Pod: failed to list *v1.Pod: pods is forbidden: User "system:kube-scheduler" cannot list resource "pods" in API group "" at the cluster scope W0423 12:30:50.125828 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User "system:kube-scheduler" cannot list resource "statefulsets" in API group "apps" at the cluster scope E0423 12:30:50.126739 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.StatefulSet: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User "system:kube-scheduler" cannot list resource "statefulsets" in API group "apps" at the cluster scope W0423 12:30:50.164799 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csistoragecapacities" in API group "storage.k8s.io" at the cluster scope E0423 12:30:50.166056 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIStorageCapacity: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csistoragecapacities" in API group "storage.k8s.io" at the cluster scope W0423 12:30:50.207485 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "storageclasses" in API group "storage.k8s.io" at the cluster scope E0423 12:30:50.209134 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.StorageClass: failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "storageclasses" in API group "storage.k8s.io" at the cluster scope W0423 12:30:50.231471 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User "system:kube-scheduler" cannot list resource "poddisruptionbudgets" in API group "policy" at the cluster scope E0423 12:30:50.232439 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PodDisruptionBudget: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User "system:kube-scheduler" cannot list resource "poddisruptionbudgets" in API group "policy" at the cluster scope W0423 12:30:50.265787 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User "system:kube-scheduler" cannot list resource "replicationcontrollers" in API group "" at the cluster scope E0423 12:30:50.266190 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.ReplicationController: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User "system:kube-scheduler" cannot list resource "replicationcontrollers" in API group "" at the cluster scope W0423 12:30:50.286606 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumes" in API group "" at the cluster scope E0423 12:30:50.287760 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolume: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumes" in API group "" at the cluster scope I0423 12:30:52.051058 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file ==> kubelet <== Apr 23 12:30:53 minikube kubelet[7651]: Perhaps ip6tables or your kernel needs to be upgraded. Apr 23 12:30:53 minikube kubelet[7651]: > table="nat" chain="KUBE-KUBELET-CANARY" Apr 23 12:30:53 minikube kubelet[7651]: I0423 12:30:53.494203 7651 topology_manager.go:215] "Topology Admit Handler" podUID="ef035c13e17a41ff337a03dab72df835" podNamespace="kube-system" podName="etcd-minikube" Apr 23 12:30:53 minikube kubelet[7651]: I0423 12:30:53.494312 7651 topology_manager.go:215] "Topology Admit Handler" podUID="d1e9444300b70f0fe72146e3750d538b" podNamespace="kube-system" podName="kube-apiserver-minikube" Apr 23 12:30:53 minikube kubelet[7651]: I0423 12:30:53.494386 7651 topology_manager.go:215] "Topology Admit Handler" podUID="b2f93c4e2ec4e1950aa41a71fd8273b0" podNamespace="kube-system" podName="kube-controller-manager-minikube" Apr 23 12:30:53 minikube kubelet[7651]: I0423 12:30:53.494644 7651 topology_manager.go:215] "Topology Admit Handler" podUID="f9c8e1d0d74b1727abdb4b4a31d3a7c1" podNamespace="kube-system" podName="kube-scheduler-minikube" Apr 23 12:30:53 minikube kubelet[7651]: E0423 12:30:53.520983 7651 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-apiserver-minikube\" already exists" pod="kube-system/kube-apiserver-minikube" Apr 23 12:30:53 minikube kubelet[7651]: I0423 12:30:53.548097 7651 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/b2f93c4e2ec4e1950aa41a71fd8273b0-k8s-certs\") pod \"kube-controller-manager-minikube\" (UID: \"b2f93c4e2ec4e1950aa41a71fd8273b0\") " pod="kube-system/kube-controller-manager-minikube" Apr 23 12:30:53 minikube kubelet[7651]: I0423 12:30:53.548579 7651 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/b2f93c4e2ec4e1950aa41a71fd8273b0-kubeconfig\") pod \"kube-controller-manager-minikube\" (UID: \"b2f93c4e2ec4e1950aa41a71fd8273b0\") " pod="kube-system/kube-controller-manager-minikube" Apr 23 12:30:53 minikube kubelet[7651]: I0423 12:30:53.548613 7651 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/b2f93c4e2ec4e1950aa41a71fd8273b0-usr-share-ca-certificates\") pod \"kube-controller-manager-minikube\" (UID: \"b2f93c4e2ec4e1950aa41a71fd8273b0\") " pod="kube-system/kube-controller-manager-minikube" Apr 23 12:30:53 minikube kubelet[7651]: I0423 12:30:53.548633 7651 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-certs\" (UniqueName: \"kubernetes.io/host-path/ef035c13e17a41ff337a03dab72df835-etcd-certs\") pod \"etcd-minikube\" (UID: \"ef035c13e17a41ff337a03dab72df835\") " pod="kube-system/etcd-minikube" Apr 23 12:30:53 minikube kubelet[7651]: I0423 12:30:53.548646 7651 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/d1e9444300b70f0fe72146e3750d538b-k8s-certs\") pod \"kube-apiserver-minikube\" (UID: \"d1e9444300b70f0fe72146e3750d538b\") " pod="kube-system/kube-apiserver-minikube" Apr 23 12:30:53 minikube kubelet[7651]: I0423 12:30:53.548662 7651 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/d1e9444300b70f0fe72146e3750d538b-usr-share-ca-certificates\") pod \"kube-apiserver-minikube\" (UID: \"d1e9444300b70f0fe72146e3750d538b\") " pod="kube-system/kube-apiserver-minikube" Apr 23 12:30:53 minikube kubelet[7651]: I0423 12:30:53.548676 7651 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/b2f93c4e2ec4e1950aa41a71fd8273b0-ca-certs\") pod \"kube-controller-manager-minikube\" (UID: \"b2f93c4e2ec4e1950aa41a71fd8273b0\") " pod="kube-system/kube-controller-manager-minikube" Apr 23 12:30:53 minikube kubelet[7651]: I0423 12:30:53.548687 7651 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-data\" (UniqueName: \"kubernetes.io/host-path/ef035c13e17a41ff337a03dab72df835-etcd-data\") pod \"etcd-minikube\" (UID: \"ef035c13e17a41ff337a03dab72df835\") " pod="kube-system/etcd-minikube" Apr 23 12:30:53 minikube kubelet[7651]: I0423 12:30:53.548705 7651 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/d1e9444300b70f0fe72146e3750d538b-ca-certs\") pod \"kube-apiserver-minikube\" (UID: \"d1e9444300b70f0fe72146e3750d538b\") " pod="kube-system/kube-apiserver-minikube" Apr 23 12:30:53 minikube kubelet[7651]: I0423 12:30:53.548723 7651 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/b2f93c4e2ec4e1950aa41a71fd8273b0-flexvolume-dir\") pod \"kube-controller-manager-minikube\" (UID: \"b2f93c4e2ec4e1950aa41a71fd8273b0\") " pod="kube-system/kube-controller-manager-minikube" Apr 23 12:30:53 minikube kubelet[7651]: I0423 12:30:53.548737 7651 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/f9c8e1d0d74b1727abdb4b4a31d3a7c1-kubeconfig\") pod \"kube-scheduler-minikube\" (UID: \"f9c8e1d0d74b1727abdb4b4a31d3a7c1\") " pod="kube-system/kube-scheduler-minikube" Apr 23 12:30:53 minikube kubelet[7651]: E0423 12:30:53.576742 7651 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-scheduler-minikube\" already exists" pod="kube-system/kube-scheduler-minikube" Apr 23 12:30:53 minikube kubelet[7651]: I0423 12:30:53.910677 7651 apiserver.go:52] "Watching apiserver" Apr 23 12:30:53 minikube kubelet[7651]: I0423 12:30:53.931261 7651 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" Apr 23 12:30:54 minikube kubelet[7651]: E0423 12:30:54.677751 7651 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"etcd-minikube\" already exists" pod="kube-system/etcd-minikube" Apr 23 12:30:54 minikube kubelet[7651]: E0423 12:30:54.677785 7651 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-minikube\" already exists" pod="kube-system/kube-controller-manager-minikube" Apr 23 12:30:54 minikube kubelet[7651]: E0423 12:30:54.677833 7651 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-scheduler-minikube\" already exists" pod="kube-system/kube-scheduler-minikube" Apr 23 12:30:54 minikube kubelet[7651]: I0423 12:30:54.712463 7651 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-minikube" podStartSLOduration=3.712445791 podStartE2EDuration="3.712445791s" podCreationTimestamp="2024-04-23 12:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-04-23 12:30:54.694973598 +0000 UTC m=+2.073423998" watchObservedRunningTime="2024-04-23 12:30:54.712445791 +0000 UTC m=+2.090896193" Apr 23 12:30:54 minikube kubelet[7651]: I0423 12:30:54.736865 7651 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/etcd-minikube" podStartSLOduration=1.736845657 podStartE2EDuration="1.736845657s" podCreationTimestamp="2024-04-23 12:30:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-04-23 12:30:54.715550198 +0000 UTC m=+2.094000606" watchObservedRunningTime="2024-04-23 12:30:54.736845657 +0000 UTC m=+2.115296058" Apr 23 12:30:54 minikube kubelet[7651]: I0423 12:30:54.767057 7651 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-minikube" podStartSLOduration=1.767041273 podStartE2EDuration="1.767041273s" podCreationTimestamp="2024-04-23 12:30:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-04-23 12:30:54.766831691 +0000 UTC m=+2.145282098" watchObservedRunningTime="2024-04-23 12:30:54.767041273 +0000 UTC m=+2.145491673" Apr 23 12:31:06 minikube kubelet[7651]: I0423 12:31:06.796644 7651 topology_manager.go:215] "Topology Admit Handler" podUID="2d7b173b-2b4e-48d5-8dca-cdee293996e3" podNamespace="kube-system" podName="kube-proxy-jx599" Apr 23 12:31:06 minikube kubelet[7651]: I0423 12:31:06.846121 7651 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/2d7b173b-2b4e-48d5-8dca-cdee293996e3-kube-proxy\") pod \"kube-proxy-jx599\" (UID: \"2d7b173b-2b4e-48d5-8dca-cdee293996e3\") " pod="kube-system/kube-proxy-jx599" Apr 23 12:31:06 minikube kubelet[7651]: I0423 12:31:06.846177 7651 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/2d7b173b-2b4e-48d5-8dca-cdee293996e3-xtables-lock\") pod \"kube-proxy-jx599\" (UID: \"2d7b173b-2b4e-48d5-8dca-cdee293996e3\") " pod="kube-system/kube-proxy-jx599" Apr 23 12:31:06 minikube kubelet[7651]: I0423 12:31:06.846514 7651 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hx5br\" (UniqueName: \"kubernetes.io/projected/2d7b173b-2b4e-48d5-8dca-cdee293996e3-kube-api-access-hx5br\") pod \"kube-proxy-jx599\" (UID: \"2d7b173b-2b4e-48d5-8dca-cdee293996e3\") " pod="kube-system/kube-proxy-jx599" Apr 23 12:31:06 minikube kubelet[7651]: I0423 12:31:06.846548 7651 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2d7b173b-2b4e-48d5-8dca-cdee293996e3-lib-modules\") pod \"kube-proxy-jx599\" (UID: \"2d7b173b-2b4e-48d5-8dca-cdee293996e3\") " pod="kube-system/kube-proxy-jx599" Apr 23 12:31:06 minikube kubelet[7651]: W0423 12:31:06.848508 7651 reflector.go:547] object-"kube-system"/"kube-proxy": failed to list *v1.ConfigMap: configmaps "kube-proxy" is forbidden: User "system:node:minikube" cannot list resource "configmaps" in API group "" in the namespace "kube-system": no relationship found between node 'minikube' and this object Apr 23 12:31:06 minikube kubelet[7651]: E0423 12:31:06.848605 7651 reflector.go:150] object-"kube-system"/"kube-proxy": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps "kube-proxy" is forbidden: User "system:node:minikube" cannot list resource "configmaps" in API group "" in the namespace "kube-system": no relationship found between node 'minikube' and this object Apr 23 12:31:06 minikube kubelet[7651]: I0423 12:31:06.857420 7651 topology_manager.go:215] "Topology Admit Handler" podUID="81363069-ef01-4b3f-8920-273e1c16014c" podNamespace="kube-system" podName="coredns-7db6d8ff4d-fqg4r" Apr 23 12:31:06 minikube kubelet[7651]: I0423 12:31:06.897585 7651 topology_manager.go:215] "Topology Admit Handler" podUID="df023805-6c34-47d2-9df9-3081da40656c" podNamespace="kube-system" podName="coredns-7db6d8ff4d-z7ds7" Apr 23 12:31:07 minikube kubelet[7651]: I0423 12:31:07.048329 7651 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df023805-6c34-47d2-9df9-3081da40656c-config-volume\") pod \"coredns-7db6d8ff4d-z7ds7\" (UID: \"df023805-6c34-47d2-9df9-3081da40656c\") " pod="kube-system/coredns-7db6d8ff4d-z7ds7" Apr 23 12:31:07 minikube kubelet[7651]: I0423 12:31:07.048768 7651 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/81363069-ef01-4b3f-8920-273e1c16014c-config-volume\") pod \"coredns-7db6d8ff4d-fqg4r\" (UID: \"81363069-ef01-4b3f-8920-273e1c16014c\") " pod="kube-system/coredns-7db6d8ff4d-fqg4r" Apr 23 12:31:07 minikube kubelet[7651]: I0423 12:31:07.048816 7651 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dhjm\" (UniqueName: \"kubernetes.io/projected/df023805-6c34-47d2-9df9-3081da40656c-kube-api-access-8dhjm\") pod \"coredns-7db6d8ff4d-z7ds7\" (UID: \"df023805-6c34-47d2-9df9-3081da40656c\") " pod="kube-system/coredns-7db6d8ff4d-z7ds7" Apr 23 12:31:07 minikube kubelet[7651]: I0423 12:31:07.048869 7651 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5g9z\" (UniqueName: \"kubernetes.io/projected/81363069-ef01-4b3f-8920-273e1c16014c-kube-api-access-v5g9z\") pod \"coredns-7db6d8ff4d-fqg4r\" (UID: \"81363069-ef01-4b3f-8920-273e1c16014c\") " pod="kube-system/coredns-7db6d8ff4d-fqg4r" Apr 23 12:31:07 minikube kubelet[7651]: E0423 12:31:07.949471 7651 configmap.go:199] Couldn't get configMap kube-system/kube-proxy: failed to sync configmap cache: timed out waiting for the condition Apr 23 12:31:07 minikube kubelet[7651]: E0423 12:31:07.950098 7651 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2d7b173b-2b4e-48d5-8dca-cdee293996e3-kube-proxy podName:2d7b173b-2b4e-48d5-8dca-cdee293996e3 nodeName:}" failed. No retries permitted until 2024-04-23 12:31:08.450052444 +0000 UTC m=+15.828502842 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-proxy" (UniqueName: "kubernetes.io/configmap/2d7b173b-2b4e-48d5-8dca-cdee293996e3-kube-proxy") pod "kube-proxy-jx599" (UID: "2d7b173b-2b4e-48d5-8dca-cdee293996e3") : failed to sync configmap cache: timed out waiting for the condition Apr 23 12:31:08 minikube kubelet[7651]: I0423 12:31:08.201608 7651 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64c37cca31f058fcec9a30d9358a7f0d32f7b4e35111b96605369b21bfd1a996" Apr 23 12:31:09 minikube kubelet[7651]: I0423 12:31:09.453615 7651 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7db6d8ff4d-z7ds7" podStartSLOduration=3.453598455 podStartE2EDuration="3.453598455s" podCreationTimestamp="2024-04-23 12:31:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-04-23 12:31:09.378861219 +0000 UTC m=+16.757311669" watchObservedRunningTime="2024-04-23 12:31:09.453598455 +0000 UTC m=+16.832048853" Apr 23 12:31:09 minikube kubelet[7651]: I0423 12:31:09.578610 7651 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7db6d8ff4d-fqg4r" podStartSLOduration=3.578591276 podStartE2EDuration="3.578591276s" podCreationTimestamp="2024-04-23 12:31:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-04-23 12:31:09.535766542 +0000 UTC m=+16.914216948" watchObservedRunningTime="2024-04-23 12:31:09.578591276 +0000 UTC m=+16.957041684" Apr 23 12:31:14 minikube kubelet[7651]: I0423 12:31:14.135989 7651 kuberuntime_manager.go:1523] "Updating runtime config through cri with podcidr" CIDR="10.244.0.0/24" Apr 23 12:31:14 minikube kubelet[7651]: I0423 12:31:14.145574 7651 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="10.244.0.0/24" Apr 23 12:31:45 minikube kubelet[7651]: E0423 12:31:45.226019 7651 kubelet.go:2511] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.109s" Apr 23 12:31:53 minikube kubelet[7651]: E0423 12:31:53.516850 7651 iptables.go:577] "Could not set up iptables canary" err=< Apr 23 12:31:53 minikube kubelet[7651]: error creating chain "KUBE-KUBELET-CANARY": exit status 3: Ignoring deprecated --wait-interval option. Apr 23 12:31:53 minikube kubelet[7651]: ip6tables v1.8.9 (legacy): can't initialize ip6tables table `nat': Table does not exist (do you need to insmod?) Apr 23 12:31:53 minikube kubelet[7651]: Perhaps ip6tables or your kernel needs to be upgraded. Apr 23 12:31:53 minikube kubelet[7651]: > table="nat" chain="KUBE-KUBELET-CANARY" Apr 23 12:32:17 minikube kubelet[7651]: E0423 12:32:17.321277 7651 kubelet.go:2511] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="3.199s" Apr 23 12:32:53 minikube kubelet[7651]: E0423 12:32:53.590300 7651 iptables.go:577] "Could not set up iptables canary" err=< Apr 23 12:32:53 minikube kubelet[7651]: error creating chain "KUBE-KUBELET-CANARY": exit status 3: Ignoring deprecated --wait-interval option. Apr 23 12:32:53 minikube kubelet[7651]: ip6tables v1.8.9 (legacy): can't initialize ip6tables table `nat': Table does not exist (do you need to insmod?) Apr 23 12:32:53 minikube kubelet[7651]: Perhaps ip6tables or your kernel needs to be upgraded. Apr 23 12:32:53 minikube kubelet[7651]: > table="nat" chain="KUBE-KUBELET-CANARY" Apr 23 12:33:16 minikube kubelet[7651]: E0423 12:33:16.221796 7651 kubelet.go:2511] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="2.072s"