Minikube: Unable to override node-name: lookup minikube on x:53: no such host

Created on 23 Mar 2020  ·  3Comments  ·  Source: kubernetes/minikube

The exact command to reproduce the issue:

$ minikube start \
  --profile profile-v1.18 \
  --kubernetes-version "v1.18.0-rc.1" \
  --vm-driver virtualbox \
  --extra-config=kubeadm.node-name=minikube \
  --extra-config=kubelet.hostname-override=minikube

The full output of the command that failed:

😄  [test-v1.18] minikube v1.8.2 on Darwin 10.15.3
    ▪ KUBECONFIG=/Users/i331370/.kube/config
✨  Using the virtualbox driver based on user configuration
🔥  Creating virtualbox VM (CPUs=2, Memory=4000MB, Disk=20000MB) ...
🐳  Preparing Kubernetes v1.18.0-rc.1 on Docker 19.03.6 ...
    ▪ kubeadm.node-name=minikube
    ▪ kubelet.hostname-override=minikube
🚀  Launching Kubernetes ...

💣  Error starting cluster: init failed. output: "-- stdout --\n[init] Using Kubernetes version: v1.18.0-rc.1\n[preflight] Running pre-flight checks\n[preflight] Pulling images required for setting up a Kubernetes cluster\n[preflight] This might take a minute or two, depending on the speed of your internet connection\n[preflight] You can also perform this action in beforehand using 'kubeadm config images pull'\n[kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n[kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n[kubelet-start] Starting the kubelet\n[certs] Using certificateDir folder \"/var/lib/minikube/certs\"\n[certs] Using existing ca certificate authority\n[certs] Using existing apiserver certificate and key on disk\n[certs] Generating \"apiserver-kubelet-client\" certificate and key\n[certs] Generating \"front-proxy-ca\" certificate and key\n[certs] Generating \"front-proxy-client\" certificate and key\n[certs] Generating \"etcd/ca\" certificate and key\n[certs] Generating \"etcd/server\" certificate and key\n[certs] etcd/server serving cert is signed for DNS names [minikube localhost] and IPs [192.168.99.105 127.0.0.1 ::1]\n[certs] Generating \"etcd/peer\" certificate and key\n[certs] etcd/peer serving cert is signed for DNS names [minikube localhost] and IPs [192.168.99.105 127.0.0.1 ::1]\n[certs] Generating \"etcd/healthcheck-client\" certificate and key\n[certs] Generating \"apiserver-etcd-client\" certificate and key\n[certs] Generating \"sa\" key and public key\n[kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n[kubeconfig] Writing \"admin.conf\" kubeconfig file\n[kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n[kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n[kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n[control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n[control-plane] Creating static Pod manifest for \"kube-apiserver\"\n[control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n[control-plane] Creating static Pod manifest for \"kube-scheduler\"\n[etcd] Creating static Pod manifest for local etcd in \"/etc/kubernetes/manifests\"\n[wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\". This can take up to 4m0s\n[apiclient] All control plane components are healthy after 13.003771 seconds\n[upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n[kubelet] Creating a ConfigMap \"kubelet-config-1.18\" in namespace kube-system with the configuration for the kubelets in the cluster\n[kubelet-check] Initial timeout of 40s passed.\n[kubelet-check] It seems like the kubelet isn't running or healthy.\n[kubelet-check] The HTTP call equal to 'curl -sSL http://localhost:10248/healthz' failed with error: Get http://localhost:10248/healthz: dial tcp 127.0.0.1:10248: connect: connection refused.\n\n-- /stdout --\n** stderr ** \nW0323 16:58:21.182181    2851 configset.go:202] WARNING: kubeadm cannot validate component configs for API groups [kubelet.config.k8s.io kubeproxy.config.k8s.io]\n\t[WARNING Hostname]: hostname \"minikube\" could not be reached\n\t[WARNING Hostname]: hostname \"minikube\": lookup minikube on 10.0.2.3:53: no such host\n\t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\nW0323 16:58:24.553548    2851 manifests.go:225] the default kube-apiserver authorization-mode is \"Node,RBAC\"; using \"Node,RBAC\"\nW0323 16:58:24.554639    2851 manifests.go:225] the default kube-apiserver authorization-mode is \"Node,RBAC\"; using \"Node,RBAC\"\nerror execution phase upload-config/kubelet: Error writing Crisocket information for the control-plane node: timed out waiting for the condition\nTo see the stack trace of this error execute with --v=5 or higher\n\n** /stderr **": /bin/bash -c "sudo env PATH=/var/lib/minikube/binaries/v1.18.0-rc.1:$PATH kubeadm init --config /var/tmp/minikube/kubeadm.yaml --node-name=minikube --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests,DirAvailable--var-lib-minikube,DirAvailable--var-lib-minikube-etcd,FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml,FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml,FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml,FileAvailable--etc-kubernetes-manifests-etcd.yaml,Port-10250,Swap,SystemVerification": Process exited with status 1
stdout:
[init] Using Kubernetes version: v1.18.0-rc.1
[preflight] Running pre-flight checks
[preflight] Pulling images required for setting up a Kubernetes cluster
[preflight] This might take a minute or two, depending on the speed of your internet connection
[preflight] You can also perform this action in beforehand using 'kubeadm config images pull'
[kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
[kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
[kubelet-start] Starting the kubelet
[certs] Using certificateDir folder "/var/lib/minikube/certs"
[certs] Using existing ca certificate authority
[certs] Using existing apiserver certificate and key on disk
[certs] Generating "apiserver-kubelet-client" certificate and key
[certs] Generating "front-proxy-ca" certificate and key
[certs] Generating "front-proxy-client" certificate and key
[certs] Generating "etcd/ca" certificate and key
[certs] Generating "etcd/server" certificate and key
[certs] etcd/server serving cert is signed for DNS names [minikube localhost] and IPs [192.168.99.105 127.0.0.1 ::1]
[certs] Generating "etcd/peer" certificate and key
[certs] etcd/peer serving cert is signed for DNS names [minikube localhost] and IPs [192.168.99.105 127.0.0.1 ::1]
[certs] Generating "etcd/healthcheck-client" certificate and key
[certs] Generating "apiserver-etcd-client" certificate and key
[certs] Generating "sa" key and public key
[kubeconfig] Using kubeconfig folder "/etc/kubernetes"
[kubeconfig] Writing "admin.conf" kubeconfig file
[kubeconfig] Writing "kubelet.conf" kubeconfig file
[kubeconfig] Writing "controller-manager.conf" kubeconfig file
[kubeconfig] Writing "scheduler.conf" kubeconfig file
[control-plane] Using manifest folder "/etc/kubernetes/manifests"
[control-plane] Creating static Pod manifest for "kube-apiserver"
[control-plane] Creating static Pod manifest for "kube-controller-manager"
[control-plane] Creating static Pod manifest for "kube-scheduler"
[etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
[wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests". This can take up to 4m0s
[apiclient] All control plane components are healthy after 13.003771 seconds
[upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
[kubelet] Creating a ConfigMap "kubelet-config-1.18" in namespace kube-system with the configuration for the kubelets in the cluster
[kubelet-check] Initial timeout of 40s passed.
[kubelet-check] It seems like the kubelet isn't running or healthy.
[kubelet-check] The HTTP call equal to 'curl -sSL http://localhost:10248/healthz' failed with error: Get http://localhost:10248/healthz: dial tcp 127.0.0.1:10248: connect: connection refused.

stderr:
W0323 16:58:21.182181    2851 configset.go:202] WARNING: kubeadm cannot validate component configs for API groups [kubelet.config.k8s.io kubeproxy.config.k8s.io]
    [WARNING Hostname]: hostname "minikube" could not be reached
    [WARNING Hostname]: hostname "minikube": lookup minikube on 10.0.2.3:53: no such host
    [WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
W0323 16:58:24.553548    2851 manifests.go:225] the default kube-apiserver authorization-mode is "Node,RBAC"; using "Node,RBAC"
W0323 16:58:24.554639    2851 manifests.go:225] the default kube-apiserver authorization-mode is "Node,RBAC"; using "Node,RBAC"
error execution phase upload-config/kubelet: Error writing Crisocket information for the control-plane node: timed out waiting for the condition
To see the stack trace of this error execute with --v=5 or higher


😿  minikube is exiting due to an error. If the above message is not useful, open an issue:
👉  https://github.com/kubernetes/minikube/issues/new/choose

The output of the minikube logs command:

$ minikube logs -p profile-v1.18

==> Docker <==
-- Logs begin at Mon 2020-03-23 16:44:38 UTC, end at Mon 2020-03-23 16:52:25 UTC. --
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.739963449Z" level=info msg="loading plugin "io.containerd.differ.v1.walking"..." type=io.containerd.differ.v1
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.740019389Z" level=info msg="loading plugin "io.containerd.gc.v1.scheduler"..." type=io.containerd.gc.v1
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.740069743Z" level=info msg="loading plugin "io.containerd.service.v1.containers-service"..." type=io.containerd.service.v1
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.740081504Z" level=info msg="loading plugin "io.containerd.service.v1.content-service"..." type=io.containerd.service.v1
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.740089997Z" level=info msg="loading plugin "io.containerd.service.v1.diff-service"..." type=io.containerd.service.v1
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.740098927Z" level=info msg="loading plugin "io.containerd.service.v1.images-service"..." type=io.containerd.service.v1
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.740107418Z" level=info msg="loading plugin "io.containerd.service.v1.leases-service"..." type=io.containerd.service.v1
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.740116625Z" level=info msg="loading plugin "io.containerd.service.v1.namespaces-service"..." type=io.containerd.service.v1
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.740124752Z" level=info msg="loading plugin "io.containerd.service.v1.snapshots-service"..." type=io.containerd.service.v1
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.740135846Z" level=info msg="loading plugin "io.containerd.runtime.v1.linux"..." type=io.containerd.runtime.v1
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.740192086Z" level=info msg="loading plugin "io.containerd.runtime.v2.task"..." type=io.containerd.runtime.v2
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.740303086Z" level=info msg="loading plugin "io.containerd.monitor.v1.cgroups"..." type=io.containerd.monitor.v1
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.740666051Z" level=info msg="loading plugin "io.containerd.service.v1.tasks-service"..." type=io.containerd.service.v1
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.740696495Z" level=info msg="loading plugin "io.containerd.internal.v1.restart"..." type=io.containerd.internal.v1
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.740726141Z" level=info msg="loading plugin "io.containerd.grpc.v1.containers"..." type=io.containerd.grpc.v1
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.740769314Z" level=info msg="loading plugin "io.containerd.grpc.v1.content"..." type=io.containerd.grpc.v1
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.740778991Z" level=info msg="loading plugin "io.containerd.grpc.v1.diff"..." type=io.containerd.grpc.v1
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.740786957Z" level=info msg="loading plugin "io.containerd.grpc.v1.events"..." type=io.containerd.grpc.v1
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.740794282Z" level=info msg="loading plugin "io.containerd.grpc.v1.healthcheck"..." type=io.containerd.grpc.v1
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.740802279Z" level=info msg="loading plugin "io.containerd.grpc.v1.images"..." type=io.containerd.grpc.v1
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.740809801Z" level=info msg="loading plugin "io.containerd.grpc.v1.leases"..." type=io.containerd.grpc.v1
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.740817049Z" level=info msg="loading plugin "io.containerd.grpc.v1.namespaces"..." type=io.containerd.grpc.v1
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.740824343Z" level=info msg="loading plugin "io.containerd.internal.v1.opt"..." type=io.containerd.internal.v1
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.740851384Z" level=info msg="loading plugin "io.containerd.grpc.v1.snapshots"..." type=io.containerd.grpc.v1
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.740861156Z" level=info msg="loading plugin "io.containerd.grpc.v1.tasks"..." type=io.containerd.grpc.v1
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.740868769Z" level=info msg="loading plugin "io.containerd.grpc.v1.version"..." type=io.containerd.grpc.v1
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.740876167Z" level=info msg="loading plugin "io.containerd.grpc.v1.introspection"..." type=io.containerd.grpc.v1
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.741012489Z" level=info msg=serving... address="/var/run/docker/containerd/containerd-debug.sock"
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.741061541Z" level=info msg=serving... address="/var/run/docker/containerd/containerd.sock"
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.741072508Z" level=info msg="containerd successfully booted in 0.004406s"
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.752227050Z" level=info msg="parsed scheme: \"unix\"" module=grpc
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.752370277Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.752495671Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/docker/containerd/containerd.sock 0  <nil>}] <nil>}" module=grpc
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.752562888Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.754992630Z" level=info msg="parsed scheme: \"unix\"" module=grpc
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.755033402Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.755058876Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/docker/containerd/containerd.sock 0  <nil>}] <nil>}" module=grpc
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.755078448Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.773694886Z" level=warning msg="Your kernel does not support cgroup blkio weight"
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.773735035Z" level=warning msg="Your kernel does not support cgroup blkio weight_device"
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.773742195Z" level=warning msg="Your kernel does not support cgroup blkio throttle.read_bps_device"
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.773747284Z" level=warning msg="Your kernel does not support cgroup blkio throttle.write_bps_device"
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.773752373Z" level=warning msg="Your kernel does not support cgroup blkio throttle.read_iops_device"
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.773757508Z" level=warning msg="Your kernel does not support cgroup blkio throttle.write_iops_device"
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.774037951Z" level=info msg="Loading containers: start."
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.891735369Z" level=info msg="Default bridge (docker0) is assigned with an IP address 172.17.0.0/16. Daemon option --bip can be used to set a preferred IP address"
Mar 23 16:45:10 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:10.926117240Z" level=info msg="Loading containers: done."
Mar 23 16:45:11 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:11.339575304Z" level=info msg="Docker daemon" commit=369ce74a3c graphdriver(s)=overlay2 version=19.03.6
Mar 23 16:45:11 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:11.339665291Z" level=info msg="Daemon has completed initialization"
Mar 23 16:45:11 test-v1.18 systemd[1]: Started Docker Application Container Engine.
Mar 23 16:45:11 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:11.376904888Z" level=info msg="API listen on /var/run/docker.sock"
Mar 23 16:45:11 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:11.377155448Z" level=info msg="API listen on [::]:2376"
Mar 23 16:45:17 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:17.622025760Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/00eaca5acffc072c53ee0db156ba7e7a015ca23357166281449370d28967c4c3/shim.sock" debug=false pid=3460
Mar 23 16:45:17 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:17.648544066Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/7c0d506ca234975d8f96214aac439b6e443ae0511141ed026a59ec9bf9230504/shim.sock" debug=false pid=3478
Mar 23 16:45:17 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:17.657949052Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/ca1f6495340577c7ffbe8a61227c3b6a479e5621f4fae5affcab2ff50e519c88/shim.sock" debug=false pid=3480
Mar 23 16:45:17 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:17.722335145Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/6d53e84e3bb280d5d37adfc7b16a427a8792402ff55038fd27780a7cb98b7b89/shim.sock" debug=false pid=3523
Mar 23 16:45:17 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:17.928491986Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/b966653912a458891549b84293140a1e74958a970b1f6bed9d1d2f6eceec0047/shim.sock" debug=false pid=3607
Mar 23 16:45:17 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:17.942994630Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/607fa8f8e39b07624ecebe530fcd2d7203549892c436806b5134301859ca4d2a/shim.sock" debug=false pid=3618
Mar 23 16:45:17 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:17.949803358Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/4eeda3a79711abc3704c6b6bee92d0367f6134c4f1523e20566fe7825c6834a6/shim.sock" debug=false pid=3631
Mar 23 16:45:18 test-v1.18 dockerd[2591]: time="2020-03-23T16:45:18.077544672Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/09afecbf4abde9c25ae301c3bde0d4626644b420bbb87b2860d0457fef06921c/shim.sock" debug=false pid=3686

==> container status <==
CONTAINER           IMAGE               CREATED             STATE               NAME                      ATTEMPT             POD ID
09afecbf4abde       bce13e0cc95a6       7 minutes ago       Running             kube-scheduler            0                   6d53e84e3bb28
4eeda3a79711a       b4f6b0bffa351       7 minutes ago       Running             kube-controller-manager   0                   ca1f649534057
607fa8f8e39b0       5347d260989ad       7 minutes ago       Running             kube-apiserver            0                   7c0d506ca2349
b966653912a45       303ce5db0e90d       7 minutes ago       Running             etcd                      0                   00eaca5acffc0

==> dmesg <==
[Mar23 16:44] You have booted with nomodeset. This means your GPU drivers are DISABLED
[  +0.000000] Any video related functionality will be severely degraded, and you may not even be able to suspend the system properly
[  +0.000001] Unless you actually understand what nomodeset does, you should reboot without enabling it
[  +0.201394] acpi PNP0A03:00: fail to add MMCONFIG information, can't access extended PCI configuration space under this bridge.
[  +2.803375] platform regulatory.0: Direct firmware load for regulatory.db failed with error -2
[ +10.296134] systemd[1]: Failed to bump fs.file-max, ignoring: Invalid argument
[  +0.008755] systemd-fstab-generator[1349]: Ignoring "noauto" for root device
[  +0.001580] systemd[1]: File /usr/lib/systemd/system/systemd-journald.service:12 configures an IP firewall (IPAddressDeny=any), but the local system does not support BPF/cgroup based firewalling.
[  +0.000002] systemd[1]: Proceeding WITHOUT firewalling in effect! (This warning is only shown for the first loaded unit using IP firewalling.)
[  +0.422280] vboxvideo: loading out-of-tree module taints kernel.
[  +0.000039] vboxvideo: Unknown symbol ttm_bo_mmap (err -2)
[  +0.000010] vboxvideo: Unknown symbol ttm_bo_global_release (err -2)
[  +0.000009] vboxvideo: Unknown symbol ttm_bo_manager_func (err -2)
[  +0.000003] vboxvideo: Unknown symbol ttm_bo_global_init (err -2)
[  +0.000007] vboxvideo: Unknown symbol ttm_bo_device_release (err -2)
[  +0.000013] vboxvideo: Unknown symbol ttm_bo_kunmap (err -2)
[  +0.000006] vboxvideo: Unknown symbol ttm_bo_del_sub_from_lru (err -2)
[  +0.000007] vboxvideo: Unknown symbol ttm_bo_device_init (err -2)
[  +0.000039] vboxvideo: Unknown symbol ttm_bo_init_mm (err -2)
[  +0.000018] vboxvideo: Unknown symbol ttm_bo_dma_acc_size (err -2)
[  +0.000004] vboxvideo: Unknown symbol ttm_tt_init (err -2)
[  +0.000001] vboxvideo: Unknown symbol ttm_bo_kmap (err -2)
[  +0.000005] vboxvideo: Unknown symbol ttm_bo_add_to_lru (err -2)
[  +0.000004] vboxvideo: Unknown symbol ttm_mem_global_release (err -2)
[  +0.000002] vboxvideo: Unknown symbol ttm_mem_global_init (err -2)
[  +0.000010] vboxvideo: Unknown symbol ttm_bo_init (err -2)
[  +0.000001] vboxvideo: Unknown symbol ttm_bo_validate (err -2)
[  +0.000004] vboxvideo: Unknown symbol ttm_bo_put (err -2)
[  +0.000003] vboxvideo: Unknown symbol ttm_tt_fini (err -2)
[  +0.000002] vboxvideo: Unknown symbol ttm_bo_eviction_valuable (err -2)
[  +0.025140] vgdrvHeartbeatInit: Setting up heartbeat to trigger every 2000 milliseconds
[  +0.000337] vboxguest: misc device minor 58, IRQ 20, I/O port d020, MMIO at 00000000f0000000 (size 0x400000)
[  +0.195456] hpet1: lost 699 rtc interrupts
[  +0.036591] VBoxService 5.2.32 r132073 (verbosity: 0) linux.amd64 (Jul 12 2019 10:32:28) release log
              00:00:00.004066 main     Log opened 2020-03-23T16:44:38.792228000Z
[  +0.000058] 00:00:00.004182 main     OS Product: Linux
[  +0.000030] 00:00:00.004217 main     OS Release: 4.19.94
[  +0.000028] 00:00:00.004246 main     OS Version: #1 SMP Fri Mar 6 11:41:28 PST 2020
[  +0.000035] 00:00:00.004273 main     Executable: /usr/sbin/VBoxService
              00:00:00.004274 main     Process ID: 2093
              00:00:00.004274 main     Package type: LINUX_64BITS_GENERIC
[  +0.000029] 00:00:00.004310 main     5.2.32 r132073 started. Verbose level = 0
[  +0.002175] 00:00:00.006476 main     Error: Service 'control' failed to initialize: VERR_INVALID_PARAMETER
[  +0.000078] 00:00:00.006561 main     Session 0 is about to close ...
[  +0.000045] 00:00:00.006607 main     Stopping all guest processes ...
[  +0.000028] 00:00:00.006636 main     Closing all guest files ...
[  +0.000703] 00:00:00.007330 main     Ended.
[  +0.422174] hpet1: lost 11 rtc interrupts
[  +0.140077] NFSD: the nfsdcld client tracking upcall will be removed in 3.10. Please transition to using nfsdcltrack.
[ +12.085718] systemd-fstab-generator[2343]: Ignoring "noauto" for root device
[  +0.153867] systemd-fstab-generator[2359]: Ignoring "noauto" for root device
[  +0.145606] systemd-fstab-generator[2375]: Ignoring "noauto" for root device
[Mar23 16:45] kauditd_printk_skb: 65 callbacks suppressed
[  +1.364183] systemd-fstab-generator[2795]: Ignoring "noauto" for root device
[  +1.348352] systemd-fstab-generator[2995]: Ignoring "noauto" for root device
[  +3.476763] kauditd_printk_skb: 107 callbacks suppressed
[Mar23 16:46] NFSD: Unable to end grace period: -110

==> kernel <==
 16:52:25 up 8 min,  0 users,  load average: 0.39, 0.54, 0.33
Linux test-v1.18 4.19.94 #1 SMP Fri Mar 6 11:41:28 PST 2020 x86_64 GNU/Linux
PRETTY_NAME="Buildroot 2019.02.9"

==> kube-apiserver [607fa8f8e39b] <==
I0323 16:45:20.148614       1 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://127.0.0.1:2379  <nil> 0 <nil>}]
I0323 16:45:20.156746       1 client.go:361] parsed scheme: "endpoint"
I0323 16:45:20.156778       1 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://127.0.0.1:2379  <nil> 0 <nil>}]
W0323 16:45:20.277248       1 genericapiserver.go:409] Skipping API batch/v2alpha1 because it has no resources.
W0323 16:45:20.284530       1 genericapiserver.go:409] Skipping API discovery.k8s.io/v1alpha1 because it has no resources.
W0323 16:45:20.294038       1 genericapiserver.go:409] Skipping API node.k8s.io/v1alpha1 because it has no resources.
W0323 16:45:20.310631       1 genericapiserver.go:409] Skipping API rbac.authorization.k8s.io/v1alpha1 because it has no resources.
W0323 16:45:20.313764       1 genericapiserver.go:409] Skipping API scheduling.k8s.io/v1alpha1 because it has no resources.
W0323 16:45:20.326577       1 genericapiserver.go:409] Skipping API storage.k8s.io/v1alpha1 because it has no resources.
W0323 16:45:20.346383       1 genericapiserver.go:409] Skipping API apps/v1beta2 because it has no resources.
W0323 16:45:20.346582       1 genericapiserver.go:409] Skipping API apps/v1beta1 because it has no resources.
I0323 16:45:20.356941       1 plugins.go:158] Loaded 12 mutating admission controller(s) successfully in the following order: NamespaceLifecycle,LimitRanger,ServiceAccount,NodeRestriction,TaintNodesByCondition,Priority,DefaultTolerationSeconds,DefaultStorageClass,StorageObjectInUseProtection,RuntimeClass,DefaultIngressClass,MutatingAdmissionWebhook.
I0323 16:45:20.357109       1 plugins.go:161] Loaded 10 validating admission controller(s) successfully in the following order: LimitRanger,ServiceAccount,Priority,PersistentVolumeClaimResize,RuntimeClass,CertificateApproval,CertificateSigning,CertificateSubjectRestriction,ValidatingAdmissionWebhook,ResourceQuota.
I0323 16:45:20.358673       1 client.go:361] parsed scheme: "endpoint"
I0323 16:45:20.358737       1 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://127.0.0.1:2379  <nil> 0 <nil>}]
I0323 16:45:20.366452       1 client.go:361] parsed scheme: "endpoint"
I0323 16:45:20.366581       1 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://127.0.0.1:2379  <nil> 0 <nil>}]
I0323 16:45:22.305946       1 dynamic_cafile_content.go:167] Starting request-header::/var/lib/minikube/certs/front-proxy-ca.crt
I0323 16:45:22.306113       1 dynamic_cafile_content.go:167] Starting client-ca-bundle::/var/lib/minikube/certs/ca.crt
I0323 16:45:22.306512       1 dynamic_serving_content.go:130] Starting serving-cert::/var/lib/minikube/certs/apiserver.crt::/var/lib/minikube/certs/apiserver.key
I0323 16:45:22.307043       1 secure_serving.go:178] Serving securely on [::]:8443
I0323 16:45:22.307874       1 tlsconfig.go:240] Starting DynamicServingCertificateController
I0323 16:45:22.308250       1 crd_finalizer.go:266] Starting CRDFinalizer
I0323 16:45:22.308316       1 apiservice_controller.go:94] Starting APIServiceRegistrationController
I0323 16:45:22.308367       1 cache.go:32] Waiting for caches to sync for APIServiceRegistrationController controller
I0323 16:45:22.310797       1 controller.go:81] Starting OpenAPI AggregationController
I0323 16:45:22.310837       1 autoregister_controller.go:141] Starting autoregister controller
I0323 16:45:22.310841       1 cache.go:32] Waiting for caches to sync for autoregister controller
I0323 16:45:22.311419       1 cluster_authentication_trust_controller.go:440] Starting cluster_authentication_trust_controller controller
I0323 16:45:22.311440       1 shared_informer.go:223] Waiting for caches to sync for cluster_authentication_trust_controller
I0323 16:45:22.311659       1 available_controller.go:387] Starting AvailableConditionController
I0323 16:45:22.311675       1 cache.go:32] Waiting for caches to sync for AvailableConditionController controller
I0323 16:45:22.312060       1 crdregistration_controller.go:111] Starting crd-autoregister controller
I0323 16:45:22.312078       1 shared_informer.go:223] Waiting for caches to sync for crd-autoregister
I0323 16:45:22.312098       1 dynamic_cafile_content.go:167] Starting client-ca-bundle::/var/lib/minikube/certs/ca.crt
I0323 16:45:22.312241       1 dynamic_cafile_content.go:167] Starting request-header::/var/lib/minikube/certs/front-proxy-ca.crt
I0323 16:45:22.372442       1 controller.go:86] Starting OpenAPI controller
I0323 16:45:22.375452       1 customresource_discovery_controller.go:209] Starting DiscoveryController
I0323 16:45:22.375723       1 naming_controller.go:291] Starting NamingConditionController
I0323 16:45:22.375872       1 establishing_controller.go:76] Starting EstablishingController
I0323 16:45:22.375994       1 nonstructuralschema_controller.go:186] Starting NonStructuralSchemaConditionController
I0323 16:45:22.376104       1 apiapproval_controller.go:186] Starting KubernetesAPIApprovalPolicyConformantConditionController
E0323 16:45:22.384750       1 controller.go:152] Unable to remove old endpoints from kubernetes service: StorageError: key not found, Code: 1, Key: /registry/masterleases/192.168.99.104, ResourceVersion: 0, AdditionalErrorMsg:
I0323 16:45:22.423638       1 shared_informer.go:230] Caches are synced for crd-autoregister
I0323 16:45:22.423813       1 cache.go:39] Caches are synced for autoregister controller
I0323 16:45:22.424130       1 shared_informer.go:230] Caches are synced for cluster_authentication_trust_controller
I0323 16:45:22.424190       1 cache.go:39] Caches are synced for AvailableConditionController controller
I0323 16:45:22.515190       1 cache.go:39] Caches are synced for APIServiceRegistrationController controller
I0323 16:45:23.305786       1 controller.go:130] OpenAPI AggregationController: action for item : Nothing (removed from the queue).
I0323 16:45:23.305827       1 controller.go:130] OpenAPI AggregationController: action for item k8s_internal_local_delegation_chain_0000000000: Nothing (removed from the queue).
I0323 16:45:23.326510       1 storage_scheduling.go:134] created PriorityClass system-node-critical with value 2000001000
I0323 16:45:23.343832       1 storage_scheduling.go:134] created PriorityClass system-cluster-critical with value 2000000000
I0323 16:45:23.343880       1 storage_scheduling.go:143] all system priority classes are created successfully or already exist.
I0323 16:45:24.230737       1 controller.go:606] quota admission added evaluator for: roles.rbac.authorization.k8s.io
I0323 16:45:24.315291       1 controller.go:606] quota admission added evaluator for: rolebindings.rbac.authorization.k8s.io
W0323 16:45:24.456316       1 lease.go:224] Resetting endpoints for master service "kubernetes" to [192.168.99.104]
I0323 16:45:24.457442       1 controller.go:606] quota admission added evaluator for: endpoints
I0323 16:45:24.462595       1 controller.go:606] quota admission added evaluator for: endpointslices.discovery.k8s.io
I0323 16:45:25.852828       1 controller.go:606] quota admission added evaluator for: leases.coordination.k8s.io
I0323 16:45:26.196790       1 controller.go:606] quota admission added evaluator for: serviceaccounts

==> kube-controller-manager [4eeda3a79711] <==
I0323 16:45:31.243291       1 controllermanager.go:533] Started "endpointslice"
I0323 16:45:31.243317       1 endpointslice_controller.go:213] Starting endpoint slice controller
I0323 16:45:31.243740       1 shared_informer.go:223] Waiting for caches to sync for endpoint_slice
I0323 16:45:31.493012       1 controllermanager.go:533] Started "podgc"
I0323 16:45:31.493332       1 gc_controller.go:89] Starting GC controller
I0323 16:45:31.493427       1 shared_informer.go:223] Waiting for caches to sync for GC
I0323 16:45:32.401120       1 controllermanager.go:533] Started "garbagecollector"
W0323 16:45:32.401159       1 controllermanager.go:525] Skipping "root-ca-cert-publisher"
I0323 16:45:32.401521       1 garbagecollector.go:133] Starting garbage collector controller
I0323 16:45:32.402545       1 shared_informer.go:223] Waiting for caches to sync for garbage collector
I0323 16:45:32.402627       1 graph_builder.go:282] GraphBuilder running
I0323 16:45:32.426455       1 controllermanager.go:533] Started "tokencleaner"
I0323 16:45:32.426483       1 tokencleaner.go:118] Starting token cleaner controller
I0323 16:45:32.426820       1 shared_informer.go:223] Waiting for caches to sync for token_cleaner
I0323 16:45:32.426826       1 shared_informer.go:230] Caches are synced for token_cleaner
I0323 16:45:32.453107       1 controllermanager.go:533] Started "daemonset"
I0323 16:45:32.453151       1 daemon_controller.go:257] Starting daemon sets controller
I0323 16:45:32.453461       1 shared_informer.go:223] Waiting for caches to sync for daemon sets
I0323 16:45:32.643392       1 controllermanager.go:533] Started "job"
I0323 16:45:32.643500       1 job_controller.go:144] Starting job controller
I0323 16:45:32.643509       1 shared_informer.go:223] Waiting for caches to sync for job
I0323 16:45:32.893515       1 controllermanager.go:533] Started "statefulset"
I0323 16:45:32.894304       1 shared_informer.go:223] Waiting for caches to sync for resource quota
I0323 16:45:32.894419       1 stateful_set.go:146] Starting stateful set controller
I0323 16:45:32.894437       1 shared_informer.go:223] Waiting for caches to sync for stateful set
I0323 16:45:32.932986       1 shared_informer.go:230] Caches are synced for namespace
I0323 16:45:32.943437       1 shared_informer.go:230] Caches are synced for HPA
I0323 16:45:32.943821       1 shared_informer.go:230] Caches are synced for expand
I0323 16:45:32.944422       1 shared_informer.go:230] Caches are synced for service account
I0323 16:45:32.945313       1 shared_informer.go:230] Caches are synced for PVC protection
I0323 16:45:32.953807       1 shared_informer.go:230] Caches are synced for ClusterRoleAggregator
I0323 16:45:32.958220       1 shared_informer.go:230] Caches are synced for ReplicationController
I0323 16:45:32.970897       1 shared_informer.go:230] Caches are synced for persistent volume
E0323 16:45:32.993642       1 clusterroleaggregation_controller.go:181] view failed with : Operation cannot be fulfilled on clusterroles.rbac.authorization.k8s.io "view": the object has been modified; please apply your changes to the latest version and try again
I0323 16:45:32.994165       1 shared_informer.go:230] Caches are synced for ReplicaSet
E0323 16:45:32.994702       1 clusterroleaggregation_controller.go:181] edit failed with : Operation cannot be fulfilled on clusterroles.rbac.authorization.k8s.io "edit": the object has been modified; please apply your changes to the latest version and try again
I0323 16:45:32.995336       1 shared_informer.go:230] Caches are synced for certificate-csrapproving
I0323 16:45:32.997506       1 shared_informer.go:230] Caches are synced for certificate-csrsigning
I0323 16:45:32.997563       1 shared_informer.go:230] Caches are synced for GC
I0323 16:45:32.997623       1 shared_informer.go:230] Caches are synced for bootstrap_signer
I0323 16:45:32.997677       1 shared_informer.go:230] Caches are synced for TTL
I0323 16:45:33.006281       1 shared_informer.go:230] Caches are synced for PV protection
I0323 16:45:33.042775       1 shared_informer.go:230] Caches are synced for disruption
I0323 16:45:33.042906       1 disruption.go:339] Sending events to api server.
I0323 16:45:33.044441       1 shared_informer.go:230] Caches are synced for deployment
E0323 16:45:33.045838       1 clusterroleaggregation_controller.go:181] edit failed with : Operation cannot be fulfilled on clusterroles.rbac.authorization.k8s.io "edit": the object has been modified; please apply your changes to the latest version and try again
I0323 16:45:33.240954       1 shared_informer.go:230] Caches are synced for endpoint
I0323 16:45:33.243992       1 shared_informer.go:230] Caches are synced for endpoint_slice
I0323 16:45:33.346439       1 shared_informer.go:230] Caches are synced for taint
I0323 16:45:33.346584       1 taint_manager.go:187] Starting NoExecuteTaintManager
I0323 16:45:33.443894       1 shared_informer.go:230] Caches are synced for job
I0323 16:45:33.453840       1 shared_informer.go:230] Caches are synced for daemon sets
I0323 16:45:33.494524       1 shared_informer.go:230] Caches are synced for stateful set
I0323 16:45:33.505156       1 shared_informer.go:230] Caches are synced for garbage collector
I0323 16:45:33.505194       1 garbagecollector.go:142] Garbage collector: all resource monitors have synced. Proceeding to collect garbage
I0323 16:45:33.532314       1 shared_informer.go:230] Caches are synced for attach detach
I0323 16:45:33.594710       1 shared_informer.go:230] Caches are synced for resource quota
I0323 16:45:33.596365       1 shared_informer.go:230] Caches are synced for resource quota
I0323 16:45:33.895099       1 shared_informer.go:223] Waiting for caches to sync for garbage collector
I0323 16:45:33.895276       1 shared_informer.go:230] Caches are synced for garbage collector

==> kube-scheduler [09afecbf4abd] <==
I0323 16:45:18.562514       1 registry.go:150] Registering EvenPodsSpread predicate and priority function
I0323 16:45:18.563527       1 registry.go:150] Registering EvenPodsSpread predicate and priority function
I0323 16:45:19.141334       1 serving.go:313] Generated self-signed cert in-memory
W0323 16:45:22.404198       1 authentication.go:349] Unable to get configmap/extension-apiserver-authentication in kube-system.  Usually fixed by 'kubectl create rolebinding -n kube-system ROLEBINDING_NAME --role=extension-apiserver-authentication-reader --serviceaccount=YOUR_NS:YOUR_SA'
W0323 16:45:22.404548       1 authentication.go:297] Error looking up in-cluster authentication configuration: configmaps "extension-apiserver-authentication" is forbidden: User "system:kube-scheduler" cannot get resource "configmaps" in API group "" in the namespace "kube-system"
W0323 16:45:22.404635       1 authentication.go:298] Continuing without authentication configuration. This may treat all requests as anonymous.
W0323 16:45:22.406313       1 authentication.go:299] To require authentication configuration lookup to succeed, set --authentication-tolerate-lookup-failure=false
I0323 16:45:22.431637       1 registry.go:150] Registering EvenPodsSpread predicate and priority function
I0323 16:45:22.432053       1 registry.go:150] Registering EvenPodsSpread predicate and priority function
W0323 16:45:22.436674       1 authorization.go:47] Authorization is disabled
W0323 16:45:22.436868       1 authentication.go:40] Authentication is disabled
I0323 16:45:22.437264       1 deprecated_insecure_serving.go:51] Serving healthz insecurely on [::]:10251
I0323 16:45:22.438644       1 configmap_cafile_content.go:202] Starting client-ca::kube-system::extension-apiserver-authentication::client-ca-file
I0323 16:45:22.438670       1 shared_informer.go:223] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file
I0323 16:45:22.439150       1 secure_serving.go:178] Serving securely on 127.0.0.1:10259
I0323 16:45:22.439276       1 tlsconfig.go:240] Starting DynamicServingCertificateController
E0323 16:45:22.441239       1 reflector.go:178] k8s.io/client-go/informers/factory.go:135: Failed to list *v1beta1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User "system:kube-scheduler" cannot list resource "poddisruptionbudgets" in API group "policy" at the cluster scope
E0323 16:45:22.441769       1 reflector.go:178] k8s.io/kubernetes/cmd/kube-scheduler/app/server.go:233: Failed to list *v1.Pod: pods is forbidden: User "system:kube-scheduler" cannot list resource "pods" in API group "" at the cluster scope
E0323 16:45:22.442078       1 reflector.go:178] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.Node: nodes is forbidden: User "system:kube-scheduler" cannot list resource "nodes" in API group "" at the cluster scope
E0323 16:45:22.442378       1 reflector.go:178] k8s.io/apiserver/pkg/server/dynamiccertificates/configmap_cafile_content.go:206: Failed to list *v1.ConfigMap: configmaps "extension-apiserver-authentication" is forbidden: User "system:kube-scheduler" cannot list resource "configmaps" in API group "" in the namespace "kube-system"
E0323 16:45:22.443141       1 reflector.go:178] k8s.io/kubernetes/cmd/kube-scheduler/app/server.go:233: Failed to list *v1.Pod: pods is forbidden: User "system:kube-scheduler" cannot list resource "pods" in API group "" at the cluster scope
E0323 16:45:22.443147       1 reflector.go:178] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.Service: services is forbidden: User "system:kube-scheduler" cannot list resource "services" in API group "" at the cluster scope
E0323 16:45:22.443419       1 reflector.go:178] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csinodes" in API group "storage.k8s.io" at the cluster scope
E0323 16:45:22.443507       1 reflector.go:178] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "storageclasses" in API group "storage.k8s.io" at the cluster scope
E0323 16:45:22.443634       1 reflector.go:178] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumeclaims" in API group "" at the cluster scope
E0323 16:45:22.443639       1 reflector.go:178] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumes" in API group "" at the cluster scope
E0323 16:45:22.444489       1 reflector.go:178] k8s.io/apiserver/pkg/server/dynamiccertificates/configmap_cafile_content.go:206: Failed to list *v1.ConfigMap: configmaps "extension-apiserver-authentication" is forbidden: User "system:kube-scheduler" cannot list resource "configmaps" in API group "" in the namespace "kube-system"
E0323 16:45:22.446369       1 reflector.go:178] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.Node: nodes is forbidden: User "system:kube-scheduler" cannot list resource "nodes" in API group "" at the cluster scope
E0323 16:45:22.449781       1 reflector.go:178] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.Service: services is forbidden: User "system:kube-scheduler" cannot list resource "services" in API group "" at the cluster scope
E0323 16:45:22.451161       1 reflector.go:178] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csinodes" in API group "storage.k8s.io" at the cluster scope
E0323 16:45:22.451956       1 reflector.go:178] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "storageclasses" in API group "storage.k8s.io" at the cluster scope
E0323 16:45:22.454103       1 reflector.go:178] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumeclaims" in API group "" at the cluster scope
E0323 16:45:22.455601       1 reflector.go:178] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumes" in API group "" at the cluster scope
E0323 16:45:23.853318       1 reflector.go:178] k8s.io/client-go/informers/factory.go:135: Failed to list *v1beta1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User "system:kube-scheduler" cannot list resource "poddisruptionbudgets" in API group "policy" at the cluster scope
I0323 16:45:24.839119       1 shared_informer.go:230] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file
I0323 16:45:25.839842       1 leaderelection.go:242] attempting to acquire leader lease  kube-system/kube-scheduler...
I0323 16:45:25.870111       1 leaderelection.go:252] successfully acquired lease kube-system/kube-scheduler

==> kubelet <==
-- Logs begin at Mon 2020-03-23 16:44:38 UTC, end at Mon 2020-03-23 16:52:25 UTC. --
Mar 23 16:52:19 test-v1.18 kubelet[11089]: E0323 16:52:19.886727   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:19 test-v1.18 kubelet[11089]: E0323 16:52:19.988040   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:20 test-v1.18 kubelet[11089]: E0323 16:52:20.088521   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:20 test-v1.18 kubelet[11089]: E0323 16:52:20.189985   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:20 test-v1.18 kubelet[11089]: E0323 16:52:20.290662   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:20 test-v1.18 kubelet[11089]: E0323 16:52:20.392106   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:20 test-v1.18 kubelet[11089]: E0323 16:52:20.492778   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:20 test-v1.18 kubelet[11089]: E0323 16:52:20.594766   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:20 test-v1.18 kubelet[11089]: E0323 16:52:20.695317   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:20 test-v1.18 kubelet[11089]: E0323 16:52:20.795792   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:20 test-v1.18 kubelet[11089]: E0323 16:52:20.896870   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:20 test-v1.18 kubelet[11089]: E0323 16:52:20.998109   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:21 test-v1.18 kubelet[11089]: E0323 16:52:21.099304   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:21 test-v1.18 kubelet[11089]: E0323 16:52:21.199756   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:21 test-v1.18 kubelet[11089]: E0323 16:52:21.301188   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:21 test-v1.18 kubelet[11089]: E0323 16:52:21.401333   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:21 test-v1.18 kubelet[11089]: E0323 16:52:21.502136   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:21 test-v1.18 kubelet[11089]: E0323 16:52:21.603383   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:21 test-v1.18 kubelet[11089]: E0323 16:52:21.704025   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:21 test-v1.18 kubelet[11089]: E0323 16:52:21.804550   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:21 test-v1.18 kubelet[11089]: E0323 16:52:21.905238   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:22 test-v1.18 kubelet[11089]: E0323 16:52:22.005891   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:22 test-v1.18 kubelet[11089]: E0323 16:52:22.106174   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:22 test-v1.18 kubelet[11089]: E0323 16:52:22.208534   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:22 test-v1.18 kubelet[11089]: E0323 16:52:22.309457   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:22 test-v1.18 kubelet[11089]: E0323 16:52:22.409953   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:22 test-v1.18 kubelet[11089]: E0323 16:52:22.510868   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:22 test-v1.18 kubelet[11089]: E0323 16:52:22.612114   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:22 test-v1.18 kubelet[11089]: E0323 16:52:22.712379   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:22 test-v1.18 kubelet[11089]: E0323 16:52:22.812665   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:22 test-v1.18 kubelet[11089]: E0323 16:52:22.912796   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:23 test-v1.18 kubelet[11089]: E0323 16:52:23.013033   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:23 test-v1.18 kubelet[11089]: E0323 16:52:23.113410   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:23 test-v1.18 kubelet[11089]: E0323 16:52:23.215934   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:23 test-v1.18 kubelet[11089]: E0323 16:52:23.316352   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:23 test-v1.18 kubelet[11089]: E0323 16:52:23.417511   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:23 test-v1.18 kubelet[11089]: E0323 16:52:23.518136   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:23 test-v1.18 kubelet[11089]: E0323 16:52:23.618957   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:23 test-v1.18 kubelet[11089]: E0323 16:52:23.719471   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:23 test-v1.18 kubelet[11089]: E0323 16:52:23.819664   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:23 test-v1.18 kubelet[11089]: E0323 16:52:23.920556   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:24 test-v1.18 kubelet[11089]: E0323 16:52:24.021664   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:24 test-v1.18 kubelet[11089]: E0323 16:52:24.122058   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:24 test-v1.18 kubelet[11089]: E0323 16:52:24.222992   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:24 test-v1.18 kubelet[11089]: E0323 16:52:24.324068   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:24 test-v1.18 kubelet[11089]: E0323 16:52:24.424375   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:24 test-v1.18 kubelet[11089]: E0323 16:52:24.524710   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:24 test-v1.18 kubelet[11089]: E0323 16:52:24.625574   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:24 test-v1.18 kubelet[11089]: E0323 16:52:24.726173   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:24 test-v1.18 kubelet[11089]: E0323 16:52:24.826945   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:24 test-v1.18 kubelet[11089]: E0323 16:52:24.927321   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:25 test-v1.18 kubelet[11089]: E0323 16:52:25.027986   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:25 test-v1.18 kubelet[11089]: E0323 16:52:25.128520   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:25 test-v1.18 kubelet[11089]: E0323 16:52:25.229171   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:25 test-v1.18 kubelet[11089]: E0323 16:52:25.329718   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:25 test-v1.18 kubelet[11089]: E0323 16:52:25.398041   11089 controller.go:136] failed to ensure node lease exists, will retry in 7s, error: leases.coordination.k8s.io "m01" is forbidden: User "system:node:minikube" cannot get resource "leases" in API group "coordination.k8s.io" in the namespace "kube-node-lease": can only access node lease with the same name as the requesting node
Mar 23 16:52:25 test-v1.18 kubelet[11089]: E0323 16:52:25.430094   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:25 test-v1.18 kubelet[11089]: E0323 16:52:25.530702   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:25 test-v1.18 kubelet[11089]: E0323 16:52:25.631054   11089 kubelet.go:2267] node "m01" not found
Mar 23 16:52:25 test-v1.18 kubelet[11089]: E0323 16:52:25.731458   11089 kubelet.go:2267] node "m01" not found

The operating system version:

$ sw_vers
ProductName:    Mac OS X
ProductVersion: 10.15.3
BuildVersion:   19D76
kinbug kinsupport

All 3 comments

/kind bug

I believe this is related to #6200

I assumed this was fixed at head, but I'm seeing the same behavior there:

[kubelet] Creating a ConfigMap "kubelet-config-1.18" in namespace kube-system with the configuration for the kubelets in the cluster
[kubelet-check] Initial timeout of 40s passed.
[kubelet-check] It seems like the kubelet isn't running or healthy.
[kubelet-check] The HTTP call equal to 'curl -sSL http://localhost:10248/healthz' failed with error: Get http://localhost:10248/healthz: dial tcp 127.0.0.1:10248: connect: connection refused.

stderr:
W0325 17:33:03.257504    2929 configset.go:202] WARNING: kubeadm cannot validate component configs for API groups [kubelet.config.k8s.io kubeproxy.config.k8s.io]
    [WARNING Service-Docker]: docker service is not enabled, please run 'systemctl enable docker.service'
    [WARNING Hostname]: hostname "minikube" could not be reached
    [WARNING Hostname]: hostname "minikube": lookup minikube on 10.0.2.3:53: no such host
    [WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
W0325 17:33:05.972259    2929 manifests.go:225] the default kube-apiserver authorization-mode is "Node,RBAC"; using "Node,RBAC"
W0325 17:33:05.973298    2929 manifests.go:225] the default kube-apiserver authorization-mode is "Node,RBAC"; using "Node,RBAC"
error execution phase upload-config/kubelet: Error writing Crisocket information for the control-plane node: timed out waiting for the condition
To see the stack trace of this error execute with --v=5 or higher
Was this page helpful?
0 / 5 - 0 ratings