The exact command to reproduce the issue:
kubectl get pods --all-namespaces
The full output of the command that failed:
NAMESPACE NAME READY STATUS RESTARTS AGE
kube-system coredns-fb8b8dccf-tn8vz 0/1 CrashLoopBackOff 4 110s
kube-system coredns-fb8b8dccf-z28dc 0/1 CrashLoopBackOff 4 110s
kube-system etcd-minikube 1/1 Running 0 58s
kube-system kube-addon-manager-minikube 1/1 Running 0 48s
kube-system kube-apiserver-minikube 1/1 Running 0 52s
kube-system kube-controller-manager-minikube 1/1 Running 0 41s
kube-system kube-proxy-wb9bj 1/1 Running 0 110s
kube-system kube-scheduler-minikube 1/1 Running 0 41s
kube-system storage-provisioner 1/1 Running 0 109s
The output of the minikube logs command:
[root@fabsnuc ~]# minikube logs
==> coredns <==
E0524 17:47:43.607851 1 reflector.go:134] github.com/coredns/coredns/plugin/kubernetes/controller.go:322: Failed to list *v1.Namespace: Get https://10.96.0.1:443/api/v1/namespaces?limit=500&resourceVersion=0: dial tcp 10.96.0.1:443: connect: no route to host
E0524 17:47:43.607851 1 reflector.go:134] github.com/coredns/coredns/plugin/kubernetes/controller.go:322: Failed to list *v1.Namespace: Get https://10.96.0.1:443/api/v1/namespaces?limit=500&resourceVersion=0: dial tcp 10.96.0.1:443: connect: no route to host
log: exiting because of error: log: cannot create log: open /tmp/coredns.coredns-fb8b8dccf-tn8vz.unknownuser.log.ERROR.20190524-174743.1: no such file or directory
==> dmesg <==
dmesg: invalid option -- '='
Usage:
dmesg [options]
Options:
-C, --clear clear the kernel ring buffer
-c, --read-clear read and clear all messages
-D, --console-off disable printing messages to console
-d, --show-delta show time delta between printed messages
-e, --reltime show local time and time delta in readable format
-E, --console-on enable printing messages to console
-F, --file <file> use the file instead of the kernel log buffer
-f, --facility <list> restrict output to defined facilities
-H, --human human readable output
-k, --kernel display kernel messages
-L, --color colorize messages
-l, --level <list> restrict output to defined levels
-n, --console-level <level> set level of messages printed to console
-P, --nopager do not pipe output into a pager
-r, --raw print the raw message buffer
-S, --syslog force to use syslog(2) rather than /dev/kmsg
-s, --buffer-size <size> buffer size to query the kernel ring buffer
-T, --ctime show human readable timestamp (could be
inaccurate if you have used SUSPEND/RESUME)
-t, --notime don't print messages timestamp
-u, --userspace display userspace messages
-w, --follow wait for new messages
-x, --decode decode facility and level to readable string
-h, --help display this help and exit
-V, --version output version information and exit
Supported log facilities:
kern - kernel messages
user - random user-level messages
mail - mail system
daemon - system daemons
auth - security/authorization messages
syslog - messages generated internally by syslogd
lpr - line printer subsystem
news - network news subsystem
Supported log levels (priorities):
emerg - system is unusable
alert - action must be taken immediately
crit - critical conditions
err - error conditions
warn - warning conditions
notice - normal but significant condition
info - informational
debug - debug-level messages
For more details see dmesg(q).
==> kernel <==
12:48:49 up 12:26, 3 users, load average: 0.21, 0.23, 0.20
Linux fabsnuc.intel.com 3.10.0-957.12.2.el7.x86_64 #1 SMP Tue May 14 21:24:32 UTC 2019 x86_64 x86_64 x86_64 GNU/Linux
==> kube-addon-manager <==
WRN: == Error getting default service account, retry in 0.5 second ==
Error from server (NotFound): serviceaccounts "default" not found
WRN: == Error getting default service account, retry in 0.5 second ==
Error from server (NotFound): serviceaccounts "default" not found
WRN: == Error getting default service account, retry in 0.5 second ==
Error from server (NotFound): serviceaccounts "default" not found
WRN: == Error getting default service account, retry in 0.5 second ==
Error from server (NotFound): serviceaccounts "default" not found
WRN: == Error getting default service account, retry in 0.5 second ==
INFO: == Default service account in the kube-system namespace has token default-token-qb9ck ==
find: '/etc/kubernetes/admission-controls': No such file or directory
INFO: == Entering periodical apply loop at 2019-05-24T17:44:29+00:00 ==
INFO: Leader is fabsnuc.intel.com
clusterrolebinding.rbac.authorization.k8s.io/storage-provisioner created
storageclass.storage.k8s.io/standard created
INFO: == Kubernetes addon ensure completed at 2019-05-24T17:44:29+00:00 ==
INFO: == Reconciling with deprecated label ==
error: no objects passed to apply
INFO: == Reconciling with addon-manager label ==
serviceaccount/storage-provisioner created
pod/storage-provisioner created
INFO: == Kubernetes addon reconcile completed at 2019-05-24T17:44:32+00:00 ==
INFO: Leader is fabsnuc.intel.com
INFO: == Kubernetes addon ensure completed at 2019-05-24T17:45:29+00:00 ==
INFO: == Reconciling with deprecated label ==
error: no objects passed to apply
INFO: == Reconciling with addon-manager label ==
serviceaccount/storage-provisioner unchanged
INFO: == Kubernetes addon reconcile completed at 2019-05-24T17:45:30+00:00 ==
INFO: Leader is fabsnuc.intel.com
INFO: == Kubernetes addon ensure completed at 2019-05-24T17:46:29+00:00 ==
INFO: == Reconciling with deprecated label ==
error: no objects passed to apply
INFO: == Reconciling with addon-manager label ==
serviceaccount/storage-provisioner unchanged
INFO: == Kubernetes addon reconcile completed at 2019-05-24T17:46:31+00:00 ==
INFO: Leader is fabsnuc.intel.com
INFO: == Kubernetes addon ensure completed at 2019-05-24T17:47:29+00:00 ==
INFO: == Reconciling with deprecated label ==
error: no objects passed to apply
INFO: == Reconciling with addon-manager label ==
serviceaccount/storage-provisioner unchanged
INFO: == Kubernetes addon reconcile completed at 2019-05-24T17:47:30+00:00 ==
INFO: Leader is fabsnuc.intel.com
INFO: == Kubernetes addon ensure completed at 2019-05-24T17:48:29+00:00 ==
INFO: == Reconciling with deprecated label ==
error: no objects passed to apply
INFO: == Reconciling with addon-manager label ==
serviceaccount/storage-provisioner unchanged
INFO: == Kubernetes addon reconcile completed at 2019-05-24T17:48:31+00:00 ==
==> kube-apiserver <==
I0524 17:44:20.171744 1 storage_rbac.go:223] created clusterrolebinding.rbac.authorization.k8s.io/system:node
I0524 17:44:20.213052 1 storage_rbac.go:223] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:attachdetach-controller
I0524 17:44:20.252888 1 storage_rbac.go:223] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:clusterrole-aggregation-controller
I0524 17:44:20.293107 1 storage_rbac.go:223] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:cronjob-controller
I0524 17:44:20.333042 1 storage_rbac.go:223] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:daemon-set-controller
I0524 17:44:20.373102 1 storage_rbac.go:223] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:deployment-controller
I0524 17:44:20.413045 1 storage_rbac.go:223] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:disruption-controller
I0524 17:44:20.453050 1 storage_rbac.go:223] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:endpoint-controller
I0524 17:44:20.493174 1 storage_rbac.go:223] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:expand-controller
I0524 17:44:20.532906 1 storage_rbac.go:223] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:generic-garbage-collector
I0524 17:44:20.572937 1 storage_rbac.go:223] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:horizontal-pod-autoscaler
I0524 17:44:20.613066 1 storage_rbac.go:223] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:job-controller
I0524 17:44:20.652852 1 storage_rbac.go:223] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:namespace-controller
I0524 17:44:20.693204 1 storage_rbac.go:223] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:node-controller
I0524 17:44:20.731879 1 storage_rbac.go:223] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:persistent-volume-binder
I0524 17:44:20.773076 1 storage_rbac.go:223] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:pod-garbage-collector
I0524 17:44:20.816392 1 storage_rbac.go:223] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:replicaset-controller
I0524 17:44:20.852923 1 storage_rbac.go:223] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:replication-controller
I0524 17:44:21.145954 1 storage_rbac.go:223] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:resourcequota-controller
I0524 17:44:21.149326 1 storage_rbac.go:223] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:route-controller
I0524 17:44:21.152939 1 storage_rbac.go:223] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:service-account-controller
I0524 17:44:21.156458 1 storage_rbac.go:223] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:service-controller
I0524 17:44:21.159898 1 storage_rbac.go:223] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:statefulset-controller
I0524 17:44:21.163538 1 storage_rbac.go:223] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:ttl-controller
I0524 17:44:21.167086 1 storage_rbac.go:223] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:certificate-controller
I0524 17:44:21.172823 1 storage_rbac.go:223] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:pvc-protection-controller
I0524 17:44:21.212745 1 storage_rbac.go:223] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:pv-protection-controller
I0524 17:44:21.251844 1 controller.go:606] quota admission added evaluator for: roles.rbac.authorization.k8s.io
I0524 17:44:21.253898 1 storage_rbac.go:254] created role.rbac.authorization.k8s.io/extension-apiserver-authentication-reader in kube-system
I0524 17:44:21.292135 1 storage_rbac.go:254] created role.rbac.authorization.k8s.io/system:controller:bootstrap-signer in kube-system
I0524 17:44:21.331692 1 storage_rbac.go:254] created role.rbac.authorization.k8s.io/system:controller:cloud-provider in kube-system
I0524 17:44:21.372887 1 storage_rbac.go:254] created role.rbac.authorization.k8s.io/system:controller:token-cleaner in kube-system
I0524 17:44:21.388219 1 controller.go:606] quota admission added evaluator for: endpoints
I0524 17:44:21.412944 1 storage_rbac.go:254] created role.rbac.authorization.k8s.io/system::leader-locking-kube-controller-manager in kube-system
I0524 17:44:21.453044 1 storage_rbac.go:254] created role.rbac.authorization.k8s.io/system::leader-locking-kube-scheduler in kube-system
I0524 17:44:21.495019 1 storage_rbac.go:254] created role.rbac.authorization.k8s.io/system:controller:bootstrap-signer in kube-public
I0524 17:44:21.531542 1 controller.go:606] quota admission added evaluator for: rolebindings.rbac.authorization.k8s.io
I0524 17:44:21.533395 1 storage_rbac.go:284] created rolebinding.rbac.authorization.k8s.io/system::extension-apiserver-authentication-reader in kube-system
I0524 17:44:21.573147 1 storage_rbac.go:284] created rolebinding.rbac.authorization.k8s.io/system::leader-locking-kube-controller-manager in kube-system
I0524 17:44:21.613109 1 storage_rbac.go:284] created rolebinding.rbac.authorization.k8s.io/system::leader-locking-kube-scheduler in kube-system
I0524 17:44:21.653065 1 storage_rbac.go:284] created rolebinding.rbac.authorization.k8s.io/system:controller:bootstrap-signer in kube-system
I0524 17:44:21.692961 1 storage_rbac.go:284] created rolebinding.rbac.authorization.k8s.io/system:controller:cloud-provider in kube-system
I0524 17:44:21.732967 1 storage_rbac.go:284] created rolebinding.rbac.authorization.k8s.io/system:controller:token-cleaner in kube-system
I0524 17:44:21.773047 1 storage_rbac.go:284] created rolebinding.rbac.authorization.k8s.io/system:controller:bootstrap-signer in kube-public
W0524 17:44:21.791574 1 lease.go:222] Resetting endpoints for master service "kubernetes" to [192.168.0.159]
I0524 17:44:22.361003 1 controller.go:606] quota admission added evaluator for: serviceaccounts
I0524 17:44:22.965355 1 controller.go:606] quota admission added evaluator for: deployments.apps
I0524 17:44:23.308886 1 controller.go:606] quota admission added evaluator for: daemonsets.apps
I0524 17:44:29.163799 1 controller.go:606] quota admission added evaluator for: controllerrevisions.apps
I0524 17:44:29.282538 1 controller.go:606] quota admission added evaluator for: replicasets.apps
==> kube-proxy <==
W0524 17:44:30.645444 1 proxier.go:498] Failed to load kernel module ip_vs_rr with modprobe. You can ignore this message when kube-proxy is running inside container without mounting /lib/modules
W0524 17:44:30.651474 1 server_others.go:267] Flag proxy-mode="" unknown, assuming iptables proxy
I0524 17:44:30.656952 1 server_others.go:146] Using iptables Proxier.
W0524 17:44:30.657029 1 proxier.go:319] clusterCIDR not specified, unable to distinguish between internal and external traffic
I0524 17:44:30.657135 1 server.go:562] Version: v1.14.2
I0524 17:44:30.669112 1 conntrack.go:52] Setting nf_conntrack_max to 262144
I0524 17:44:30.669356 1 config.go:202] Starting service config controller
I0524 17:44:30.669383 1 controller_utils.go:1027] Waiting for caches to sync for service config controller
I0524 17:44:30.670178 1 config.go:102] Starting endpoints config controller
I0524 17:44:30.670192 1 controller_utils.go:1027] Waiting for caches to sync for endpoints config controller
I0524 17:44:30.769721 1 controller_utils.go:1034] Caches are synced for service config controller
I0524 17:44:30.770421 1 controller_utils.go:1034] Caches are synced for endpoints config controller
==> kube-scheduler <==
I0524 17:44:15.952574 1 serving.go:319] Generated self-signed cert in-memory
W0524 17:44:16.280112 1 authentication.go:249] No authentication-kubeconfig provided in order to lookup client-ca-file in configmap/extension-apiserver-authentication in kube-system, so client certificate authentication won't work.
W0524 17:44:16.280124 1 authentication.go:252] No authentication-kubeconfig provided in order to lookup requestheader-client-ca-file in configmap/extension-apiserver-authentication in kube-system, so request-header client certificate authentication won't work.
W0524 17:44:16.280153 1 authorization.go:146] No authorization-kubeconfig provided, so SubjectAccessReview of authorization tokens won't work.
I0524 17:44:16.281836 1 server.go:142] Version: v1.14.2
I0524 17:44:16.281876 1 defaults.go:87] TaintNodesByCondition is enabled, PodToleratesNodeTaints predicate is mandatory
W0524 17:44:16.283073 1 authorization.go:47] Authorization is disabled
W0524 17:44:16.283082 1 authentication.go:55] Authentication is disabled
I0524 17:44:16.283093 1 deprecated_insecure_serving.go:49] Serving healthz insecurely on [::]:10251
I0524 17:44:16.283369 1 secure_serving.go:116] Serving securely on 127.0.0.1:10259
E0524 17:44:18.410541 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumes" in API group "" at the cluster scope
E0524 17:44:18.410618 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User "system:kube-scheduler" cannot list resource "replicationcontrollers" in API group "" at the cluster scope
E0524 17:44:18.411874 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1beta1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User "system:kube-scheduler" cannot list resource "poddisruptionbudgets" in API group "policy" at the cluster scope
E0524 17:44:18.414829 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.Node: nodes is forbidden: User "system:kube-scheduler" cannot list resource "nodes" in API group "" at the cluster scope
E0524 17:44:18.414876 1 reflector.go:126] k8s.io/kubernetes/cmd/kube-scheduler/app/server.go:223: Failed to list *v1.Pod: pods is forbidden: User "system:kube-scheduler" cannot list resource "pods" in API group "" at the cluster scope
E0524 17:44:18.426899 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "storageclasses" in API group "storage.k8s.io" at the cluster scope
E0524 17:44:18.426920 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.Service: services is forbidden: User "system:kube-scheduler" cannot list resource "services" in API group "" at the cluster scope
E0524 17:44:18.427736 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User "system:kube-scheduler" cannot list resource "statefulsets" in API group "apps" at the cluster scope
E0524 17:44:18.428376 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumeclaims" in API group "" at the cluster scope
E0524 17:44:18.428377 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User "system:kube-scheduler" cannot list resource "replicasets" in API group "apps" at the cluster scope
E0524 17:44:19.411860 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumes" in API group "" at the cluster scope
E0524 17:44:19.412766 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User "system:kube-scheduler" cannot list resource "replicationcontrollers" in API group "" at the cluster scope
E0524 17:44:19.414768 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1beta1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User "system:kube-scheduler" cannot list resource "poddisruptionbudgets" in API group "policy" at the cluster scope
E0524 17:44:19.415861 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.Node: nodes is forbidden: User "system:kube-scheduler" cannot list resource "nodes" in API group "" at the cluster scope
E0524 17:44:19.423611 1 reflector.go:126] k8s.io/kubernetes/cmd/kube-scheduler/app/server.go:223: Failed to list *v1.Pod: pods is forbidden: User "system:kube-scheduler" cannot list resource "pods" in API group "" at the cluster scope
E0524 17:44:19.428104 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "storageclasses" in API group "storage.k8s.io" at the cluster scope
E0524 17:44:19.429304 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.Service: services is forbidden: User "system:kube-scheduler" cannot list resource "services" in API group "" at the cluster scope
E0524 17:44:19.430785 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User "system:kube-scheduler" cannot list resource "statefulsets" in API group "apps" at the cluster scope
E0524 17:44:19.432002 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumeclaims" in API group "" at the cluster scope
E0524 17:44:19.432981 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User "system:kube-scheduler" cannot list resource "replicasets" in API group "apps" at the cluster scope
I0524 17:44:21.284704 1 controller_utils.go:1027] Waiting for caches to sync for scheduler controller
I0524 17:44:21.384829 1 controller_utils.go:1034] Caches are synced for scheduler controller
I0524 17:44:21.384968 1 leaderelection.go:217] attempting to acquire leader lease kube-system/kube-scheduler...
I0524 17:44:21.389935 1 leaderelection.go:227] successfully acquired lease kube-system/kube-scheduler
==> kubelet <==
-- Logs begin at Fri 2019-05-24 00:22:03 CDT, end at Fri 2019-05-24 12:48:49 CDT. --
May 24 12:44:33 fabsnuc.intel.com kubelet[26430]: E0524 12:44:33.120690 26430 pod_workers.go:190] Error syncing pod 94a8990f-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 10s restarting failed container=coredns pod=coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"
May 24 12:44:34 fabsnuc.intel.com kubelet[26430]: E0524 12:44:34.147984 26430 pod_workers.go:190] Error syncing pod 94a8990f-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 10s restarting failed container=coredns pod=coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"
May 24 12:44:34 fabsnuc.intel.com kubelet[26430]: E0524 12:44:34.159168 26430 pod_workers.go:190] Error syncing pod 94a7de57-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 10s restarting failed container=coredns pod=coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"
May 24 12:44:35 fabsnuc.intel.com kubelet[26430]: E0524 12:44:35.176560 26430 pod_workers.go:190] Error syncing pod 94a7de57-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 10s restarting failed container=coredns pod=coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"
May 24 12:44:36 fabsnuc.intel.com kubelet[26430]: E0524 12:44:36.125943 26430 pod_workers.go:190] Error syncing pod 94a8990f-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 10s restarting failed container=coredns pod=coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"
May 24 12:44:38 fabsnuc.intel.com kubelet[26430]: E0524 12:44:38.222372 26430 pod_workers.go:190] Error syncing pod 94a7de57-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 10s restarting failed container=coredns pod=coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"
May 24 12:44:52 fabsnuc.intel.com kubelet[26430]: E0524 12:44:52.325022 26430 pod_workers.go:190] Error syncing pod 94a7de57-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 20s restarting failed container=coredns pod=coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"
May 24 12:44:53 fabsnuc.intel.com kubelet[26430]: E0524 12:44:53.345496 26430 pod_workers.go:190] Error syncing pod 94a8990f-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 20s restarting failed container=coredns pod=coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"
May 24 12:44:56 fabsnuc.intel.com kubelet[26430]: E0524 12:44:56.126216 26430 pod_workers.go:190] Error syncing pod 94a8990f-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 20s restarting failed container=coredns pod=coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"
May 24 12:44:58 fabsnuc.intel.com kubelet[26430]: E0524 12:44:58.222897 26430 pod_workers.go:190] Error syncing pod 94a7de57-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 20s restarting failed container=coredns pod=coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"
May 24 12:45:09 fabsnuc.intel.com kubelet[26430]: E0524 12:45:09.417015 26430 pod_workers.go:190] Error syncing pod 94a8990f-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 20s restarting failed container=coredns pod=coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"
May 24 12:45:09 fabsnuc.intel.com kubelet[26430]: E0524 12:45:09.417037 26430 pod_workers.go:190] Error syncing pod 94a7de57-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 20s restarting failed container=coredns pod=coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"
May 24 12:45:22 fabsnuc.intel.com kubelet[26430]: E0524 12:45:22.754070 26430 pod_workers.go:190] Error syncing pod 94a8990f-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 40s restarting failed container=coredns pod=coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"
May 24 12:45:22 fabsnuc.intel.com kubelet[26430]: E0524 12:45:22.776061 26430 pod_workers.go:190] Error syncing pod 94a7de57-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 40s restarting failed container=coredns pod=coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"
May 24 12:45:23 fabsnuc.intel.com kubelet[26430]: E0524 12:45:23.792888 26430 pod_workers.go:190] Error syncing pod 94a8990f-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 40s restarting failed container=coredns pod=coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"
May 24 12:45:23 fabsnuc.intel.com kubelet[26430]: E0524 12:45:23.804715 26430 pod_workers.go:190] Error syncing pod 94a7de57-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 40s restarting failed container=coredns pod=coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"
May 24 12:45:26 fabsnuc.intel.com kubelet[26430]: E0524 12:45:26.126308 26430 pod_workers.go:190] Error syncing pod 94a8990f-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 40s restarting failed container=coredns pod=coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"
May 24 12:45:28 fabsnuc.intel.com kubelet[26430]: E0524 12:45:28.222872 26430 pod_workers.go:190] Error syncing pod 94a7de57-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 40s restarting failed container=coredns pod=coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"
May 24 12:45:38 fabsnuc.intel.com kubelet[26430]: E0524 12:45:38.417057 26430 pod_workers.go:190] Error syncing pod 94a8990f-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 40s restarting failed container=coredns pod=coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"
May 24 12:45:42 fabsnuc.intel.com kubelet[26430]: E0524 12:45:42.416826 26430 pod_workers.go:190] Error syncing pod 94a7de57-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 40s restarting failed container=coredns pod=coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"
May 24 12:45:51 fabsnuc.intel.com kubelet[26430]: E0524 12:45:51.416901 26430 pod_workers.go:190] Error syncing pod 94a8990f-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 40s restarting failed container=coredns pod=coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"
May 24 12:45:55 fabsnuc.intel.com kubelet[26430]: E0524 12:45:55.416863 26430 pod_workers.go:190] Error syncing pod 94a7de57-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 40s restarting failed container=coredns pod=coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"
May 24 12:46:06 fabsnuc.intel.com kubelet[26430]: E0524 12:46:06.134784 26430 pod_workers.go:190] Error syncing pod 94a8990f-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 1m20s restarting failed container=coredns pod=coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"
May 24 12:46:08 fabsnuc.intel.com kubelet[26430]: E0524 12:46:08.174227 26430 pod_workers.go:190] Error syncing pod 94a7de57-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 1m20s restarting failed container=coredns pod=coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"
May 24 12:46:09 fabsnuc.intel.com kubelet[26430]: E0524 12:46:09.187954 26430 pod_workers.go:190] Error syncing pod 94a7de57-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 1m20s restarting failed container=coredns pod=coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"
May 24 12:46:16 fabsnuc.intel.com kubelet[26430]: E0524 12:46:16.126184 26430 pod_workers.go:190] Error syncing pod 94a8990f-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 1m20s restarting failed container=coredns pod=coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"
May 24 12:46:22 fabsnuc.intel.com kubelet[26430]: E0524 12:46:22.417038 26430 pod_workers.go:190] Error syncing pod 94a7de57-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 1m20s restarting failed container=coredns pod=coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"
May 24 12:46:31 fabsnuc.intel.com kubelet[26430]: E0524 12:46:31.416842 26430 pod_workers.go:190] Error syncing pod 94a8990f-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 1m20s restarting failed container=coredns pod=coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"
May 24 12:46:34 fabsnuc.intel.com kubelet[26430]: E0524 12:46:34.418387 26430 pod_workers.go:190] Error syncing pod 94a7de57-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 1m20s restarting failed container=coredns pod=coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"
May 24 12:46:42 fabsnuc.intel.com kubelet[26430]: E0524 12:46:42.416809 26430 pod_workers.go:190] Error syncing pod 94a8990f-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 1m20s restarting failed container=coredns pod=coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"
May 24 12:46:46 fabsnuc.intel.com kubelet[26430]: E0524 12:46:46.416890 26430 pod_workers.go:190] Error syncing pod 94a7de57-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 1m20s restarting failed container=coredns pod=coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"
May 24 12:46:53 fabsnuc.intel.com kubelet[26430]: E0524 12:46:53.416860 26430 pod_workers.go:190] Error syncing pod 94a8990f-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 1m20s restarting failed container=coredns pod=coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"
May 24 12:47:01 fabsnuc.intel.com kubelet[26430]: E0524 12:47:01.417019 26430 pod_workers.go:190] Error syncing pod 94a7de57-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 1m20s restarting failed container=coredns pod=coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"
May 24 12:47:04 fabsnuc.intel.com kubelet[26430]: E0524 12:47:04.417086 26430 pod_workers.go:190] Error syncing pod 94a8990f-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 1m20s restarting failed container=coredns pod=coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"
May 24 12:47:13 fabsnuc.intel.com kubelet[26430]: E0524 12:47:13.416857 26430 pod_workers.go:190] Error syncing pod 94a7de57-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 1m20s restarting failed container=coredns pod=coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"
May 24 12:47:18 fabsnuc.intel.com kubelet[26430]: E0524 12:47:18.416919 26430 pod_workers.go:190] Error syncing pod 94a8990f-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 1m20s restarting failed container=coredns pod=coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"
May 24 12:47:27 fabsnuc.intel.com kubelet[26430]: E0524 12:47:27.416850 26430 pod_workers.go:190] Error syncing pod 94a7de57-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 1m20s restarting failed container=coredns pod=coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"
May 24 12:47:31 fabsnuc.intel.com kubelet[26430]: E0524 12:47:31.759366 26430 pod_workers.go:190] Error syncing pod 94a8990f-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 2m40s restarting failed container=coredns pod=coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"
May 24 12:47:36 fabsnuc.intel.com kubelet[26430]: E0524 12:47:36.126234 26430 pod_workers.go:190] Error syncing pod 94a8990f-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 2m40s restarting failed container=coredns pod=coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"
May 24 12:47:43 fabsnuc.intel.com kubelet[26430]: E0524 12:47:43.868686 26430 pod_workers.go:190] Error syncing pod 94a7de57-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 2m40s restarting failed container=coredns pod=coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"
May 24 12:47:48 fabsnuc.intel.com kubelet[26430]: E0524 12:47:48.222803 26430 pod_workers.go:190] Error syncing pod 94a7de57-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 2m40s restarting failed container=coredns pod=coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"
May 24 12:47:48 fabsnuc.intel.com kubelet[26430]: E0524 12:47:48.416945 26430 pod_workers.go:190] Error syncing pod 94a8990f-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 2m40s restarting failed container=coredns pod=coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"
May 24 12:47:59 fabsnuc.intel.com kubelet[26430]: E0524 12:47:59.416959 26430 pod_workers.go:190] Error syncing pod 94a7de57-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 2m40s restarting failed container=coredns pod=coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"
May 24 12:48:01 fabsnuc.intel.com kubelet[26430]: E0524 12:48:01.416976 26430 pod_workers.go:190] Error syncing pod 94a8990f-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 2m40s restarting failed container=coredns pod=coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"
May 24 12:48:11 fabsnuc.intel.com kubelet[26430]: E0524 12:48:11.417023 26430 pod_workers.go:190] Error syncing pod 94a7de57-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 2m40s restarting failed container=coredns pod=coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"
May 24 12:48:15 fabsnuc.intel.com kubelet[26430]: E0524 12:48:15.416908 26430 pod_workers.go:190] Error syncing pod 94a8990f-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 2m40s restarting failed container=coredns pod=coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"
May 24 12:48:26 fabsnuc.intel.com kubelet[26430]: E0524 12:48:26.416923 26430 pod_workers.go:190] Error syncing pod 94a7de57-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 2m40s restarting failed container=coredns pod=coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"
May 24 12:48:27 fabsnuc.intel.com kubelet[26430]: E0524 12:48:27.416851 26430 pod_workers.go:190] Error syncing pod 94a8990f-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 2m40s restarting failed container=coredns pod=coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"
May 24 12:48:37 fabsnuc.intel.com kubelet[26430]: E0524 12:48:37.416896 26430 pod_workers.go:190] Error syncing pod 94a7de57-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 2m40s restarting failed container=coredns pod=coredns-fb8b8dccf-tn8vz_kube-system(94a7de57-7e4b-11e9-82df-54b20311c753)"
May 24 12:48:40 fabsnuc.intel.com kubelet[26430]: E0524 12:48:40.416996 26430 pod_workers.go:190] Error syncing pod 94a8990f-7e4b-11e9-82df-54b20311c753 ("coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"), skipping: failed to "StartContainer" for "coredns" with CrashLoopBackOff: "Back-off 2m40s restarting failed container=coredns pod=coredns-fb8b8dccf-z28dc_kube-system(94a8990f-7e4b-11e9-82df-54b20311c753)"
==> storage-provisioner <==
The operating system version:
[root@fabsnuc ~]# cat /etc/os-release
NAME="CentOS Linux"
VERSION="7 (Core)"
ID="centos"
ID_LIKE="rhel fedora"
VERSION_ID="7"
PRETTY_NAME="CentOS Linux 7 (Core)"
ANSI_COLOR="0;31"
CPE_NAME="cpe:/o:centos:centos:7"
HOME_URL="https://www.centos.org/"
BUG_REPORT_URL="https://bugs.centos.org/"
CENTOS_MANTISBT_PROJECT="CentOS-7"
CENTOS_MANTISBT_PROJECT_VERSION="7"
REDHAT_SUPPORT_PRODUCT="centos"
REDHAT_SUPPORT_PRODUCT_VERSION="7"
[root@fabsnuc ~]# docker version
Client:
Version: 18.09.6
API version: 1.39
Go version: go1.10.8
Git commit: 481bc77156
Built: Sat May 4 02:34:58 2019
OS/Arch: linux/amd64
Experimental: false
Server: Docker Engine - Community
Engine:
Version: 18.09.6
API version: 1.39 (minimum version 1.12)
Go version: go1.10.8
Git commit: 481bc77
Built: Sat May 4 02:02:43 2019
OS/Arch: linux/amd64
Experimental: false
[root@fabsnuc ~]#
[root@fabsnuc ~]# getenforce
Disabled
[root@fabsnuc ~]#
[root@fabsnuc ~]# kubectl logs deployment/coredns -n kube-system --previous
Found 2 pods, using pod/coredns-fb8b8dccf-5tpm6
E0524 18:06:42.120194 1 reflector.go:134] github.com/coredns/coredns/plugin/kubernetes/controller.go:315: Failed to list *v1.Service: Get https://10.96.0.1:443/api/v1/services?limit=500&resourceVersion=0: dial tcp 10.96.0.1:443: connect: no route to host
E0524 18:06:42.120194 1 reflector.go:134] github.com/coredns/coredns/plugin/kubernetes/controller.go:315: Failed to list *v1.Service: Get https://10.96.0.1:443/api/v1/services?limit=500&resourceVersion=0: dial tcp 10.96.0.1:443: connect: no route to host
log: exiting because of error: log: cannot create log: open /tmp/coredns.coredns-fb8b8dccf-5tpm6.unknownuser.log.ERROR.20190524-180642.1: no such file or directory
This message isn't normal either:
dial tcp 10.96.0.1:443: connect: no route to host
Some other folks have similar coredns failures outside of minikube when the apiserver isn't available: https://github.com/kubernetes/kubernetes/issues/75414
Why wouldn't the apiserver be available though? Here's one possible hint from kube-proxy:
W0524 17:44:30.645444 1 proxier.go:498] Failed to load kernel module ip_vs_rr with modprobe. You can ignore this message when kube-proxy is running inside container without mounting /lib/modules
This may be a red herring, but do you mind seeing what sudo modprobe ip_vs_rr outputs?
@fabstao I was facing the exact same issue on my CentOS VM. I got it fixed by following the instructions in this comment: https://github.com/kubernetes/kubeadm/issues/193#issuecomment-330060848 to flush the iptables
Good to know. We should update the error message for this failure to mention flushing iptables then. Thanks!
@fabstao I was facing the exact same issue on my CentOS VM. I got it fixed by following the instructions in this comment: kubernetes/kubeadm#193 (comment) to flush the iptables
It's not solved this problem...
I got exactly the same issue too and was solved by @albinsuresh 's reply.
thx a lot.
Unfortunately, the fix in @albinsuresh 's reply is a work-around. Does anyone know what he true fix is if you're running a customized local firewall? I'll do some digging and post again if I find it.
@slalomnut could you please provide logs from the newest minikube version ?
both minikube logs and start output and also the kubectl get pods -o wide -n kube-system
and also
kubectl describe pod coredns -n kube-system
in the latest version we provide a better logging
and I wonder has anyone checked to see if this comment helps them ? (if the issue still exists with 1.3.1) https://github.com/kubernetes/kubeadm/issues/193#issuecomment-462649752
I can confirm that it was a firewall issue on my side.
I was running kubernetes 1.15.3 on my local machine(Ubuntu 18.04.3), and I had ufw enabled, and because of that, it was unable to communicate with 10.96.0.1:443, and after I disabled ufw, coredns pods were up and running.
This seems solved, but I will leave it open for anyone else who runs into this.
use command systemctl stop firewalld
I've worked out the problem.
Upgraded to latest v1.5.1 and seeing same issue but because of a different error now - /etc/coredns/Corefile:4 - Error during parsing: Unknown directive 'ready'
Using none driver on Ubuntu 18.04.3
Happening only on the v1.4.0 and above, when I switch back to v1.3.1 and use --extra-config=kubelet.resolv-conf=/run/systemd/resolve/resolv.conf it works fine
Issues go stale after 90d of inactivity.
Mark the issue as fresh with /remove-lifecycle stale.
Stale issues rot after an additional 30d of inactivity and eventually close.
If this issue is safe to close now please do so with /close.
Send feedback to sig-testing, kubernetes/test-infra and/or fejta.
/lifecycle stale
In my case it was an issue with dashboard.
馃挘 http://127.0.0.1:37577/api/v1/namespaces/kubernetes-dashboard/services/http:kubernetes-dashboard:/proxy/ is not accessible: Temporary Error: unexpected response code: 503
if you have firewalld enabled you can add docker0 bridge interface to trusted zone which should allow docker containers to host communication
$ sudo minikube start --vm-driver=none
$ sudo firewall-cmd --permanent --zone=trusted --add-interface=docker0
$ sudo firewall-cmd --reload
$ sudo firewall-cmd --get-active-zones
$ sudo firewall-cmd --list-all --zone=trusted
$ sudo chown -R $USER $HOME/.kube $HOME/.minikube
$ minikube dashboard &
$ minikube version
minikube version: v1.7.2
commit: 50d543b5fcb0e1c0d7c27b1398a9a9790df09dfb
$ minikube status
host: Running
kubelet: Running
apiserver: Running
kubeconfig: Configured
$ sudo firewall-cmd --state
running
$ sudo firewall-cmd --get-active-zones
public
interfaces: wlp1s0
trusted
interfaces: docker0
$ sudo firewall-cmd --list-all --zone=trusted
trusted (active)
target: ACCEPT
icmp-block-inversion: no
interfaces: docker0
sources:
services:
ports:
protocols:
masquerade: no
forward-ports:
source-ports:
icmp-blocks:
rich rules:
$ kubectl get pods --all-namespaces
NAMESPACE NAME READY STATUS RESTARTS AGE
kube-system coredns-6955765f44-7zjth 1/1 Running 0 67m
kube-system coredns-6955765f44-b9gbq 1/1 Running 0 67m
kube-system etcd-venga 1/1 Running 0 67m
kube-system kube-apiserver-venga 1/1 Running 0 67m
kube-system kube-controller-manager-venga 1/1 Running 0 67m
kube-system kube-proxy-7xv6h 1/1 Running 0 67m
kube-system kube-scheduler-venga 1/1 Running 0 67m
kube-system storage-provisioner 1/1 Running 0 67m
kubernetes-dashboard dashboard-metrics-scraper-7b64584c5c-nw82r 1/1 Running 0 65m
kubernetes-dashboard kubernetes-dashboard-79d9cd965-zlbl2 1/1 Running 13 65m
Stale issues rot after 30d of inactivity.
Mark the issue as fresh with /remove-lifecycle rotten.
Rotten issues close after an additional 30d of inactivity.
If this issue is safe to close now please do so with /close.
Send feedback to sig-testing, kubernetes/test-infra and/or fejta.
/lifecycle rotten
Rotten issues close after 30d of inactivity.
Reopen the issue with /reopen.
Mark the issue as fresh with /remove-lifecycle rotten.
Send feedback to sig-testing, kubernetes/test-infra and/or fejta.
/close
@fejta-bot: Closing this issue.
In response to this:
Rotten issues close after 30d of inactivity.
Reopen the issue with/reopen.
Mark the issue as fresh with/remove-lifecycle rotten.Send feedback to sig-testing, kubernetes/test-infra and/or fejta.
/close
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository.
Just adding my experience. I had the same problem. In my case it has been enough to add the Masquerading option in the default host outbound interface and then the communication started to work.
Just adding my experience. I had the same problem. In my case it has been enough to add the Masquerading option in the default host outbound interface and then the communication started to work.
I have this problem in prod env after runing some days, server in pod can not access outer network , exececute blow commands reloved, i want to know how happened before encounter this problem , and how to prevent this ;
systemctl stop kubelet
systemctl stop docker
iptables --flush
iptables -tnat --flush
systemctl start kubelet
systemctl start docker
Most helpful comment
use command
systemctl stop firewalldI've worked out the problem.