BUG REPORT
Environment:
printf "$(uname -srm)\n$(cat /etc/os-release)\n"): Linux 4.13.0-45-generic x86_64
NAME="Ubuntu"
VERSION="16.04.4 LTS (Xenial Xerus)"
ID=ubuntu
ID_LIKE=debian
PRETTY_NAME="Ubuntu 16.04.4 LTS"
VERSION_ID="16.04"
HOME_URL="http://www.ubuntu.com/"
SUPPORT_URL="http://help.ubuntu.com/"
BUG_REPORT_URL="http://bugs.launchpad.net/ubuntu/"
VERSION_CODENAME=xenial
UBUNTU_CODENAME=xenial
ansible --version): ansible 2.5.4
config file = /root/kubespray/ansible.cfg
configured module search path = [u'/root/kubespray/library']
ansible python module location = /usr/lib/python2.7/dist-packages/ansible
executable location = /usr/bin/ansible
python version = 2.7.12 (default, Dec 4 2017, 14:50:18) [GCC 5.4.0 20160609]
Kubespray version (commit) (git rev-parse --short HEAD): 0686b84
Network plugin used: calico
Copy of your inventory file:
[all:vars]
[kube-master]
heku1 ansible_host=ip1 ansible_ssh_user=root ansible_ssh_pass=pass ansible_connection=ssh ansible_port=122
heku2 ansible_host=ip2 ansible_ssh_user=root ansible_ssh_pass=pass ansible_connection=ssh ansible_port=122
heku3 ansible_host=ip3 ansible_ssh_user=root ansible_ssh_pass=pass ansible_connection=ssh ansible_port=122
[etcd]
heku1 ansible_host=ip1 ansible_ssh_user=root ansible_ssh_pass=pass ansible_connection=ssh ansible_port=122
heku2 ansible_host=ip2 ansible_ssh_user=root ansible_ssh_pass=pass ansible_connection=ssh ansible_port=122
heku3 ansible_host=ip3 ansible_ssh_user=root ansible_ssh_pass=pass ansible_connection=ssh ansible_port=122
[kube-node]
heku1 ansible_host=ip1 ansible_ssh_user=root ansible_ssh_pass=pass ansible_connection=ssh ansible_port=122
heku2 ansible_host=ip2 ansible_ssh_user=root ansible_ssh_pass=pass ansible_connection=ssh ansible_port=122
heku3 ansible_host=ip3 ansible_ssh_user=root ansible_ssh_pass=pass ansible_connection=ssh ansible_port=122
heku4 ansible_host=newIP ansible_ssh_user=root ansible_ssh_pass=pass ansible_connection=ssh ansible_port=122
[k8s-cluster:children]
kube-node
kube-master
Command used to invoke ansible:
ansible-playbook -i inventory/inventory.ip cluster.yml --limit=heku4 -e "docker_dns_servers_strict=no" -v
Output of ansible run:
TASK [network_plugin/calico : Calico | Link etcd certificates for calico-node] ************************************************************************************************************************************
Thursday 14 June 2018 13:48:55 +0200 (0:00:00.136) 0:00:58.112 *********
changed: [heku4] => (item={u's': u'ca.pem', u'd': u'ca_cert.crt'}) => {"changed": true, "dest": "/etc/calico/certs/ca_cert.crt", "gid": 0, "group": "root", "item": {"d": "ca_cert.crt", "s": "ca.pem"}, "mode": "0640", "owner": "kube", "size": 1090, "src": "/etc/ssl/etcd/ssl/ca.pem", "state": "hard", "uid": 999}
failed: [heku4] (item={u's': u'node-heku4.pem', u'd': u'cert.crt'}) => {"changed": false, "item": {"d": "cert.crt", "s": "node-heku4.pem"}, "msg": "Error while linking: [Errno 2] No such file or directory", "path": "/etc/calico/certs/cert.crt", "state": "absent"}
failed: [heku4] (item={u's': u'node-heku4-key.pem', u'd': u'key.pem'}) => {"changed": false, "item": {"d": "key.pem", "s": "node-heku4-key.pem"}, "msg": "Error while linking: [Errno 2] No such file or directory", "path": "/etc/calico/certs/key.pem", "state": "absent"}
root@heku4 # ls /etc/ssl/etcd/ssl/
ca.pem
Anything else do we need to know:
I've found the issue #2867 and commented out the line, but I've get the same error.
I dont' know, I have also another three k8s nodes, can I copy the certs from one of them?
this #2820 solved my issue
merge #3215 should fix this issue. ticket could be closed.
I still have this problem on master branch. Log attached.
stdout-cluster.yml-2018-10-11T16_00_29+10_00.log
facing the same problem on both master branch and v2.7.0 on a new installation
I still have this problem on master branch. Log attached.
stdout-cluster.yml-2018-10-11T16_00_29+10_00.log
Problem solved with ba5c0fa364533b78cec4330383584546dc67f9b8
Issues go stale after 90d of inactivity.
Mark the issue as fresh with /remove-lifecycle stale.
Stale issues rot after an additional 30d of inactivity and eventually close.
If this issue is safe to close now please do so with /close.
Send feedback to sig-testing, kubernetes/test-infra and/or fejta.
/lifecycle stale
Stale issues rot after 30d of inactivity.
Mark the issue as fresh with /remove-lifecycle rotten.
Rotten issues close after an additional 30d of inactivity.
If this issue is safe to close now please do so with /close.
Send feedback to sig-testing, kubernetes/test-infra and/or fejta.
/lifecycle rotten
Rotten issues close after 30d of inactivity.
Reopen the issue with /reopen.
Mark the issue as fresh with /remove-lifecycle rotten.
Send feedback to sig-testing, kubernetes/test-infra and/or fejta.
/close
@fejta-bot: Closing this issue.
In response to this:
Rotten issues close after 30d of inactivity.
Reopen the issue with/reopen.
Mark the issue as fresh with/remove-lifecycle rotten.Send feedback to sig-testing, kubernetes/test-infra and/or fejta.
/close
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository.
Most helpful comment
I still have this problem on master branch. Log attached.
stdout-cluster.yml-2018-10-11T16_00_29+10_00.log