I’ve successfully created the Gravity tarfile for the cluster deployment, and have installed the cluster on a fresh Debian 9 image in GCP with: sudo ./gravity install --advertise-addr=10.168.0.10 --token=secret
(I confirm I can gravity enter
and then kubectl get pods
shows the active pods).
But I then spin up an identically configured GCP instance on the same VPC, in the same region/AZ, with firewall rules enabling full networking between those instances. And I attempt to join the existing cluster with: sudo ./gravity join 10.168.0.10 --advertise-addr=10.168.0.11 --token=secret
However, this times out or fails:
Wed Aug 7 17:44:27 UTC Still waiting for the planet to start (8 minutes elapsed)
Wed Aug 7 17:44:27 UTC Saving debug report to /home/user/mattermost/crashreport.tgz
And, on the second host, after sudo gravity enter
, I attempt to diagnose with kubectl get -o pods
but that eventually produces: Unable to connect to the server: dial tcp 10.100.0.1:443: i/o timeout
Within the second host’s environment (after I run sudo gravity enter
, that is), there is no indication of any running containers via docker ps -a
. journalctl -xe
gives: kube-proxy[109]: E0807 18:59:57.805298 109 proxier.go:1402] Failed to execute iptables-restore: exit status 2 (iptables-restore v1.6.0: Couldn't load target
KUBE-MARK-DROP’:No such file or directory, and then
kube-proxy[109]: Error occurred at line: 81`
I’m wondering what could cause all of this. Where would I look for further error messages? Is there a way to invoke a more failsafe or stable instance known to work?