Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Any known problems with GCP Compute Engine #384

Closed
Jonathan-Eid opened this issue Jun 14, 2024 · 6 comments
Closed

Any known problems with GCP Compute Engine #384

Jonathan-Eid opened this issue Jun 14, 2024 · 6 comments

Comments

@Jonathan-Eid
Copy link

Jonathan-Eid commented Jun 14, 2024

I cannot get master to worker communication if both nodes are in GCP and set to the same location, trying to get logs from worker nodes times out and the pods can't communicate with the kube apiserver. I tried putting the worker nodes in a different subnet as well and set their locations unique. When I do the latter, I can connect to the leader of that new subnet, but the same problems arise with any followers. What's going on? External AWS nodes connected to our GCP master work fine.

@squat
Copy link
Owner

squat commented Jun 14, 2024

Hi @Jonathan-Eid is this a GKE cluster? And are you running Kilo in add-on mode or as the only CNI for the cluster? If it's add-on mode then this is probably a limitation of the compatibility with the other CNI and you'll need to run Kilo in full-mesh mode to get full connectivity.

@Jonathan-Eid
Copy link
Author

We're not running on GKE, we setup vanilla K8s on fresh GCP Instances with Kilo as the main CNI, we're not running it on add-on mode. Please let me know if there's any more information I should send.

@Jonathan-Eid
Copy link
Author

Jonathan-Eid commented Jun 17, 2024

Here I'm calling tracepath from the master to two different pod ips

pod ip 192.168.7.3 is a pod on the worker that is the follower of the location i setup for workers, it hits the wg ip of the its leader @ 10.4.0.3 then times out

pod ip 192.168.8.2 is a pod on the leader of the location i setup for workers, it hits the wg ip and then resolves to the pod ip

image

@Jonathan-Eid
Copy link
Author

Seems like a GCP firewall issue actually, things started working when i opened up all ports and ip sources.

Do you have recommendations on which ports and ip sources i need to open on the masters and workers?

I opened up their external and private subnet ips to each other on the kubelet api port, kilo port, kube server api port, wasn't sure what i was missing honestly

@Jonathan-Eid
Copy link
Author

I allowed the ipip protocol on the firewall between nodes, now things are working normally

@squat
Copy link
Owner

squat commented Jun 17, 2024

Glad you got it working 💫

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants