Pods on Worker Node Can't Communicating with API Server or Service in general #11728
Replies: 3 comments
-
|
Beta Was this translation helpful? Give feedback.
-
Thank you for the suggestion! Regarding the VXLAN kernel module, I confirmed that it's loaded on the master node (lsmod | grep vxlan shows the module) and that the CONFIG_VXLAN=y setting is enabled on the worker node. For the second suggestion, I’ve disabled the hardware checksum offload on the flannel interface using ethtool -K flannel.1 tx-checksum-ip-generic off, but unfortunately, I’m still facing the same issue. Additionally, I tried performing a cluster reset, but I'm encountering the following errors:
Do you have any further recommendations? |
Beta Was this translation helpful? Give feedback.
-
Hi all, I’m providing an update on the previously reported issue regarding the routing problem between the pods and services in my Kubernetes cluster. Summary of the Issue:
Current Observations: Pods in the cluster:
Services in the cluster:
Traceroute from the worker-pod to the myservice service (IP: 10.43.109.4):
As observed, the traffic reaches the 192.168.1.1 router but is not routed properly to the service.
The traffic correctly reaches the destination pod in this case.
Actions Taken So Far: I have added an explicit route for the 10.43.0.0/24 network in the worker-pod using:
However, the issue persists and traffic towards the service is still not routed correctly. |
Beta Was this translation helpful? Give feedback.
-
Environmental Info:
K3s Version:
Node(s) CPU architecture, OS, and Version:
Master:
Worker:
Cluster Configuration:
1 server, 1 agent
Describe the bug:
I'm encountering an issue when installing K3s on my server and worker nodes. Any pod on the worker node that attempts to communicate with the API server experiences a timeout. I suspect that the issue might be related to CoreDNS. I am installing K3s without any additional configuration options, and I have also tried replacing Flannel with Calico, but the problem persists.
As an example, when attempting to install the NVIDIA Device Plugin, I get the following error:
This is just an example, as the same issue occurs with other applications like Prometheus.
I have also ensured that I have opened all the required ports in iptables according to the K3s requirements, and I have disabled ufw on both the worker and server nodes.
Additionally, I am not sure if it is relevant, but the pods on the worker node are unable to ping the API server, and DNS is functioning correctly on the master node.
Steps To Reproduce:
Expected behavior:
Pods on the worker node should be able to communicate with the API server without experiencing timeouts. The installation of additional components like the NVIDIA Device Plugin or Prometheus should work without issues.
Actual behavior:
Pods on the worker node that try to communicate with the API server are timing out.
Additional context / logs:
Logs coreDns
nodes:
Pods in the kube-system namespace:
Beta Was this translation helpful? Give feedback.
All reactions