-
Notifications
You must be signed in to change notification settings - Fork 2.4k
Open
Description
Hello
I'm experiencing DNS resolution failures between pods on a specific Kubernetes node that has a complex network bonding configuration. Other nodes in the cluster work perfectly.
Environment:
- Kubernetes 1.32.5
- Calico CNI with VXLAN backend
- RHEL 9.4 nodes
- Kubespray deployment
Problem Node Configuration: - Bond interface
bond0with two slaves (eno1,eno2) in active-backup mode - VLAN interface
bond0.114with IP10.172.114.1/26 - VLAN interface
bond0.1083with IP10.172.108.129/26 - Kubernetes node internal IP:
10.172.108.129
Symptoms: - DNS resolution times out in pods scheduled on this node
nslookup kubernetes.default.svc.cluster.localfails with "connection timed out; no servers could be reached"- Pods can communicate via IP but not via DNS
- CoreDNS pods are healthy and working on other nodes
What I've Tried:
- Verified CoreDNS pods are running and healthy
- Checked Calico daemonset configuration
- Compared with working nodes (which have simple
ens192interface without bonding) - Tested network connectivity - basic IP communication works
Current Calico Configuration:
env:
- name: IP_AUTODETECTION_METHOD
value: can-reach=$(NODEIP)Key Observations:
- Working nodes have simple network interface (
ens192) - Problem node has bond interface with VLAN tagging
- The bond0 interface itself has no IP address
- IP addresses are assigned to VLAN subinterfaces (
bond0.114,bond0.1083) - One VLAN interface has an incorrect /2 subnet mask
Questions:
- How should I configure Calico's IP autodetection for this bond+VLAN setup?
- Should the node IP be on the bond interface itself or can it remain on the VLAN interface?
- What's the recommended approach for Calico with bonded interfaces and VLANs?
- Could the incorrect /2 subnet mask be causing the DNS issues even if basic IP communication works?
Debug Information:
# On problem node
$ nmcli device status
bond0.1083 vlan connected
bond0 bond connected
eno1 ethernet connected
eno2 ethernet connected
# IP addresses
$ ip -4 addr show bond0
# (no IP address)
$ ip -4 addr show bond0.114
inet 10.172.114.1/26
$ ip -4 addr show bond0.1083
inet 10.172.108.129/26
# Kubernetes node
$ kubectl get node dc2spk8sprdma001 -o wide
INTERNAL-IP: 10.172.108.129Any guidance on proper Calico configuration for this bonded network setup would be greatly appreciated!
Metadata
Metadata
Assignees
Labels
No labels