-
Hi there, I'm setting up a Talos cluster on bare metal on a few blade servers. I've got the first control node up, but the first worker node gets stuck with the error in the title, and continuously reboots. I haven't changed the Snippet of `talos logs kubelet` on the worker
controlplane.yaml (the useful bits)version: v1alpha1
debug: false
persist: true
machine:
type: controlplane
token: ...
ca:
crt: ...
key: ...
certSANs: []
kubelet:
image: ghcr.io/siderolabs/kubelet:v1.26.1
defaultRuntimeSeccompProfileEnabled: true
disableManifestsDirectory: true
network:
interfaces:
- interface: eth0
addresses:
- 192.168.3.3/24
routes:
- network: 0.0.0.0/0
gateway: 192.168.3.1
vip:
ip: 192.168.3.19
nameservers:
- 8.8.8.8
install:
disk: /dev/sda
image: ghcr.io/siderolabs/installer:v1.3.3
bootloader: true
wipe: false
registries: {}
features:
rbac: true
stableHostname: true
apidCheckExtKeyUsage: true
systemDiskEncryption:
ephemeral:
provider: luks2
keys:
- nodeID: {}
slot: 0
state:
provider: luks2
keys:
- nodeID: {}
slot: 0
cluster:
id: ...
secret: ...
controlPlane:
endpoint: https://192.168.3.19:6443
clusterName: m1000e
network:
dnsDomain: cluster.local
podSubnets:
- 10.244.0.0/16
serviceSubnets:
- 10.96.0.0/12
discovery:
enabled: true
registries:
kubernetes: {}
[... omitted a bunch of auto-generated and untouched config ...]
worker.yamlversion: v1alpha1
debug: false
persist: true
machine:
type: worker
token: ...
ca:
crt: ...
key: ""
certSANs: []
kubelet:
image: ghcr.io/siderolabs/kubelet:v1.26.1
defaultRuntimeSeccompProfileEnabled: true
disableManifestsDirectory: true
network:
interfaces:
- interface: eth0
addresses:
- 192.168.3.4/24
routes:
- network: 0.0.0.0/0
gateway: 192.168.3.1
nameservers:
- 8.8.8.8
install:
disk: /dev/sda
image: ghcr.io/siderolabs/installer:v1.3.3
bootloader: true
wipe: false
registries: {}
features:
rbac: true
stableHostname: true
apidCheckExtKeyUsage: true
systemDiskEncryption:
ephemeral:
provider: luks2
keys:
- nodeID: {}
slot: 0
state:
provider: luks2
keys:
- nodeID: {}
slot: 0
cluster:
id: ...
secret: ...
controlPlane:
endpoint: https://192.168.3.19:6443
network:
dnsDomain: cluster.local
podSubnets:
- 10.244.0.0/16
serviceSubnets:
- 10.96.0.0/12
ca:
crt: ...
key: ""
discovery:
enabled: true
registries:
kubernetes: {} |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 1 reply
-
The error in the end means that Kubelet uses bootstrap token to provision the initial client certificate and register the node. |
Beta Was this translation helpful? Give feedback.
The error in the end means that
kubelet
on the worker node can't join the controlplane. The exact reason is hard to guess, but it might be bootstrap token being wrong? controlplane endpoint being wrong? ...Kubelet uses bootstrap token to provision the initial client certificate and register the node.