Replies: 2 comments 1 reply
-
This sounds like a cloud init template issue. This playbook doesn't do anything to the machine's networking. I've seen similar things when machines are cloned but not cleaned. |
Beta Was this translation helpful? Give feedback.
-
So far I have had success adding the following to my netplan configs. seems weird that setting these values on a pysical install were I set static IP that it still uses DHCP. Note this issue happened when I manually run through the ubuntu 22.04 live install media, and select manual for ip settings.
network:
version: 2
renderer: networkd
ethernets:
eth0:
dhcp4: no
addresses:
- 10.0.99.4/24
match:
macaddress: 26:e6:21:2a:6a:80
nameservers:
addresses:
- 10.0.44.50
- 10.0.44.52
search:
- turtleware.au
routes:
- to: default
via: 10.0.99.1
set-name: eth0 |
Beta Was this translation helpful? Give feedback.
-
I have 2 clusters
Production - 6 node (3 etcd, 3 worker)
Development - 3 node (3 etcd).
Production is cloud-init backed Ubuntu 22.04 machines with static ip's set in cloud init. When building the cluster the first time its online and runs without issues. When I get a power outage or reboot. all nodes come back online but cluster is not available.
On reboot of development my nodes get ip's
1 - 10.0.99.104/24
2 - 10.0.99.105/24 + 10.0.99.104/32
3 - 10.0.99.106/24
example
ip a
on the node which kills the clusterExpected Behavior
Cluster starts up without issues
Current Behavior
Cluster is unable to start, restart required on all non first node to trigger IP move.
Steps to Reproduce
Context (variables)
Operating system: Ubuntu 22.04 | Debian 12
Hardware: Lenovo Tiny m900 (Production) | Lenovo Tiny M703 (Ubuntu 22.04 Server)
Variables Used
all.yml
Hosts
host.ini
Possible Solution
I am wondering if this is a cloud-init | proxmox | k3s issue. I am only seeing this issue on nodes 1/2 of my clusters. It started happening about 2 months ago when I was using Debian 12, saw there was a cloud-init bug about ip's. Switched to Ubuntu 22.04 and seeing the same issues.
Terraform VM clone config
Extra Testing 6 March 2024
I have set DHCP leasing to start at 10.0.99.50 10.0.99.90 however that still does not solve the issue.
I have had some success with the following
Adding the following line to the cloud-init config on my secondary node moved the ip address issue to the third node. I don't see this as being a viable option but maybe it will help.
Beta Was this translation helpful? Give feedback.
All reactions