Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Flatcar 3975.2.1 Bonding Config Bug #1580

Open
DFYT42 opened this issue Nov 13, 2024 · 1 comment
Open

Flatcar 3975.2.1 Bonding Config Bug #1580

DFYT42 opened this issue Nov 13, 2024 · 1 comment
Labels
kind/bug Something isn't working

Comments

@DFYT42
Copy link

DFYT42 commented Nov 13, 2024

Description

We've encountered a problem with bonding configs., after our most recent Flatcar upgrade from v3760.2.0 to v3975.2.1. The behavior is very weird, in that the bond0 interface actor churn does not always begin after the initial upgrade reboot. Instead, the bond0 interface actor churn most frequently appears after a subsequent reboot.

  • We can commonly recover from this by rebooting, but that does not always fix it
  • We have tried downing and reupping the effected Bond0 interface but that doesn't seem to have any effect
  • We tried to upgrade to the next known stable, 3975.2.2, but we see the same problem
  • We tried downgrading to v3760.2.0 and that worked-- the interface no longer enters churn
  • We then tried upgrading back to 3975.2.1, rebooting after the upgrade reboot, and churn reappeared

Impact

Nodes, rebooted after the initial upgrade reboot, go into churn on the secondary Bond0 interface and are subsequently unable to communicate with other nodes in the cluster.

Environment and steps to reproduce

  1. Set-up: [ describe the environment Flatcar/Nebraska etc was running in when encountering the bug; Platform etc. ]
    a. Baremetal Flatcar OS 3760.2.0 upgraded via Nebraska to Flatcar OS 3975.2.1
  2. Task: [ describe the task performing when encountering the bug ]
    a. After the node is upgraded and rebooted, the node is then rebooted a second time, and churn appears, which causes lag during node login and commands being run
  3. Action(s): [ sequence of actions that triggered the bug, see example below ]
    a. Rebooted the node, after the initial upgrade reboot
    b. Node login and commands begin to hang and take many seconds to minutes to complete
    c. /proc/net/bonding/bond0 shows churn on the secondary interface, and has no system mac address present
  4. Error: [describe the error that was triggered]
    a. Nodes were unable to communicate with effected node

Expected behavior

Expect nodes to communicate with other nodes in the cluster

Additional information

Please add any information here that does not fit the above format.

@DFYT42
Copy link
Author

DFYT42 commented Nov 14, 2024

We were asked for the following:

  • underlying hardware (network devices / virt environments etc.)
    • Bare metal
    • Problematic Flatcar OS 3975.2.1 & 3975.2.2
    • Working Flatcar OS 3760.2.0
    • Systemd 252
    • Kubernetes versions 1.28.x to 1.31.x
    • Server: Dell R6515
    • Switch: Juniper EX4300-48T
  • dmesg
     [   18.646307] ice 0000:41:00.1 enp65s0f1np1: NIC Link is up 25 Gbps Full Duplex, Requested FEC: RS-FEC, Negotiated FEC: FC-FEC/BASE-R, Autoneg Advertised: On, Autoneg Negotiated: False, Flow Control: None
[   18.666616] bond0: (slave enp65s0f1np1): Enslaving as a backup interface with an up link
[   18.675470] bond0: Warning: No 802.3ad response from the link partner for any adapters in the bond
[   18.686452] ice 0000:41:00.1 enp65s0f1np1: Error ADDING CP rule for fail-over
[   18.693782] ice 0000:41:00.1 enp65s0f1np1: Shared SR-IOV resources in bond are active
[   18.702648] ice 0000:41:00.0: Primary interface not in switchdev mode - VF LAG disabled

We were asked to try the following but are still seeing issues:

  • Create /etc/systemd/network/98-bond-mac.link
    • Add the following to the newly created /etc/systemd/network/98-bond-mac.link
[Match]
Type=bond

[Link]
MACAddressPolicy=none

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
kind/bug Something isn't working
Projects
Status: 📝 Needs Triage
Development

No branches or pull requests

1 participant