-
Notifications
You must be signed in to change notification settings - Fork 670
Captured frame from MAC associated with another peer #3713
Comments
That makes sense! |
This is from syslog on the node
|
One thing I noticed it starts in fastdp mode and then switch to sleeve mode and remains in sleeve
|
Given this log message:
It's quite believable that promiscuous mode would cause packets to appear where they aren't expected. Any idea why those devices are entering promiscuous mode? We've seen some strange symptoms in the past when |
that promiscuous event happens when the weave pods are restarted.
|
Sorry, I think I wanted |
|
OK, we'll let systemd off this time. And I looked at my own machine, and I have the same "promiscuous" messages, so that's probably a red herring also. |
I thought this was just cosmetic in my case until recently, when I observed the same issue on a different cluster. It was reported that there were some connection issues with a specific pod, and i could correlate this event around the same time. Is it possible that the actual frames are being dropped? I really have no clue what is happening, it may be just simpler to switch to a different plugin until this has been resolved. |
I may have some progress here. I followed the clue with fastdp failing back to sleeve. I found this #3378 and investigated 'certain MTU conditions' I am running Kube nodes on Openstack instances. Default MTU size on Ubuntu was 1500. I lowered Weave MTU from default 1376 to 1366 just to test and after restarting pods, I saw the fastdp mode did not fall back to sleeve. At this point i had all kinds of issues with my apps, so ultimately I started looking to enabling jumbo frames. Jumbo frames are enabled on Openstack compute nodes so I enabled MTU 8000 on the Kubernetes nodes (Ubuntu) and set MTU 7000 on Weave. I restarted Weave and have not seen the "captured frame" messages since the restart. I will keep an eye on this. My question - what could cause the frames associated with another peer, being observed by weave in sleeve mode and not in fastdp mode? |
I can confirm all the issues are resolved. |
Just a note: since this is used for ingress traffic from outside where I dont have control over MTU size, i had to revert back from jumbo frames. The node MTU size is now 1454 and Weave MTU size is 1364. (Not 1366 as I mentioned previously, since this needs to be divisible by four for fastdp mode(jeez even my cat knows it)) |
What happened?
My logs are flooded with the following messages:
ERRO: 2019/10/10 17:16:55.967023 Captured frame from MAC (be:ed:6f:3c:e5:6f) to (92:43:ab:22:f9:76) associated with another peer ba:93:aa:cc:34:a8(knode27-hf)
How to reproduce it?
This issue is ongoing and never stops.
Anything else we need to know?
Kubernetes installed by Kubespray on Openstack.
I am looking for help to troubleshoot this. I read some older issues here how this can be innocuous in some cases, but here its tens of millions event messages in 12 hours.
Edit: the mac addresses in question are always associated with Elasticsearch/Fluentd pods
Versions:
Logs:
ERRO: 2019/10/10 17:16:55.967023 Captured frame from MAC (be:ed:6f:3c:e5:6f) to (92:43:ab:22:f9:76) associated with another peer ba:93:aa:cc:34:a8(knode27-hf)
The text was updated successfully, but these errors were encountered: