<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: RX out of buffer drops on 25/40/100G (Mellanox) interfaces in Firewall and Security Management</title>
    <link>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/190886#M35232</link>
    <description>&lt;P&gt;Thank you for the replay also....&lt;/P&gt;&lt;P&gt;I have added a file with the cmd outputs. My type of drops seems not to be related to the sk166424. I agree regarding the really small percentage of them - they are not affecting the production traffic, or at least no one is noticing them &lt;span class="lia-unicode-emoji" title=":slightly_smiling_face:"&gt;🙂&lt;/span&gt; I'm just trying to find some logic in them. Eventually I'll increase the RX buffer.&lt;/P&gt;&lt;P&gt;Some time ago I had a similar problem on a 21k appliance where in the end a NIC driver update solved the problem...but it was a really long time ago.&lt;/P&gt;</description>
    <pubDate>Tue, 29 Aug 2023 13:00:12 GMT</pubDate>
    <dc:creator>Tilen_Pegan</dc:creator>
    <dc:date>2023-08-29T13:00:12Z</dc:date>
    <item>
      <title>RX out of buffer drops on 25/40/100G (Mellanox) interfaces</title>
      <link>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/190699#M35191</link>
      <description>&lt;P&gt;After having an SR open for quite a while and doing a loop on same questions and suggestions for a few times, I’m slowly losing hope in getting anywhere near a logical explanation.&lt;/P&gt;&lt;P&gt;I would kindly ask if anyone is/was having the same “problem” as I am.&lt;/P&gt;&lt;P&gt;I’ll try to be short in the explanation:&lt;/P&gt;&lt;P&gt;After migrating from an older open server HW (minor SW change R81.10 JHF82 to JHF94) &amp;nbsp;from a 10G Intel i340-t2 copper to a 25G Mellanox Connect-X4, we started to observe RX drops on both bond slave interfaces (on the previous hardware it was 0). There’s not really a big amount of them (100 – 200 frames every 30s) but they are constantly increasing. (RX packets:710147729181 errors:0 dropped:11891177)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Details:&lt;/P&gt;&lt;P&gt;Connection: Logical Bond (LACP) 2x 25G&lt;/P&gt;&lt;P&gt;Implementation type: datacentre FW (only FW blade currently) with a 10/6 CoreXL/SND split. Most of the traffic is accelerated.&lt;/P&gt;&lt;P&gt;Accelerated pkts/Total pkts&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; : 1452240020060/1468567378332 (98%)&lt;/P&gt;&lt;P&gt;6 RX queues on the mentioned interfaces.&lt;/P&gt;&lt;P&gt;RX-ring size: 1024&lt;/P&gt;&lt;P&gt;Traffic rates: average: 1G - 2G, PPS 200 – 300K; Bond traffic balance is in average a 10% difference. Most of the traffic is on those 25G data interfaces.&lt;/P&gt;&lt;P&gt;CPU load: SND CPU cores: 10 – 25%&lt;/P&gt;&lt;P&gt;CoreXL CPU cores: 5 – 15%&lt;/P&gt;&lt;P&gt;No CPU spikes not even near 100%. Drops are observed on the 10 – 25% CPU load.&lt;/P&gt;&lt;P&gt;Interrupt distribution is more or less normal:&lt;/P&gt;&lt;P&gt;68: 2121940581&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0 eth4-0&lt;/P&gt;&lt;P&gt;&amp;nbsp; 69:&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 1 1614573236&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 2&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0 eth4-1&lt;/P&gt;&lt;P&gt;&amp;nbsp; 70:&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 1&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0&amp;nbsp; 351831748&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 2&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0 eth4-2&lt;/P&gt;&lt;P&gt;&amp;nbsp; 71:&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 1&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0 2941469923&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0 eth4-3&lt;/P&gt;&lt;P&gt;&amp;nbsp; 72:&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 1&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0 1388895963&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0 eth4-4&lt;/P&gt;&lt;P&gt;&amp;nbsp; 73:&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;1&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0&amp;nbsp; 761200896 eth4-5&lt;/P&gt;&lt;P&gt;&amp;nbsp;112: 3360525067&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0 eth6-0&lt;/P&gt;&lt;P&gt;&amp;nbsp;113:&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 1 4282830995&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 2&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0 eth6-1&lt;/P&gt;&lt;P&gt;&amp;nbsp;114:&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 1&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0 3228875557&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 2&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0 eth6-2&lt;/P&gt;&lt;P&gt;&amp;nbsp;115:&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 1&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0&amp;nbsp; 330241849&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0 eth6-3&lt;/P&gt;&lt;P&gt;&amp;nbsp;116:&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 1&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0&amp;nbsp; 687642450&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0 eth6-4&lt;/P&gt;&lt;P&gt;&amp;nbsp;117:&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 1&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 0 1837144760 eth6-5&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Driver version:&lt;/P&gt;&lt;P&gt;driver: mlx5_core&lt;/P&gt;&lt;P&gt;version: 5.5-1.0.3 (07 Dec 21)&lt;/P&gt;&lt;P&gt;firmware-version: 14.32.2004 (DEL2810000034)&lt;/P&gt;&lt;P&gt;Jumbo frames enabled (no change from previous HW)&lt;/P&gt;&lt;P&gt;RX drop type: rx_out_of_buffer: 11891177&lt;/P&gt;&lt;P&gt;Which should translate to …. CPU or OS was too lazy or overloaded to process all the frames listed in the RX ring buffer.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Default coalesce parameters:&lt;/P&gt;&lt;P&gt;Coalesce parameters for eth4:&lt;/P&gt;&lt;P&gt;Adaptive RX: on&amp;nbsp; TX: on&lt;/P&gt;&lt;P&gt;rx-usecs: 32&lt;/P&gt;&lt;P&gt;rx-frames: 64&lt;/P&gt;&lt;P&gt;rx-usecs-irq: 0&lt;/P&gt;&lt;P&gt;rx-frames-irq: 0&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;In desperation tried to optimize those without any noticeable difference:&lt;/P&gt;&lt;P&gt;Coalesce parameters for eth4:&lt;/P&gt;&lt;P&gt;Adaptive RX: off&amp;nbsp; TX: off&lt;/P&gt;&lt;P&gt;rx-usecs: 8&lt;/P&gt;&lt;P&gt;rx-frames: 32&lt;/P&gt;&lt;P&gt;rx-usecs-irq: 0&lt;/P&gt;&lt;P&gt;rx-frames-irq: 0&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I could be wrong with the following: a “dummy” calculation of a maximal possible number of frames in the RX ring buffer (if CPU is coping with its tasks): Traffic rate 300K PPS /2 per bond slave= 150K PPS&lt;/P&gt;&lt;P&gt;Convert to micro seconds: 0.15 PPus&lt;/P&gt;&lt;P&gt;rx-usec=8 / rx-frames=32 ….a frame descriptor in the RX buffer should wait 8us before an Interrupt is generated and sent to one of the 6 available CPUs and a max of 32 descriptors should be written before an Interrupt.&lt;/P&gt;&lt;P&gt;With my traffic rate:&lt;/P&gt;&lt;P&gt;8us/0.15ppus = 53 frames which is more than 32 so in theory an interrupt is generated every 4.8us (32 x 0.15). More than 32 descriptors should not be present in the RX buffer (of course, there are also other parameters that affect that)…far from the current 1024….&lt;/P&gt;&lt;P&gt;OS weight and budget are fixed values (64, 300)&lt;/P&gt;&lt;P&gt;I’m probably missing something here &lt;span class="lia-unicode-emoji" title=":smiling_face_with_smiling_eyes:"&gt;😊&lt;/span&gt;&lt;/P&gt;&lt;P&gt;To not make it just an “open server” thing, I’m observing the same behavior on a Maestro Backplane 100G interfaces (same Mellanox driver) on 16600 modules. CPU load is also not much different than the one mentioned here.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;The obvious solution: increase the RX ring size. I’m trying to find a logical explanation for this.&lt;/P&gt;&lt;P&gt;Would appreciate any thoughts or experience on that matter.&lt;/P&gt;</description>
      <pubDate>Mon, 28 Aug 2023 08:06:45 GMT</pubDate>
      <guid>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/190699#M35191</guid>
      <dc:creator>Tilen_Pegan</dc:creator>
      <dc:date>2023-08-28T08:06:45Z</dc:date>
    </item>
    <item>
      <title>Re: RX out of buffer drops on 25/40/100G (Mellanox) interfaces</title>
      <link>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/190704#M35193</link>
      <description>&lt;P&gt;I assume you started with default value of 256 RX buffer? Sometimes 1024 is not enough you can double it to 2048 and even 4096.&lt;/P&gt;&lt;P&gt;Please be aware that counters only reset after reboot. So make a note of the value of drops before the change and check if they still increase after the change. Give it couple of hours during peak hours.&lt;/P&gt;&lt;P&gt;Also maybe the output this on the relevant interfaces could show some more relevant information:&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;ethtool -S &amp;lt;IF name&amp;gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Mon, 28 Aug 2023 08:55:01 GMT</pubDate>
      <guid>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/190704#M35193</guid>
      <dc:creator>Lesley</dc:creator>
      <dc:date>2023-08-28T08:55:01Z</dc:date>
    </item>
    <item>
      <title>Re: RX out of buffer drops on 25/40/100G (Mellanox) interfaces</title>
      <link>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/190716#M35194</link>
      <description>&lt;P&gt;Please provide output of &lt;STRONG&gt;netstat -ni&lt;/STRONG&gt; counters for the relevant interface along with its full&amp;nbsp;&lt;STRONG&gt;ethtool -S (interface)&lt;/STRONG&gt; output.&amp;nbsp; Not all RX-DRPs are traffic getting lost that you actually care about; starting in Gaia 3.10 receipt of unknown EtherTypes and improper VLAN tags will cause RX-DRP to increment.&amp;nbsp;&amp;nbsp;&lt;A href="https://support.checkpoint.com/results/sk/sk166424" target="_blank" rel="noopener"&gt;&lt;SPAN&gt;sk166424: Number of RX packet drops on interfaces increases on a Security Gateway R80.30 and higher with Gaia kernel 3.10&lt;/SPAN&gt;&lt;/A&gt;&lt;/P&gt;
&lt;P&gt;If RX-DRP is &amp;lt;0.1% of RX-OK for an interface, generally that is OK and you should not increase ring buffer size in a quest to make RX-DRP absolutely zero.&amp;nbsp; Such a quest can lead you to &lt;A href="https://en.wikipedia.org/wiki/Bufferbloat" target="_self"&gt;Bufferbloat&lt;/A&gt; and actually make performance worse, and in some cases much worse.&amp;nbsp; We can discuss some other options after seeing the command outputs.&lt;/P&gt;</description>
      <pubDate>Mon, 28 Aug 2023 11:55:56 GMT</pubDate>
      <guid>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/190716#M35194</guid>
      <dc:creator>Timothy_Hall</dc:creator>
      <dc:date>2023-08-28T11:55:56Z</dc:date>
    </item>
    <item>
      <title>Re: RX out of buffer drops on 25/40/100G (Mellanox) interfaces</title>
      <link>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/190882#M35230</link>
      <description>&lt;P&gt;Thanks for the replay. Well on the new HW I already started with 1024 - on the previous one I had the default 256 (on the 10G int). With the same traffic and same number of RX queues I would suppose the 1024 should be enough in regards to the old HW didn't have any RX drops.&lt;/P&gt;&lt;P&gt;ethtoll -S shows only "out of buffer" drops.&lt;/P&gt;</description>
      <pubDate>Tue, 29 Aug 2023 12:47:03 GMT</pubDate>
      <guid>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/190882#M35230</guid>
      <dc:creator>Tilen_Pegan</dc:creator>
      <dc:date>2023-08-29T12:47:03Z</dc:date>
    </item>
    <item>
      <title>Re: RX out of buffer drops on 25/40/100G (Mellanox) interfaces</title>
      <link>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/190886#M35232</link>
      <description>&lt;P&gt;Thank you for the replay also....&lt;/P&gt;&lt;P&gt;I have added a file with the cmd outputs. My type of drops seems not to be related to the sk166424. I agree regarding the really small percentage of them - they are not affecting the production traffic, or at least no one is noticing them &lt;span class="lia-unicode-emoji" title=":slightly_smiling_face:"&gt;🙂&lt;/span&gt; I'm just trying to find some logic in them. Eventually I'll increase the RX buffer.&lt;/P&gt;&lt;P&gt;Some time ago I had a similar problem on a 21k appliance where in the end a NIC driver update solved the problem...but it was a really long time ago.&lt;/P&gt;</description>
      <pubDate>Tue, 29 Aug 2023 13:00:12 GMT</pubDate>
      <guid>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/190886#M35232</guid>
      <dc:creator>Tilen_Pegan</dc:creator>
      <dc:date>2023-08-29T13:00:12Z</dc:date>
    </item>
    <item>
      <title>Re: RX out of buffer drops on 25/40/100G (Mellanox) interfaces</title>
      <link>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/190933#M35249</link>
      <description>&lt;P&gt;Those are legit ring buffer drops but the RX-DRP rate is so low I wouldn't worry about it.&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Are the RX-DRPs incrementing slowly &amp;amp; continuously?&amp;nbsp; Or do you get&amp;nbsp; a clump of them during a policy installation?&amp;nbsp; &lt;STRONG&gt;sar -n EDEV&lt;/STRONG&gt; can help reveal if the RX-DRPs are continuous or only happening en masse during policy installations.&lt;/P&gt;
&lt;P&gt;I was under the impression that the default ring buffer sizes for Intel 1GB is 256, Intel 10GB is 512, and Mellanox/NVIDIA 25GB+ is 1024 so you may already be at the default ring buffer setting.&amp;nbsp;&lt;/P&gt;
&lt;P&gt;If you still want to increase it I'd suggest doubling the default size (do not crank it to the maximum) and going from there, but both before and after doing so I'd strongly advise running this bufferbloat test through the interface with the enlarged ring buffer to validate you haven't made jitter substantially worse:&lt;/P&gt;
&lt;P&gt;&lt;A href="https://www.waveform.com/tools/bufferbloat" target="_blank" rel="noopener"&gt;https://www.waveform.com/tools/bufferbloat&lt;/A&gt;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Tue, 29 Aug 2023 16:36:53 GMT</pubDate>
      <guid>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/190933#M35249</guid>
      <dc:creator>Timothy_Hall</dc:creator>
      <dc:date>2023-08-29T16:36:53Z</dc:date>
    </item>
    <item>
      <title>Re: RX out of buffer drops on 25/40/100G (Mellanox) interfaces</title>
      <link>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/191042#M35275</link>
      <description>&lt;P&gt;RX-DRPs are incrementing continuously...but in clumps of i.e. 500 - 1000 packets get dropped every 60s. They are not related to policy installs. (attached log and short drop statistics)&lt;/P&gt;&lt;P&gt;Regarding the defaults; my bad, I have gone over the configuration again and apparently while I have increased the RX-buffer on the 1G interfaces to 1024, I did the same for 25G without checking the default, which is as you said 1024....so, I'm on defaults on the 25G. Sorry for the misleading info.&lt;/P&gt;&lt;P&gt;But still, why do you think those drop happen? If we look at the CPU usage, it should easily handle the current packet rate and never go over the current RX ring size with the current coalesce timings. Unless my bond split over two physical NICs ( 2 x (2x 25G NIC)) is worsening the thing.&lt;/P&gt;&lt;P&gt;Tasks: 423 total, 2 running, 421 sleeping, 0 stopped, 0 zombie&lt;BR /&gt;%Cpu0 : 0.0 us, 0.0 sy, 0.0 ni, 88.0 id, 0.0 wa, 2.0 hi, 10.0 si, 0.0 st&lt;BR /&gt;%Cpu1 : 0.0 us, 0.0 sy, 0.0 ni, 85.9 id, 0.0 wa, 2.0 hi, 12.1 si, 0.0 st&lt;BR /&gt;%Cpu2 : 0.0 us, 0.0 sy, 0.0 ni, 87.9 id, 0.0 wa, 2.0 hi, 10.1 si, 0.0 st&lt;BR /&gt;%Cpu3 : 0.0 us, 0.0 sy, 0.0 ni, 84.2 id, 0.0 wa, 3.0 hi, 12.9 si, 0.0 st&lt;BR /&gt;%Cpu4 : 0.0 us, 0.0 sy, 0.0 ni, 89.0 id, 0.0 wa, 2.0 hi, 9.0 si, 0.0 st&lt;BR /&gt;%Cpu5 : 0.0 us, 1.0 sy, 0.0 ni, 90.1 id, 0.0 wa, 2.0 hi, 6.9 si, 0.0 st&lt;BR /&gt;%Cpu6 : 1.0 us, 4.9 sy, 0.0 ni, 94.1 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st&lt;BR /&gt;%Cpu7 : 0.0 us, 3.1 sy, 0.0 ni, 96.9 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st&lt;BR /&gt;%Cpu8 : 0.0 us, 4.0 sy, 0.0 ni, 95.0 id, 0.0 wa, 0.0 hi, 1.0 si, 0.0 st&lt;BR /&gt;%Cpu9 : 2.0 us, 3.0 sy, 0.0 ni, 95.0 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st&lt;BR /&gt;%Cpu10 : 0.0 us, 6.0 sy, 0.0 ni, 94.0 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st&lt;BR /&gt;%Cpu11 : 1.0 us, 4.0 sy, 0.0 ni, 95.0 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st&lt;BR /&gt;%Cpu12 : 0.0 us, 3.0 sy, 0.0 ni, 95.0 id, 0.0 wa, 1.0 hi, 1.0 si, 0.0 st&lt;BR /&gt;%Cpu13 : 1.0 us, 4.0 sy, 0.0 ni, 95.0 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st&lt;BR /&gt;%Cpu14 : 0.0 us, 6.0 sy, 0.0 ni, 93.0 id, 0.0 wa, 1.0 hi, 0.0 si, 0.0 st&lt;BR /&gt;%Cpu15 : 1.0 us, 3.0 sy, 0.0 ni, 95.0 id, 0.0 wa, 1.0 hi, 0.0 si, 0.0 st&lt;BR /&gt;KiB Mem : 64903508 total, 46446860 free, 12609436 used, 5847212 buff/cache&lt;BR /&gt;KiB Swap: 33551748 total, 33551748 free, 0 used. 51006384 avail Mem&lt;/P&gt;&lt;P&gt;I don't want to waste your time on a thing that is actually not critical. As I said, I'm just trying to make some logic out of those drops.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Wed, 30 Aug 2023 07:21:13 GMT</pubDate>
      <guid>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/191042#M35275</guid>
      <dc:creator>Tilen_Pegan</dc:creator>
      <dc:date>2023-08-30T07:21:13Z</dc:date>
    </item>
    <item>
      <title>Re: RX out of buffer drops on 25/40/100G (Mellanox) interfaces</title>
      <link>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/191064#M35277</link>
      <description>&lt;P&gt;As there are just&amp;nbsp;RX-DRP but 0 RX-OVR it's IMHO not a CPU/hw-limit problem. The network sends packages to the system which are not related. Maybe not configured VLAN(s) on GAiA....&lt;/P&gt;</description>
      <pubDate>Wed, 30 Aug 2023 09:06:46 GMT</pubDate>
      <guid>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/191064#M35277</guid>
      <dc:creator>Daniel_</dc:creator>
      <dc:date>2023-08-30T09:06:46Z</dc:date>
    </item>
    <item>
      <title>Re: RX out of buffer drops on 25/40/100G (Mellanox) interfaces</title>
      <link>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/191090#M35279</link>
      <description>&lt;P&gt;I agree the SND CPUs should be able to keep up given their low load.&amp;nbsp; Once again the RX-DRP rate is negligible and fixing it won't make anywhere close to a noticeable difference.&lt;/P&gt;
&lt;P&gt;However since you are on open server I assume you are using a static CoreXL split.&amp;nbsp; You could try increasing the number of SNDs, as doing so will create additional ring buffer queues to help spread out the traffic more and prevent RX-DRPs.&amp;nbsp; However if the issue is being caused by elephant/heavy flows (&lt;STRONG&gt;fw ctl multik print_heavy_conn&lt;/STRONG&gt;), the packets of a single elephant flow will always be "stuck" to the same SND core and its single ring buffer for that interface, so adding more SNDs and queues for an interface will probably not help.&lt;/P&gt;
&lt;P&gt;So there are two things that may be happening:&lt;/P&gt;
&lt;P&gt;1) When a SoftIRQ run starts on the dispatcher, it is able to completely empty all ring buffers and does not have to stop short.&amp;nbsp; However the ring buffer completely fills up and overflows before the next SoftIRQ run starts.&amp;nbsp; This is the less likely of the two scenarios and if it is occurring I don't think there is much you can do about it, other than perhaps adjusting something to make SoftIRQ runs happen more often.&lt;/P&gt;
&lt;P&gt;2) The SoftIRQ run is being stopped short due to the maximum allowable number of frames having been processed or two jiffies going by.&amp;nbsp; Because the emptying was incomplete, an overflow of the ring buffer is more likely before the next SoftIRQ run can start.&amp;nbsp; See the following 2 pages from my book below, in the age of 25Gbps+ interfaces perhaps the default value of &lt;STRONG&gt;net.core.netdev_budget&lt;/STRONG&gt; is insufficient and should be increased.&amp;nbsp; But be warned that doing so could upset the delicate balance of CPU utilization between SoftIRQ runs and CPU utilization by the sim (SecureXL) driver on a SND core.&amp;nbsp;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="budget1.png" style="width: 765px;"&gt;&lt;img src="https://community.checkpoint.com/t5/image/serverpage/image-id/22273i593BA9FC2652D9DC/image-size/large?v=v2&amp;amp;px=999" role="button" title="budget1.png" alt="budget1.png" /&gt;&lt;/span&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="budget2.png" style="width: 773px;"&gt;&lt;img src="https://community.checkpoint.com/t5/image/serverpage/image-id/22274iF8BC9506C56B62F0/image-size/large?v=v2&amp;amp;px=999" role="button" title="budget2.png" alt="budget2.png" /&gt;&lt;/span&gt;&lt;/P&gt;</description>
      <pubDate>Wed, 30 Aug 2023 13:54:29 GMT</pubDate>
      <guid>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/191090#M35279</guid>
      <dc:creator>Timothy_Hall</dc:creator>
      <dc:date>2023-08-30T13:54:29Z</dc:date>
    </item>
    <item>
      <title>Re: RX out of buffer drops on 25/40/100G (Mellanox) interfaces</title>
      <link>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/191195#M35296</link>
      <description>&lt;P&gt;I have already checked the softnet_stats but was unable to correlate the output to my drops. Although the third column is showing that some SoftIRQ stopped early those numbers do not change in correlation to the increasing number of RX-DRP:&lt;/P&gt;&lt;P&gt;i.e.:&lt;/P&gt;&lt;P&gt;when I first checked a few days ago:&lt;/P&gt;&lt;P&gt;526d9b30 00000000 0000001b&lt;BR /&gt;2373b998 00000000 00000046&lt;BR /&gt;ffb1af7e 00000000 00000039&lt;BR /&gt;9623181c 00000000 0000003c&lt;BR /&gt;27ed8578 00000000 00000037&lt;BR /&gt;260510b2 00000000 00000039&lt;BR /&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;and from today:&lt;/P&gt;&lt;P&gt;ab2bae59 00000000 0000001b&lt;BR /&gt;58b6c52c 00000000 00000046&lt;BR /&gt;757e1d0b 00000000 0000003a&lt;BR /&gt;ff2fa192 00000000 0000003e&lt;BR /&gt;b58f086d 00000000 00000037&lt;BR /&gt;d51737f5 00000000 00000039&lt;BR /&gt;00b951bc 00000000 00000000&lt;BR /&gt;00a862d8 00000000 00000000&lt;BR /&gt;00bb74b7 00000000 00000000&lt;BR /&gt;00cc18e3 00000000 00000000&lt;BR /&gt;00de0e4b 00000000 00000000&lt;BR /&gt;00d1e151 00000000 00000000&lt;BR /&gt;00c31fc0 00000000 00000000&lt;BR /&gt;00bcc946 00000000 00000000&lt;BR /&gt;00ac045d 00000000 00000000&lt;BR /&gt;00c0c86b 00000000 00000000&lt;/P&gt;&lt;P&gt;I would interpret this output that although some SoftIRQ are eventually terminated earlier, not packets are getting dropped - is this correct?&lt;/P&gt;&lt;P&gt;Regarding adding more CPUs to SND and eventually lower the number of packets in each RX buffer would make sense if I had reasonably high SND CPU usage. As you wrote, wouldn't help with elephant flows which are the usual reason for CPU spikes...but fortunately I'm not in the elephant flow territory.&lt;/P&gt;&lt;P&gt;I wouldn't even be digging into those drops if I had them also on the previous hardware with 10G interfaces (I must correct myself as you said, the default buffer on 10G Intel was 512 and not 256)....&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thanks a lot for your opinions/suggestions. I'll probably go and increase the RX buffer to 2048 and stop digging into this matter.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Thu, 31 Aug 2023 08:40:03 GMT</pubDate>
      <guid>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/191195#M35296</guid>
      <dc:creator>Tilen_Pegan</dc:creator>
      <dc:date>2023-08-31T08:40:03Z</dc:date>
    </item>
    <item>
      <title>Re: RX out of buffer drops on 25/40/100G (Mellanox) interfaces</title>
      <link>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/191197#M35297</link>
      <description>&lt;P&gt;Yes, indeed I though about those drops but still they should be shown under different stats rather than my "rx_out_of_buffer" which at least by name, should not point to "not related packets". I also should have had them on the previous hardware - as in regards to network there was no change - drops appeared right after I switched to the new servers.&lt;/P&gt;</description>
      <pubDate>Thu, 31 Aug 2023 08:50:08 GMT</pubDate>
      <guid>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/191197#M35297</guid>
      <dc:creator>Tilen_Pegan</dc:creator>
      <dc:date>2023-08-31T08:50:08Z</dc:date>
    </item>
    <item>
      <title>Re: RX out of buffer drops on 25/40/100G (Mellanox) interfaces</title>
      <link>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/191248#M35306</link>
      <description>&lt;P&gt;Right some SoftIRQ runs are ending early based on that output but those are very low numbers.&amp;nbsp; However even through a few runs were ended early, that does not necessarily mean that RX-DRPs will occur as a result, as the next SoftIRQ run may commence and start emptying the ring buffer again before it fills up completely.&amp;nbsp; I guess if you want to increase the ring buffer size a reasonable amount for peace of mind that will be fine.&lt;/P&gt;
&lt;P&gt;After thinking about this some more, I believe the reason you started seeing this when you moved from Intel to Mellanox/NVIDIA adapters is due to the difference in how these adapters handle a full ring buffer.&amp;nbsp; Allow me to explain.&lt;/P&gt;
&lt;P&gt;When certain advanced Intel NIC hardware/drivers attempt to DMA a frame from the NIC hardware buffer into a Gaia memory socket buffer (referenced with descriptors contained in the ring buffer), if the ring buffer is full at that time the frame is "pulled back", the NIC waits a moment, then tries to transfer it again hoping a slot has opened up.&amp;nbsp;However suppose that the ring buffer is perennially full and frames are constantly being pulled back and having to try again.&amp;nbsp; Eventually enough incoming frames pile up in the NIC hardware buffer that it too overruns and incoming frames from the wire begin to be lost.&amp;nbsp; The NIC indicates this condition by incrementing both RX-DRP and RX-OVR in "lock step" for each lost frame, in other words these two counters always have exactly the same value for that interface.&amp;nbsp; Even though technically the loss was due to an overrun of the NIC's hardware buffer, it was actually caused by backpressure from the perennially full ring buffer.&lt;/P&gt;
&lt;P&gt;On older Intel NIC hardware/drivers the ability to "pull back" the frame and try again did not exist, and the frame would simply be lost if the ring buffer was full with RX-DRP incremented, along with the appropriate fifo/missed error counter.&amp;nbsp; I would speculate that the Mellanox/NVIDIA adapters do not have the "pull back" capability which I find a bit surprising considering how supposedly advanced these NICs are, or perhaps they have some kind of limit on how many times a single frame can be pulled back, and when they eventually give up and drop the frame RX-DRP is incremented.&amp;nbsp;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Thu, 31 Aug 2023 15:14:16 GMT</pubDate>
      <guid>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/191248#M35306</guid>
      <dc:creator>Timothy_Hall</dc:creator>
      <dc:date>2023-08-31T15:14:16Z</dc:date>
    </item>
    <item>
      <title>Re: RX out of buffer drops on 25/40/100G (Mellanox) interfaces</title>
      <link>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/191328#M35316</link>
      <description>&lt;P&gt;That theory would explain why there's no RX-OVR drops - could be, thanks. I would also find strange that Mellanox is not having the "pull back" function implemented or it really reaches the "retry count"....&lt;/P&gt;&lt;P&gt;But every reason for those drops points to the system not being able to pull the packets out of the ring buffer in time - but this is in contradiction with&amp;nbsp;softnet_stats which are saying that the NAPI poll did not leave any packets behind - at least from the stats point of view.&lt;/P&gt;&lt;P&gt;I still having trouble to imagine that after the NIC writes the frames into RAM via DMA and in my case (at least from my understanding) NIC with RSS (6x queue) has 6 x 1024 frames of space to write and read from it and with a really low loaded CPUs not able to cope with that. Unless there's something with ring buffer space reservation for each queue - I would suppose that RX memory reservation size is in line with the (possible) accepted max frame size on the interface, where mine is Jumbo 9216 - does not really make much sense &lt;span class="lia-unicode-emoji" title=":slightly_smiling_face:"&gt;🙂&lt;/span&gt;&lt;/P&gt;&lt;P&gt;Unless the counter rx-out-of-buf is named inappropriately and means also other than just RX buffer on a queue was exhausted.&lt;/P&gt;&lt;P&gt;Thanks Timothy for your explanations. I will stop here in digging deeper as after all the articles red and talks, I'm starting to find myself in a loop.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Fri, 01 Sep 2023 11:06:44 GMT</pubDate>
      <guid>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/191328#M35316</guid>
      <dc:creator>Tilen_Pegan</dc:creator>
      <dc:date>2023-09-01T11:06:44Z</dc:date>
    </item>
    <item>
      <title>Re: RX out of buffer drops on 25/40/100G (Mellanox) interfaces</title>
      <link>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/192513#M35585</link>
      <description>&lt;P&gt;We have very similar problems with the 25G NICs from Mellanox.&lt;BR /&gt;We have had the high error rate since migrating to these systems.&lt;BR /&gt;However, the situation became significantly worse with the update to R81.20 and the associated update of the NIC drivers.&lt;BR /&gt;&lt;A href="https://network.nvidia.com/pdf/firmware/ConnectX4-FW-12_22_1002-release_notes.pdf" target="_blank"&gt;https://network.nvidia.com/pdf/firmware/ConnectX4-FW-12_22_1002-release_notes.pdf&lt;/A&gt;&lt;BR /&gt;Since then, after a policy push, we have had the problem of losing a number of applications in monitoring.&lt;BR /&gt;Case is pending.&lt;/P&gt;</description>
      <pubDate>Wed, 13 Sep 2023 10:48:24 GMT</pubDate>
      <guid>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/192513#M35585</guid>
      <dc:creator>LST</dc:creator>
      <dc:date>2023-09-13T10:48:24Z</dc:date>
    </item>
    <item>
      <title>Re: RX out of buffer drops on 25/40/100G (Mellanox) interfaces</title>
      <link>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/192544#M35590</link>
      <description>&lt;P&gt;Exactly, the impact of NIC driver updates is covered in my &lt;A href="http://www.maxpowerfirewalls.com/gw-optimization-course.html" target="_blank" rel="noopener"&gt;Gateway Performance Optimization Class&lt;/A&gt; as something to watch out for upon an upgrade or Jumbo HFA application.&amp;nbsp; It does not look like the mlx* driver version shipped with R81.20 GA has been updated yet, at least as of Take 24.&amp;nbsp; Here is the relevant course content which you may find useful:&lt;/P&gt;
&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="nic_drivers.png" style="width: 999px;"&gt;&lt;img src="https://community.checkpoint.com/t5/image/serverpage/image-id/22444i174A1892034C48E9/image-size/large?v=v2&amp;amp;px=999" role="button" title="nic_drivers.png" alt="nic_drivers.png" /&gt;&lt;/span&gt;&lt;/P&gt;</description>
      <pubDate>Wed, 13 Sep 2023 14:03:36 GMT</pubDate>
      <guid>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/192544#M35590</guid>
      <dc:creator>Timothy_Hall</dc:creator>
      <dc:date>2023-09-13T14:03:36Z</dc:date>
    </item>
    <item>
      <title>Re: RX out of buffer drops on 25/40/100G (Mellanox) interfaces</title>
      <link>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/192615#M35606</link>
      <description>&lt;P&gt;I'm not entirely sure if I understood the difference between version and firmware (ethtool -i) correctly, but it's correct, we also have the version from your table (kernel driver?).&lt;BR /&gt;However, the firmware has changed, as announced in the release notes of R81.20.&lt;BR /&gt;&lt;A href="https://downloads.checkpoint.com/fileserver/SOURCE/direct/ID/125378/FILE/CP_R81.20_ReleaseNotes.pdf" target="_blank" rel="noopener"&gt;https://downloads.checkpoint.com/fileserver/SOURCE/direct/ID/125378/FILE/CP_R81.20_ReleaseNotes.pdf&lt;/A&gt;&lt;/P&gt;&lt;P&gt;driver: mlx5_core&lt;BR /&gt;version: 5.5-1.0.3 (13 May 22)&lt;BR /&gt;firmware-version: 12.22.1002 (CP_0000000001)&lt;/P&gt;&lt;P&gt;Unfortunately, we did not analyze the version or parameters in detail before the update.&lt;BR /&gt;But it could also be that other parameters have changed. For example, flow control is now activated on the 25G Mellanox cards, but it is off on the 10G Intel cards.&lt;BR /&gt;(ethtool -a):&lt;BR /&gt;RX: on&lt;BR /&gt;TX: on&lt;/P&gt;&lt;P&gt;Unhappily, after the last push we deleted the counters with a reboot, so we are not sure whether a pause flag was sent to the switch during the push (or shortly after) (flow control is switched off here).&lt;BR /&gt;Regrettably, we are currently not allowed to carry out any further tests.&lt;/P&gt;&lt;P&gt;The SK for permanently switching off Flow Control also sounds a bit strange.&lt;BR /&gt;&lt;A href="https://support.checkpoint.com/results/sk/sk61922" target="_blank" rel="noopener"&gt;https://support.checkpoint.com/results/sk/sk61922&lt;/A&gt;&lt;BR /&gt;Especially since it is also disabled for all other cards in the system.&lt;/P&gt;&lt;P&gt;So we are currently relatively unsure what caused our problem and what is the best way to fix it.&lt;BR /&gt;Back to R81.10? Actually unfortunate and the firmware of the SFP remains up to date. Because of the same driver,&amp;nbsp;&lt;SPAN&gt;shouldn't this be a problem?&lt;/SPAN&gt;&lt;BR /&gt;Use an old SFP (with old firmware)? Update to the R81.20 Take26 (from 24)?&lt;BR /&gt;Are there other parameters like ring buffer or something similar?&lt;/P&gt;</description>
      <pubDate>Thu, 14 Sep 2023 05:58:46 GMT</pubDate>
      <guid>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/192615#M35606</guid>
      <dc:creator>LST</dc:creator>
      <dc:date>2023-09-14T05:58:46Z</dc:date>
    </item>
    <item>
      <title>Re: RX out of buffer drops on 25/40/100G (Mellanox) interfaces</title>
      <link>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/192661#M35616</link>
      <description>&lt;P&gt;"Version" is the version of the Gaia kernel driver used to access the NIC hardware.&amp;nbsp; "firmware-version" is the level of software in the BIOS on the NIC hardware itself.&amp;nbsp; Actually it does appear that the firmware can be downgraded, see here:&lt;A href="https://support.checkpoint.com/results/sk/sk141812" target="_blank" rel="noopener"&gt;&lt;SPAN&gt;sk141812: Installing and Configuring Dual 40GbE and Dual 100/25GbE I/O card in Applicable Check Point Appliances&lt;/SPAN&gt;&lt;/A&gt;&lt;/P&gt;
&lt;P&gt;However I have found it a bit distressing that the Mellanox/NVIDIA cards have required firmware updates on the NIC card itself after being shipped to a customer, and that the mlx driver seems to be constantly getting major updates as shown in my table above.&amp;nbsp; I can't recall ever needing to update the firmware on any Intel-based NICs for a gateway, and the kernel driver versions for Intel-based NICs are rarely updated, and when they are it is very minor.&amp;nbsp; I understand that the Mellanox cards are on the bleeding edge as far as speed &amp;amp; capabilities, but whatever that firmware update fixed must have been pretty major, as firmware updates always run the risk of bricking the card if anything goes wrong.&lt;/P&gt;
&lt;P&gt;Generally pause frames (EthernetFlow Control) are disabled by default for Intel-based NICs but enabled for Mellanox NICs.&amp;nbsp; Not really a fan of pause frames myself as they can to some degree interfere with TCP's congestion control mechanism and degrade performance in certain situations; would be very interested to hear the justification for why they are enabled by default for Mellanox cards.&amp;nbsp; However even if pause frames are enabled on the NIC, the switch will ignore them unless it is configured to honor them.&lt;/P&gt;
&lt;P&gt;As far as where to go from here, please post the output of the following expert mode command so we can see what Multi-Queue is doing:&lt;/P&gt;
&lt;P&gt;&lt;STRONG&gt;mq_mng -o -v&lt;/STRONG&gt;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Thu, 14 Sep 2023 16:01:41 GMT</pubDate>
      <guid>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/192661#M35616</guid>
      <dc:creator>Timothy_Hall</dc:creator>
      <dc:date>2023-09-14T16:01:41Z</dc:date>
    </item>
    <item>
      <title>Re: RX out of buffer drops on 25/40/100G (Mellanox) interfaces</title>
      <link>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/192707#M35635</link>
      <description>&lt;P&gt;&lt;SPAN class=""&gt;&lt;SPAN class=""&gt;Thank you very much for the explanation.&lt;/SPAN&gt;&lt;/SPAN&gt;&lt;BR /&gt;&lt;SPAN class=""&gt;&lt;SPAN class=""&gt;Attached is the output of the command.&lt;BR /&gt;&lt;/SPAN&gt;&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Fri, 15 Sep 2023 05:34:06 GMT</pubDate>
      <guid>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/192707#M35635</guid>
      <dc:creator>LST</dc:creator>
      <dc:date>2023-09-15T05:34:06Z</dc:date>
    </item>
    <item>
      <title>Re: RX out of buffer drops on 25/40/100G (Mellanox) interfaces</title>
      <link>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/192794#M35657</link>
      <description>&lt;P&gt;It looks like Multi-Queue is doing a pretty good job of balancing the traffic between the various dispatcher cores.&amp;nbsp;&lt;/P&gt;
&lt;P&gt;One thing that is slightly unusual: there are 42 queues (dispatcher cores) available for Multi-Queue but the mlx driver is only using 30 cores, while the Intel i40e driver is utilizing 38 (supposedly i40e can use up to 64, while igb can only use 2-16 depending on NIC hardware and has 8 in use).&amp;nbsp; I was under the impression that the Mellanox cards could utilize up to 60 queues/dispatchers, but on your system it is stopping short at 30.&amp;nbsp; This may be a limitation of the 25Gbps cards you are using, whereas perhaps the Mellanox 40/100 Gbps cards can utilize 60.&amp;nbsp; That would be something ask TAC about.&lt;/P&gt;
&lt;P&gt;Only other thing is that Multi-Queue thinks it can utilize 42 cores, are you sure you have that many dispatchers in your CoreXL split?&amp;nbsp; Are you sure that a Firewall Worker Instance is not also accidentally assigned to one of these 42 dispatcher cores?&amp;nbsp; This will cause issues and is a common problem on VSX systems.&lt;/P&gt;</description>
      <pubDate>Fri, 15 Sep 2023 20:22:36 GMT</pubDate>
      <guid>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/192794#M35657</guid>
      <dc:creator>Timothy_Hall</dc:creator>
      <dc:date>2023-09-15T20:22:36Z</dc:date>
    </item>
    <item>
      <title>Re: RX out of buffer drops on 25/40/100G (Mellanox) interfaces</title>
      <link>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/192850#M35668</link>
      <description>&lt;P&gt;Hello Timothy,&lt;/P&gt;&lt;P&gt;thank you very much for these tips.&lt;BR /&gt;We had a maintenance window yesterday where we tried out a lot.&lt;BR /&gt;Checkpoint itself pointed us to this SK: &lt;A href="https://support.checkpoint.com/results/sk/sk180437" target="_blank"&gt;https://support.checkpoint.com/results/sk/sk180437&lt;/A&gt;&lt;BR /&gt;Since this unfortunately didn't work, we updated to the latest take (26), also without any improvement.&lt;BR /&gt;Then we looked at the balancing again and quickly reset it to default, which unfortunately didn't help either.&lt;BR /&gt;We also took another close look at the rx_err rate and considered whether a ring buffer change would help.&lt;BR /&gt;Since the error rate is well below 1%, we think that's fine.&lt;BR /&gt;However, we have noticed that the error rate on the passive is always higher.&lt;BR /&gt;So today we're going to investigate whether the switches are sending things that Checkpoint doesn't support or something like that.&lt;/P&gt;&lt;P&gt;Unfortunately, we still have our actual problem, which is that some applications fail after a policy push.&lt;BR /&gt;The fact that it is due to the bonds with the 25G cards is just a theory because we have not seen any problems on the bonds with the 10G cards so far (here the RX_ERR rate is always 0).&lt;/P&gt;&lt;P&gt;Best regards&lt;BR /&gt;Lars&lt;/P&gt;</description>
      <pubDate>Mon, 18 Sep 2023 05:39:23 GMT</pubDate>
      <guid>https://community.checkpoint.com/t5/Firewall-and-Security-Management/RX-out-of-buffer-drops-on-25-40-100G-Mellanox-interfaces/m-p/192850#M35668</guid>
      <dc:creator>LST</dc:creator>
      <dc:date>2023-09-18T05:39:23Z</dc:date>
    </item>
  </channel>
</rss>

