- CheckMates
- :
- Products
- :
- Quantum
- :
- Management
- :
- Re: SND/FW changes due to core and OS upgrade?
- Subscribe to RSS Feed
- Mark Topic as New
- Mark Topic as Read
- Float this Topic for Current User
- Bookmark
- Subscribe
- Mute
- Printer Friendly Page
Are you a member of CheckMates?
×- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
SND/FW changes due to core and OS upgrade?
Hi –
We currently have an 8 core open server HA cluster running 77.30 and are upgrading those servers to 16 core servers and R80.10. I’d like to understand what, if any, CoreXL changes I should make (or perhaps any other performance enhancements?). Unfortunately I don’t believe I can use the cpsizeme tool being these are Dell open servers. My initial thought is go to with four SNDs and 12 firewall works only because almost all interfaces are 10Gb.
There is a moderate amount of VoIP traffic, no VPN traffic and some IPS protections enabled, no other blades are enabled.
Some snapshots of command output is shown as a reference if anyone has thoughts, suggestions or experience with a change such as this. Thanks in advance.
gatewayA> fw ctl affinity -l -r
CPU 0: s0p3 s0p2 eth0 eth1 s7p1 s7p2
CPU 1: fw_3
CPU 2: s5p2 s0p4 s0p1 s4p1 s4p2
CPU 3: fw_2
CPU 4: fw_5
CPU 5: fw_1
CPU 6: fw_4
CPU 7: fw_0
All: fwd rtmd mpdaemon cpd cprid
gatewayA> fw ctl multik stat
ID | Active | CPU | Connections | Peak
----------------------------------------------
0 | Yes | 7 | 7588 | 144409
1 | Yes | 5 | 13856 | 125878
2 | Yes | 3 | 8876 | 136021
3 | Yes | 1 | 14764 | 147924
4 | Yes | 6 | 15675 | 149068
5 | Yes | 4 | 15052 | 150053
TOP
top - 23:06:32 up 20 days, 1:25, 1 user, load average: 0.63, 0.43, 0.38
Tasks: 171 total, 1 running, 170 sleeping, 0 stopped, 0 zombie
Cpu0 : 0.0%us, 0.3%sy, 0.0%ni, 80.3%id, 0.0%wa, 1.0%hi, 18.3%si, 0.0%st
Cpu1 : 0.7%us, 1.3%sy, 0.0%ni, 92.4%id, 0.0%wa, 0.0%hi, 5.6%si, 0.0%st
Cpu2 : 0.3%us, 0.3%sy, 0.0%ni, 71.7%id, 0.0%wa, 1.0%hi, 26.7%si, 0.0%st
Cpu3 : 0.0%us, 0.3%sy, 0.0%ni, 94.3%id, 0.0%wa, 0.0%hi, 5.3%si, 0.0%st
Cpu4 : 0.0%us, 0.0%sy, 0.0%ni, 94.7%id, 0.0%wa, 0.0%hi, 5.3%si, 0.0%st
Cpu5 : 0.0%us, 1.0%sy, 0.0%ni, 94.3%id, 0.0%wa, 0.0%hi, 4.7%si, 0.0%st
Cpu6 : 0.0%us, 0.3%sy, 0.0%ni, 92.3%id, 0.0%wa, 0.0%hi, 7.3%si, 0.0%st
Cpu7 : 0.3%us, 2.3%sy, 0.0%ni, 91.7%id, 0.0%wa, 0.0%hi, 5.7%si, 0.0%st
Mem: 32778728k total, 12168000k used, 20610728k free, 349588k buffers
Swap: 8385920k total, 0k used, 8385920k free, 6004236k cached
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
8004 admin 15 0 0 0 0 S 8 0.0 1988:24 fw_worker_4
8000 admin 15 0 0 0 0 S 6 0.0 2126:14 fw_worker_0
8003 admin 15 0 0 0 0 S 6 0.0 2135:56 fw_worker_3
8001 admin 15 0 0 0 0 S 6 0.0 2156:24 fw_worker_1
8002 admin 15 0 0 0 0 S 6 0.0 2105:59 fw_worker_2
8005 admin 15 0 0 0 0 S 5 0.0 1962:16 fw_worker_5
Netstat –ni (up for 20 days)
Iface | MTU_Met | NA | RX-OK | RX-ERR | RX-DRP | RX-OVR | TX-OK | TX-ERR | TX-DRP | TX-OVR | Flg |
bond11 | 1500 | 0 | 130349881574 | 0 | 9296 | 0 | 80152928665 | 0 | 0 | 0 | BMmRU |
bond12 | 1500 | 0 | 78527374619 | 0 | 939 | 0 | 133113696739 | 0 | 0 | 0 | BMmRU |
bond13 | 1500 | 0 | 27729450377 | 0 | 25 | 0 | 37792884917 | 0 | 0 | 0 | BMmRU |
bond13.3 | 1500 | 0 | 1637665830 | 0 | 0 | 0 | 729217784 | 0 | 0 | 0 | BMmRU |
bond13.11 | 1500 | 0 | 174927162 | 0 | 0 | 0 | 141145382 | 0 | 0 | 0 | BMmRU |
bond13.12 | 1500 | 0 | 355043413 | 0 | 0 | 0 | 528118604 | 0 | 0 | 0 | BMmRU |
bond13.16 | 1500 | 0 | 604802057 | 0 | 0 | 0 | 729538625 | 0 | 0 | 0 | BMmRU |
bond13.17 | 1500 | 0 | 1333 | 0 | 0 | 0 | 27573 | 0 | 0 | 0 | BMmRU |
bond13.18 | 1500 | 0 | 2296 | 0 | 0 | 0 | 31674 | 0 | 0 | 0 | BMmRU |
bond13.22 | 1500 | 0 | 162695764 | 0 | 0 | 0 | 149482123 | 0 | 0 | 0 | BMmRU |
bond13.23 | 1500 | 0 | 333894296 | 0 | 0 | 0 | 174112873 | 0 | 0 | 0 | BMmRU |
bond13.24 | 1500 | 0 | 12881847684 | 0 | 0 | 0 | 25321641913 | 0 | 0 | 0 | BMmRU |
bond13.26 | 1500 | 0 | 29909098 | 0 | 0 | 0 | 20838220 | 0 | 0 | 0 | BMmRU |
bond13.27 | 1500 | 0 | 1520095225 | 0 | 0 | 0 | 692780450 | 0 | 0 | 0 | BMmRU |
bond13.30 | 1500 | 0 | 42886857 | 0 | 0 | 0 | 66776815 | 0 | 0 | 0 | BMmRU |
bond13.36 | 1500 | 0 | 9970300137 | 0 | 0 | 0 | 9219791498 | 0 | 0 | 0 | BMmRU |
bond13.42 | 1500 | 0 | 733944 | 0 | 0 | 0 | 1146823 | 0 | 0 | 0 | BMmRU |
bond13.43 | 1500 | 0 | 13283229 | 0 | 0 | 0 | 13369836 | 0 | 0 | 0 | BMmRU |
eth0 | 1500 | 0 | 37812616839 | 0 | 444 | 0 | 76189885995 | 0 | 0 | 0 | BMsRU |
eth1 | 1500 | 0 | 40714758548 | 0 | 495 | 0 | 56923813966 | 0 | 0 | 0 | BMsRU |
lo | 16436 | 0 | 104048 | 0 | 0 | 0 | 104048 | 0 | 0 | 0 | LRU |
s0p1 | 1500 | 0 | 16459725358 | 0 | 164 | 0 | 1449884045 | 0 | 0 | 0 | BMRU |
s0p1.4 | 1500 | 0 | 569145883 | 0 | 0 | 0 | 13351065 | 0 | 0 | 0 | BMRU |
s0p1.13 | 1500 | 0 | 15832888325 | 0 | 0 | 0 | 1393461365 | 0 | 0 | 0 | BMRU |
s0p1.31 | 1500 | 0 | 5064618 | 0 | 0 | 0 | 3418652 | 0 | 0 | 0 | BMRU |
s0p1.32 | 1500 | 0 | 6229135 | 0 | 0 | 0 | 4716002 | 0 | 0 | 0 | BMRU |
s0p1.33 | 1500 | 0 | 867958 | 0 | 0 | 0 | 27576 | 0 | 0 | 0 | BMRU |
s0p1.34 | 1500 | 0 | 1205573 | 0 | 0 | 0 | 85171 | 0 | 0 | 0 | BMRU |
s0p1.39 | 1500 | 0 | 1955359 | 0 | 0 | 0 | 939969 | 0 | 0 | 0 | BMRU |
s0p1.92 | 1500 | 0 | 38392944 | 0 | 0 | 0 | 33858440 | 0 | 0 | 0 | BMRU |
s0p2 | 1500 | 0 | 564176135 | 0 | 0 | 0 | 63106917 | 0 | 0 | 0 | BMRU |
s0p2.2 | 1500 | 0 | 30781935 | 0 | 0 | 0 | 13345212 | 0 | 0 | 0 | BMRU |
s0p2.100 | 1500 | 0 | 4722643 | 0 | 0 | 0 | 1079796 | 0 | 0 | 0 | BMRU |
s0p2.101 | 1500 | 0 | 1035852 | 0 | 0 | 0 | 1027773 | 0 | 0 | 0 | BMRU |
s0p2.102 | 1500 | 0 | 4161432 | 0 | 0 | 0 | 2436836 | 0 | 0 | 0 | BMRU |
s0p2.103 | 1500 | 0 | 0 | 0 | 0 | 0 | 27573 | 0 | 0 | 0 | BMRU |
s0p2.105 | 1500 | 0 | 1271313 | 0 | 0 | 0 | 820712 | 0 | 0 | 0 | BMRU |
s0p2.106 | 1500 | 0 | 13370017 | 0 | 0 | 0 | 12000147 | 0 | 0 | 0 | BMRU |
s0p2.107 | 1500 | 0 | 4758716 | 0 | 0 | 0 | 4223971 | 0 | 0 | 0 | BMRU |
s0p2.108 | 1500 | 0 | 3208956 | 0 | 0 | 0 | 2833171 | 0 | 0 | 0 | BMRU |
s0p2.110 | 1500 | 0 | 1180982 | 0 | 0 | 0 | 829131 | 0 | 0 | 0 | BMRU |
s0p2.111 | 1500 | 0 | 25420965 | 0 | 0 | 0 | 24482520 | 0 | 0 | 0 | BMRU |
s0p3 | 1500 | 0 | 203957021 | 0 | 1 | 0 | 2022423755 | 0 | 0 | 0 | BMRU |
s0p4 | 1500 | 0 | 79603650 | 0 | 0 | 0 | 672326345 | 0 | 0 | 0 | BMRU |
s4p1 | 1500 | 0 | 71335693881 | 0 | 7463 | 0 | 38876131270 | 0 | 0 | 0 | BMsRU |
s4p2 | 1500 | 0 | 59014194533 | 0 | 1833 | 0 | 41276800116 | 0 | 0 | 0 | BMsRU |
s5p2 | 1500 | 0 | 234755003 | 0 | 0 | 0 | 180788381 | 0 | 0 | 0 | BMRU |
s5p2.6 | 1500 | 0 | 13934961 | 0 | 0 | 0 | 13715248 | 0 | 0 | 0 | BMRU |
s5p2.7 | 1500 | 0 | 6035826 | 0 | 0 | 0 | 4093516 | 0 | 0 | 0 | BMRU |
s5p2.11 | 1500 | 0 | 153695023 | 0 | 0 | 0 | 121650121 | 0 | 0 | 0 | BMRU |
s5p2.12 | 1500 | 0 | 20878604 | 0 | 0 | 0 | 27577 | 0 | 0 | 0 | BMRU |
s5p2.13 | 1500 | 0 | 3792193 | 0 | 0 | 0 | 5771594 | 0 | 0 | 0 | BMRU |
s5p2.14 | 1500 | 0 | 5364843 | 0 | 0 | 0 | 9187536 | 0 | 0 | 0 | BMRU |
s5p2.51 | 1500 | 0 | 31053557 | 0 | 0 | 0 | 26348111 | 0 | 0 | 0 | BMRU |
s7p1 | 1500 | 0 | 14330536424 | 0 | 0 | 0 | 19082335534 | 0 | 0 | 0 | BMsRU |
s7p2 | 1500 | 0 | 13398915077 | 0 | 25 | 0 | 18710550994 | 0 | 0 | 0 | BMsRU |
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
It's usually better to start with the defaults and adjust from there versus going in with a specific configuration, particularly if you're changing hardware.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Hi Kevin Vargo,
The default 2/14 split for a 16-core firewall will probably be fine as Dameon Welch Abernathy suggests, however given the command outputs & other info you have provided, I'd say doing a 4/12 split is a quite reasonable starting point as long as no other manual affinities are set in fwaffinity.conf or sim_aff.conf. The Dynamic Dispatcher will be on by default in R80.10 so no need to worry about that.
Once thing that may change the distribution of load on the cores for R80.10 is that processes may only execute on the Firewall Worker cores, and are no longer allowed to grab spare cycles on the SND/IRQ cores as they were in R77.30. I'm assuming this was done to keep the CPU fast caches as "hot" as possible on the SND/IRQ cores. So keep an eye on the Firewall Worker cores as they may have slightly higher utilization. There is also a possibility based on the blades you have enabled that much more traffic could be accelerated on R80.10, and also more rulebase lookups will be templated by SecureXL which will increase the load on the SND/IRQ cores, and is why I'd estimate 4 of them is a good starting point.
You are seeing a negligible number of RX-DRPs currently (not RX-OVRs as your network statistics columns are off by one to the right vertically). You shouldn't pile up RX-DRPs in the new config, but if you do you may need to enable Multi-Queue assuming that all SND/IRQ cores are not topping out at the same time. I don't recommend turning on Multi-Queue unless you need it due to the increased overhead involved to "stick" connections to the same queue every time.
--
Second Edition of my "Max Power" Firewall Book
Now Available at http://www.maxpowerfirewalls.com
CET (Europe) Timezone Course Scheduled for July 1-2
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Thank you, Tim. I appreciate the post and your recommendations. I just made an edit to my column heading so output lines up, I missed that. Thanks again.
