-
Notifications
You must be signed in to change notification settings - Fork 87
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
cpu-load on Checkpoint R81.20 #326
Comments
'check_nwc_health' '--mode' 'cpu-load' '--protocol' '2c' -vvvvvvvvvvv Tue Feb 6 15:19:37 2024: i know package Monitoring::GLPlugin::SNMP::MibsAndOids::MIB2MIB Tue Feb 6 15:19:37 2024: I am a Linux 3.10.0-1160.15.2cpx86_64 #1 SMP Sun Nov 12 09:27:02 IST 2023 x86_64 Tue Feb 6 15:19:37 2024: i know package Monitoring::GLPlugin::SNMP::MibsAndOids::SYNOPTICSROOTMIB Tue Feb 6 15:19:37 2024: AUTOLOAD Monitoring::GLPlugin::Commandline::override_opt Tue Feb 6 15:19:37 2024: AUTOLOAD CheckNwcHealth::CheckPoint::check_messages Tue Feb 6 15:19:37 2024: i know package Monitoring::GLPlugin::SNMP::MibsAndOids::CHECKPOINTMIB Tue Feb 6 15:19:37 2024: i know package Monitoring::GLPlugin::SNMP::MibsAndOids::CHECKPOINTMIB Tue Feb 6 15:19:37 2024: get_table returned 24 oids in 0s Tue Feb 6 15:19:37 2024: get_matching_oids returns 24 from 33 oids Tue Feb 6 15:19:37 2024: AUTOLOAD CheckNwcHealth::CheckPoint::Firewall1::Component::CpuSubsystem::set_thresholds Tue Feb 6 15:19:37 2024: AUTOLOAD CheckNwcHealth::CheckPoint::Firewall1::Component::CpuSubsystem::check_thresholds Tue Feb 6 15:19:37 2024: AUTOLOAD CheckNwcHealth::CheckPoint::Firewall1::Component::CpuSubsystem::add_perfdata Tue Feb 6 15:19:37 2024: AUTOLOAD CheckNwcHealth::CheckPoint::Firewall1::Component::CpuSubsystem::MultiProc::set_thresholds Tue Feb 6 15:19:37 2024: AUTOLOAD CheckNwcHealth::CheckPoint::Firewall1::Component::CpuSubsystem::MultiProc::check_thresholds Tue Feb 6 15:19:37 2024: AUTOLOAD CheckNwcHealth::CheckPoint::Firewall1::Component::CpuSubsystem::MultiProc::add_perfdata Tue Feb 6 15:19:37 2024: AUTOLOAD CheckNwcHealth::CheckPoint::Firewall1::Component::CpuSubsystem::MultiProc::set_thresholds Tue Feb 6 15:19:37 2024: AUTOLOAD CheckNwcHealth::CheckPoint::Firewall1::Component::CpuSubsystem::MultiProc::check_thresholds Tue Feb 6 15:19:37 2024: AUTOLOAD CheckNwcHealth::CheckPoint::Firewall1::Component::CpuSubsystem::MultiProc::add_perfdata Tue Feb 6 15:19:37 2024: AUTOLOAD CheckNwcHealth::CheckPoint::Firewall1::Component::CpuSubsystem::MultiProc::set_thresholds Tue Feb 6 15:19:37 2024: AUTOLOAD CheckNwcHealth::CheckPoint::Firewall1::Component::CpuSubsystem::MultiProc::check_thresholds Tue Feb 6 15:19:37 2024: AUTOLOAD CheckNwcHealth::CheckPoint::Firewall1::Component::CpuSubsystem::MultiProc::add_perfdata Tue Feb 6 15:19:37 2024: AUTOLOAD CheckNwcHealth::CheckPoint::Firewall1::Component::CpuSubsystem::MultiProc::set_thresholds Tue Feb 6 15:19:37 2024: AUTOLOAD CheckNwcHealth::CheckPoint::Firewall1::Component::CpuSubsystem::MultiProc::check_thresholds Tue Feb 6 15:19:37 2024: AUTOLOAD CheckNwcHealth::CheckPoint::Firewall1::Component::CpuSubsystem::MultiProc::add_perfdata [CPUSUBSYSTEM] [MULTIPROC_2.0] [MULTIPROC_3.0] [MULTIPROC_4.0] Tue Feb 6 15:19:37 2024: AUTOLOAD CheckNwcHealth::CheckPoint::Firewall1::check_messages Tue Feb 6 15:19:37 2024: AUTOLOAD CheckNwcHealth::CheckPoint::Firewall1::check_messages Tue Feb 6 15:19:37 2024: AUTOLOAD CheckNwcHealth::CheckPoint::Firewall1::nagios_exit CRITICAL - cpu core 1 usage is 100.00%, cpu usage is 1.00%, cpu core 2 usage is 0.00%, cpu core 3 usage is 0.00%, cpu core 4 usage is 0.00% |
You are not the only one. Can you put this code in a file mininwc.pl please?
then run it with |
This is the output: `$VAR1 = { ` |
Ah, you have 4 cpus. So you can change |
I'm glad to contribute. OK - cpu usage is 10.00%, cpu core 1 usage is 0.00%, cpu core 2 usage is 4.00%, cpu core 3 usage is 15.00%, cpu core 4 usage is 11.00% | 'cpu_usage'=10%;80;90;0;100 'cpu_core_1_usage'=0%;80;90;0;100 'cpu_core_2_usage'=4%;80;90;0;100 'cpu_core_3_usage'=15%;80;90;0;100 'cpu_core_4_usage'=11%;80;90;0;100 OK - cpu usage is 10.00%, cpu core 1 usage is 0.00%, cpu core 2 usage is 4.00%, cpu core 3 usage is 15.00%, cpu core 4 usage is 11.00% | 'cpu_usage'=10%;80;90;0;100 'cpu_core_1_usage'=0%;80;90;0;100 'cpu_core_2_usage'=4%;80;90;0;100 'cpu_core_3_usage'=15%;80;90;0;100 'cpu_core_4_usage'=11%;80;90;0;100 CRITICAL - cpu core 1 usage is 100.00%, cpu core 3 usage is 100.00%, cpu core 4 usage is 100.00%, cpu usage is 3.00%, cpu core 2 usage is 0.00% | 'cpu_usage'=3%;80;90;0;100 'cpu_core_1_usage'=100%;80;90;0;100 'cpu_core_2_usage'=0%;80;90;0;100 'cpu_core_3_usage'=100%;80;90;0;100 'cpu_core_4_usage'=100%;80;90;0;100 CRITICAL - cpu core 1 usage is 100.00%, cpu core 3 usage is 100.00%, cpu core 4 usage is 100.00%, cpu usage is 3.00%, cpu core 2 usage is 0.00% | 'cpu_usage'=3%;80;90;0;100 'cpu_core_1_usage'=100%;80;90;0;100 'cpu_core_2_usage'=0%;80;90;0;100 'cpu_core_3_usage'=100%;80;90;0;100 'cpu_core_4_usage'=100%;80;90;0;100 CRITICAL - cpu core 1 usage is 100.00%, cpu core 3 usage is 100.00%, cpu core 4 usage is 100.00%, cpu usage is 3.00%, cpu core 2 usage is 0.00% | 'cpu_usage'=3%;80;90;0;100 'cpu_core_1_usage'=100%;80;90;0;100 'cpu_core_2_usage'=0%;80;90;0;100 'cpu_core_3_usage'=100%;80;90;0;100 'cpu_core_4_usage'=100%;80;90;0;100 CRITICAL - cpu core 1 usage is 100.00%, cpu core 3 usage is 100.00%, cpu core 4 usage is 100.00%, cpu usage is 3.00%, cpu core 2 usage is 0.00% | 'cpu_usage'=3%;80;90;0;100 'cpu_core_1_usage'=100%;80;90;0;100 'cpu_core_2_usage'=0%;80;90;0;100 'cpu_core_3_usage'=100%;80;90;0;100 'cpu_core_4_usage'=100%;80;90;0;100 |
When you let it run 100 times, can you estimate if there are more good results than bad results? And for how many cycles the bad result is returned before there are usable metrics? |
Except for 10:29 it all looks like crap. On the command line there was at least a 50:50 ratio. |
There are also cpu load metrics from the HOST-RESOURCES-MIB. Let's compare them. Can you replace the old mininwc.pl with the above code? |
That was because i've filtered the output for a loop of 100 iterations. |
looks the same found 4 cpus |
Is there a way to check the validity of the hrProcessorLoad with a command line tool on a console? |
It's hard to say as the CPU load is very low (<5%) on the device and hrProcessorLoad shows the average over the last min. |
Even if the hrProcessorLoad are not identical to the multiProcUsage (in the cases when these are not [0, 20, 50, 100]), it would be sufficient when they just are approximately high and low to what the cli tool shows. |
I did some load test on all 4 cores. hrProcessorLoad works as designed with the delay so it may provide a suitable workaround. found 4 cpus |
Is R81.20 the latest version? I googled around for "R81.20 cpu" and found a lot of people complaining about the general quality of checkpoint and also some cpu-related posts, but non of them affected snmp. If there is an update or hotfix (or whatever it's called) i would be very interested if the metrics are better then. |
Yes. 81.20 is the latest version. We've currently deploying the latest hotfix/jumbo fix and it's still present. But i'll try future versions with your first mininwc.pl script. |
Can you send me an email, so i can attach a plugin for testing to my reply? I can't attach it here. [email protected] |
with the latest update for Checkpoint R81.20 (Take 41) check_nwc_health shows not the correct values:
'/usr/lib64/nagios/plugins/contrib/check_nwc_health' '--mode' 'cpu-load' '--protocol' '2c' -v
CRITICAL - cpu core 1 usage is 100.00%, cpu core 2 usage is 100.00%, cpu core 3 usage is 100.00%, cpu usage is 1.00%, cpu core 4 usage is 0.00%
checking cpus
cpu usage is 1.00%
checking cpu cores
cpu core 1 usage is 100.00%
cpu core 2 usage is 100.00%
cpu core 3 usage is 100.00%
cpu core 4 usage is 0.00% | 'cpu_usage'=1%;80;90;0;100 'cpu_core_1_usage'=100%;80;90;0;100 'cpu_core_2_usage'=100%;80;90;0;100 'cpu_core_3_usage'=100%;80;90;0;100 'cpu_core_4_usage'=0%;80;90;0;100
snmpwalk -v 2c .1.3.6.1.4.1.2620.1.6.7.5.1.5
SNMPv2-SMI::enterprises.2620.1.6.7.5.1.5.1.0 = Gauge32: 0
SNMPv2-SMI::enterprises.2620.1.6.7.5.1.5.2.0 = Gauge32: 6
SNMPv2-SMI::enterprises.2620.1.6.7.5.1.5.3.0 = Gauge32: 10
SNMPv2-SMI::enterprises.2620.1.6.7.5.1.5.4.0 = Gauge32: 9
top shows the same values as snmpwalk and the OID are the same as on older versions.
The text was updated successfully, but these errors were encountered: