Rules

blackbox.rules

25.588s ago

21.13ms

Rule State Error Last Evaluation Evaluation Time
alert: Ssl Certificate Has Expired Soon expr: probe_ssl_earliest_cert_expiry - time() < 86400 * 25 for: 1d labels: severity: critical annotations: summary: SSL сертификат на {{ $labels.instance }} закончится через 25 дней ok 25.589s ago 2.041ms
alert: Ssl Certificate Expired expr: probe_ssl_earliest_cert_expiry - time() <= 0 for: 1d labels: severity: critical annotations: summary: SSL certificate has expired already on {{ $labels.instance }} ok 25.587s ago 1.741ms
alert: ATS down expr: probe_success{job=~"ATS"} == 0 for: 5m labels: severity: critical annotations: summary: '{{ $labels.instance }} АТС не отвечает больше 5 минут' ok 25.585s ago 291.9us
alert: Radioactive Down expr: probe_success{job=~"radioactive"} == 0 for: 5m labels: severity: radioactive annotations: summary: Радиоактивная комната {{ $labels.instance }} не отвечает больше 5 минут ok 25.585s ago 258.3us
alert: Slow Http expr: avg_over_time(probe_http_duration_seconds[5m]) > 5 for: 5m labels: severity: critical annotations: summary: Очень медленная обработка HTTP-запросов на {{ $labels.instance }} ok 25.585s ago 13.74ms
alert: Blackbox Slow Probe expr: avg_over_time(probe_duration_seconds[5m]) > 10 for: 5m labels: severity: critical annotations: summary: 'Blackbox очень долго обрабатывает сайт {{ $labels.instance }}. Время обработки: {{ $value }}' ok 25.572s ago 3.016ms

custom.rules

2.367s ago

22.77ms

Rule State Error Last Evaluation Evaluation Time
alert: Switch Down expr: up{job=~"upTime|dmc1002"} == 0 for: 1m labels: severity: critical annotations: summary: '{{ $labels.instance }} not responding for more than 1 minutes' ok 2.368s ago 7.587ms
alert: Squid Server Down expr: squid_server_up == 0 for: 10m labels: severity: critical annotations: summary: Squid on {{ $labels.instance }} not responding for more than 10 minutes ok 2.36s ago 249.5us
alert: Bind Server Down expr: bind_up == 0 for: 5m labels: severity: critical annotations: summary: '{{ $labels.job }} on {{ $labels.instance }} not responding for more than 5 minutes' ok 2.36s ago 265.8us
alert: Instance Down expr: up{job=~"squid|wpad|etcd|dhcpd6|bind"} == 0 for: 5m labels: severity: critical annotations: summary: '{{ $labels.job }} {{ $labels.instance }} not responding for more than 5 minutes.' ok 2.36s ago 514.9us
alert: Error in dhcpserv config expr: rate(dhcpserv_build_errors[1m]) > 1 for: 2m labels: severity: critical annotations: summary: '{{ $labels.instance }} have too many error in dhcpserv generate config: {{ $value }}' ok 2.36s ago 323.8us
alert: Cisco Netflow Cache High expr: netflow_active_entries:ratio > 0.8 for: 5m labels: severity: ctritical ok 2.36s ago 194.1us
alert: Postfix Queue High expr: postfix_showq_message_size_bytes_count > 200 for: 10m labels: severity: critical annotations: summary: Postfix query overflow. Query exceeded {{ $value }} mails ok 2.36s ago 315.9us
alert: Borg Backup Missing expr: time() - borgbackup_last_modified > 86400 * 8 for: 30m labels: severity: critical annotations: summary: Backup on {{ $labels.instance }} not created last 3 days ok 2.36s ago 575.8us
alert: Omada device down expr: omada_device_uptime_seconds == 0 for: 5m labels: severity: critical annotations: summary: Device {{ $labels.device_type }} {{ $labels.device }} from site {{ $labels.site }} is down ok 2.359s ago 2.735ms
alert: Unifi device down expr: unifipoller_device_uptime_seconds == 0 for: 5m labels: severity: critical annotations: summary: Device {{ $labels.type }} {{ $labels.name }} from site {{ $labels.site_name }} is down ok 2.357s ago 4.247ms
alert: Nginx is down expr: nginx_up == 0 for: 5m labels: severity: critical annotations: summary: Nginx is down on the host {{ $labels.instance }} for more than 5 minutes ok 2.353s ago 486.7us
alert: Nginx not all connections are handled expr: rate(nginx_connections_handled{job="nginx"}[5m]) / rate(nginx_connections_accepted{job="nginx"}[5m]) < 1 for: 3m labels: severity: critical annotations: summary: Nginx does not handle all accept connections on the host {{ $labels.instance }} for more than 3 minutes ok 2.353s ago 945.4us
alert: Nginx high active connections expr: nginx_connections_active{job="nginx"} > 300 for: 5m labels: severity: critical annotations: summary: Большое коллчиество активных подключений в Nginx на {{ $labels.instance }} в течении 5 минут ok 2.352s ago 325.8us
alert: Certificate expired expr: nginx_cert_exporter_file_expired - time() < 86400 * 25 for: 12h labels: severity: critical annotations: summary: Локальный сертификат {{ $labels.name }} на {{ $labels.instance }} просрочиться в течении 25 дней ok 2.352s ago 2.544ms
alert: DDOS expr: (rate(nginx_http_requests_total[5m]) / rate(nginx_http_requests_total[5m] offset 5m) > 5) and rate(nginx_http_requests_total[5m]) > 100 for: 5m labels: severity: critical annotations: summary: '{{ $labels.instance }} находится под DDOS атакой' ok 2.349s ago 980.4us
alert: Fail2Ban DDOS expr: avg by(instance, jail) (f2b_jail_banned_current) > 20 for: 5m labels: severity: critical annotations: summary: 'Слишком часто срабатывает правило {{ $labels.jail }}: {{ $value }} на сервере {{ $labels.instance }}' ok 2.349s ago 411us

mysql.rules

14.329s ago

3.217ms

Rule State Error Last Evaluation Evaluation Time
alert: MysqlDown expr: mysql_up == 0 for: 1m labels: severity: critical annotations: description: MySQL instance is down on {{ $labels.instance }} summary: MySQL down (instance {{ $labels.instance }}) ok 14.329s ago 664.7us
alert: MysqlTooManyConnections expr: avg by(instance) (rate(mysql_global_status_threads_connected[5m])) / avg by(instance) (mysql_global_variables_max_connections) * 100 > 60 for: 5m labels: severity: critical annotations: description: More than 60% of MySQL connections are in use on {{ $labels.instance }} summary: MySQL too many connections (> 60%) (instance {{ $labels.instance }}) ok 14.329s ago 1.109ms
alert: MysqlHighThreadsRunning expr: avg by(instance) (rate(mysql_global_status_threads_running[1m])) / avg by(instance) (mysql_global_variables_max_connections) * 100 > 60 for: 5m labels: severity: critical annotations: description: More than 60% of MySQL connections are in running state on {{ $labels.instance }} summary: MySQL high threads running (instance {{ $labels.instance }}) ok 14.328s ago 680.2us
alert: MysqlSlowQueries expr: increase(mysql_global_status_slow_queries[2m]) > 10 for: 5m labels: severity: critical annotations: description: MySQL server mysql has some new slow query. {{ $value }} summary: MySQL slow queries (instance {{ $labels.instance }}) ok 14.327s ago 329us
alert: MysqlInnodbLogWaits expr: rate(mysql_global_status_innodb_log_waits[15m]) > 10 for: 5m labels: severity: critical annotations: description: MySQL innodb log writes stalling {{ $value }} summary: MySQL InnoDB log waits (instance {{ $labels.instance }}) ok 14.327s ago 407.6us

node.rules

26.646s ago

208.6ms

Rule State Error Last Evaluation Evaluation Time
alert: Node Down expr: up{job=~"node"} == 0 for: 1m labels: severity: critical annotations: summary: Node {{ $labels.instance }} not responding for more than 1 minutes. ok 26.646s ago 1.29ms
alert: Service Down expr: node_systemd_unit_state{name!="dnf-makecache.service",state=~"failed"} == 1 for: 5m labels: severity: critical annotations: summary: Service {{ $labels.name }} is not responding for 5m on {{ $labels.instance }} with status {{ $labels.state }} ok 26.645s ago 85.16ms
alert: Node Low Disk Space expr: (node_filesystem_avail_bytes * 100) / node_filesystem_size_bytes{mountpoint!~".mnt.*"} < 10 and on(instance, device, mountpoint) node_filesystem_readonly == 0 for: 5m labels: severity: critical annotations: summary: '{{ $labels.instance }} low disk space. {{ $value }} percent of the {{ $labels.mountpoint }} is free left' ok 26.56s ago 8.108ms
alert: Host Oom Kill Detected expr: increase(node_vmstat_oom_kill[1m]) > 0 labels: severity: warning annotations: summary: OOM kill detected on {{ $labels.instance }} ok 26.552s ago 392us
alert: Node Out Of Memory expr: node_memory_MemAvailable_bytes / node_memory_MemTotal_bytes * 100 < 20 for: 5m labels: severity: critical annotations: summary: 'Node {{ $labels.instance }} memory is filling up (< 20% left). Free now: {{ $value }}%' ok 26.552s ago 1.586ms
alert: Node Memory Fill Up Soon expr: predict_linear(node_memory_MemAvailable_bytes[2h], 1 * 3600) <= 0 for: 2h labels: severity: critical annotations: summary: '{{ $labels.instance }} memory will fiil up soon in 1h' ok 26.551s ago 2.747ms
alert: Node SWAP Out Of Memory expr: node_memory_SwapFree_bytes / node_memory_SwapTotal_bytes * 100 < 20 for: 10m labels: severity: critical annotations: summary: 'Node {{ $labels.instance }} swap memory is filing up (< 20% left). Free now: {{ $value }}%' ok 26.548s ago 2.179ms
alert: Node Memory Under Pressure expr: (node_memory_MemAvailable_bytes / node_memory_MemTotal_bytes * 100) < 20 and rate(node_vmstat_pgmajfault[5m]) > 100 for: 5m labels: severity: critical annotations: summary: Node {{ $labels.instance }} is under heavy memory pressure. High rate of major page faults. VALUE = {{ $value }} ok 26.546s ago 2.338ms
alert: Node Unusual Network ThroughputIn expr: sum by(instance) (rate(node_network_receive_bytes_total[30m])) / 1024 / 1024 > 100 for: 15m labels: severity: critical annotations: summary: Host unusual network throughput in {{ $labels.instance }} - {{ $value }} ok 26.544s ago 5.011ms
alert: Node Unusual Network ThroughputOut expr: sum by(instance) (rate(node_network_transmit_bytes_total[30m])) / 1024 / 1024 > 100 for: 15m labels: severity: critical annotations: summary: Host unusual network throughput out {{ $labels.instance }} - {{ $value }} ok 26.539s ago 4.708ms
alert: Node Unusual Disk Read Rate expr: sum by(instance) (rate(node_disk_read_bytes_total[30m])) / 1024 / 1024 > 200 for: 30m labels: severity: critical annotations: summary: Host unusual disk read rate last 30 minutes {{ $labels.instance }} - {{ $value }} mb/s ok 26.535s ago 5.948ms
alert: Node Unusual Disk Write Rate expr: sum by(instance) (rate(node_disk_written_bytes_total[30m])) / 1024 / 1024 > 200 for: 30m labels: severity: critical annotations: summary: Host unusual disk write rate last 30 minutes {{ $labels.instance }} - {{ $value }} mb/s ok 26.529s ago 6.324ms
alert: Node Unusual Disk ReadLatency expr: rate(node_disk_read_time_seconds_total[30m]) / rate(node_disk_reads_completed_total[30m]) > 0.3 and rate(node_disk_reads_completed_total[30m]) > 0 for: 30m labels: severity: critical annotations: summary: Host unusual disk read latency {{ $labels.instance }} - {{ $value }} s ok 26.523s ago 18.56ms
alert: Node Unusual Disk WriteLatency expr: rate(node_disk_write_time_seconds_total[30m]) / rate(node_disk_writes_completed_total[30m]) > 0.15 and rate(node_disk_writes_completed_total[30m]) > 0 for: 30m labels: severity: critical annotations: summary: Host unusual disk write latency {{ $labels.instance }} - {{ $value }} s ok 26.505s ago 18.46ms
alert: Big IOWait expr: avg by(instance) (irate(node_cpu_seconds_total{mode="iowait"}[1m])) * 100 > 15 for: 2m labels: severity: critical annotations: summary: '{{ $labels.instance }} have big iowait query: {{ $value }}%' ok 26.487s ago 9.794ms
alert: Node High Cpu Load expr: 100 - (avg by(instance) (rate(node_cpu_seconds_total{job="node",mode="idle"}[15m])) * 100) > 80 for: 15m labels: severity: critical annotations: summary: Host high CPU load {{ $labels.instance }} - {{ $value }}% ok 26.477s ago 10.46ms
alert: Node Physical Component TooHot expr: node_hwmon_temp_celsius{chip!~"nvme.+"} > 78 for: 10m labels: severity: critical annotations: summary: Host physical component too hot (instance {{ $labels.instance }}) {{ $labels.chip }} - {{ $value }}C ok 26.467s ago 4.561ms
alert: Node NVMe Drive TooHot expr: node_hwmon_temp_celsius{chip=~"nvme.+"} > 100 for: 10m labels: severity: critical annotations: summary: Host physical component too hot (instance {{ $labels.instance }}) {{ $labels.chip }} - {{ $value }}C ok 26.462s ago 629.7us
alert: Node Overtemperature Alarm expr: node_hwmon_temp_crit_alarm_celsius == 1 labels: severity: critical annotations: summary: Host node overtemperature alarm (instance {{ $labels.instance }}) - {{ $value }} ok 26.462s ago 5.663ms
alert: Node Raid Array Got Inactive expr: node_md_state{state="inactive"} > 0 for: 10m labels: severity: critical annotations: summary: RAID array {{ $labels.device }} is in degraded state due to one or more disks failures. Number of spare drives is insufficient to fix issue automatically ok 26.456s ago 593.9us
alert: Node Raid Disk Failure expr: node_md_disks{state="failed"} > 0 for: 10m labels: severity: critical annotations: summary: At least one device in RAID array on {{ $labels.instance }} failed. Array {{ $labels.md_device }} needs attention and possibly a disk swap ok 26.456s ago 529.1us
alert: Node Edac Correctable Errors Detected expr: increase(node_edac_correctable_errors_total[1m]) > 0 for: 10m labels: severity: critical annotations: summary: Host EDAC Correctable Errors detected (instance {{ $labels.instance }}) {{ $value }} ok 26.456s ago 1.23ms
alert: Node Edac Uncorrectable Errors Detected expr: node_edac_uncorrectable_errors_total > 0 labels: severity: critical annotations: summary: Host EDAC Uncorrectable Errors detected (instance {{ $labels.instance }}) {{ $value }} ok 26.455s ago 1.054ms
alert: Relocated Sectors expr: smart_attribute_raw{attribute_id=~"197|198"} > 0 for: 3h labels: severity: critical annotations: summary: '{{ $labels.instance }} имеет {{ $value }} переназначенных секторов. {{ $labels.device_model_family }} {{ $labels.device_model_name }} под именем {{ $labels.device_name }} с серийным номером {{ $labels.device_serial_number }}' ok 26.454s ago 4.652ms
alert: Node Clock Not Synchronising expr: min_over_time(node_timex_sync_status[1m]) == 0 and node_timex_maxerror_seconds >= 16 for: 2m labels: severity: critical annotations: summary: Clock not synchronising. Ensure NTP is configured on this host {{ $labels.instance }} ok 26.449s ago 1.658ms
alert: HostClockSkew expr: (node_timex_offset_seconds > 0.05 and deriv(node_timex_offset_seconds[5m]) >= 0) or (node_timex_offset_seconds < -0.05 and deriv(node_timex_offset_seconds[5m]) <= 0) for: 2m labels: severity: warning annotations: summary: Clock skew detected. Clock is out of sync. Ensure NTP is configured correctly on this host {{ $labels.instance }} ok 26.448s ago 2.421ms
alert: Node Disk Is Missing expr: node_btrfs_device_size_bytes == 0 for: 1m labels: severity: critical annotations: summary: '{{ $labels.instance }} has a disk {{ $labels.device }} that is not in the file system' ok 26.446s ago 2.43ms

self.rules

820ms ago

5.179ms

Rule State Error Last Evaluation Evaluation Time
alert: Prometheus Configuration Failure expr: prometheus_config_last_reload_successful != 1 labels: severity: critical annotations: summary: Prometheus configuration reload failure (instance {{ $labels.instance }}) ok 820ms ago 502.7us
alert: AlertManager Configuration Failure expr: alertmanager_config_last_reload_successful != 1 labels: severity: critical annotations: summary: AlertManager configuration reload failure (instance {{ $labels.iinstance }}) ok 820ms ago 275us
alert: Prometheus NotConnected To Alertmanager expr: prometheus_notifications_alertmanagers_discovered < 1 labels: severity: critical annotations: summary: Prometheus not connected to alertmanager (instance {{ $labels.instance }}) ok 820ms ago 209.9us
alert: Prometheus Rule Evaluation Failures expr: increase(prometheus_rule_evaluation_failures_total[3m]) > 0 labels: severity: critical annotations: summary: Prometheus rule evaluation failures (instance {{ $labels.instance }}) ok 820ms ago 547.4us
alert: Prometheus Template Text Expansion Failures expr: increase(prometheus_template_text_expansion_failures_total[3m]) > 0 labels: severity: critical annotations: summary: Prometheus template text expansion failures (instance {{ $labels.instance }}) ok 819ms ago 284.5us
alert: Prometheus Rule Evaluation Slow expr: prometheus_rule_group_last_duration_seconds > prometheus_rule_group_interval_seconds for: 5m labels: severity: critical annotations: summary: Prometheus rule evaluation slow (instance {{ $labels.instance }}) ok 819ms ago 775.2us
alert: Prometheus Target Empty expr: prometheus_sd_discovered_targets == 0 labels: severity: critical annotations: description: Prometheus has no target in service discovery summary: Prometheus target empty (instance {{ $labels.instance }}) ok 819ms ago 1.854ms
alert: Prometheus Large Scrape expr: increase(prometheus_target_scrapes_exceeded_sample_limit_total[10m]) > 10 for: 5m labels: severity: critical annotations: description: Prometheus has many scrapes that exceed the sample limit summary: Prometheus large scrape (instance {{ $labels.instance }}) ok 817ms ago 421.9us
alert: Prometheus Target Scrape Duplicate expr: increase(prometheus_target_scrapes_sample_duplicate_timestamp_total[5m]) > 0 labels: severity: critical annotations: description: Prometheus has many samples rejected due to duplicate timestamps but different values summary: Prometheus target scrape duplicate (instance {{ $labels.instance }}) ok 817ms ago 272.3us

ups.rules

20.746s ago

8.555ms

Rule State Error Last Evaluation Evaluation Time
alert: UPS not response expr: up{job=~"apcsnmp|upssnmp"} == 0 for: 5m labels: severity: critical annotations: summary: UPS {{ $labels.instance }} not responding for more than 5 minutes. ok 20.746s ago 2.123ms
alert: Server Room In FIRE! expr: iemStatusProbeCurrentTemp{instance="Symmetra.ups"} > 24 for: 10m labels: severity: critical annotations: summary: 'Server room too hot: {{ $value }}C' ok 20.744s ago 410.6us
alert: There is no ELECTRICITY in the Server Room! expr: upsBasicInputPhase{instance="Symmetra.ups"} < 3 for: 5m labels: severity: critical annotations: summary: 'Working phases in Symmetra at the moment: {{ $value }}' ok 20.744s ago 331.7us
alert: APC UPS too hot! expr: iemStatusProbeCurrentTemp > 30 for: 5m labels: severity: critical annotations: summary: Too hot in {{ $labels.instance }}. Temperatur is {{ $value }}C ok 20.744s ago 255.4us
alert: UPS too hot! expr: upsBatteryTemperature > 45 for: 5m labels: severity: critical annotations: summary: Too hot in {{ $labels.instance }}. Temperatur is {{ $value }}C ok 20.744s ago 1.437ms
alert: APC UPS low Input voltage expr: upsAdvInputLineVoltage < 200 for: 5m labels: severity: critical annotations: summary: 'APC UPS {{ $labels.instance }} has low voltage: {{ $value }}V' ok 20.743s ago 305.2us
alert: UPS low Input voltage expr: upsInputVoltage < 200 for: 5m labels: severity: critical annotations: summary: 'UPS {{ $labels.instance }} has low input voltage: {{ $value }}V' ok 20.743s ago 846.4us
alert: UPS low Output voltage expr: upsOutputVoltage{upsOutputLineIndex="1"} < 200 for: 5m labels: severity: critical annotations: summary: 'UPS {{ $labels.instance }} has low output voltage: {{ $value }}V' ok 20.742s ago 727.9us
alert: UPS big load expr: upsOutputPercentLoad{upsOutputLineIndex="1"} > 75 for: 5m labels: severity: critical annotations: summary: 'UPS {{ $labels.instance }} has big load: {{ $value }}%' ok 20.741s ago 735us
alert: Low APC UPS battery capacity expr: upsAdvBatteryCapacity < 75 for: 5m labels: severity: critical annotations: summary: '{{ $labels.instance }} has low battery capacity: {{ $value }}%' ok 20.741s ago 207.3us
alert: Low UPS battery capacity expr: upsEstimatedChargeRemaining < 75 for: 5m labels: severity: critical annotations: summary: '{{ $labels.instance }} has low battery capacity: {{ $value }}%' ok 20.741s ago 1.13ms

windows.rules

25.944s ago

50.39ms

Rule State Error Last Evaluation Evaluation Time
alert: Windows Server Collector Error expr: windows_exporter_collector_success{job="node"} == 0 labels: severity: critical annotations: summary: Windows Server collector Error on {{ $labels.instance }}. Collector {{ $labels.collector }} was not successful ok 25.945s ago 1.966ms
alert: Windows Server Service Status expr: windows_service_status{job="node",status="ok"} != 1 for: 1m labels: severity: critical annotations: summary: Windows Server service Status on {{ $labels.instance }}. {{ $labels.name }} state is not OK ok 25.943s ago 37.63ms
alert: Windows Server Cpu Usage expr: 100 - (avg by(instance) (rate(windows_cpu_time_total{job="node",mode="idle"}[10m])) * 100) > 80 for: 10m labels: severity: critical annotations: summary: Windows Server CPU Usage on {{ $labels.instance }} is more than 80% ok 25.905s ago 5.722ms
alert: Windows Server Memory Usage expr: 100 - ((windows_os_physical_memory_free_bytes{job="node"} / windows_cs_physical_memory_bytes{job="node"}) * 100) > 80 for: 5m labels: severity: critical annotations: summary: Windows Server memory Usage on {{ $labels.instance }} is more than 80% ok 25.9s ago 1.734ms
alert: Windows Server Disk Space Usage expr: 100 - 100 * ((windows_logical_disk_free_bytes{job="node"} / 1024 / 1024) / ((windows_logical_disk_size_bytes{job="node"} / 1024 / 1024) > 20000)) > 90 for: 10m labels: severity: critical annotations: summary: Windows Server disk Space Usage on {{ $labels.instance }} is more than 90% on {{ $labels.volume }} ok 25.899s ago 3.307ms