[00:00:32] alerting : [FIRING:1] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [00:02:37] PROBLEM - cp31 Stunnel HTTP for mw101 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:03:02] PROBLEM - mw101 MediaWiki Rendering on mw101 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:03:06] PROBLEM - test101 Current Load on test101 is CRITICAL: CRITICAL - load average: 2.83, 1.93, 1.42 [00:03:25] RECOVERY - gluster121 Current Load on gluster121 is OK: OK - load average: 2.04, 3.14, 3.13 [00:03:44] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 11.22, 8.22, 6.77 [00:03:48] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 4.13, 3.69, 3.58 [00:03:50] PROBLEM - mw101 Puppet on mw101 is CRITICAL: CRITICAL: Failed to apply catalog, zero resources tracked by Puppet. It might be a dependency cycle. [00:04:37] RECOVERY - cp31 Stunnel HTTP for mw101 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 0.968 second response time [00:05:00] RECOVERY - mw101 MediaWiki Rendering on mw101 is OK: HTTP OK: HTTP/1.1 200 OK - 22334 bytes in 2.901 second response time [00:05:48] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.27, 3.39, 3.47 [00:07:48] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 4.03, 3.30, 3.41 [00:09:16] PROBLEM - cp21 Stunnel HTTP for mw111 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:09:24] PROBLEM - cp30 Stunnel HTTP for mw111 on cp30 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:09:25] PROBLEM - mw112 MediaWiki Rendering on mw112 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:09:31] PROBLEM - cp21 Stunnel HTTP for mw112 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:09:46] RECOVERY - gluster101 Current Load on gluster101 is OK: OK - load average: 2.32, 2.79, 3.35 [00:09:50] PROBLEM - cp31 Stunnel HTTP for mw111 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:09:51] PROBLEM - cp31 Stunnel HTTP for mw112 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:09:52] PROBLEM - cp20 Stunnel HTTP for mw111 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:10:02] PROBLEM - mw111 MediaWiki Rendering on mw111 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:10:08] PROBLEM - cp20 Stunnel HTTP for mw112 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:10:14] PROBLEM - cp31 Stunnel HTTP for mw121 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:10:34] PROBLEM - cp30 Stunnel HTTP for mw112 on cp30 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:10:45] PROBLEM - mw121 MediaWiki Rendering on mw121 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:11:32] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 4.47, 7.17, 7.01 [00:11:36] PROBLEM - cp20 Stunnel HTTP for mw121 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:11:44] PROBLEM - cp30 Stunnel HTTP for mw121 on cp30 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:11:48] RECOVERY - gluster111 Current Load on gluster111 is OK: OK - load average: 1.30, 2.44, 3.05 [00:11:57] PROBLEM - cp21 Stunnel HTTP for mw121 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:11:58] PROBLEM - cp21 Stunnel HTTP for mw102 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:12:08] PROBLEM - cp30 Stunnel HTTP for mw101 on cp30 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:12:26] PROBLEM - cp21 Stunnel HTTP for mw101 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:12:33] PROBLEM - cp20 Stunnel HTTP for mw122 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:12:38] PROBLEM - cp30 Stunnel HTTP for mw102 on cp30 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:12:41] PROBLEM - mw102 MediaWiki Rendering on mw102 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:12:45] PROBLEM - cp20 Stunnel HTTP for mw101 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:12:53] PROBLEM - cp21 Stunnel HTTP for mw122 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:13:04] PROBLEM - cp31 Stunnel HTTP for mw102 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:13:06] PROBLEM - test101 Current Load on test101 is WARNING: WARNING - load average: 1.78, 1.99, 1.74 [00:13:10] PROBLEM - cp20 Stunnel HTTP for mw102 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:13:11] PROBLEM - cp31 Stunnel HTTP for mw101 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:13:20] PROBLEM - mw101 MediaWiki Rendering on mw101 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:13:32] PROBLEM - cp30 Stunnel HTTP for mw122 on cp30 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:13:36] PROBLEM - mw122 MediaWiki Rendering on mw122 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:13:39] PROBLEM - cp31 Stunnel HTTP for mw122 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:13:59] RECOVERY - cp21 Stunnel HTTP for mw102 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 5.615 second response time [00:14:26] RECOVERY - cp31 Stunnel HTTP for mw121 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 9.059 second response time [00:14:45] RECOVERY - mw102 MediaWiki Rendering on mw102 is OK: HTTP OK: HTTP/1.1 200 OK - 22334 bytes in 9.922 second response time [00:14:46] RECOVERY - cp30 Stunnel HTTP for mw102 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 7.819 second response time [00:15:06] PROBLEM - test101 Current Load on test101 is CRITICAL: CRITICAL - load average: 2.06, 2.03, 1.79 [00:15:07] RECOVERY - cp31 Stunnel HTTP for mw102 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 7.323 second response time [00:15:18] RECOVERY - cp20 Stunnel HTTP for mw102 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14548 bytes in 8.322 second response time [00:15:32] RECOVERY - db101 Current Load on db101 is OK: OK - load average: 4.32, 6.14, 6.67 [00:17:08] !log DELETE FROM incidents WHERE i_id='50'; duplicate report created by mistake [00:17:54] RECOVERY - cp30 Stunnel HTTP for mw111 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 8.685 second response time [00:18:04] RECOVERY - cp31 Stunnel HTTP for mw111 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14548 bytes in 7.599 second response time [00:18:07] RECOVERY - cp20 Stunnel HTTP for mw111 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 6.043 second response time [00:18:18] RECOVERY - mw111 MediaWiki Rendering on mw111 is OK: HTTP OK: HTTP/1.1 200 OK - 22335 bytes in 4.513 second response time [00:18:21] RECOVERY - cp30 Stunnel HTTP for mw101 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14548 bytes in 6.202 second response time [00:18:21] PROBLEM - cp21 Stunnel HTTP for mw102 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:18:38] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [00:18:40] RECOVERY - cp21 Stunnel HTTP for mw101 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14548 bytes in 2.984 second response time [00:18:52] PROBLEM - cp31 Stunnel HTTP for mw121 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:19:01] PROBLEM - wiki.simorgh.me - reverse DNS on sslhost is WARNING: Traceback (most recent call last): File "/usr/lib/nagios/plugins/check_reverse_dns.py", line 155, in main() File "/usr/lib/nagios/plugins/check_reverse_dns.py", line 136, in main records = check_records(args.hostname) File "/usr/lib/nagios/plugins/check_reverse_dns.py", line 86, in check_records cname = str(dns_resolver.resolve(hostname, 'CNAME')[0]) File "/usr/lib/py [00:19:01] st-packages/dns/resolver.py", line 1040, in resolve (nameserver, port, tcp, backoff) = resolution.next_nameserver() File "/usr/lib/python3/dist-packages/dns/resolver.py", line 598, in next_nameserver raise NoNameservers(request=self.request, errors=self.errors)dns.resolver.NoNameservers: All nameservers failed to answer the query wiki.simorgh.me. IN CNAME: Server 2606:4700:4700::1111 UDP port 53 answered SERVFAIL [00:19:03] RECOVERY - cp20 Stunnel HTTP for mw101 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 0.095 second response time [00:19:07] !log DELETE FROM incidents_log WHERE log_incident='50'; duplicate report created by mistake [00:19:15] PROBLEM - cp30 Stunnel HTTP for mw102 on cp30 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:19:27] RECOVERY - mw101 MediaWiki Rendering on mw101 is OK: HTTP OK: HTTP/1.1 200 OK - 22334 bytes in 2.407 second response time [00:19:29] PROBLEM - cp31 Stunnel HTTP for mw102 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:19:30] RECOVERY - cp31 Stunnel HTTP for mw101 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 1.670 second response time [00:19:33] RECOVERY - cp21 Stunnel HTTP for mw111 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 0.484 second response time [00:20:29] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [00:20:44] !log DELETE FROM incidents_reviewer WHERE r_incident='50'; duplicate report created by mistake [00:21:10] PROBLEM - cp31 Current Load on cp31 is CRITICAL: CRITICAL - load average: 2.01, 1.72, 1.13 [00:22:32] RECOVERY - cp21 Stunnel HTTP for mw102 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 9.760 second response time [00:22:40] PROBLEM - mw111 MediaWiki Rendering on mw111 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:22:44] PROBLEM - cp30 Stunnel HTTP for mw101 on cp30 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:22:54] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [00:23:01] PROBLEM - cp21 Stunnel HTTP for mw101 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:23:08] RECOVERY - cp31 Current Load on cp31 is OK: OK - load average: 0.71, 1.32, 1.05 [00:23:31] RECOVERY - cp30 Stunnel HTTP for mw102 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 5.969 second response time [00:23:31] PROBLEM - cp20 Stunnel HTTP for mw101 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:23:34] RECOVERY - cp31 Stunnel HTTP for mw102 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 4.977 second response time [00:23:37] PROBLEM - mw101 MediaWiki Rendering on mw101 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:23:45] PROBLEM - cp21 Stunnel HTTP for mw111 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:23:57] PROBLEM - cp31 Stunnel HTTP for mw101 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:24:07] PROBLEM - cp30 Stunnel HTTP for mw111 on cp30 is CRITICAL: HTTP CRITICAL - No data received from host [00:24:20] PROBLEM - cp31 Stunnel HTTP for mw111 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:24:27] PROBLEM - cp20 Stunnel HTTP for mw111 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:24:29] PROBLEM - cp30 Current Load on cp30 is CRITICAL: CRITICAL - load average: 1.74, 2.67, 1.79 [00:25:33] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 4.72, 3.70, 3.29 [00:25:42] RECOVERY - cp21 Stunnel HTTP for mw111 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 0.019 second response time [00:25:44] PROBLEM - dnd.bellinrattin.it - reverse DNS on sslhost is WARNING: Traceback (most recent call last): File "/usr/lib/nagios/plugins/check_reverse_dns.py", line 155, in main() File "/usr/lib/nagios/plugins/check_reverse_dns.py", line 136, in main records = check_records(args.hostname) File "/usr/lib/nagios/plugins/check_reverse_dns.py", line 73, in check_records nameserversans = dns_resolver.resolve(root_domain, 'NS') File "/usr/l [00:25:44] n3/dist-packages/dns/resolver.py", line 1040, in resolve (nameserver, port, tcp, backoff) = resolution.next_nameserver() File "/usr/lib/python3/dist-packages/dns/resolver.py", line 598, in next_nameserver raise NoNameservers(request=self.request, errors=self.errors)dns.resolver.NoNameservers: All nameservers failed to answer the query bellinrattin.it. IN NS: Server 2606:4700:4700::1111 UDP port 53 answered SERVFAIL [00:25:55] ugh, this is the same sort of problem we were having last night, dont think there is anything I can do about it [00:26:05] RECOVERY - cp30 Stunnel HTTP for mw111 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 0.475 second response time [00:26:15] RECOVERY - cp31 Stunnel HTTP for mw111 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 0.346 second response time [00:26:21] RECOVERY - cp20 Stunnel HTTP for mw111 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 0.019 second response time [00:26:28] PROBLEM - cp30 Current Load on cp30 is WARNING: WARNING - load average: 0.64, 1.98, 1.65 [00:26:41] RECOVERY - mw111 MediaWiki Rendering on mw111 is OK: HTTP OK: HTTP/1.1 200 OK - 22335 bytes in 1.133 second response time [00:27:31] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.05, 3.51, 3.27 [00:27:44] PROBLEM - cp30 Varnish Backends on cp30 is CRITICAL: 2 backends are down. mw102 mw111 [00:28:27] RECOVERY - cp30 Current Load on cp30 is OK: OK - load average: 0.99, 1.63, 1.56 [00:29:39] RECOVERY - cp30 Varnish Backends on cp30 is OK: All 12 backends are healthy [00:29:42] RECOVERY - mw101 MediaWiki Rendering on mw101 is OK: HTTP OK: HTTP/1.1 200 OK - 22334 bytes in 1.066 second response time [00:29:48] RECOVERY - cp20 Stunnel HTTP for mw101 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 0.045 second response time [00:30:05] RECOVERY - cp31 Stunnel HTTP for mw101 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 0.318 second response time [00:30:51] RECOVERY - cp30 Stunnel HTTP for mw101 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 0.313 second response time [00:31:10] PROBLEM - mw102 MediaWiki Rendering on mw102 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:31:12] PROBLEM - cp20 Varnish Backends on cp20 is CRITICAL: 2 backends are down. mw101 mw112 [00:31:18] PROBLEM - cp20 Stunnel HTTP for mw102 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:31:20] RECOVERY - cp21 Stunnel HTTP for mw101 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 1.560 second response time [00:31:36] PROBLEM - cp21 Varnish Backends on cp21 is CRITICAL: 2 backends are down. mw101 mw121 [00:31:50] RECOVERY - mw101 Puppet on mw101 is OK: OK: Puppet is currently enabled, last run 26 seconds ago with 0 failures [00:32:09] RECOVERY - cp21 Stunnel HTTP for mw112 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 0.018 second response time [00:32:25] RECOVERY - cp30 Stunnel HTTP for mw121 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14548 bytes in 6.967 second response time [00:32:31] RECOVERY - cp20 Stunnel HTTP for mw121 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14548 bytes in 7.081 second response time [00:32:45] RECOVERY - cp31 Stunnel HTTP for mw112 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 8.920 second response time [00:33:02] !log restart php7.4-fpm and nginx on mw* [00:33:25] RECOVERY - cp21 Stunnel HTTP for mw121 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 4.245 second response time [00:33:25] RECOVERY - mw121 MediaWiki Rendering on mw121 is OK: HTTP OK: HTTP/1.1 200 OK - 22334 bytes in 4.701 second response time [00:33:26] RECOVERY - gluster101 Current Load on gluster101 is OK: OK - load average: 2.75, 3.22, 3.22 [00:33:28] RECOVERY - cp21 Stunnel HTTP for mw122 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 0.021 second response time [00:33:31] RECOVERY - cp20 Stunnel HTTP for mw122 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 0.015 second response time [00:34:07] RECOVERY - cp30 Stunnel HTTP for mw122 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 3.876 second response time [00:34:22] RECOVERY - mw122 MediaWiki Rendering on mw122 is OK: HTTP OK: HTTP/1.1 200 OK - 22334 bytes in 6.864 second response time [00:34:29] PROBLEM - cp31 Stunnel HTTP for mw111 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:34:34] RECOVERY - cp31 Stunnel HTTP for mw122 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 8.079 second response time [00:34:39] PROBLEM - cp20 Stunnel HTTP for mw111 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:34:53] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [00:35:06] PROBLEM - cp31 Varnish Backends on cp31 is CRITICAL: 1 backends are down. mw102 [00:35:08] RECOVERY - mw102 MediaWiki Rendering on mw102 is OK: HTTP OK: HTTP/1.1 200 OK - 22334 bytes in 1.065 second response time [00:35:20] ._. [00:35:27] RECOVERY - cp20 Stunnel HTTP for mw102 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 0.016 second response time [00:35:59] PROBLEM - mw101 MediaWiki Rendering on mw101 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:36:05] PROBLEM - cp20 Stunnel HTTP for mw101 on cp20 is CRITICAL: HTTP CRITICAL: HTTP/1.1 502 Bad Gateway - 328 bytes in 0.011 second response time [00:36:21] PROBLEM - cp31 Stunnel HTTP for mw101 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:36:26] RECOVERY - cp31 Stunnel HTTP for mw111 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 2.575 second response time [00:36:28] PROBLEM - cp21 Stunnel HTTP for mw112 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:36:36] RECOVERY - cp20 Stunnel HTTP for mw111 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 3.161 second response time [00:37:01] i suspect this will resolve itself eventually, but the logs don't show anything helpful, meaning i cant determine why this is happening [00:37:05] PROBLEM - test101 Current Load on test101 is WARNING: WARNING - load average: 1.86, 1.99, 1.99 [00:37:06] PROBLEM - cp31 Stunnel HTTP for mw112 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:37:07] PROBLEM - cp30 Stunnel HTTP for mw101 on cp30 is CRITICAL: HTTP CRITICAL: HTTP/1.1 502 Bad Gateway - 328 bytes in 1.518 second response time [00:37:28] RECOVERY - cp30 Stunnel HTTP for mw112 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 9.652 second response time [00:37:30] RECOVERY - cp20 Stunnel HTTP for mw112 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 9.848 second response time [00:37:32] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 9.36, 7.41, 6.61 [00:37:47] PROBLEM - cp21 Stunnel HTTP for mw101 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:37:48] PROBLEM - cp21 Stunnel HTTP for mw122 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:37:58] PROBLEM - cp20 Stunnel HTTP for mw122 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:38:02] RECOVERY - mw101 MediaWiki Rendering on mw101 is OK: HTTP OK: HTTP/1.1 200 OK - 22335 bytes in 9.101 second response time [00:38:09] RECOVERY - cp20 Stunnel HTTP for mw101 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14548 bytes in 5.103 second response time [00:38:19] PROBLEM - cp30 Stunnel HTTP for mw122 on cp30 is CRITICAL: HTTP CRITICAL - No data received from host [00:38:21] RECOVERY - cp31 Stunnel HTTP for mw101 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 0.987 second response time [00:38:30] RECOVERY - cp21 Stunnel HTTP for mw112 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 7.686 second response time [00:38:37] !sre we are having the same issue that we had last night [00:38:46] PROBLEM - mw122 MediaWiki Rendering on mw122 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:39:01] PROBLEM - cp31 Stunnel HTTP for mw122 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:39:05] RECOVERY - cp30 Stunnel HTTP for mw101 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 2.082 second response time [00:39:05] PROBLEM - test101 Current Load on test101 is CRITICAL: CRITICAL - load average: 2.15, 2.03, 2.01 [00:39:06] RECOVERY - cp31 Stunnel HTTP for mw112 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 5.399 second response time [00:39:10] RECOVERY - mw112 MediaWiki Rendering on mw112 is OK: HTTP OK: HTTP/1.1 200 OK - 22334 bytes in 3.672 second response time [00:39:32] RECOVERY - db101 Current Load on db101 is OK: OK - load average: 4.99, 6.46, 6.36 [00:39:45] RECOVERY - cp21 Stunnel HTTP for mw101 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 0.376 second response time [00:40:25] PROBLEM - cp21 Stunnel HTTP for mw102 on cp21 is CRITICAL: HTTP CRITICAL: HTTP/1.1 502 Bad Gateway - 328 bytes in 0.014 second response time [00:40:47] PROBLEM - cp31 Stunnel HTTP for mw102 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:40:50] PROBLEM - gluster121 Current Load on gluster121 is WARNING: WARNING - load average: 3.66, 3.21, 2.90 [00:40:51] PROBLEM - mw121 MediaWiki Rendering on mw121 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:41:01] PROBLEM - cp30 Stunnel HTTP for mw121 on cp30 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:41:06] RECOVERY - cp31 Varnish Backends on cp31 is OK: All 12 backends are healthy [00:41:08] PROBLEM - cp21 Stunnel HTTP for mw121 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:41:08] PROBLEM - cp20 Stunnel HTTP for mw121 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:41:16] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 4.13, 3.39, 3.21 [00:41:20] PROBLEM - mw102 MediaWiki Rendering on mw102 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:41:25] PROBLEM - cp30 Stunnel HTTP for mw102 on cp30 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:41:46] PROBLEM - cp20 Stunnel HTTP for mw102 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:41:49] PROBLEM - cp30 Stunnel HTTP for mw111 on cp30 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:41:50] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 4.03, 3.46, 3.07 [00:42:13] PROBLEM - mw111 MediaWiki Rendering on mw111 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:42:17] PROBLEM - cp21 Stunnel HTTP for mw111 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:42:43] PROBLEM - cp31 Stunnel HTTP for mw111 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:42:50] RECOVERY - gluster121 Current Load on gluster121 is OK: OK - load average: 3.17, 3.06, 2.87 [00:42:51] PROBLEM - cp20 Stunnel HTTP for mw111 on cp20 is CRITICAL: HTTP CRITICAL: HTTP/1.1 502 Bad Gateway - 328 bytes in 0.010 second response time [00:43:14] RECOVERY - gluster101 Current Load on gluster101 is OK: OK - load average: 3.12, 3.25, 3.18 [00:43:18] PROBLEM - cp30 Stunnel HTTP for mw101 on cp30 is CRITICAL: HTTP CRITICAL: HTTP/1.1 502 Bad Gateway - 328 bytes in 0.561 second response time [00:43:19] RECOVERY - mw102 MediaWiki Rendering on mw102 is OK: HTTP OK: HTTP/1.1 200 OK - 22334 bytes in 5.695 second response time [00:43:25] RECOVERY - cp30 Stunnel HTTP for mw102 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 0.334 second response time [00:43:47] RECOVERY - cp20 Stunnel HTTP for mw102 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 1.137 second response time [00:43:48] RECOVERY - gluster111 Current Load on gluster111 is OK: OK - load average: 2.61, 3.05, 2.96 [00:43:58] PROBLEM - cp21 Stunnel HTTP for mw101 on cp21 is CRITICAL: HTTP CRITICAL: HTTP/1.1 502 Bad Gateway - 328 bytes in 0.010 second response time [00:44:10] RECOVERY - cp31 Stunnel HTTP for mw121 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 5.041 second response time [00:44:19] PROBLEM - mw101 MediaWiki Rendering on mw101 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:44:26] RECOVERY - cp21 Stunnel HTTP for mw102 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 0.019 second response time [00:44:36] PROBLEM - cp20 Stunnel HTTP for mw101 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:44:38] PROBLEM - cp31 Stunnel HTTP for mw101 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:44:46] PROBLEM - cp21 Stunnel HTTP for mw112 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:44:48] RECOVERY - cp31 Stunnel HTTP for mw102 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 0.327 second response time [00:45:22] PROBLEM - cp31 Stunnel HTTP for mw112 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:45:38] PROBLEM - mw112 MediaWiki Rendering on mw112 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:45:51] PROBLEM - cp30 Stunnel HTTP for mw112 on cp30 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:45:53] PROBLEM - cp30 Varnish Backends on cp30 is CRITICAL: 1 backends are down. mw102 [00:46:02] PROBLEM - cp20 Stunnel HTTP for mw112 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:47:48] RECOVERY - cp30 Varnish Backends on cp30 is OK: All 12 backends are healthy [00:48:26] PROBLEM - cp31 Stunnel HTTP for mw121 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:48:48] RECOVERY - cp20 Stunnel HTTP for mw101 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 3.445 second response time [00:48:48] PROBLEM - cp21 Stunnel HTTP for mw102 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:48:49] RECOVERY - cp31 Stunnel HTTP for mw101 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 2.955 second response time [00:48:59] RECOVERY - wiki.simorgh.me - reverse DNS on sslhost is OK: SSL OK - wiki.simorgh.me reverse DNS resolves to cp21.miraheze.org - CNAME OK [00:49:17] RECOVERY - mw122 MediaWiki Rendering on mw122 is OK: HTTP OK: HTTP/1.1 200 OK - 22334 bytes in 9.261 second response time [00:49:31] RECOVERY - cp31 Stunnel HTTP for mw122 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 6.479 second response time [00:49:34] PROBLEM - mw102 MediaWiki Rendering on mw102 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:50:02] RECOVERY - cp21 Stunnel HTTP for mw122 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14548 bytes in 6.702 second response time [00:50:17] RECOVERY - cp21 Stunnel HTTP for mw101 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14548 bytes in 8.609 second response time [00:50:30] RECOVERY - mw101 MediaWiki Rendering on mw101 is OK: HTTP OK: HTTP/1.1 200 OK - 22334 bytes in 7.404 second response time [00:50:32] RECOVERY - cp20 Stunnel HTTP for mw122 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 0.971 second response time [00:50:40] RECOVERY - cp30 Stunnel HTTP for mw122 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 0.423 second response time [00:50:49] PROBLEM - cp31 Current Load on cp31 is CRITICAL: CRITICAL - load average: 2.22, 2.36, 1.49 [00:51:12] RECOVERY - cp20 Varnish Backends on cp20 is OK: All 12 backends are healthy [00:51:12] PROBLEM - cp31 Stunnel HTTP for mw102 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:51:14] PROBLEM - cp30 Stunnel HTTP for mw102 on cp30 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:51:18] PROBLEM - cp20 Stunnel HTTP for mw102 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:51:36] RECOVERY - cp21 Varnish Backends on cp21 is OK: All 12 backends are healthy [00:51:48] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 5.94, 3.78, 3.18 [00:52:33] RECOVERY - cp31 Stunnel HTTP for mw121 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 3.665 second response time [00:52:47] PROBLEM - cp31 Current Load on cp31 is WARNING: WARNING - load average: 0.58, 1.72, 1.36 [00:53:12] RECOVERY - mw121 MediaWiki Rendering on mw121 is OK: HTTP OK: HTTP/1.1 200 OK - 22334 bytes in 0.664 second response time [00:53:19] RECOVERY - cp30 Stunnel HTTP for mw121 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 0.313 second response time [00:53:32] PROBLEM - cp30 Varnish Backends on cp30 is CRITICAL: 1 backends are down. mw121 [00:53:43] RECOVERY - cp21 Stunnel HTTP for mw121 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 1.169 second response time [00:53:47] RECOVERY - cp20 Stunnel HTTP for mw121 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 1.342 second response time [00:54:38] PROBLEM - cp21 Stunnel HTTP for mw101 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:54:46] RECOVERY - cp31 Current Load on cp31 is OK: OK - load average: 0.51, 1.34, 1.26 [00:54:50] PROBLEM - gluster121 Current Load on gluster121 is WARNING: WARNING - load average: 3.56, 3.31, 2.85 [00:54:51] PROBLEM - mw101 MediaWiki Rendering on mw101 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:55:01] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 4.75, 3.92, 3.36 [00:55:12] PROBLEM - cp20 Stunnel HTTP for mw101 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:55:12] PROBLEM - cp20 Varnish Backends on cp20 is CRITICAL: 1 backends are down. mw112 [00:55:13] PROBLEM - cp31 Stunnel HTTP for mw101 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:55:27] RECOVERY - cp30 Varnish Backends on cp30 is OK: All 12 backends are healthy [00:55:44] PROBLEM - dnd.bellinrattin.it - reverse DNS on sslhost is CRITICAL: rDNS CRITICAL - dnd.bellinrattin.it All nameservers failed to answer the query. [00:56:10] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 7.49, 6.90, 6.44 [00:56:25] PROBLEM - cp21 Stunnel HTTP for mw122 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:56:50] RECOVERY - gluster121 Current Load on gluster121 is OK: OK - load average: 3.39, 3.36, 2.92 [00:56:59] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.75, 3.98, 3.45 [00:57:03] PROBLEM - cp20 Stunnel HTTP for mw122 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:57:04] PROBLEM - cp30 Stunnel HTTP for mw122 on cp30 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:57:10] RECOVERY - cp31 Stunnel HTTP for mw111 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 7.134 second response time [00:57:11] RECOVERY - cp20 Stunnel HTTP for mw111 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14548 bytes in 5.693 second response time [00:57:28] RECOVERY - cp31 Stunnel HTTP for mw102 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 9.735 second response time [00:57:36] PROBLEM - cp21 Varnish Backends on cp21 is CRITICAL: 1 backends are down. mw111 [00:57:42] RECOVERY - cp30 Stunnel HTTP for mw102 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 8.821 second response time [00:57:44] RECOVERY - cp20 Stunnel HTTP for mw102 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 6.833 second response time [00:57:45] PROBLEM - mw122 MediaWiki Rendering on mw122 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:57:47] RECOVERY - mw102 MediaWiki Rendering on mw102 is OK: HTTP OK: HTTP/1.1 200 OK - 22334 bytes in 7.410 second response time [00:57:48] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.47, 3.92, 3.47 [00:58:03] PROBLEM - cp31 Stunnel HTTP for mw122 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:58:07] RECOVERY - db101 Current Load on db101 is OK: OK - load average: 5.83, 6.64, 6.41 [00:58:23] RECOVERY - cp30 Stunnel HTTP for mw111 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14548 bytes in 0.470 second response time [00:58:26] RECOVERY - mw111 MediaWiki Rendering on mw111 is OK: HTTP OK: HTTP/1.1 200 OK - 22334 bytes in 0.642 second response time [00:58:45] RECOVERY - cp21 Stunnel HTTP for mw111 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 0.362 second response time [00:59:03] PROBLEM - cp31 Stunnel HTTP for mw121 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:59:10] RECOVERY - cp21 Stunnel HTTP for mw102 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14548 bytes in 5.284 second response time [00:59:12] RECOVERY - cp20 Varnish Backends on cp20 is OK: All 12 backends are healthy [00:59:37] PROBLEM - mw121 MediaWiki Rendering on mw121 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:59:47] PROBLEM - cp30 Stunnel HTTP for mw121 on cp30 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:59:48] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 4.86, 4.16, 3.60 [01:00:11] PROBLEM - cp20 Stunnel HTTP for mw121 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:00:13] PROBLEM - cp21 Stunnel HTTP for mw121 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:00:50] PROBLEM - gluster121 Current Load on gluster121 is WARNING: WARNING - load average: 3.45, 3.84, 3.24 [01:01:36] RECOVERY - cp21 Varnish Backends on cp21 is OK: All 12 backends are healthy [01:01:48] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.11, 3.74, 3.52 [01:02:50] RECOVERY - gluster121 Current Load on gluster121 is OK: OK - load average: 2.26, 3.36, 3.14 [01:02:54] RECOVERY - gluster101 Current Load on gluster101 is OK: OK - load average: 2.48, 3.37, 3.39 [01:03:39] PROBLEM - cp31 Stunnel HTTP for mw102 on cp31 is CRITICAL: HTTP CRITICAL - No data received from host [01:03:48] RECOVERY - gluster111 Current Load on gluster111 is OK: OK - load average: 2.37, 3.24, 3.36 [01:04:04] PROBLEM - mw102 MediaWiki Rendering on mw102 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:04:17] PROBLEM - cp30 Stunnel HTTP for mw102 on cp30 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:04:18] PROBLEM - cp20 Stunnel HTTP for mw102 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:04:46] PROBLEM - mw111 MediaWiki Rendering on mw111 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:04:47] PROBLEM - cp30 Stunnel HTTP for mw111 on cp30 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:05:33] PROBLEM - cp21 Stunnel HTTP for mw102 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:06:07] RECOVERY - cp31 Stunnel HTTP for mw112 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 2.238 second response time [01:06:35] RECOVERY - cp30 Stunnel HTTP for mw112 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 0.682 second response time [01:06:39] RECOVERY - mw112 MediaWiki Rendering on mw112 is OK: HTTP OK: HTTP/1.1 200 OK - 22335 bytes in 1.256 second response time [01:07:11] RECOVERY - cp21 Stunnel HTTP for mw112 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 0.017 second response time [01:07:18] RECOVERY - cp20 Stunnel HTTP for mw112 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 0.249 second response time [01:08:36] RECOVERY - cp31 Stunnel HTTP for mw122 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 9.969 second response time [01:09:29] PROBLEM - cp21 Stunnel HTTP for mw111 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:09:30] PROBLEM - cp31 Stunnel HTTP for mw111 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:09:31] RECOVERY - cp30 Stunnel HTTP for mw122 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 4.210 second response time [01:09:35] PROBLEM - cp20 Stunnel HTTP for mw111 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:09:47] RECOVERY - cp20 Stunnel HTTP for mw122 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 0.093 second response time [01:10:08] PROBLEM - cp31 Stunnel HTTP for mw112 on cp31 is CRITICAL: HTTP CRITICAL: HTTP/1.1 502 Bad Gateway - 328 bytes in 0.239 second response time [01:10:13] RECOVERY - mw122 MediaWiki Rendering on mw122 is OK: HTTP OK: HTTP/1.1 200 OK - 22334 bytes in 2.596 second response time [01:10:46] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 4.28, 3.59, 3.42 [01:10:52] [discord] getting 15-30 sec response delays on non-cached content [01:10:52] PROBLEM - mw112 MediaWiki Rendering on mw112 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:10:56] PROBLEM - cp30 Stunnel HTTP for mw112 on cp30 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:10:59] RECOVERY - cp21 Stunnel HTTP for mw122 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14548 bytes in 7.331 second response time [01:11:31] PROBLEM - cp21 Stunnel HTTP for mw112 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:11:41] [discord] it is quite unfortunate if nothing is on the logs and apache-status also doesn't show anything [01:11:44] PROBLEM - cp20 Stunnel HTTP for mw112 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:12:13] [discord] it is quite unfortunate if nothing is on the logs and apache-status also doesn't show anything worthwhile (edited) [01:12:44] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.87, 3.74, 3.50 [01:12:52] MacFan4000: are we still having issues again? [01:12:52] [discord] getting 15-30 sec response delays on non-cached content, cached is instant (edited) [01:13:33] I even tried rebooting the mw servers yesterday and the issues just kept coming back. [01:13:46] RECOVERY - cp20 Stunnel HTTP for mw112 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 3.051 second response time [01:13:52] PROBLEM - cp30 Stunnel HTTP for mw122 on cp30 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:14:08] RECOVERY - cp31 Stunnel HTTP for mw112 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 0.312 second response time [01:14:15] PROBLEM - cp20 Stunnel HTTP for mw122 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:14:33] This is getting ridiculous. [01:14:36] PROBLEM - mw122 MediaWiki Rendering on mw122 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:14:42] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 6.02, 4.37, 3.74 [01:14:54] How and why is this happening...? [01:14:58] RECOVERY - cp30 Stunnel HTTP for mw112 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14548 bytes in 1.595 second response time [01:15:00] RECOVERY - mw112 MediaWiki Rendering on mw112 is OK: HTTP OK: HTTP/1.1 200 OK - 22334 bytes in 2.331 second response time [01:15:02] PROBLEM - cp31 Stunnel HTTP for mw122 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:15:09] PROBLEM - cp21 Stunnel HTTP for mw122 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:15:13] RECOVERY - cp21 Stunnel HTTP for mw121 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 7.856 second response time [01:15:30] RECOVERY - cp21 Stunnel HTTP for mw112 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 0.016 second response time [01:15:32] RECOVERY - cp31 Stunnel HTTP for mw121 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 5.094 second response time [01:15:48] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.81, 3.51, 3.33 [01:16:03] RECOVERY - cp30 Stunnel HTTP for mw121 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 0.727 second response time [01:16:08] RECOVERY - mw121 MediaWiki Rendering on mw121 is OK: HTTP OK: HTTP/1.1 200 OK - 22334 bytes in 0.748 second response time [01:16:45] RECOVERY - cp20 Stunnel HTTP for mw121 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 0.014 second response time [01:17:01] PROBLEM - cp30 Varnish Backends on cp30 is CRITICAL: 1 backends are down. mw112 [01:17:12] PROBLEM - cp20 Varnish Backends on cp20 is CRITICAL: 1 backends are down. mw102 [01:17:45] MacFan4000: WHY is this happening again??? [01:18:00] Same time as yesterday also. [01:18:41] RECOVERY - cp31 Stunnel HTTP for mw101 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 8.844 second response time [01:18:42] RECOVERY - cp30 Stunnel HTTP for mw101 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 8.707 second response time [01:19:01] RECOVERY - cp30 Varnish Backends on cp30 is OK: All 12 backends are healthy [01:19:10] [discord] it was getting kinda slow an hour ago, but now is [01:19:11] [discord] https://cdn.discordapp.com/attachments/808001911868489748/962159711002697769/unknown.png [01:19:18] RECOVERY - cp20 Stunnel HTTP for mw102 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 3.792 second response time [01:19:20] RECOVERY - mw101 MediaWiki Rendering on mw101 is OK: HTTP OK: HTTP/1.1 200 OK - 22335 bytes in 2.208 second response time [01:19:20] RECOVERY - cp30 Stunnel HTTP for mw102 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14548 bytes in 4.366 second response time [01:19:32] RECOVERY - cp21 Stunnel HTTP for mw101 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 2.323 second response time [01:19:36] PROBLEM - cp21 Varnish Backends on cp21 is CRITICAL: 1 backends are down. mw101 [01:19:48] RECOVERY - gluster111 Current Load on gluster111 is OK: OK - load average: 3.18, 3.34, 3.30 [01:19:48] PROBLEM - cp31 Stunnel HTTP for mw121 on cp31 is CRITICAL: HTTP CRITICAL: HTTP/1.1 502 Bad Gateway - 328 bytes in 0.235 second response time [01:19:50] RECOVERY - cp21 Stunnel HTTP for mw102 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 6.121 second response time [01:20:06] RECOVERY - cp31 Stunnel HTTP for mw102 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 4.194 second response time [01:20:22] PROBLEM - cp30 Stunnel HTTP for mw121 on cp30 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:20:24] RECOVERY - cp20 Stunnel HTTP for mw101 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 2.193 second response time [01:20:29] PROBLEM - mw121 MediaWiki Rendering on mw121 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:20:34] RECOVERY - mw102 MediaWiki Rendering on mw102 is OK: HTTP OK: HTTP/1.1 200 OK - 22334 bytes in 5.644 second response time [01:20:44] [discord] hmm, reloaded and is totally inconsistent [01:20:45] [discord] https://cdn.discordapp.com/attachments/808001911868489748/962160106026434570/unknown.png [01:20:49] PROBLEM - cp20 Stunnel HTTP for mw121 on cp20 is CRITICAL: HTTP CRITICAL: HTTP/1.1 502 Bad Gateway - 328 bytes in 0.010 second response time [01:20:57] [discord] but still with 25 sec spikes [01:21:11] !log reboot mw* [01:21:12] RECOVERY - cp20 Varnish Backends on cp20 is OK: All 12 backends are healthy [01:21:36] RECOVERY - cp21 Varnish Backends on cp21 is OK: All 12 backends are healthy [01:21:43] PROBLEM - cp21 Stunnel HTTP for mw112 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:21:49] PROBLEM - cp21 Stunnel HTTP for mw121 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:22:10] PROBLEM - cp20 Stunnel HTTP for mw112 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:23:40] RECOVERY - cp21 Stunnel HTTP for mw112 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 2.679 second response time [01:23:47] PROBLEM - cp20 Stunnel HTTP for mw102 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:23:49] PROBLEM - cp30 Stunnel HTTP for mw102 on cp30 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:24:09] RECOVERY - cp20 Stunnel HTTP for mw112 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 0.017 second response time [01:24:12] PROBLEM - cp21 Stunnel HTTP for mw102 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:24:24] And logbot quit... [01:24:28] PROBLEM - cp31 Stunnel HTTP for mw102 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:24:33] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.73, 4.00, 3.87 [01:24:48] Ugh.... [01:24:51] PROBLEM - mw102 MediaWiki Rendering on mw102 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:24:58] There's nothing more I can do right now... [01:25:05] PROBLEM - test101 Current Load on test101 is WARNING: WARNING - load average: 0.83, 1.53, 1.87 [01:25:27] [discord] yeah, the times are just a little bit better... retrieving a simple css file... [01:25:43] RECOVERY - dnd.bellinrattin.it - reverse DNS on sslhost is OK: SSL OK - dnd.bellinrattin.it reverse DNS resolves to cp21.miraheze.org - CNAME OK [01:25:54] HOW??? What is pointing to test101??? test101 is currently down for MediaWiki so how is it having high loads?? [01:26:01] !log reboot mw* [01:26:31] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 4.62, 4.19, 3.96 [01:26:53] I have to go now again also.... [01:27:14] [discord] i am sorry for having to disturb you and making you lose sanity Cosmic [01:27:48] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 4.47, 3.76, 3.41 [01:28:08] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 7.42, 6.67, 6.22 [01:28:29] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 2.64, 3.61, 3.77 [01:29:05] RECOVERY - test101 Current Load on test101 is OK: OK - load average: 1.08, 1.24, 1.67 [01:29:22] @Kozd No problem at all, I don't understand what is happening here... I tried rebooting the servers, restarting fpm, everything I can think of, and can not figure this out. I'm going to recommend to the rest of SRE that if these issues continue again tomorrow, we focus our entire efforts for a resolution as this is getting ridiculous now. [01:29:48] RECOVERY - gluster111 Current Load on gluster111 is OK: OK - load average: 2.48, 3.32, 3.29 [01:30:04] RECOVERY - db101 Current Load on db101 is OK: OK - load average: 4.94, 6.18, 6.10 [01:30:08] Very sorry for the issues. [01:30:08] [discord] yeah i am not sure if installing a real profiler is worth it at this time, even if it would show long times waiting on db or something [01:30:22] PROBLEM - cp31 Stunnel HTTP for mw101 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:30:47] PROBLEM - mw101 MediaWiki Rendering on mw101 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:31:23] [discord] no problem, at least the cached responses are working great... hahaha. [01:32:02] PROBLEM - cp21 Stunnel HTTP for mw101 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:32:05] PROBLEM - cp20 Stunnel HTTP for mw101 on cp20 is CRITICAL: HTTP CRITICAL: HTTP/1.1 502 Bad Gateway - 328 bytes in 0.010 second response time [01:32:09] RECOVERY - cp30 Stunnel HTTP for mw102 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 0.321 second response time [01:32:14] PROBLEM - cp30 Stunnel HTTP for mw101 on cp30 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:32:15] RECOVERY - cp20 Stunnel HTTP for mw102 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 0.124 second response time [01:32:23] RECOVERY - cp21 Stunnel HTTP for mw102 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 0.017 second response time [01:32:26] RECOVERY - gluster101 Current Load on gluster101 is OK: OK - load average: 2.25, 2.81, 3.40 [01:32:47] RECOVERY - cp31 Stunnel HTTP for mw102 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 4.574 second response time [01:32:59] RECOVERY - mw102 MediaWiki Rendering on mw102 is OK: HTTP OK: HTTP/1.1 200 OK - 22334 bytes in 2.951 second response time [01:33:36] PROBLEM - cp20 Stunnel HTTP for mw112 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:33:41] PROBLEM - cp31 Stunnel HTTP for mw112 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:33:49] PROBLEM - cp30 Stunnel HTTP for mw112 on cp30 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:33:57] PROBLEM - mw112 MediaWiki Rendering on mw112 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:34:10] PROBLEM - cp21 Stunnel HTTP for mw112 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:37:36] PROBLEM - cp21 Varnish Backends on cp21 is CRITICAL: 1 backends are down. mw111 [01:38:29] Yeah no idea why this is happening [01:39:07] PROBLEM - cp31 Stunnel HTTP for mw102 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:39:14] PROBLEM - mw102 MediaWiki Rendering on mw102 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:39:44] PROBLEM - cp30 Stunnel HTTP for mw102 on cp30 is CRITICAL: HTTP CRITICAL - No data received from host [01:40:35] RECOVERY - cp30 Stunnel HTTP for mw122 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 3.991 second response time [01:40:50] PROBLEM - cp20 Stunnel HTTP for mw102 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:40:51] PROBLEM - cp21 Stunnel HTTP for mw102 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:41:18] RECOVERY - mw122 MediaWiki Rendering on mw122 is OK: HTTP OK: HTTP/1.1 200 OK - 22334 bytes in 0.578 second response time [01:41:36] RECOVERY - cp21 Varnish Backends on cp21 is OK: All 12 backends are healthy [01:41:57] RECOVERY - cp21 Stunnel HTTP for mw122 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 0.227 second response time [01:41:59] RECOVERY - cp20 Stunnel HTTP for mw122 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 0.016 second response time [01:42:20] RECOVERY - cp31 Stunnel HTTP for mw122 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 0.329 second response time [01:43:06] PROBLEM - cp31 Varnish Backends on cp31 is CRITICAL: 1 backends are down. mw112 [01:43:12] PROBLEM - cp20 Varnish Backends on cp20 is CRITICAL: 1 backends are down. mw102 [01:43:16] mw111 has been down consistently for 40 minutes. [01:43:49] RECOVERY - cp31 Stunnel HTTP for mw112 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 2.163 second response time [01:43:51] RECOVERY - cp30 Stunnel HTTP for mw102 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 7.157 second response time [01:43:59] RECOVERY - cp20 Stunnel HTTP for mw112 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 0.682 second response time [01:44:06] RECOVERY - cp30 Stunnel HTTP for mw112 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 0.516 second response time [01:44:17] RECOVERY - mw112 MediaWiki Rendering on mw112 is OK: HTTP OK: HTTP/1.1 200 OK - 22334 bytes in 0.827 second response time [01:44:21] RECOVERY - cp21 Stunnel HTTP for mw112 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 0.017 second response time [01:45:06] RECOVERY - cp31 Varnish Backends on cp31 is OK: All 12 backends are healthy [01:45:12] RECOVERY - cp20 Varnish Backends on cp20 is OK: All 12 backends are healthy [01:46:17] PROBLEM - cp21 Stunnel HTTP for mw122 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:46:26] PROBLEM - cp20 Stunnel HTTP for mw122 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:46:42] PROBLEM - cp30 Stunnel HTTP for mw122 on cp30 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:46:47] PROBLEM - cp31 Stunnel HTTP for mw122 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:47:09] .op [01:47:10] Attempting to OP... [01:47:21] RECOVERY - mw101 MediaWiki Rendering on mw101 is OK: HTTP OK: HTTP/1.1 200 OK - 22334 bytes in 5.617 second response time [01:47:27] RECOVERY - cp31 Stunnel HTTP for mw101 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 2.554 second response time [01:47:36] PROBLEM - cp21 Varnish Backends on cp21 is CRITICAL: 2 backends are down. mw101 mw122 [01:47:47] PROBLEM - mw122 MediaWiki Rendering on mw122 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:47:51] PROBLEM - cp30 Varnish Backends on cp30 is CRITICAL: 1 backends are down. mw101 [01:48:08] PROBLEM - hypotheticalhurricanes.com - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['ns2.hostknox.com.', 'ns1.hostknox.com.'], 'CNAME': None} [01:48:21] PROBLEM - cp30 Stunnel HTTP for mw102 on cp30 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:48:27] RECOVERY - cp21 Stunnel HTTP for mw101 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 0.019 second response time [01:48:31] RECOVERY - cp30 Stunnel HTTP for mw101 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 0.609 second response time [01:48:54] RECOVERY - cp20 Stunnel HTTP for mw101 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 5.166 second response time [01:49:45] RECOVERY - cp30 Varnish Backends on cp30 is OK: All 12 backends are healthy [01:49:51] RECOVERY - mw122 MediaWiki Rendering on mw122 is OK: HTTP OK: HTTP/1.1 200 OK - 22334 bytes in 8.332 second response time [01:50:02] PROBLEM - cp31 Stunnel HTTP for mw112 on cp31 is CRITICAL: HTTP CRITICAL: HTTP/1.1 502 Bad Gateway - 328 bytes in 0.257 second response time [01:50:16] RECOVERY - cp21 Stunnel HTTP for mw122 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 0.020 second response time [01:50:27] PROBLEM - cp30 Stunnel HTTP for mw112 on cp30 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:50:30] PROBLEM - cp20 Stunnel HTTP for mw112 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:50:36] PROBLEM - cp21 Stunnel HTTP for mw112 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:50:39] RECOVERY - cp20 Stunnel HTTP for mw122 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 5.017 second response time [01:50:41] PROBLEM - mw112 MediaWiki Rendering on mw112 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:50:51] RECOVERY - cp30 Stunnel HTTP for mw122 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 8.503 second response time [01:50:57] RECOVERY - cp31 Stunnel HTTP for mw122 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 6.620 second response time [01:51:12] PROBLEM - cp20 Varnish Backends on cp20 is CRITICAL: 1 backends are down. mw111 [01:51:17] PROBLEM - cp30 Current Load on cp30 is CRITICAL: CRITICAL - load average: 2.02, 1.36, 0.98 [01:51:36] RECOVERY - cp21 Varnish Backends on cp21 is OK: All 12 backends are healthy [01:51:41] PROBLEM - mw101 MediaWiki Rendering on mw101 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:51:54] PROBLEM - cp31 Stunnel HTTP for mw101 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:52:01] RECOVERY - cp31 Stunnel HTTP for mw112 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 3.318 second response time [01:52:27] RECOVERY - cp30 Stunnel HTTP for mw112 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14548 bytes in 4.866 second response time [01:52:35] RECOVERY - cp20 Stunnel HTTP for mw112 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14548 bytes in 6.180 second response time [01:52:37] RECOVERY - cp21 Stunnel HTTP for mw112 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 6.779 second response time [01:52:46] RECOVERY - mw112 MediaWiki Rendering on mw112 is OK: HTTP OK: HTTP/1.1 200 OK - 22334 bytes in 7.399 second response time [01:52:47] PROBLEM - cp21 Stunnel HTTP for mw101 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:52:54] PROBLEM - cp30 Stunnel HTTP for mw101 on cp30 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:53:12] RECOVERY - cp20 Varnish Backends on cp20 is OK: All 12 backends are healthy [01:53:17] RECOVERY - cp30 Current Load on cp30 is OK: OK - load average: 1.05, 1.28, 1.00 [01:53:21] PROBLEM - cp20 Stunnel HTTP for mw101 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:53:34] RECOVERY - cp20 Stunnel HTTP for mw102 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 6.742 second response time [01:53:37] RECOVERY - mw102 MediaWiki Rendering on mw102 is OK: HTTP OK: HTTP/1.1 200 OK - 22335 bytes in 7.739 second response time [01:53:41] RECOVERY - cp31 Stunnel HTTP for mw102 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 6.939 second response time [01:54:20] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 8.34, 7.27, 6.38 [01:54:35] PROBLEM - cp21 Stunnel HTTP for mw122 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:54:47] RECOVERY - cp30 Stunnel HTTP for mw102 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 7.489 second response time [01:55:06] PROBLEM - cp20 Stunnel HTTP for mw122 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:55:13] PROBLEM - cp30 Stunnel HTTP for mw122 on cp30 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:55:13] RECOVERY - cp21 Stunnel HTTP for mw102 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 4.349 second response time [01:55:26] PROBLEM - cp31 Stunnel HTTP for mw122 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:56:16] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 6.51, 7.03, 6.40 [01:56:19] PROBLEM - mw122 MediaWiki Rendering on mw122 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:58:12] RECOVERY - db101 Current Load on db101 is OK: OK - load average: 4.53, 6.16, 6.16 [01:59:22] RECOVERY - cp20 Stunnel HTTP for mw122 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14548 bytes in 8.637 second response time [01:59:22] RECOVERY - cp30 Stunnel HTTP for mw122 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14548 bytes in 9.198 second response time [01:59:24] PROBLEM - cp31 Stunnel HTTP for mw112 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:59:38] RECOVERY - cp31 Stunnel HTTP for mw122 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14548 bytes in 8.182 second response time [02:00:01] PROBLEM - cp30 Stunnel HTTP for mw112 on cp30 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [02:00:04] PROBLEM - cp21 Stunnel HTTP for mw112 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [02:00:15] PROBLEM - cp20 Stunnel HTTP for mw112 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [02:00:16] PROBLEM - mw112 MediaWiki Rendering on mw112 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [02:00:44] RECOVERY - cp31 Stunnel HTTP for mw111 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 6.914 second response time [02:00:48] RECOVERY - cp20 Stunnel HTTP for mw121 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 7.599 second response time [02:00:55] RECOVERY - mw111 MediaWiki Rendering on mw111 is OK: HTTP OK: HTTP/1.1 200 OK - 22334 bytes in 4.289 second response time [02:01:08] RECOVERY - cp20 Stunnel HTTP for mw111 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 1.455 second response time [02:01:17] PROBLEM - cp30 Stunnel HTTP for mw102 on cp30 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [02:01:21] RECOVERY - cp31 Stunnel HTTP for mw112 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 2.041 second response time [02:01:33] PROBLEM - cp21 Stunnel HTTP for mw102 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [02:01:36] RECOVERY - cp30 Stunnel HTTP for mw121 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 4.825 second response time [02:01:49] PROBLEM - mw102 MediaWiki Rendering on mw102 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [02:01:50] RECOVERY - cp31 Stunnel HTTP for mw121 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14548 bytes in 4.873 second response time [02:01:56] RECOVERY - cp30 Stunnel HTTP for mw112 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 0.333 second response time [02:01:59] RECOVERY - cp21 Stunnel HTTP for mw112 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 0.222 second response time [02:02:00] PROBLEM - cp31 Stunnel HTTP for mw102 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [02:02:03] RECOVERY - cp30 Stunnel HTTP for mw111 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 9.477 second response time [02:02:05] RECOVERY - cp21 Stunnel HTTP for mw111 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14548 bytes in 9.034 second response time [02:02:07] PROBLEM - cp20 Stunnel HTTP for mw102 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [02:02:07] RECOVERY - mw121 MediaWiki Rendering on mw121 is OK: HTTP OK: HTTP/1.1 200 OK - 22334 bytes in 8.443 second response time [02:02:17] RECOVERY - cp20 Stunnel HTTP for mw112 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14548 bytes in 2.750 second response time [02:02:17] RECOVERY - mw112 MediaWiki Rendering on mw112 is OK: HTTP OK: HTTP/1.1 200 OK - 22334 bytes in 3.285 second response time [02:02:36] RECOVERY - cp21 Stunnel HTTP for mw121 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 6.662 second response time [02:03:24] RECOVERY - cp30 Stunnel HTTP for mw102 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14548 bytes in 6.794 second response time [02:03:35] RECOVERY - cp21 Stunnel HTTP for mw102 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 3.018 second response time [02:03:44] PROBLEM - cp30 Stunnel HTTP for mw122 on cp30 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [02:03:46] RECOVERY - mw102 MediaWiki Rendering on mw102 is OK: HTTP OK: HTTP/1.1 200 OK - 22334 bytes in 3.378 second response time [02:03:49] PROBLEM - cp20 Stunnel HTTP for mw122 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [02:03:55] PROBLEM - cp31 Stunnel HTTP for mw122 on cp31 is CRITICAL: HTTP CRITICAL: HTTP/1.1 502 Bad Gateway - 328 bytes in 0.239 second response time [02:03:59] RECOVERY - cp31 Stunnel HTTP for mw102 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 3.116 second response time [02:04:09] RECOVERY - cp20 Stunnel HTTP for mw102 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 3.015 second response time [02:05:17] PROBLEM - cp30 Current Load on cp30 is CRITICAL: CRITICAL - load average: 2.74, 1.76, 1.28 [02:06:27] PROBLEM - cp30 Stunnel HTTP for mw111 on cp30 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [02:07:03] PROBLEM - cp21 Stunnel HTTP for mw121 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [02:07:04] PROBLEM - cp31 Stunnel HTTP for mw111 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [02:07:16] PROBLEM - cp20 Stunnel HTTP for mw121 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [02:07:17] RECOVERY - cp30 Current Load on cp30 is OK: OK - load average: 1.12, 1.43, 1.21 [02:07:29] PROBLEM - cp20 Stunnel HTTP for mw111 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [02:07:48] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.50, 2.83, 2.62 [02:07:50] PROBLEM - cp30 Stunnel HTTP for mw121 on cp30 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [02:07:53] PROBLEM - cp30 Stunnel HTTP for mw102 on cp30 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [02:07:58] PROBLEM - cp21 Stunnel HTTP for mw102 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [02:08:03] PROBLEM - mw102 MediaWiki Rendering on mw102 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [02:08:14] PROBLEM - cp31 Stunnel HTTP for mw121 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [02:08:21] PROBLEM - cp31 Stunnel HTTP for mw102 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [02:08:26] PROBLEM - mw121 MediaWiki Rendering on mw121 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [02:08:38] PROBLEM - cp20 Stunnel HTTP for mw102 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [02:09:32] RECOVERY - cp20 Stunnel HTTP for mw111 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 7.410 second response time [02:09:48] RECOVERY - gluster111 Current Load on gluster111 is OK: OK - load average: 2.13, 2.44, 2.49 [02:09:54] RECOVERY - cp21 Stunnel HTTP for mw102 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 0.504 second response time [02:09:54] RECOVERY - cp30 Stunnel HTTP for mw102 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 1.157 second response time [02:09:55] RECOVERY - cp30 Stunnel HTTP for mw121 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14548 bytes in 9.943 second response time [02:09:58] RECOVERY - mw102 MediaWiki Rendering on mw102 is OK: HTTP OK: HTTP/1.1 200 OK - 22334 bytes in 0.775 second response time [02:10:12] RECOVERY - mw101 MediaWiki Rendering on mw101 is OK: HTTP OK: HTTP/1.1 200 OK - 22334 bytes in 7.884 second response time [02:10:15] RECOVERY - cp31 Stunnel HTTP for mw121 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 4.320 second response time [02:10:18] RECOVERY - cp31 Stunnel HTTP for mw102 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 2.101 second response time [02:10:26] RECOVERY - mw121 MediaWiki Rendering on mw121 is OK: HTTP OK: HTTP/1.1 200 OK - 22334 bytes in 4.875 second response time [02:10:26] RECOVERY - cp20 Stunnel HTTP for mw101 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 1.662 second response time [02:10:33] RECOVERY - cp30 Stunnel HTTP for mw111 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 0.337 second response time [02:10:38] RECOVERY - cp20 Stunnel HTTP for mw102 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 0.847 second response time [02:10:52] PROBLEM - cp20 Stunnel HTTP for mw112 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [02:11:03] RECOVERY - cp31 Stunnel HTTP for mw111 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 0.350 second response time [02:11:06] RECOVERY - cp31 Stunnel HTTP for mw101 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 0.312 second response time [02:11:13] RECOVERY - cp21 Stunnel HTTP for mw121 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 1.418 second response time [02:11:22] RECOVERY - cp20 Stunnel HTTP for mw121 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 0.322 second response time [02:11:24] PROBLEM - cp21 Stunnel HTTP for mw112 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [02:11:24] PROBLEM - cp30 Stunnel HTTP for mw112 on cp30 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [02:11:42] RECOVERY - cp30 Stunnel HTTP for mw101 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 5.008 second response time [02:11:43] RECOVERY - cp21 Stunnel HTTP for mw101 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 4.421 second response time [02:11:56] PROBLEM - mw112 MediaWiki Rendering on mw112 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [02:12:02] RECOVERY - cp30 Stunnel HTTP for mw122 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 6.969 second response time [02:12:15] RECOVERY - cp31 Stunnel HTTP for mw122 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 3.440 second response time [02:12:18] RECOVERY - cp20 Stunnel HTTP for mw122 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 3.695 second response time [02:12:54] RECOVERY - cp20 Stunnel HTTP for mw112 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 3.577 second response time [02:12:58] RECOVERY - mw122 MediaWiki Rendering on mw122 is OK: HTTP OK: HTTP/1.1 200 OK - 22334 bytes in 1.364 second response time [02:13:06] RECOVERY - cp21 Stunnel HTTP for mw122 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 0.040 second response time [02:14:04] RECOVERY - mw112 MediaWiki Rendering on mw112 is OK: HTTP OK: HTTP/1.1 200 OK - 22334 bytes in 9.776 second response time [02:14:40] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 5.78, 4.17, 3.22 [02:14:44] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 7.57, 7.16, 6.72 [02:15:32] RECOVERY - cp21 Stunnel HTTP for mw112 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 9.265 second response time [02:15:34] RECOVERY - cp30 Stunnel HTTP for mw112 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14548 bytes in 9.542 second response time [02:16:50] PROBLEM - gluster121 Current Load on gluster121 is WARNING: WARNING - load average: 3.45, 3.28, 2.78 [02:21:48] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 4.40, 4.01, 3.32 [02:22:50] PROBLEM - gluster121 Current Load on gluster121 is CRITICAL: CRITICAL - load average: 4.38, 3.81, 3.17 [02:26:20] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 9.48, 7.85, 7.21 [02:29:48] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.42, 3.87, 3.59 [02:30:50] PROBLEM - gluster121 Current Load on gluster121 is WARNING: WARNING - load average: 3.26, 3.99, 3.62 [02:32:24] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.66, 4.00, 4.00 [02:32:50] PROBLEM - gluster121 Current Load on gluster121 is CRITICAL: CRITICAL - load average: 4.32, 3.94, 3.63 [02:33:08] PROBLEM - cp21 Stunnel HTTP for mw101 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [02:33:19] PROBLEM - mw101 MediaWiki Rendering on mw101 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [02:34:50] PROBLEM - gluster121 Current Load on gluster121 is WARNING: WARNING - load average: 2.67, 3.53, 3.52 [02:35:12] RECOVERY - cp21 Stunnel HTTP for mw101 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14548 bytes in 6.854 second response time [02:35:20] RECOVERY - mw101 MediaWiki Rendering on mw101 is OK: HTTP OK: HTTP/1.1 200 OK - 22334 bytes in 6.821 second response time [02:35:48] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 6.12, 4.08, 3.67 [02:36:20] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 6.29, 4.85, 4.30 [02:36:50] PROBLEM - gluster121 Current Load on gluster121 is CRITICAL: CRITICAL - load average: 4.04, 3.82, 3.63 [02:36:51] PROBLEM - mw101 Current Load on mw101 is WARNING: WARNING - load average: 10.96, 10.02, 8.71 [02:38:11] [02dns] 07MacFan4000 opened pull request 03#264: Remove 2 domains - 13https://github.com/miraheze/dns/pull/264 [02:38:50] RECOVERY - mw101 Current Load on mw101 is OK: OK - load average: 8.00, 9.30, 8.62 [02:39:47] PROBLEM - cp31 Current Load on cp31 is CRITICAL: CRITICAL - load average: 1.37, 2.56, 1.62 [02:41:47] PROBLEM - cp31 Current Load on cp31 is WARNING: WARNING - load average: 0.40, 1.80, 1.45 [02:43:47] RECOVERY - cp31 Current Load on cp31 is OK: OK - load average: 1.62, 1.70, 1.45 [02:47:48] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 2.62, 3.79, 3.92 [02:48:50] PROBLEM - gluster121 Current Load on gluster121 is WARNING: WARNING - load average: 2.89, 3.67, 3.81 [02:50:30] PROBLEM - hypotheticalhurricanes.com - LetsEncrypt on sslhost is CRITICAL: connect to address hypotheticalhurricanes.com and port 443: Network is unreachableHTTP CRITICAL - Unable to open TCP socket [02:50:50] PROBLEM - gluster121 Current Load on gluster121 is CRITICAL: CRITICAL - load average: 4.45, 4.14, 3.97 [02:53:48] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 4.44, 3.67, 3.82 [02:54:50] PROBLEM - gluster121 Current Load on gluster121 is WARNING: WARNING - load average: 3.77, 3.99, 3.97 [02:55:48] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 2.99, 3.39, 3.70 [02:57:48] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 4.20, 3.72, 3.78 [03:00:50] PROBLEM - gluster121 Current Load on gluster121 is CRITICAL: CRITICAL - load average: 5.01, 4.05, 3.96 [03:01:48] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.93, 3.61, 3.70 [03:02:50] PROBLEM - gluster121 Current Load on gluster121 is WARNING: WARNING - load average: 2.46, 3.53, 3.79 [03:03:54] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 2.19, 3.25, 3.92 [03:05:48] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 4.05, 3.26, 3.49 [03:07:48] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.30, 3.13, 3.41 [03:08:50] RECOVERY - gluster121 Current Load on gluster121 is OK: OK - load average: 2.43, 2.68, 3.30 [03:09:32] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 7.20, 7.60, 7.99 [03:09:48] RECOVERY - gluster111 Current Load on gluster111 is OK: OK - load average: 3.13, 3.04, 3.33 [03:13:46] RECOVERY - gluster101 Current Load on gluster101 is OK: OK - load average: 1.99, 2.60, 3.35 [03:21:32] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 9.23, 7.46, 7.38 [03:27:32] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 7.32, 7.45, 7.42 [03:33:32] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 8.44, 7.78, 7.55 [03:39:32] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 6.16, 7.46, 7.56 [03:39:46] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.41, 3.36, 3.08 [03:40:34] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.62, 3.04, 2.83 [03:41:32] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 8.15, 7.86, 7.70 [03:42:28] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 4.52, 3.42, 2.99 [03:43:32] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 7.51, 7.67, 7.65 [03:43:46] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 4.02, 3.57, 3.21 [03:44:22] RECOVERY - gluster111 Current Load on gluster111 is OK: OK - load average: 2.28, 3.00, 2.89 [03:45:46] RECOVERY - gluster101 Current Load on gluster101 is OK: OK - load average: 2.57, 3.22, 3.12 [03:47:32] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 9.21, 8.14, 7.82 [03:55:32] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 6.91, 7.82, 7.87 [03:59:48] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.21, 3.44, 3.15 [04:00:32] alerting : [FIRING:1] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [04:01:48] RECOVERY - gluster111 Current Load on gluster111 is OK: OK - load average: 2.29, 3.12, 3.07 [04:03:32] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 8.19, 7.42, 7.56 [04:05:32] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 6.93, 7.34, 7.53 [04:07:48] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 4.35, 3.63, 3.25 [04:07:55] PROBLEM - mw101 MediaWiki Rendering on mw101 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [04:08:01] PROBLEM - cp30 Stunnel HTTP for mw101 on cp30 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [04:08:19] PROBLEM - cp21 Stunnel HTTP for mw101 on cp21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [04:08:33] PROBLEM - cp20 Stunnel HTTP for mw101 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [04:08:36] PROBLEM - cp31 Stunnel HTTP for mw101 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [04:10:00] RECOVERY - mw101 MediaWiki Rendering on mw101 is OK: HTTP OK: HTTP/1.1 200 OK - 22334 bytes in 9.628 second response time [04:10:03] RECOVERY - cp30 Stunnel HTTP for mw101 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 7.381 second response time [04:10:24] RECOVERY - cp21 Stunnel HTTP for mw101 on cp21 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 8.134 second response time [04:11:48] RECOVERY - gluster111 Current Load on gluster111 is OK: OK - load average: 2.83, 3.36, 3.24 [04:12:44] RECOVERY - cp20 Stunnel HTTP for mw101 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14548 bytes in 2.579 second response time [04:12:46] RECOVERY - cp31 Stunnel HTTP for mw101 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 1.950 second response time [04:16:16] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.53, 3.39, 3.18 [04:17:31] PROBLEM - gluster101 Puppet on gluster101 is CRITICAL: CRITICAL: Failed to apply catalog, zero resources tracked by Puppet. It might be a dependency cycle. [04:17:32] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 9.21, 7.64, 7.40 [04:18:14] RECOVERY - gluster101 Current Load on gluster101 is OK: OK - load average: 1.74, 2.72, 2.96 [04:19:32] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 6.88, 7.19, 7.26 [04:21:48] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.55, 3.44, 3.29 [04:23:48] RECOVERY - gluster111 Current Load on gluster111 is OK: OK - load average: 2.63, 3.20, 3.22 [04:25:05] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.39, 3.67, 3.37 [04:27:03] RECOVERY - gluster101 Current Load on gluster101 is OK: OK - load average: 1.73, 3.01, 3.16 [04:33:54] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 4.38, 3.77, 3.43 [04:35:52] RECOVERY - gluster101 Current Load on gluster101 is OK: OK - load average: 2.39, 3.30, 3.30 [04:43:07] PROBLEM - cloud12 Puppet on cloud12 is CRITICAL: CRITICAL: Puppet has 1 failures. Last run 2 minutes ago with 1 failures. Failed resources (up to 3 shown): Service[ulogd2] [04:43:31] PROBLEM - cp20 Current Load on cp20 is WARNING: WARNING - load average: 0.64, 1.86, 1.13 [04:43:32] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 8.01, 7.71, 7.60 [04:45:31] RECOVERY - cp20 Current Load on cp20 is OK: OK - load average: 0.31, 1.33, 1.02 [04:45:32] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 6.68, 7.44, 7.52 [04:45:32] RECOVERY - gluster101 Puppet on gluster101 is OK: OK: Puppet is currently enabled, last run 1 minute ago with 0 failures [04:51:32] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 9.28, 7.93, 7.66 [04:52:35] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.66, 3.57, 3.41 [04:53:32] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 7.34, 7.69, 7.60 [04:54:33] RECOVERY - gluster101 Current Load on gluster101 is OK: OK - load average: 2.29, 3.20, 3.30 [04:55:32] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 8.75, 7.90, 7.68 [04:57:32] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 7.24, 7.64, 7.61 [05:00:58] PROBLEM - cp31 Current Load on cp31 is WARNING: WARNING - load average: 1.97, 1.70, 1.13 [05:02:57] PROBLEM - cp31 Current Load on cp31 is CRITICAL: CRITICAL - load average: 2.69, 2.19, 1.38 [05:03:32] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 8.17, 8.08, 7.80 [05:04:55] RECOVERY - cp31 Current Load on cp31 is OK: OK - load average: 0.64, 1.60, 1.26 [05:11:07] RECOVERY - cloud12 Puppet on cloud12 is OK: OK: Puppet is currently enabled, last run 1 minute ago with 0 failures [05:11:32] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 7.31, 7.95, 7.92 [05:13:32] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 9.46, 8.27, 8.02 [05:14:47] PROBLEM - cp31 Current Load on cp31 is WARNING: WARNING - load average: 1.26, 1.74, 1.44 [05:15:57] PROBLEM - test101 Disk Space on test101 is CRITICAL: DISK CRITICAL - free space: / 1057 MB (5% inode=56%); [05:16:09] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.53, 3.41, 3.16 [05:16:46] RECOVERY - cp31 Current Load on cp31 is OK: OK - load average: 0.78, 1.36, 1.33 [05:18:07] RECOVERY - gluster101 Current Load on gluster101 is OK: OK - load average: 2.93, 3.11, 3.07 [05:20:32] ok : [RESOLVED] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [05:24:58] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 4.01, 3.78, 3.38 [05:26:56] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.56, 3.80, 3.44 [05:33:41] PROBLEM - mw101 Current Load on mw101 is WARNING: WARNING - load average: 11.36, 9.92, 8.85 [05:34:48] RECOVERY - gluster101 Current Load on gluster101 is OK: OK - load average: 1.98, 2.92, 3.21 [05:35:40] RECOVERY - mw101 Current Load on mw101 is OK: OK - load average: 9.92, 9.78, 8.93 [05:36:31] alerting : [FIRING:1] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [05:47:32] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 6.48, 7.19, 7.93 [05:49:32] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 9.76, 8.31, 8.25 [06:01:32] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 7.75, 7.47, 7.97 [06:05:32] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 8.93, 7.86, 7.96 [06:08:04] PROBLEM - test101 Current Load on test101 is WARNING: WARNING - load average: 1.95, 1.58, 1.26 [06:10:03] PROBLEM - test101 Current Load on test101 is CRITICAL: CRITICAL - load average: 2.12, 1.77, 1.37 [06:11:32] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 7.76, 7.88, 7.98 [06:25:56] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.11, 3.43, 3.08 [06:25:57] [02ssl] 07Universal-Omega opened pull request 03#502: T9054: Remove hypotheticalhurricanes.com - 13https://github.com/miraheze/ssl/pull/502 [06:26:24] [02ssl] 07Universal-Omega synchronize pull request 03#502: T9054: Remove hypotheticalhurricanes.com - 13https://github.com/miraheze/ssl/pull/502 [06:27:50] [02dns] 07Universal-Omega opened pull request 03#265: T9054: Remove hypotheticalhurricanes.com zone - 13https://github.com/miraheze/dns/pull/265 [06:27:54] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 4.13, 3.81, 3.27 [06:29:52] RECOVERY - gluster101 Current Load on gluster101 is OK: OK - load average: 2.17, 3.30, 3.16 [06:39:32] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 8.29, 7.49, 7.27 [06:41:32] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 7.68, 7.59, 7.33 [06:43:32] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 8.74, 7.93, 7.48 [06:47:32] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 7.45, 7.81, 7.57 [06:51:32] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 8.98, 8.39, 7.85 [06:57:32] PROBLEM - cp20 Current Load on cp20 is WARNING: WARNING - load average: 1.78, 1.18, 0.77 [06:59:31] RECOVERY - cp20 Current Load on cp20 is OK: OK - load average: 1.06, 1.12, 0.80 [07:13:07] PROBLEM - cloud12 Puppet on cloud12 is CRITICAL: CRITICAL: Puppet has 1 failures. Last run 3 minutes ago with 1 failures. Failed resources (up to 3 shown): Service[ulogd2] [07:13:56] PROBLEM - db111 Current Load on db111 is CRITICAL: CRITICAL - load average: 17.61, 11.21, 6.75 [07:17:54] PROBLEM - mw101 Current Load on mw101 is WARNING: WARNING - load average: 10.84, 9.96, 9.15 [07:19:53] RECOVERY - mw101 Current Load on mw101 is OK: OK - load average: 8.36, 9.28, 9.00 [07:19:56] PROBLEM - db111 Current Load on db111 is WARNING: WARNING - load average: 4.79, 7.15, 6.39 [07:21:56] RECOVERY - db111 Current Load on db111 is OK: OK - load average: 3.73, 5.92, 6.03 [07:22:58] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.73, 3.00, 2.58 [07:24:56] RECOVERY - gluster101 Current Load on gluster101 is OK: OK - load average: 3.03, 3.04, 2.65 [07:25:56] PROBLEM - db111 Current Load on db111 is WARNING: WARNING - load average: 6.37, 7.42, 6.71 [07:27:56] PROBLEM - db111 Current Load on db111 is CRITICAL: CRITICAL - load average: 11.72, 9.24, 7.45 [07:31:56] PROBLEM - db111 Current Load on db111 is WARNING: WARNING - load average: 3.67, 6.89, 6.95 [07:33:56] RECOVERY - db111 Current Load on db111 is OK: OK - load average: 4.56, 6.06, 6.63 [07:37:35] PROBLEM - test101 Current Load on test101 is WARNING: WARNING - load average: 1.33, 1.67, 1.98 [07:39:07] RECOVERY - cloud12 Puppet on cloud12 is OK: OK: Puppet is currently enabled, last run 36 seconds ago with 0 failures [07:39:56] PROBLEM - db111 Current Load on db111 is CRITICAL: CRITICAL - load average: 10.27, 8.72, 7.47 [07:43:33] RECOVERY - test101 Current Load on test101 is OK: OK - load average: 1.11, 1.24, 1.67 [07:43:38] [02CreateWiki] 07lens0021 commented on pull request 03#315: Inject CreateWikiHookRunner - 13https://github.com/miraheze/CreateWiki/pull/315#issuecomment-1093781165 [07:45:56] PROBLEM - db111 Current Load on db111 is WARNING: WARNING - load average: 3.17, 6.31, 7.01 [07:46:31] ok : [RESOLVED] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [07:47:56] PROBLEM - db111 Current Load on db111 is CRITICAL: CRITICAL - load average: 38.61, 18.25, 11.19 [07:55:32] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 6.08, 7.10, 7.92 [08:01:56] PROBLEM - db111 Current Load on db111 is WARNING: WARNING - load average: 2.11, 5.25, 7.58 [08:03:33] alerting : [FIRING:1] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [08:04:14] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.96, 3.56, 3.03 [08:05:56] PROBLEM - db111 Current Load on db111 is CRITICAL: CRITICAL - load average: 21.44, 9.61, 8.46 [08:06:12] RECOVERY - gluster101 Current Load on gluster101 is OK: OK - load average: 2.43, 3.04, 2.90 [08:07:56] PROBLEM - db111 Current Load on db111 is WARNING: WARNING - load average: 5.60, 7.43, 7.79 [08:11:32] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 9.27, 7.88, 7.74 [08:13:32] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 7.07, 7.78, 7.74 [08:13:56] PROBLEM - db111 Current Load on db111 is CRITICAL: CRITICAL - load average: 11.11, 6.39, 6.98 [08:15:56] PROBLEM - db111 Current Load on db111 is WARNING: WARNING - load average: 5.92, 6.67, 7.07 [08:17:56] RECOVERY - db111 Current Load on db111 is OK: OK - load average: 4.51, 5.82, 6.70 [08:18:35] ok : [RESOLVED] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [08:30:32] alerting : [FIRING:1] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [08:31:32] RECOVERY - db101 Current Load on db101 is OK: OK - load average: 5.21, 6.05, 6.70 [08:32:31] PROBLEM - db111 Current Load on db111 is WARNING: WARNING - load average: 7.33, 6.29, 6.19 [08:34:27] RECOVERY - db111 Current Load on db111 is OK: OK - load average: 5.41, 5.79, 6.00 [08:35:32] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 8.36, 7.29, 7.06 [08:37:32] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 7.70, 7.29, 7.07 [08:39:32] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 9.43, 7.85, 7.29 [08:45:05] PROBLEM - db111 Current Load on db111 is CRITICAL: CRITICAL - load average: 7.63, 9.46, 7.64 [08:47:32] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 6.40, 7.75, 7.57 [08:49:32] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 8.81, 8.04, 7.69 [08:50:32] ok : [RESOLVED] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [08:50:51] PROBLEM - db111 Current Load on db111 is WARNING: WARNING - load average: 2.75, 6.22, 6.85 [08:52:47] PROBLEM - db111 Current Load on db111 is CRITICAL: CRITICAL - load average: 16.04, 9.34, 7.85 [08:53:32] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 7.74, 7.88, 7.72 [08:56:38] PROBLEM - db111 Current Load on db111 is WARNING: WARNING - load average: 5.31, 6.90, 7.19 [09:02:25] RECOVERY - db111 Current Load on db111 is OK: OK - load average: 5.89, 6.18, 6.78 [09:08:02] alerting : [FIRING:1] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [09:13:32] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 8.29, 7.47, 7.41 [09:15:32] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 5.93, 6.99, 7.24 [09:18:02] ok : [RESOLVED] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [09:23:04] alerting : [FIRING:1] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [09:29:32] RECOVERY - db101 Current Load on db101 is OK: OK - load average: 4.70, 5.88, 6.67 [09:38:04] ok : [RESOLVED] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [09:40:08] PROBLEM - db111 Current Load on db111 is WARNING: WARNING - load average: 4.80, 7.10, 6.68 [09:42:03] RECOVERY - db111 Current Load on db111 is OK: OK - load average: 3.34, 5.82, 6.26 [09:43:32] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 7.47, 6.82, 6.59 [09:45:32] alerting : [FIRING:1] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [09:49:20] [02miraheze/puppet] 07JohnFLewis pushed 031 commit to 03master [+0/-0/±1] 13https://github.com/miraheze/puppet/compare/554521acee88...885eb3d63a5e [09:49:21] [02miraheze/puppet] 07JohnFLewis 03885eb3d - absent db backups, high IO usage [09:50:32] ok : [RESOLVED] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [09:55:32] RECOVERY - db101 Current Load on db101 is OK: OK - load average: 5.18, 6.50, 6.67 [09:56:01] alerting : [FIRING:1] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [10:05:56] PROBLEM - db111 Current Load on db111 is CRITICAL: CRITICAL - load average: 8.87, 11.94, 8.42 [10:16:01] ok : [RESOLVED] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [10:20:52] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 7.54, 7.29, 6.71 [10:26:40] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 9.45, 7.76, 7.03 [10:37:56] PROBLEM - db111 Current Load on db111 is WARNING: WARNING - load average: 3.71, 5.04, 7.73 [10:41:56] RECOVERY - db111 Current Load on db111 is OK: OK - load average: 2.86, 4.01, 6.71 [10:46:04] .in 5mins . [10:46:04] RhinosF1: Okay, will remind at 2022-04-09 - 11:51:04BST [10:51:05] RhinosF1: . [10:53:45] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 6.62, 7.62, 7.93 [10:57:37] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 8.22, 8.10, 8.06 [10:59:33] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 6.78, 7.59, 7.87 [11:02:01] alerting : [FIRING:1] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [11:03:32] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 9.35, 8.19, 8.03 [11:05:32] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 7.70, 7.99, 7.97 [11:11:32] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 9.00, 7.78, 7.81 [11:12:01] ok : [RESOLVED] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [11:13:32] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 7.53, 7.45, 7.67 [11:23:17] PROBLEM - cp30 Current Load on cp30 is WARNING: WARNING - load average: 1.58, 1.77, 1.05 [11:25:17] RECOVERY - cp30 Current Load on cp30 is OK: OK - load average: 0.90, 1.45, 1.02 [11:27:31] alerting : [FIRING:1] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [11:34:28] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.51, 3.20, 2.89 [11:35:29] PROBLEM - gluster121 Current Load on gluster121 is WARNING: WARNING - load average: 3.50, 3.23, 2.78 [11:35:32] RECOVERY - db101 Current Load on db101 is OK: OK - load average: 5.98, 6.27, 6.78 [11:37:25] RECOVERY - gluster121 Current Load on gluster121 is OK: OK - load average: 2.13, 2.87, 2.70 [11:38:16] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 4.08, 3.70, 3.16 [11:39:32] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 7.36, 6.65, 6.80 [11:40:09] RECOVERY - gluster111 Current Load on gluster111 is OK: OK - load average: 2.46, 3.27, 3.07 [11:41:32] RECOVERY - db101 Current Load on db101 is OK: OK - load average: 6.45, 6.56, 6.75 [11:45:32] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 7.07, 6.98, 6.88 [11:49:32] RECOVERY - db101 Current Load on db101 is OK: OK - load average: 5.93, 6.61, 6.78 [11:57:32] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 8.36, 7.36, 6.96 [11:59:32] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 6.87, 7.10, 6.91 [11:59:50] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 4.41, 3.33, 2.97 [12:01:48] RECOVERY - gluster101 Current Load on gluster101 is OK: OK - load average: 2.85, 3.17, 2.97 [12:07:31] ok : [RESOLVED] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [12:13:21] PROBLEM - dnd.bellinrattin.it - reverse DNS on sslhost is CRITICAL: rDNS CRITICAL - dnd.bellinrattin.it All nameservers failed to answer the query. [12:13:32] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 9.37, 8.30, 7.47 [12:18:31] alerting : [FIRING:1] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [12:23:32] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 5.83, 7.67, 7.76 [12:23:37] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 4.13, 3.42, 3.11 [12:25:01] PROBLEM - cp21 Current Load on cp21 is CRITICAL: CRITICAL - load average: 14.13, 8.62, 4.16 [12:25:35] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.94, 3.69, 3.26 [12:27:34] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 5.67, 4.43, 3.59 [12:28:31] ok : [RESOLVED] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [12:29:48] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 4.04, 3.67, 3.19 [12:30:11] PROBLEM - cp20 Current Load on cp20 is CRITICAL: CRITICAL - load average: 13.02, 7.59, 3.65 [12:31:32] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 9.03, 8.02, 7.84 [12:31:48] RECOVERY - gluster111 Current Load on gluster111 is OK: OK - load average: 2.33, 3.29, 3.12 [12:33:28] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.57, 3.93, 3.66 [12:38:01] alerting : [FIRING:1] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [12:39:22] RECOVERY - gluster101 Current Load on gluster101 is OK: OK - load average: 2.29, 3.05, 3.37 [12:43:01] PROBLEM - cp21 Current Load on cp21 is WARNING: WARNING - load average: 0.66, 1.06, 1.99 [12:43:20] PROBLEM - dnd.bellinrattin.it - reverse DNS on sslhost is WARNING: Traceback (most recent call last): File "/usr/lib/nagios/plugins/check_reverse_dns.py", line 155, in main() File "/usr/lib/nagios/plugins/check_reverse_dns.py", line 136, in main records = check_records(args.hostname) File "/usr/lib/nagios/plugins/check_reverse_dns.py", line 73, in check_records nameserversans = dns_resolver.resolve(root_domain, 'NS') File "/usr/l [12:43:20] n3/dist-packages/dns/resolver.py", line 1040, in resolve (nameserver, port, tcp, backoff) = resolution.next_nameserver() File "/usr/lib/python3/dist-packages/dns/resolver.py", line 598, in next_nameserver raise NoNameservers(request=self.request, errors=self.errors)dns.resolver.NoNameservers: All nameservers failed to answer the query bellinrattin.it. IN NS: Server 2606:4700:4700::1111 UDP port 53 answered SERVFAIL [12:45:34] PROBLEM - cp20 Current Load on cp20 is WARNING: WARNING - load average: 0.62, 1.10, 1.95 [12:47:01] RECOVERY - cp21 Current Load on cp21 is OK: OK - load average: 0.63, 0.79, 1.65 [12:49:11] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.42, 3.36, 3.34 [12:49:31] RECOVERY - cp20 Current Load on cp20 is OK: OK - load average: 1.00, 0.91, 1.67 [12:49:48] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.61, 3.41, 3.07 [12:51:09] RECOVERY - gluster101 Current Load on gluster101 is OK: OK - load average: 2.64, 3.18, 3.28 [12:51:48] RECOVERY - gluster111 Current Load on gluster111 is OK: OK - load average: 2.13, 3.07, 2.99 [12:53:32] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 6.26, 7.48, 7.85 [12:58:01] ok : [RESOLVED] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [12:58:50] PROBLEM - gluster121 Current Load on gluster121 is WARNING: WARNING - load average: 3.61, 3.41, 3.00 [12:59:48] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 5.23, 3.99, 3.33 [12:59:58] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.70, 3.51, 3.38 [13:00:19] PROBLEM - mw101 Current Load on mw101 is WARNING: WARNING - load average: 10.26, 9.59, 9.03 [13:01:48] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.52, 3.79, 3.34 [13:02:18] RECOVERY - mw101 Current Load on mw101 is OK: OK - load average: 10.19, 9.71, 9.14 [13:03:31] alerting : [FIRING:1] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [13:03:48] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 4.46, 3.99, 3.46 [13:03:54] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 5.54, 4.45, 3.77 [13:05:48] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.22, 3.65, 3.40 [13:07:48] RECOVERY - gluster111 Current Load on gluster111 is OK: OK - load average: 2.83, 3.36, 3.32 [13:08:31] ok : [RESOLVED] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [13:09:49] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 2.55, 3.64, 3.69 [13:11:48] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.13, 3.51, 3.41 [13:12:15] PROBLEM - docse.tk - LetsEncrypt on sslhost is WARNING: WARNING - Certificate 'docse.tk' expires in 15 day(s) (Mon 25 Apr 2022 12:59:53 GMT +0000). [13:12:50] RECOVERY - gluster121 Current Load on gluster121 is OK: OK - load average: 2.28, 3.22, 3.28 [13:13:19] PROBLEM - dnd.bellinrattin.it - reverse DNS on sslhost is CRITICAL: rDNS CRITICAL - dnd.bellinrattin.it All nameservers failed to answer the query. [13:13:46] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 4.04, 3.76, 3.72 [13:15:46] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 2.92, 3.43, 3.61 [13:15:48] RECOVERY - gluster111 Current Load on gluster111 is OK: OK - load average: 2.99, 3.40, 3.40 [13:17:46] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 5.69, 4.00, 3.77 [13:19:12] [02miraheze/ssl] 07MirahezeSSLBot pushed 031 commit to 03master [+0/-0/±1] 13https://github.com/miraheze/ssl/compare/a6fa8f4a1ba4...48656f1705c7 [13:19:13] [02miraheze/ssl] 07MirahezeSSLBot 0348656f1 - Bot: Update SSL cert for docse.tk [13:19:32] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 8.21, 7.37, 7.26 [13:19:46] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 2.54, 3.42, 3.59 [13:19:48] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 2.73, 3.54, 3.48 [13:21:32] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 7.11, 7.21, 7.21 [13:21:48] RECOVERY - gluster111 Current Load on gluster111 is OK: OK - load average: 2.30, 3.20, 3.37 [13:23:01] alerting : [FIRING:1] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [13:24:18] PROBLEM - id.altilunium.xyz - LetsEncrypt on sslhost is WARNING: WARNING - Certificate 'id.altilunium.xyz' expires in 15 day(s) (Mon 25 Apr 2022 13:00:35 GMT +0000). [13:25:46] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 4.79, 3.56, 3.54 [13:27:23] [02ssl] 07MacFan4000 closed pull request 03#502: T9054: Remove hypotheticalhurricanes.com - 13https://github.com/miraheze/ssl/pull/502 [13:27:24] [02miraheze/ssl] 07MacFan4000 pushed 031 commit to 03master [+0/-1/±1] 13https://github.com/miraheze/ssl/compare/48656f1705c7...27d3b534ec1c [13:27:26] [02miraheze/ssl] 07Universal-Omega 0327d3b53 - T9054: Remove hypotheticalhurricanes.com (#502) [13:27:46] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 2.82, 3.43, 3.51 [13:31:04] PROBLEM - mw101 Current Load on mw101 is CRITICAL: CRITICAL - load average: 12.12, 10.59, 9.87 [13:31:46] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 4.80, 3.84, 3.64 [13:33:03] PROBLEM - mw101 Current Load on mw101 is WARNING: WARNING - load average: 10.59, 10.50, 9.92 [13:33:46] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.75, 3.74, 3.62 [13:34:22] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 4.82, 3.87, 3.48 [13:35:02] RECOVERY - mw101 Current Load on mw101 is OK: OK - load average: 8.69, 9.78, 9.73 [13:36:16] RECOVERY - gluster111 Current Load on gluster111 is OK: OK - load average: 2.13, 3.23, 3.29 [13:36:47] PROBLEM - notes.mridulpm.in - LetsEncrypt on sslhost is WARNING: WARNING - Certificate 'notes.mridulpm.in' expires in 15 day(s) (Mon 25 Apr 2022 13:23:29 GMT +0000). [13:37:32] RECOVERY - db101 Current Load on db101 is OK: OK - load average: 5.72, 6.12, 6.65 [13:38:04] [02miraheze/ssl] 07MirahezeSSLBot pushed 031 commit to 03master [+0/-0/±1] 13https://github.com/miraheze/ssl/compare/27d3b534ec1c...81c6105097de [13:38:05] [02miraheze/ssl] 07MirahezeSSLBot 0381c6105 - Bot: Update SSL cert for notes.mridulpm.in [13:43:46] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 4.62, 3.71, 3.58 [13:47:48] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 5.22, 3.95, 3.48 [13:49:32] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 6.78, 7.20, 7.00 [13:51:32] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 9.04, 8.03, 7.33 [13:52:54] PROBLEM - gluster121 Current Load on gluster121 is CRITICAL: CRITICAL - load average: 4.62, 3.81, 3.43 [13:53:32] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 6.28, 7.43, 7.20 [13:53:48] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.61, 3.75, 3.53 [13:55:30] PROBLEM - wiki.alathramc.com - LetsEncrypt on sslhost is WARNING: WARNING - Certificate 'wiki.alathramc.com' expires in 15 day(s) (Mon 25 Apr 2022 13:31:38 GMT +0000). [13:55:32] PROBLEM - cp20 Current Load on cp20 is CRITICAL: CRITICAL - load average: 2.49, 1.54, 1.08 [13:55:48] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 4.09, 3.81, 3.58 [13:56:50] RECOVERY - gluster121 Current Load on gluster121 is OK: OK - load average: 2.43, 3.37, 3.35 [13:56:52] [02miraheze/ssl] 07MirahezeSSLBot pushed 031 commit to 03master [+0/-0/±1] 13https://github.com/miraheze/ssl/compare/81c6105097de...8f08b0066669 [13:56:54] [02miraheze/ssl] 07MirahezeSSLBot 038f08b00 - Bot: Update SSL cert for wiki.alathramc.com [13:57:32] RECOVERY - cp20 Current Load on cp20 is OK: OK - load average: 1.23, 1.33, 1.05 [13:57:46] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.07, 3.69, 3.88 [14:00:52] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.58, 3.64, 3.57 [14:00:53] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 9.11, 8.15, 7.58 [14:01:11] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 4.54, 3.99, 3.94 [14:02:21] PROBLEM - vrcdev.wiki - LetsEncrypt on sslhost is WARNING: WARNING - Certificate 'vrcdev.wiki' expires in 15 day(s) (Mon 25 Apr 2022 13:39:05 GMT +0000). [14:02:39] PROBLEM - gluster121 Current Load on gluster121 is WARNING: WARNING - load average: 3.08, 3.39, 3.42 [14:05:03] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 2.92, 3.83, 3.91 [14:05:30] [02miraheze/ssl] 07MirahezeSSLBot pushed 031 commit to 03master [+0/-0/±1] 13https://github.com/miraheze/ssl/compare/8f08b0066669...5b07e720aa0c [14:05:31] [02miraheze/ssl] 07MirahezeSSLBot 035b07e72 - Bot: Update SSL cert for vrcdev.wiki [14:05:35] RECOVERY - notes.mridulpm.in - LetsEncrypt on sslhost is OK: OK - Certificate 'notes.mridulpm.in' will expire on Fri 08 Jul 2022 12:37:58 GMT +0000. [14:06:42] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 4.46, 3.79, 3.64 [14:06:59] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 4.41, 4.22, 4.05 [14:08:39] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.00, 3.62, 3.60 [14:08:39] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 6.33, 7.88, 7.82 [14:11:43] RECOVERY - docse.tk - LetsEncrypt on sslhost is OK: OK - Certificate 'docse.tk' will expire on Fri 08 Jul 2022 12:19:04 GMT +0000. [14:12:27] RECOVERY - gluster121 Current Load on gluster121 is OK: OK - load average: 2.35, 3.08, 3.31 [14:12:32] RECOVERY - gluster111 Current Load on gluster111 is OK: OK - load average: 2.46, 3.07, 3.38 [14:12:48] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.49, 3.88, 3.97 [14:13:01] ok : [RESOLVED] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [14:13:18] RECOVERY - dnd.bellinrattin.it - reverse DNS on sslhost is OK: SSL OK - dnd.bellinrattin.it reverse DNS resolves to cp20.miraheze.org - CNAME OK [14:14:29] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 8.02, 7.76, 7.79 [14:14:44] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 4.69, 4.47, 4.19 [14:16:25] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 6.45, 7.43, 7.68 [14:17:52] PROBLEM - mw102 Current Load on mw102 is WARNING: WARNING - load average: 11.25, 9.79, 8.97 [14:21:31] alerting : [FIRING:1] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [14:27:04] PROBLEM - gluster121 Current Load on gluster121 is WARNING: WARNING - load average: 3.98, 3.48, 3.36 [14:27:28] PROBLEM - mw102 Current Load on mw102 is CRITICAL: CRITICAL - load average: 12.54, 11.54, 10.22 [14:28:57] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.25, 3.54, 3.44 [14:29:04] PROBLEM - gluster121 Current Load on gluster121 is CRITICAL: CRITICAL - load average: 4.37, 3.73, 3.46 [14:30:01] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 8.10, 7.16, 7.30 [14:31:04] PROBLEM - gluster121 Current Load on gluster121 is WARNING: WARNING - load average: 2.80, 3.57, 3.45 [14:31:09] RECOVERY - vrcdev.wiki - LetsEncrypt on sslhost is OK: OK - Certificate 'vrcdev.wiki' will expire on Fri 08 Jul 2022 13:05:23 GMT +0000. [14:31:58] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 7.40, 7.43, 7.40 [14:32:57] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 4.56, 4.13, 3.69 [14:33:04] PROBLEM - gluster121 Current Load on gluster121 is CRITICAL: CRITICAL - load average: 4.83, 4.00, 3.61 [14:33:14] PROBLEM - mw102 Current Load on mw102 is WARNING: WARNING - load average: 10.88, 11.47, 10.66 [14:34:57] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.47, 3.82, 3.62 [14:36:57] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 5.64, 4.61, 3.94 [14:39:04] PROBLEM - gluster121 Current Load on gluster121 is WARNING: WARNING - load average: 2.56, 3.64, 3.62 [14:41:51] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 9.20, 7.97, 7.54 [14:42:57] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.31, 3.82, 3.83 [14:43:04] RECOVERY - gluster121 Current Load on gluster121 is OK: OK - load average: 2.57, 3.02, 3.38 [14:43:51] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 7.57, 7.63, 7.46 [14:48:38] RECOVERY - mw102 Current Load on mw102 is OK: OK - load average: 7.16, 8.93, 10.03 [14:48:41] PROBLEM - mw101 Current Load on mw101 is WARNING: WARNING - load average: 10.33, 9.90, 9.40 [14:49:51] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 8.93, 7.73, 7.48 [14:51:51] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 6.99, 7.48, 7.42 [14:52:38] RECOVERY - mw101 Current Load on mw101 is OK: OK - load average: 9.55, 9.94, 9.54 [14:54:47] RECOVERY - wiki.alathramc.com - LetsEncrypt on sslhost is OK: OK - Certificate 'wiki.alathramc.com' will expire on Fri 08 Jul 2022 12:56:47 GMT +0000. [14:59:32] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 2.21, 3.18, 3.77 [15:00:30] PROBLEM - mw101 Current Load on mw101 is CRITICAL: CRITICAL - load average: 12.32, 10.76, 9.98 [15:01:10] PROBLEM - cp30 Stunnel HTTP for mw101 on cp30 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [15:01:29] PROBLEM - cp20 Stunnel HTTP for mw101 on cp20 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [15:01:35] PROBLEM - mw101 MediaWiki Rendering on mw101 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [15:02:28] PROBLEM - mw101 Current Load on mw101 is WARNING: WARNING - load average: 11.10, 10.92, 10.14 [15:02:57] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 4.60, 3.49, 3.57 [15:03:10] RECOVERY - cp30 Stunnel HTTP for mw101 on cp30 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 2.907 second response time [15:03:24] RECOVERY - cp20 Stunnel HTTP for mw101 on cp20 is OK: HTTP OK: HTTP/1.1 200 OK - 14562 bytes in 1.080 second response time [15:03:31] RECOVERY - mw101 MediaWiki Rendering on mw101 is OK: HTTP OK: HTTP/1.1 200 OK - 22334 bytes in 1.327 second response time [15:03:32] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 4.19, 3.63, 3.79 [15:06:57] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.20, 3.33, 3.48 [15:08:57] RECOVERY - gluster111 Current Load on gluster111 is OK: OK - load average: 2.23, 2.82, 3.26 [15:09:32] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 2.96, 3.53, 3.75 [15:09:51] RECOVERY - db101 Current Load on db101 is OK: OK - load average: 5.59, 6.33, 6.80 [15:10:22] RECOVERY - mw101 Current Load on mw101 is OK: OK - load average: 7.70, 9.91, 10.09 [15:13:51] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 7.47, 7.19, 7.07 [15:15:24] PROBLEM - cp21 Current Load on cp21 is WARNING: WARNING - load average: 0.68, 1.83, 1.48 [15:16:31] ok : [RESOLVED] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [15:17:23] RECOVERY - cp21 Current Load on cp21 is OK: OK - load average: 0.43, 1.34, 1.34 [15:17:51] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 8.93, 7.82, 7.33 [15:19:51] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 5.86, 7.27, 7.20 [15:21:32] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 4.17, 3.76, 3.73 [15:23:32] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.15, 3.53, 3.65 [15:25:32] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 5.13, 4.17, 3.86 [15:27:26] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 4.05, 3.78, 3.34 [15:29:03] alerting : [FIRING:1] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [15:31:06] PROBLEM - gluster121 Current Load on gluster121 is WARNING: WARNING - load average: 3.07, 3.52, 3.16 [15:33:04] RECOVERY - gluster121 Current Load on gluster121 is OK: OK - load average: 2.35, 3.30, 3.14 [15:33:16] RECOVERY - gluster111 Current Load on gluster111 is OK: OK - load average: 2.02, 3.27, 3.31 [15:33:32] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 2.86, 3.64, 3.78 [15:35:51] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 8.11, 7.86, 7.50 [15:39:32] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 4.11, 3.50, 3.62 [15:39:52] PROBLEM - mw101 Current Load on mw101 is WARNING: WARNING - load average: 11.27, 10.55, 10.08 [15:41:32] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 2.70, 3.22, 3.50 [15:41:50] RECOVERY - mw101 Current Load on mw101 is OK: OK - load average: 8.69, 10.15, 10.01 [15:43:32] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 5.58, 3.77, 3.65 [15:49:32] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 2.99, 3.65, 3.68 [15:50:57] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.60, 3.58, 3.30 [15:52:57] RECOVERY - gluster111 Current Load on gluster111 is OK: OK - load average: 2.77, 3.29, 3.23 [15:57:32] RECOVERY - gluster101 Current Load on gluster101 is OK: OK - load average: 2.67, 2.93, 3.31 [16:00:32] PROBLEM - mw101 Current Load on mw101 is WARNING: WARNING - load average: 11.39, 10.86, 10.31 [16:01:09] [02miraheze/puppet] 07paladox pushed 031 commit to 03paladox-patch-8 [+0/-0/±1] 13https://github.com/miraheze/puppet/commit/a678ce781e9b [16:01:11] [02miraheze/puppet] 07paladox 03a678ce7 - Switch cp20/21/30/31 to rsyslog [16:01:12] [02puppet] 07paladox created branch 03paladox-patch-8 - 13https://github.com/miraheze/puppet [16:01:14] [02puppet] 07paladox opened pull request 03#2478: Switch cp20/21/30/31 to rsyslog - 13https://github.com/miraheze/puppet/pull/2478 [16:01:32] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 6.13, 4.31, 3.76 [16:01:34] [02miraheze/puppet] 07paladox pushed 031 commit to 03paladox-patch-8 [+0/-0/±1] 13https://github.com/miraheze/puppet/compare/a678ce781e9b...8582ef17471a [16:01:36] [02miraheze/puppet] 07paladox 038582ef1 - Update cp21.yaml [16:01:37] [02puppet] 07paladox synchronize pull request 03#2478: Switch cp20/21/30/31 to rsyslog - 13https://github.com/miraheze/puppet/pull/2478 [16:01:42] [02miraheze/puppet] 07paladox pushed 031 commit to 03paladox-patch-8 [+0/-0/±1] 13https://github.com/miraheze/puppet/compare/8582ef17471a...1c4d0f4f684e [16:01:44] [02miraheze/puppet] 07paladox 031c4d0f4 - Update cp30.yaml [16:01:45] [02puppet] 07paladox synchronize pull request 03#2478: Switch cp20/21/30/31 to rsyslog - 13https://github.com/miraheze/puppet/pull/2478 [16:01:50] [02miraheze/puppet] 07paladox pushed 031 commit to 03paladox-patch-8 [+0/-0/±1] 13https://github.com/miraheze/puppet/compare/1c4d0f4f684e...d295b68f5139 [16:01:51] [02miraheze/puppet] 07paladox 03d295b68 - Update cp31.yaml [16:01:53] [02puppet] 07paladox synchronize pull request 03#2478: Switch cp20/21/30/31 to rsyslog - 13https://github.com/miraheze/puppet/pull/2478 [16:01:58] [02puppet] 07paladox closed pull request 03#2478: Switch cp20/21/30/31 to rsyslog - 13https://github.com/miraheze/puppet/pull/2478 [16:02:00] [02miraheze/puppet] 07paladox pushed 031 commit to 03master [+0/-0/±4] 13https://github.com/miraheze/puppet/compare/885eb3d63a5e...2cd5705c4374 [16:02:01] [02miraheze/puppet] 07paladox 032cd5705 - Switch cp20/21/30/31 to rsyslog (#2478) [16:02:03] [02puppet] 07paladox deleted branch 03paladox-patch-8 - 13https://github.com/miraheze/puppet [16:02:04] [02miraheze/puppet] 07paladox deleted branch 03paladox-patch-8 [16:03:36] [02miraheze/puppet] 07paladox pushed 031 commit to 03paladox-patch-8 [+0/-0/±1] 13https://github.com/miraheze/puppet/commit/6c837dc6d617 [16:03:38] [02miraheze/puppet] 07paladox 036c837dc - Switch mw* to rsyslog [16:03:39] [02puppet] 07paladox created branch 03paladox-patch-8 - 13https://github.com/miraheze/puppet [16:03:41] [02puppet] 07paladox opened pull request 03#2479: Switch mw* to rsyslog - 13https://github.com/miraheze/puppet/pull/2479 [16:03:54] [02miraheze/puppet] 07paladox pushed 031 commit to 03paladox-patch-8 [+0/-0/±1] 13https://github.com/miraheze/puppet/compare/6c837dc6d617...97cfb00233e8 [16:03:56] [02miraheze/puppet] 07paladox 0397cfb00 - Update mw102.yaml [16:03:57] [02puppet] 07paladox synchronize pull request 03#2479: Switch mw* to rsyslog - 13https://github.com/miraheze/puppet/pull/2479 [16:04:03] [02miraheze/puppet] 07paladox pushed 031 commit to 03paladox-patch-8 [+0/-0/±1] 13https://github.com/miraheze/puppet/compare/97cfb00233e8...a9a9a456ef6b [16:04:05] [02miraheze/puppet] 07paladox 03a9a9a45 - Update mw111.yaml [16:04:06] [02puppet] 07paladox synchronize pull request 03#2479: Switch mw* to rsyslog - 13https://github.com/miraheze/puppet/pull/2479 [16:04:14] [02miraheze/puppet] 07paladox pushed 031 commit to 03paladox-patch-8 [+0/-0/±1] 13https://github.com/miraheze/puppet/compare/a9a9a456ef6b...9acdf0b4720f [16:04:16] [02miraheze/puppet] 07paladox 039acdf0b - Update mw112.yaml [16:04:17] [02puppet] 07paladox synchronize pull request 03#2479: Switch mw* to rsyslog - 13https://github.com/miraheze/puppet/pull/2479 [16:04:24] [02miraheze/puppet] 07paladox pushed 031 commit to 03paladox-patch-8 [+0/-0/±1] 13https://github.com/miraheze/puppet/compare/9acdf0b4720f...cbdbe7b5d1d5 [16:04:25] [02miraheze/puppet] 07paladox 03cbdbe7b - Update mw121.yaml [16:04:27] [02puppet] 07paladox synchronize pull request 03#2479: Switch mw* to rsyslog - 13https://github.com/miraheze/puppet/pull/2479 [16:04:29] RECOVERY - mw101 Current Load on mw101 is OK: OK - load average: 7.07, 9.53, 9.96 [16:04:36] [02miraheze/puppet] 07paladox pushed 031 commit to 03paladox-patch-8 [+0/-0/±1] 13https://github.com/miraheze/puppet/compare/cbdbe7b5d1d5...f0f9b1f39f31 [16:04:37] [02miraheze/puppet] 07paladox 03f0f9b1f - Update mw122.yaml [16:04:39] [02puppet] 07paladox synchronize pull request 03#2479: Switch mw* to rsyslog - 13https://github.com/miraheze/puppet/pull/2479 [16:04:46] [02miraheze/puppet] 07paladox pushed 031 commit to 03paladox-patch-8 [+0/-0/±1] 13https://github.com/miraheze/puppet/compare/f0f9b1f39f31...fb9d7685fd43 [16:04:48] [02miraheze/puppet] 07paladox 03fb9d768 - Update mwtask111.yaml [16:04:49] [02puppet] 07paladox synchronize pull request 03#2479: Switch mw* to rsyslog - 13https://github.com/miraheze/puppet/pull/2479 [16:05:32] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.45, 3.93, 3.74 [16:06:15] [02puppet] 07paladox closed pull request 03#2479: Switch mw* to rsyslog - 13https://github.com/miraheze/puppet/pull/2479 [16:06:16] [02puppet] 07paladox deleted branch 03paladox-patch-8 - 13https://github.com/miraheze/puppet [16:06:18] [02miraheze/puppet] 07paladox deleted branch 03paladox-patch-8 [16:06:19] [02miraheze/puppet] 07paladox pushed 031 commit to 03master [+0/-0/±7] 13https://github.com/miraheze/puppet/compare/2cd5705c4374...f31696754849 [16:06:21] [02miraheze/puppet] 07paladox 03f316967 - Switch mw* to rsyslog (#2479) [16:08:58] [02miraheze/mw-config] 07paladox pushed 031 commit to 03master [+0/-0/±1] 13https://github.com/miraheze/mw-config/compare/e6712ae2a615...170e2b370f66 [16:08:59] [02miraheze/mw-config] 07paladox 03170e2b3 - Set wmgSyslogHandler to rsyslog for all [16:09:24] !log [@mwtask111] starting deploy of {'config': True} to all [16:09:26] !log [paladox@mwtask111] starting deploy of {'pull': 'config', 'config': True} to all [16:09:31] PROBLEM - mw112 Current Load on mw112 is WARNING: WARNING - load average: 11.26, 10.12, 9.12 [16:09:33] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 4.15, 3.78, 3.69 [16:09:37] !log [paladox@mwtask111] finished deploy of {'pull': 'config', 'config': True} to all - SUCCESS in 11s [16:09:38] !log [@mwtask111] finished deploy of {'config': True} to all - SUCCESS in 13s [16:09:42] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [16:09:53] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [16:10:00] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [16:10:05] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [16:10:09] miraheze/mw-config - paladox the build passed. [16:11:27] RECOVERY - mw112 Current Load on mw112 is OK: OK - load average: 9.02, 9.53, 9.01 [16:11:30] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 4.69, 3.49, 3.22 [16:12:11] PROBLEM - cloud10 Puppet on cloud10 is CRITICAL: CRITICAL: Puppet has 1 failures. Last run 3 minutes ago with 1 failures. Failed resources (up to 3 shown): Service[ulogd2] [16:13:27] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.63, 3.56, 3.28 [16:13:45] PROBLEM - gluster121 Current Load on gluster121 is WARNING: WARNING - load average: 3.75, 3.52, 3.22 [16:17:20] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 4.01, 3.90, 3.48 [16:19:03] ok : [RESOLVED] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [16:19:17] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.95, 3.81, 3.49 [16:19:38] PROBLEM - gluster121 Current Load on gluster121 is CRITICAL: CRITICAL - load average: 4.85, 4.03, 3.51 [16:21:36] RECOVERY - gluster121 Current Load on gluster121 is OK: OK - load average: 2.17, 3.28, 3.30 [16:23:09] PROBLEM - mw101 Current Load on mw101 is WARNING: WARNING - load average: 9.97, 10.22, 10.10 [16:24:31] alerting : [FIRING:1] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [16:25:06] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 5.41, 4.11, 3.66 [16:25:07] RECOVERY - mw101 Current Load on mw101 is OK: OK - load average: 8.57, 9.82, 9.99 [16:27:09] !log puppet111: upgrade puppet-agent puppetdb puppetdb-termini puppetserver [16:27:18] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [16:27:27] PROBLEM - gluster121 Current Load on gluster121 is CRITICAL: CRITICAL - load average: 4.43, 3.79, 3.50 [16:29:00] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 2.77, 3.37, 3.45 [16:29:32] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.02, 3.70, 3.95 [16:30:21] paladox: feel free to reset config on test101 if you need to use beta for anything that your doing. If you do you may need to reset databases though by removing /srv/mediawiki/cache/*.json to fix the 404 error. [16:30:57] RECOVERY - gluster111 Current Load on gluster111 is OK: OK - load average: 2.77, 3.21, 3.39 [16:31:01] PROBLEM - mw101 Current Load on mw101 is WARNING: WARNING - load average: 11.53, 10.37, 10.09 [16:31:22] RECOVERY - gluster121 Current Load on gluster121 is OK: OK - load average: 2.66, 3.34, 3.39 [16:34:58] RECOVERY - mw101 Current Load on mw101 is OK: OK - load average: 7.74, 9.40, 9.80 [16:35:32] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 4.81, 3.85, 3.85 [16:35:53] PROBLEM - mw122 Puppet on mw122 is CRITICAL: CRITICAL: Failed to apply catalog, zero resources tracked by Puppet. It might be a dependency cycle. [16:36:02] PROBLEM - bast101 Puppet on bast101 is CRITICAL: CRITICAL: Failed to apply catalog, zero resources tracked by Puppet. It might be a dependency cycle. [16:36:40] ok [16:36:57] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.00, 3.47, 3.43 [16:37:32] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.63, 3.80, 3.83 [16:38:04] PROBLEM - mw111 Puppet on mw111 is CRITICAL: CRITICAL: Puppet has 743 failures. Last run 2 minutes ago with 743 failures. Failed resources (up to 3 shown): File[/etc/ssl/localcerts/wiki.meeusen.net.crt],File[/etc/ssl/private/wiki.meeusen.net.key],File[/etc/ssl/localcerts/speleo.wiki.crt],File[/etc/ssl/private/speleo.wiki.key] [16:39:32] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 4.48, 4.00, 3.89 [16:40:11] RECOVERY - cloud10 Puppet on cloud10 is OK: OK: Puppet is currently enabled, last run 33 seconds ago with 0 failures [16:40:57] RECOVERY - gluster111 Current Load on gluster111 is OK: OK - load average: 2.90, 3.24, 3.34 [16:41:08] PROBLEM - mw112 Current Load on mw112 is WARNING: WARNING - load average: 11.22, 9.69, 9.03 [16:41:32] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.07, 3.76, 3.82 [16:43:06] PROBLEM - gluster121 Current Load on gluster121 is WARNING: WARNING - load average: 3.21, 3.41, 3.29 [16:43:08] RECOVERY - mw112 Current Load on mw112 is OK: OK - load average: 8.81, 9.37, 9.00 [16:43:32] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 4.07, 4.01, 3.91 [16:44:48] PROBLEM - mw101 Current Load on mw101 is WARNING: WARNING - load average: 10.32, 10.20, 9.95 [16:45:04] RECOVERY - gluster121 Current Load on gluster121 is OK: OK - load average: 3.07, 3.18, 3.21 [16:45:32] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.53, 3.93, 3.89 [16:46:26] PROBLEM - dnd.bellinrattin.it - reverse DNS on sslhost is CRITICAL: rDNS CRITICAL - dnd.bellinrattin.it All nameservers failed to answer the query. [16:46:46] RECOVERY - mw101 Current Load on mw101 is OK: OK - load average: 9.87, 9.82, 9.83 [16:50:41] PROBLEM - mw101 Current Load on mw101 is WARNING: WARNING - load average: 10.50, 10.54, 10.14 [16:51:08] PROBLEM - mw112 Current Load on mw112 is WARNING: WARNING - load average: 10.48, 9.90, 9.33 [16:51:32] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 5.84, 4.51, 4.08 [16:51:51] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 7.40, 7.61, 7.94 [16:53:08] RECOVERY - mw112 Current Load on mw112 is OK: OK - load average: 9.84, 9.93, 9.42 [16:53:51] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 8.44, 7.73, 7.93 [16:55:04] PROBLEM - gluster121 Current Load on gluster121 is CRITICAL: CRITICAL - load average: 4.39, 3.74, 3.40 [16:55:30] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 5.86, 4.19, 3.60 [16:55:51] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 7.70, 7.53, 7.82 [16:57:26] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.72, 3.97, 3.59 [16:58:35] RECOVERY - mw101 Current Load on mw101 is OK: OK - load average: 9.44, 9.77, 10.02 [16:59:04] PROBLEM - gluster121 Current Load on gluster121 is WARNING: WARNING - load average: 3.12, 3.48, 3.37 [16:59:23] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 4.18, 3.87, 3.58 [17:01:04] RECOVERY - gluster121 Current Load on gluster121 is OK: OK - load average: 3.36, 3.27, 3.29 [17:01:20] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 2.82, 3.39, 3.43 [17:01:51] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 8.05, 7.77, 7.83 [17:02:30] PROBLEM - mw101 Current Load on mw101 is WARNING: WARNING - load average: 9.85, 10.25, 10.18 [17:04:02] RECOVERY - bast101 Puppet on bast101 is OK: OK: Puppet is currently enabled, last run 56 seconds ago with 0 failures [17:05:32] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.05, 3.74, 3.99 [17:05:51] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 7.41, 7.58, 7.74 [17:05:53] RECOVERY - mw122 Puppet on mw122 is OK: OK: Puppet is currently enabled, last run 1 minute ago with 0 failures [17:06:26] RECOVERY - mw101 Current Load on mw101 is OK: OK - load average: 9.19, 9.88, 10.06 [17:06:39] RECOVERY - mw111 Puppet on mw111 is OK: OK: Puppet is currently enabled, last run 1 minute ago with 0 failures [17:07:32] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 4.98, 4.02, 4.05 [17:09:51] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 8.02, 7.49, 7.64 [17:11:32] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 1.89, 3.27, 3.77 [17:11:51] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 7.50, 7.37, 7.57 [17:12:38] PROBLEM - mw102 Current Load on mw102 is WARNING: WARNING - load average: 10.50, 10.38, 9.53 [17:12:59] RECOVERY - gluster111 Current Load on gluster111 is OK: OK - load average: 2.61, 3.11, 3.36 [17:13:51] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 9.69, 8.08, 7.80 [17:14:38] RECOVERY - mw102 Current Load on mw102 is OK: OK - load average: 7.56, 9.26, 9.22 [17:16:06] PROBLEM - dnd.bellinrattin.it - reverse DNS on sslhost is WARNING: Traceback (most recent call last): File "/usr/lib/nagios/plugins/check_reverse_dns.py", line 155, in main() File "/usr/lib/nagios/plugins/check_reverse_dns.py", line 136, in main records = check_records(args.hostname) File "/usr/lib/nagios/plugins/check_reverse_dns.py", line 86, in check_records cname = str(dns_resolver.resolve(hostname, 'CNAME')[0]) File "/usr/li [17:16:06] 3/dist-packages/dns/resolver.py", line 1040, in resolve (nameserver, port, tcp, backoff) = resolution.next_nameserver() File "/usr/lib/python3/dist-packages/dns/resolver.py", line 598, in next_nameserver raise NoNameservers(request=self.request, errors=self.errors)dns.resolver.NoNameservers: All nameservers failed to answer the query dnd.bellinrattin.it. IN CNAME: Server 2606:4700:4700::1111 UDP port 53 answered SERVFAIL [17:17:32] RECOVERY - gluster101 Current Load on gluster101 is OK: OK - load average: 2.22, 2.75, 3.39 [17:21:12] PROBLEM - mw101 Current Load on mw101 is CRITICAL: CRITICAL - load average: 12.54, 10.97, 10.27 [17:23:10] PROBLEM - mw101 Current Load on mw101 is WARNING: WARNING - load average: 9.94, 10.59, 10.22 [17:24:31] ok : [RESOLVED] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [17:31:03] PROBLEM - mw101 Current Load on mw101 is CRITICAL: CRITICAL - load average: 12.67, 11.05, 10.47 [17:32:01] alerting : [FIRING:1] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [17:33:02] PROBLEM - mw101 Current Load on mw101 is WARNING: WARNING - load average: 9.91, 10.75, 10.44 [17:35:00] RECOVERY - mw101 Current Load on mw101 is OK: OK - load average: 8.06, 9.61, 10.05 [17:37:01] ok : [RESOLVED] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [17:40:53] PROBLEM - mw101 Current Load on mw101 is WARNING: WARNING - load average: 10.56, 10.27, 10.20 [17:41:06] PROBLEM - mw102 Current Load on mw102 is WARNING: WARNING - load average: 11.20, 9.92, 9.33 [17:42:31] alerting : [FIRING:1] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [17:42:52] RECOVERY - mw101 Current Load on mw101 is OK: OK - load average: 9.09, 9.95, 10.10 [17:43:02] RECOVERY - mw102 Current Load on mw102 is OK: OK - load average: 8.28, 9.24, 9.15 [17:45:46] RECOVERY - dnd.bellinrattin.it - reverse DNS on sslhost is OK: SSL OK - dnd.bellinrattin.it reverse DNS resolves to cp20.miraheze.org - CNAME OK [17:46:06] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.90, 3.83, 3.36 [17:48:02] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 7.10, 4.72, 3.72 [17:50:23] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.18, 3.53, 3.21 [17:53:51] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 5.37, 6.80, 7.80 [17:54:16] RECOVERY - gluster111 Current Load on gluster111 is OK: OK - load average: 2.75, 3.31, 3.22 [17:55:46] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 2.73, 3.56, 3.62 [17:59:38] RECOVERY - gluster101 Current Load on gluster101 is OK: OK - load average: 2.80, 3.00, 3.36 [18:02:32] PROBLEM - mw101 Current Load on mw101 is WARNING: WARNING - load average: 11.47, 10.93, 10.24 [18:04:30] RECOVERY - mw101 Current Load on mw101 is OK: OK - load average: 6.14, 9.27, 9.73 [18:06:17] PROBLEM - cp31 Stunnel HTTP for mw101 on cp31 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [18:07:04] PROBLEM - mw101 MediaWiki Rendering on mw101 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [18:08:18] RECOVERY - cp31 Stunnel HTTP for mw101 on cp31 is OK: HTTP OK: HTTP/1.1 200 OK - 14556 bytes in 1.056 second response time [18:09:01] RECOVERY - mw101 MediaWiki Rendering on mw101 is OK: HTTP OK: HTTP/1.1 200 OK - 22334 bytes in 2.264 second response time [18:09:04] PROBLEM - test101 Current Load on test101 is CRITICAL: CRITICAL - load average: 2.17, 1.67, 1.28 [18:09:51] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 11.44, 8.79, 7.89 [18:13:51] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 7.44, 8.00, 7.76 [18:14:57] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 4.12, 3.43, 3.08 [18:15:13] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 5.26, 4.25, 3.67 [18:15:38] PROBLEM - gluster121 Current Load on gluster121 is WARNING: WARNING - load average: 3.63, 3.54, 3.15 [18:16:57] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.27, 3.51, 3.17 [18:17:22] PROBLEM - dnd.bellinrattin.it - reverse DNS on sslhost is WARNING: Traceback (most recent call last): File "/usr/lib/nagios/plugins/check_reverse_dns.py", line 155, in main() File "/usr/lib/nagios/plugins/check_reverse_dns.py", line 136, in main records = check_records(args.hostname) File "/usr/lib/nagios/plugins/check_reverse_dns.py", line 73, in check_records nameserversans = dns_resolver.resolve(root_domain, 'NS') File "/usr/l [18:17:22] n3/dist-packages/dns/resolver.py", line 1040, in resolve (nameserver, port, tcp, backoff) = resolution.next_nameserver() File "/usr/lib/python3/dist-packages/dns/resolver.py", line 598, in next_nameserver raise NoNameservers(request=self.request, errors=self.errors)dns.resolver.NoNameservers: All nameservers failed to answer the query bellinrattin.it. IN NS: Server 2606:4700:4700::1111 UDP port 53 answered SERVFAIL [18:17:36] RECOVERY - gluster121 Current Load on gluster121 is OK: OK - load average: 2.76, 3.28, 3.10 [18:18:57] RECOVERY - gluster111 Current Load on gluster111 is OK: OK - load average: 2.83, 3.36, 3.16 [18:24:54] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 2.59, 3.61, 3.66 [18:26:50] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 7.11, 4.48, 3.95 [18:28:57] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 4.71, 4.05, 3.52 [18:29:20] PROBLEM - gluster121 Current Load on gluster121 is WARNING: WARNING - load average: 3.43, 3.23, 3.08 [18:31:40] PROBLEM - cp31 Current Load on cp31 is CRITICAL: CRITICAL - load average: 2.65, 2.22, 1.33 [18:31:51] RECOVERY - db101 Current Load on db101 is OK: OK - load average: 6.14, 5.92, 6.68 [18:32:57] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.44, 3.78, 3.55 [18:33:40] PROBLEM - cp31 Current Load on cp31 is WARNING: WARNING - load average: 0.81, 1.73, 1.26 [18:34:34] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 2.97, 3.96, 3.97 [18:35:13] RECOVERY - gluster121 Current Load on gluster121 is OK: OK - load average: 2.48, 3.11, 3.12 [18:35:40] RECOVERY - cp31 Current Load on cp31 is OK: OK - load average: 0.71, 1.37, 1.18 [18:36:30] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 4.69, 4.48, 4.17 [18:36:57] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 4.38, 3.90, 3.63 [18:38:57] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.23, 3.73, 3.60 [18:40:23] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.73, 3.93, 3.99 [18:40:57] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 7.34, 4.76, 3.97 [18:42:19] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 5.24, 4.66, 4.26 [18:43:04] PROBLEM - gluster121 Current Load on gluster121 is WARNING: WARNING - load average: 2.91, 3.62, 3.39 [18:45:04] PROBLEM - gluster121 Current Load on gluster121 is CRITICAL: CRITICAL - load average: 4.76, 3.92, 3.51 [18:46:57] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.29, 3.85, 3.81 [18:47:04] PROBLEM - gluster121 Current Load on gluster121 is WARNING: WARNING - load average: 2.56, 3.46, 3.40 [18:49:04] RECOVERY - gluster121 Current Load on gluster121 is OK: OK - load average: 2.09, 3.08, 3.27 [18:49:08] PROBLEM - mw112 Current Load on mw112 is WARNING: WARNING - load average: 10.22, 9.66, 9.21 [18:50:57] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 4.77, 3.82, 3.77 [18:54:57] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.31, 3.36, 3.60 [18:55:08] RECOVERY - mw112 Current Load on mw112 is OK: OK - load average: 8.99, 9.63, 9.40 [18:55:52] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 2.91, 3.44, 3.87 [18:57:31] ok : [RESOLVED] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [18:58:57] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 4.41, 3.81, 3.71 [18:59:08] PROBLEM - mw112 Current Load on mw112 is WARNING: WARNING - load average: 10.52, 10.27, 9.71 [19:00:57] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 2.23, 3.36, 3.57 [19:01:08] RECOVERY - mw112 Current Load on mw112 is OK: OK - load average: 7.53, 9.30, 9.43 [19:02:57] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 4.16, 3.74, 3.68 [19:03:36] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 4.67, 3.99, 3.92 [19:04:57] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.72, 3.81, 3.71 [19:05:01] alerting : [FIRING:1] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [19:07:32] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 2.40, 3.41, 3.73 [19:08:57] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 4.69, 3.90, 3.75 [19:10:57] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.51, 3.74, 3.71 [19:11:32] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 4.45, 3.89, 3.85 [19:13:32] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.21, 3.78, 3.82 [19:14:57] RECOVERY - gluster111 Current Load on gluster111 is OK: OK - load average: 1.90, 2.86, 3.36 [19:15:01] ok : [RESOLVED] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [19:16:42] PROBLEM - dnd.bellinrattin.it - reverse DNS on sslhost is CRITICAL: rDNS CRITICAL - dnd.bellinrattin.it All nameservers failed to answer the query. [19:24:36] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 7.87, 6.90, 6.09 [19:26:32] PROBLEM - db101 Current Load on db101 is CRITICAL: CRITICAL - load average: 8.01, 7.36, 6.37 [19:27:32] RECOVERY - gluster101 Current Load on gluster101 is OK: OK - load average: 2.60, 2.97, 3.35 [19:28:29] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 6.06, 6.97, 6.34 [19:30:26] RECOVERY - db101 Current Load on db101 is OK: OK - load average: 5.25, 6.27, 6.15 [19:33:04] PROBLEM - test101 Current Load on test101 is WARNING: WARNING - load average: 1.26, 1.60, 2.00 [19:33:31] alerting : [FIRING:1] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [19:35:32] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.64, 3.50, 3.48 [19:37:32] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 4.03, 3.49, 3.46 [19:38:31] ok : [RESOLVED] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [19:39:04] RECOVERY - test101 Current Load on test101 is OK: OK - load average: 0.95, 1.16, 1.67 [19:39:32] RECOVERY - gluster101 Current Load on gluster101 is OK: OK - load average: 2.17, 2.90, 3.25 [19:39:40] PROBLEM - cp31 Current Load on cp31 is WARNING: WARNING - load average: 1.32, 1.73, 1.29 [19:41:40] RECOVERY - cp31 Current Load on cp31 is OK: OK - load average: 0.71, 1.36, 1.21 [19:46:21] RECOVERY - dnd.bellinrattin.it - reverse DNS on sslhost is OK: SSL OK - dnd.bellinrattin.it reverse DNS resolves to cp21.miraheze.org - CNAME OK [19:47:08] PROBLEM - cp20 Current Load on cp20 is CRITICAL: CRITICAL - load average: 7.22, 6.64, 3.22 [19:52:07] PROBLEM - cp30 Current Load on cp30 is WARNING: WARNING - load average: 1.73, 1.54, 1.16 [19:54:07] RECOVERY - cp30 Current Load on cp30 is OK: OK - load average: 0.92, 1.32, 1.13 [19:54:31] alerting : [FIRING:1] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [19:56:57] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 6.92, 5.04, 3.76 [19:57:06] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.73, 3.81, 3.45 [19:57:57] PROBLEM - gluster121 Current Load on gluster121 is WARNING: WARNING - load average: 2.99, 3.54, 3.20 [19:59:54] PROBLEM - gluster121 Current Load on gluster121 is CRITICAL: CRITICAL - load average: 4.48, 3.63, 3.26 [20:00:58] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 6.85, 4.57, 3.78 [20:03:08] PROBLEM - cp20 Current Load on cp20 is WARNING: WARNING - load average: 0.58, 1.37, 2.00 [20:07:08] RECOVERY - cp20 Current Load on cp20 is OK: OK - load average: 0.55, 0.98, 1.69 [20:11:41] PROBLEM - gluster121 Current Load on gluster121 is WARNING: WARNING - load average: 2.85, 3.74, 3.69 [20:12:57] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 2.83, 3.60, 3.90 [20:13:38] PROBLEM - gluster121 Current Load on gluster121 is CRITICAL: CRITICAL - load average: 4.04, 3.76, 3.69 [20:14:57] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 5.78, 4.69, 4.27 [20:19:31] PROBLEM - gluster121 Current Load on gluster121 is WARNING: WARNING - load average: 2.79, 3.82, 3.82 [20:27:22] PROBLEM - gluster121 Current Load on gluster121 is CRITICAL: CRITICAL - load average: 4.79, 3.67, 3.63 [20:28:57] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.67, 3.65, 4.00 [20:31:18] PROBLEM - gluster121 Current Load on gluster121 is WARNING: WARNING - load average: 2.75, 3.35, 3.53 [20:33:15] RECOVERY - gluster121 Current Load on gluster121 is OK: OK - load average: 1.75, 2.75, 3.29 [20:38:57] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 4.33, 3.35, 3.52 [20:41:04] PROBLEM - gluster121 Current Load on gluster121 is CRITICAL: CRITICAL - load average: 4.29, 3.72, 3.45 [20:41:37] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 2.69, 3.62, 4.00 [20:42:57] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.07, 3.38, 3.51 [20:43:04] RECOVERY - gluster121 Current Load on gluster121 is OK: OK - load average: 2.16, 3.17, 3.28 [20:46:57] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 4.38, 3.69, 3.59 [20:48:57] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.08, 3.35, 3.47 [20:49:31] ok : [RESOLVED] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [20:50:57] RECOVERY - gluster111 Current Load on gluster111 is OK: OK - load average: 2.88, 3.07, 3.34 [20:54:40] alerting : [FIRING:1] (!sre High Job Queue Backlog yes mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [20:59:37] ok : [RESOLVED] (!sre High Job Queue Backlog yes mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [21:01:32] RECOVERY - gluster101 Current Load on gluster101 is OK: OK - load average: 2.51, 2.70, 3.34 [21:01:51] PROBLEM - db101 Current Load on db101 is WARNING: WARNING - load average: 6.85, 6.29, 5.46 [21:02:01] alerting : [FIRING:1] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [21:03:51] RECOVERY - db101 Current Load on db101 is OK: OK - load average: 5.08, 6.14, 5.53 [21:05:32] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 2.98, 3.32, 3.49 [21:09:36] PROBLEM - franchise.franchising.org.ua - reverse DNS on sslhost is WARNING: Traceback (most recent call last): File "/usr/lib/nagios/plugins/check_reverse_dns.py", line 155, in main() File "/usr/lib/nagios/plugins/check_reverse_dns.py", line 136, in main records = check_records(args.hostname) File "/usr/lib/nagios/plugins/check_reverse_dns.py", line 73, in check_records nameserversans = dns_resolver.resolve(root_domain, 'NS') Fil [21:09:36] lib/python3/dist-packages/dns/resolver.py", line 1040, in resolve (nameserver, port, tcp, backoff) = resolution.next_nameserver() File "/usr/lib/python3/dist-packages/dns/resolver.py", line 598, in next_nameserver raise NoNameservers(request=self.request, errors=self.errors)dns.resolver.NoNameservers: All nameservers failed to answer the query franchising.org.ua. IN NS: Server 2606:4700:4700::1111 UDP port 53 answered SERVFAIL [21:11:32] RECOVERY - gluster101 Current Load on gluster101 is OK: OK - load average: 2.63, 3.13, 3.36 [21:15:32] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 6.00, 4.34, 3.78 [21:17:02] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.61, 3.59, 3.37 [21:22:01] ok : [RESOLVED] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [21:22:57] RECOVERY - gluster111 Current Load on gluster111 is OK: OK - load average: 3.12, 3.32, 3.34 [21:23:32] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 2.53, 3.53, 3.66 [21:27:32] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 4.12, 3.59, 3.65 [21:28:57] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.45, 3.38, 3.32 [21:30:57] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 5.58, 4.14, 3.60 [21:31:34] PROBLEM - gluster121 Current Load on gluster121 is CRITICAL: CRITICAL - load average: 4.02, 3.52, 3.20 [21:33:31] PROBLEM - gluster121 Current Load on gluster121 is WARNING: WARNING - load average: 3.75, 3.62, 3.28 [21:34:57] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.36, 3.80, 3.60 [21:35:29] RECOVERY - gluster121 Current Load on gluster121 is OK: OK - load average: 2.65, 3.27, 3.19 [21:35:32] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.08, 3.81, 3.79 [21:36:31] alerting : [FIRING:1] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [21:38:28] RECOVERY - franchise.franchising.org.ua - reverse DNS on sslhost is OK: SSL OK - franchise.franchising.org.ua reverse DNS resolves to cp21.miraheze.org - CNAME OK [21:38:57] RECOVERY - gluster111 Current Load on gluster111 is OK: OK - load average: 2.86, 3.21, 3.40 [21:39:32] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 4.05, 3.82, 3.78 [21:41:32] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.00, 3.71, 3.76 [21:45:13] PROBLEM - gluster121 Current Load on gluster121 is WARNING: WARNING - load average: 3.31, 3.41, 3.31 [21:47:10] RECOVERY - gluster121 Current Load on gluster121 is OK: OK - load average: 2.07, 2.85, 3.12 [21:47:55] PROBLEM - dnd.bellinrattin.it - reverse DNS on sslhost is WARNING: Traceback (most recent call last): File "/usr/lib/nagios/plugins/check_reverse_dns.py", line 155, in main() File "/usr/lib/nagios/plugins/check_reverse_dns.py", line 136, in main records = check_records(args.hostname) File "/usr/lib/nagios/plugins/check_reverse_dns.py", line 73, in check_records nameserversans = dns_resolver.resolve(root_domain, 'NS') File "/usr/l [21:47:55] n3/dist-packages/dns/resolver.py", line 1040, in resolve (nameserver, port, tcp, backoff) = resolution.next_nameserver() File "/usr/lib/python3/dist-packages/dns/resolver.py", line 598, in next_nameserver raise NoNameservers(request=self.request, errors=self.errors)dns.resolver.NoNameservers: All nameservers failed to answer the query bellinrattin.it. IN NS: Server 2606:4700:4700::1111 UDP port 53 answered SERVFAIL [21:51:31] ok : [RESOLVED] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [21:51:36] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 4.36, 3.39, 3.29 [21:53:32] RECOVERY - gluster101 Current Load on gluster101 is OK: OK - load average: 2.86, 2.99, 3.37 [21:53:33] RECOVERY - gluster111 Current Load on gluster111 is OK: OK - load average: 2.82, 3.17, 3.22 [21:59:19] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.95, 3.61, 3.38 [22:00:31] alerting : [FIRING:1] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [22:01:16] RECOVERY - gluster111 Current Load on gluster111 is OK: OK - load average: 2.62, 3.35, 3.32 [22:01:32] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 4.26, 3.83, 3.61 [22:03:32] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.70, 3.80, 3.63 [22:05:31] ok : [RESOLVED] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [22:07:32] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 4.11, 3.86, 3.66 [22:11:01] alerting : [FIRING:1] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [22:11:32] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.67, 3.75, 3.67 [22:11:57] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 4.38, 3.64, 3.43 [22:13:53] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.64, 3.60, 3.44 [22:15:32] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 5.32, 4.05, 3.77 [22:17:35] PROBLEM - dnd.bellinrattin.it - reverse DNS on sslhost is CRITICAL: rDNS CRITICAL - dnd.bellinrattin.it All nameservers failed to answer the query. [22:17:46] PROBLEM - gluster111 Current Load on gluster111 is CRITICAL: CRITICAL - load average: 4.66, 4.11, 3.67 [22:19:32] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.27, 3.97, 3.84 [22:19:43] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 2.69, 3.55, 3.51 [22:21:01] ok : [RESOLVED] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [22:23:36] RECOVERY - gluster111 Current Load on gluster111 is OK: OK - load average: 2.87, 3.25, 3.39 [22:25:32] RECOVERY - gluster101 Current Load on gluster101 is OK: OK - load average: 1.68, 2.82, 3.37 [22:28:06] alerting : [FIRING:1] (!sre High Job Queue Backlog yes mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [22:34:20] [02miraheze/landing] 07Universal-Omega pushed 031 commit to 03master [+0/-0/±1] 13https://github.com/miraheze/landing/compare/27da9cff965a...d4aed874c240 [22:34:22] [02miraheze/landing] 07Universal-Omega 03d4aed87 - Update continuousIntegration.yml [22:35:12] [02miraheze/ErrorPages] 07Universal-Omega pushed 031 commit to 03master [+0/-0/±1] 13https://github.com/miraheze/ErrorPages/compare/162b6c39c392...c7634bfb5e2b [22:35:13] [02miraheze/ErrorPages] 07Universal-Omega 03c7634bf - Update continuousIntegration.yml [22:35:14] miraheze/landing - Universal-Omega the build passed. [22:36:03] miraheze/ErrorPages - Universal-Omega the build passed. [22:39:04] PROBLEM - gluster121 Current Load on gluster121 is CRITICAL: CRITICAL - load average: 4.09, 3.23, 2.98 [22:41:04] RECOVERY - gluster121 Current Load on gluster121 is OK: OK - load average: 1.99, 2.77, 2.85 [22:43:06] alerting : [FIRING:1] (mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [22:47:15] RECOVERY - dnd.bellinrattin.it - reverse DNS on sslhost is OK: SSL OK - dnd.bellinrattin.it reverse DNS resolves to cp20.miraheze.org - CNAME OK [22:52:09] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.70, 3.46, 3.23 [22:53:06] ok : [RESOLVED] (!sre High Job Queue Backlog yes mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [22:54:05] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 4.26, 3.53, 3.27 [22:57:04] PROBLEM - gluster121 Current Load on gluster121 is WARNING: WARNING - load average: 3.82, 3.29, 2.91 [22:58:29] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.83, 3.88, 3.32 [23:00:01] !log [@mwtask111] starting deploy of {'l10nupdate': True} to all [23:00:02] !log [@test101] starting deploy of {'l10nupdate': True} to all [23:00:04] !log [@test101] DEPLOY ABORTED: Non-Zero Exit Code in prep, see output. [23:00:14] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [23:00:18] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [23:00:25] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [23:01:04] PROBLEM - gluster121 Current Load on gluster121 is CRITICAL: CRITICAL - load average: 5.09, 3.73, 3.14 [23:02:01] alerting : [FIRING:1] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [23:03:04] RECOVERY - gluster121 Current Load on gluster121 is OK: OK - load average: 2.42, 3.25, 3.04 [23:04:19] RECOVERY - gluster111 Current Load on gluster111 is OK: OK - load average: 2.45, 3.25, 3.25 [23:06:12] !log [@mwtask111] finished deploy of {'l10nupdate': True} to all - SUCCESS in 370s [23:06:22] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [23:07:38] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 2.81, 3.89, 3.82 [23:17:01] ok : [RESOLVED] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [23:17:32] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 4.26, 3.41, 3.51 [23:20:57] PROBLEM - gluster111 Current Load on gluster111 is WARNING: WARNING - load average: 3.22, 3.57, 3.33 [23:21:32] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 2.98, 3.53, 3.56 [23:22:57] RECOVERY - gluster111 Current Load on gluster111 is OK: OK - load average: 3.00, 3.39, 3.29 [23:25:32] RECOVERY - gluster101 Current Load on gluster101 is OK: OK - load average: 2.67, 3.11, 3.38 [23:26:01] alerting : [FIRING:1] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [23:36:01] ok : [RESOLVED] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [23:38:51] PROBLEM - cp30 Current Load on cp30 is WARNING: WARNING - load average: 1.63, 1.71, 1.35 [23:40:46] RECOVERY - cp30 Current Load on cp30 is OK: OK - load average: 0.82, 1.41, 1.27 [23:44:27] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 4.85, 3.57, 3.21 [23:48:19] PROBLEM - gluster101 Current Load on gluster101 is WARNING: WARNING - load average: 3.12, 3.58, 3.31 [23:49:27] PROBLEM - dnd.bellinrattin.it - reverse DNS on sslhost is WARNING: Traceback (most recent call last): File "/usr/lib/nagios/plugins/check_reverse_dns.py", line 155, in main() File "/usr/lib/nagios/plugins/check_reverse_dns.py", line 136, in main records = check_records(args.hostname) File "/usr/lib/nagios/plugins/check_reverse_dns.py", line 86, in check_records cname = str(dns_resolver.resolve(hostname, 'CNAME')[0]) File "/usr/li [23:49:27] 3/dist-packages/dns/resolver.py", line 1040, in resolve (nameserver, port, tcp, backoff) = resolution.next_nameserver() File "/usr/lib/python3/dist-packages/dns/resolver.py", line 598, in next_nameserver raise NoNameservers(request=self.request, errors=self.errors)dns.resolver.NoNameservers: All nameservers failed to answer the query dnd.bellinrattin.it. IN CNAME: Server 2606:4700:4700::1111 UDP port 53 answered SERVFAIL [23:50:15] RECOVERY - gluster101 Current Load on gluster101 is OK: OK - load average: 2.25, 3.24, 3.23 [23:51:31] alerting : [FIRING:1] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [23:56:05] PROBLEM - gluster101 Current Load on gluster101 is CRITICAL: CRITICAL - load average: 4.70, 4.26, 3.66 [23:56:31] ok : [RESOLVED] (PHP-FPM Worker Usage High mediawiki) https://grafana.miraheze.org/d/dsHv5-4nz/mediawiki [23:57:04] PROBLEM - gluster121 Current Load on gluster121 is WARNING: WARNING - load average: 3.66, 3.95, 3.43 [23:59:04] PROBLEM - gluster121 Current Load on gluster121 is CRITICAL: CRITICAL - load average: 4.95, 4.39, 3.65