[00:00:34] PROBLEM - newcascadia.net - reverse DNS on sslhost is WARNING: LifetimeTimeout: The resolution lifetime expired after 5.404 seconds: Server 1.1.1.1 UDP port 53 answered The DNS operation timed out.; Server 1.1.1.1 UDP port 53 answered The DNS operation timed out.; Server 1.1.1.1 UDP port 53 answered The DNS operation timed out. [00:01:15] PROBLEM - cp2 HTTP 4xx/5xx ERROR Rate on cp2 is UNKNOWN: UNKNOWN - NGINX Error Rate is UNKNOWN [00:06:15] RECOVERY - cp3 HTTP 4xx/5xx ERROR Rate on cp3 is OK: OK - NGINX Error Rate is 38% [00:10:12] PROBLEM - cp3 HTTP 4xx/5xx ERROR Rate on cp3 is WARNING: WARNING - NGINX Error Rate is 53% [00:11:36] PROBLEM - cp2 HTTPS on cp2 is CRITICAL: HTTP CRITICAL - Invalid HTTP response received from host on port 443: cURL returned 7 - Failed to connect to cp2.wikitide.net port 443 after 31 ms: Couldn't connect to server [00:11:39] PROBLEM - cp2 Disk Space on cp2 is CRITICAL: CHECK_NRPE: Error - Could not connect to 51.79.55.151: Connection reset by peer [00:11:40] PROBLEM - cp2 conntrack_table_size on cp2 is CRITICAL: CHECK_NRPE: Error - Could not connect to 51.79.55.151: Connection reset by peer [00:11:45] PROBLEM - cp2 Varnish Backends on cp2 is CRITICAL: CHECK_NRPE: Error - Could not connect to 51.79.55.151: Connection reset by peer [00:11:45] PROBLEM - cp2 PowerDNS Recursor on cp2 is CRITICAL: CHECK_NRPE: Error - Could not connect to 51.79.55.151: Connection reset by peer [00:11:48] PROBLEM - cp2 Nginx Backend for matomo1 on cp2 is CRITICAL: CHECK_NRPE: Error - Could not connect to 51.79.55.151: Connection reset by peer [00:11:49] PROBLEM - cp2 ferm_active on cp2 is CRITICAL: CHECK_NRPE: Error - Could not connect to 51.79.55.151: Connection reset by peer [00:11:50] PROBLEM - cp2 Nginx Backend for mw2 on cp2 is CRITICAL: CHECK_NRPE: Error - Could not connect to 51.79.55.151: Connection reset by peer [00:11:52] PROBLEM - cp2 Nginx Backend for puppet1 on cp2 is CRITICAL: CHECK_NRPE: Error - Could not connect to 51.79.55.151: Connection reset by peer [00:12:01] PROBLEM - cp2 Puppet on cp2 is CRITICAL: CHECK_NRPE: Error - Could not connect to 51.79.55.151: Connection reset by peer [00:12:03] PROBLEM - cp2 Nginx Backend for mw1 on cp2 is CRITICAL: CHECK_NRPE: Error - Could not connect to 51.79.55.151: Connection reset by peer [00:12:04] PROBLEM - cp2 HTTP 4xx/5xx ERROR Rate on cp2 is CRITICAL: CHECK_NRPE: Error - Could not connect to 51.79.55.151: Connection reset by peer [00:12:08] PROBLEM - cp2 Nginx Backend for phorge1 on cp2 is CRITICAL: CHECK_NRPE: Error - Could not connect to 51.79.55.151: Connection reset by peer [00:12:09] PROBLEM - cp2 Current Load on cp2 is CRITICAL: CHECK_NRPE: Error - Could not connect to 51.79.55.151: Connection reset by peer [00:12:13] PROBLEM - cp2 APT on cp2 is CRITICAL: CHECK_NRPE: Error - Could not connect to 51.79.55.151: Connection reset by peer [00:12:19] PROBLEM - cp2 Nginx Backend for mail1 on cp2 is CRITICAL: CHECK_NRPE: Error - Could not connect to 51.79.55.151: Connection reset by peer [00:12:19] PROBLEM - cp2 Nginx Backend for test1 on cp2 is CRITICAL: CHECK_NRPE: Error - Could not connect to 51.79.55.151: Connection reset by peer [00:12:20] PROBLEM - cp2 NTP time on cp2 is CRITICAL: CHECK_NRPE: Error - Could not connect to 51.79.55.151: Connection reset by peer [00:15:58] PROBLEM - cp5 HTTP 4xx/5xx ERROR Rate on cp5 is WARNING: WARNING - NGINX Error Rate is 52% [00:16:12] PROBLEM - cp3 HTTP 4xx/5xx ERROR Rate on cp3 is CRITICAL: CRITICAL - NGINX Error Rate is 64% [00:17:54] RECOVERY - cp5 HTTP 4xx/5xx ERROR Rate on cp5 is OK: OK - NGINX Error Rate is 38% [00:18:12] PROBLEM - cp3 HTTP 4xx/5xx ERROR Rate on cp3 is WARNING: WARNING - NGINX Error Rate is 58% [00:20:12] PROBLEM - cp3 HTTP 4xx/5xx ERROR Rate on cp3 is CRITICAL: CRITICAL - NGINX Error Rate is 79% [00:20:21] [02WikiTideOrg/puppet] 07Universal-Omega pushed 031 commit to 03master [+0/-0/±1] 13https://github.com/WikiTideOrg/puppet/compare/4ed62e68c939...b47ae05dde81 [00:20:24] [02WikiTideOrg/puppet] 07Universal-Omega 03b47ae05 - Update ports [00:20:33] PROBLEM - mw2 MediaWiki Rendering on mw2 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [00:22:32] PROBLEM - mw2 MediaWiki Rendering on mw2 is WARNING: HTTP WARNING: HTTP/1.1 404 Not Found - 8191 bytes in 0.133 second response time [00:23:28] PROBLEM - cp2 Nginx Backend for mail21 on cp2 is CRITICAL: connect to address localhost and port 8202: Connection refused [00:23:39] [02WikiTideOrg/puppet] 07Universal-Omega pushed 031 commit to 03master [+0/-0/±1] 13https://github.com/WikiTideOrg/puppet/compare/b47ae05dde81...a8723ef1fd8c [00:23:41] [02WikiTideOrg/puppet] 07Universal-Omega 03a8723ef - Revert [00:24:56] PROBLEM - cp2 APT on cp2 is WARNING: APT WARNING: 0 packages available for upgrade (0 critical updates). warnings detected, errors detected. [00:24:56] PROBLEM - cp2 Nginx Backend for matomo21 on cp2 is CRITICAL: connect to address localhost and port 8204: Connection refused [00:25:24] RECOVERY - cp2 Nginx Backend for mail21 on cp2 is OK: TCP OK - 0.000 second response time on localhost port 8202 [00:26:12] PROBLEM - cp3 HTTP 4xx/5xx ERROR Rate on cp3 is WARNING: WARNING - NGINX Error Rate is 41% [00:26:48] PROBLEM - cp5 HTTP 4xx/5xx ERROR Rate on cp5 is WARNING: WARNING - NGINX Error Rate is 40% [00:26:51] RECOVERY - cp2 Nginx Backend for matomo21 on cp2 is OK: TCP OK - 0.000 second response time on localhost port 8204 [00:28:58] PROBLEM - cp2 APT on cp2 is CRITICAL: APT CRITICAL: 32 packages available for upgrade (3 critical updates). [00:30:57] PROBLEM - cp2 APT on cp2 is WARNING: APT WARNING: 0 packages available for upgrade (0 critical updates). warnings detected, errors detected. [00:32:12] PROBLEM - cp3 HTTP 4xx/5xx ERROR Rate on cp3 is CRITICAL: CRITICAL - NGINX Error Rate is 69% [00:32:58] PROBLEM - cp2 APT on cp2 is CRITICAL: APT CRITICAL: 32 packages available for upgrade (3 critical updates). [00:34:12] PROBLEM - cp3 HTTP 4xx/5xx ERROR Rate on cp3 is WARNING: WARNING - NGINX Error Rate is 47% [00:35:00] RECOVERY - cp2 APT on cp2 is OK: APT OK: 0 packages available for upgrade (0 critical updates). [00:36:12] PROBLEM - cp3 HTTP 4xx/5xx ERROR Rate on cp3 is CRITICAL: CRITICAL - NGINX Error Rate is 74% [00:36:44] RECOVERY - cp5 HTTP 4xx/5xx ERROR Rate on cp5 is OK: OK - NGINX Error Rate is 38% [00:41:27] PROBLEM - cp2 Varnish Backends on cp2 is WARNING: No backends detected. If this is an error, see readme.txt [00:42:44] PROBLEM - cp5 HTTP 4xx/5xx ERROR Rate on cp5 is CRITICAL: CRITICAL - NGINX Error Rate is 88% [00:43:27] PROBLEM - cp2 Varnish Backends on cp2 is CRITICAL: 3 backends are down. mw1 mw2 mediawiki [00:44:12] RECOVERY - cp3 HTTP 4xx/5xx ERROR Rate on cp3 is OK: OK - NGINX Error Rate is 36% [00:44:15] RECOVERY - cp2 Puppet on cp2 is OK: OK: Puppet is currently enabled, last run 13 seconds ago with 0 failures [00:45:08] PROBLEM - cp2 Nginx Backend for matomo21 on cp2 is CRITICAL: connect to address localhost and port 8204: Connection refused [00:45:15] RECOVERY - cp2 HTTP 4xx/5xx ERROR Rate on cp2 is OK: OK - NGINX Error Rate is 0% [00:46:10] PROBLEM - cp2 Nginx Backend for mail21 on cp2 is CRITICAL: connect to address localhost and port 8202: Connection refused [00:46:44] PROBLEM - cp5 HTTP 4xx/5xx ERROR Rate on cp5 is WARNING: WARNING - NGINX Error Rate is 52% [00:48:12] PROBLEM - cp3 HTTP 4xx/5xx ERROR Rate on cp3 is WARNING: WARNING - NGINX Error Rate is 53% [00:48:44] PROBLEM - cp5 HTTP 4xx/5xx ERROR Rate on cp5 is CRITICAL: CRITICAL - NGINX Error Rate is 70% [00:50:45] PROBLEM - cp5 HTTP 4xx/5xx ERROR Rate on cp5 is WARNING: WARNING - NGINX Error Rate is 41% [00:52:12] PROBLEM - cp3 HTTP 4xx/5xx ERROR Rate on cp3 is CRITICAL: CRITICAL - NGINX Error Rate is 63% [00:52:35] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is WARNING: WARNING - NGINX Error Rate is 40% [00:54:12] PROBLEM - cp3 HTTP 4xx/5xx ERROR Rate on cp3 is WARNING: WARNING - NGINX Error Rate is 50% [00:54:35] RECOVERY - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is OK: OK - NGINX Error Rate is 38% [00:56:12] PROBLEM - cp3 HTTP 4xx/5xx ERROR Rate on cp3 is CRITICAL: CRITICAL - NGINX Error Rate is 61% [00:56:46] PROBLEM - cp5 HTTP 4xx/5xx ERROR Rate on cp5 is CRITICAL: CRITICAL - NGINX Error Rate is 60% [00:58:12] PROBLEM - cp3 HTTP 4xx/5xx ERROR Rate on cp3 is WARNING: WARNING - NGINX Error Rate is 51% [00:58:44] PROBLEM - cp5 HTTP 4xx/5xx ERROR Rate on cp5 is WARNING: WARNING - NGINX Error Rate is 54% [01:00:12] RECOVERY - cp3 HTTP 4xx/5xx ERROR Rate on cp3 is OK: OK - NGINX Error Rate is 29% [01:00:33] PROBLEM - newcascadia.net - reverse DNS on sslhost is CRITICAL: rDNS CRITICAL - newcascadia.net All nameservers failed to answer the query. [01:00:44] RECOVERY - cp5 HTTP 4xx/5xx ERROR Rate on cp5 is OK: OK - NGINX Error Rate is 38% [01:01:27] I am getting database timeouts on https://rosettacode.org/ [01:14:49] PROBLEM - meta.sagan4.org - reverse DNS on sslhost is UNKNOWN: PROBLEM - www.greatamerica.wiki - reverse DNS on sslhost is UNKNOWN: [01:15:50] PROBLEM - beta.sagan4.org - reverse DNS on sslhost is UNKNOWN: PROBLEM - www.polandballwiki.com - reverse DNS on sslhost is UNKNOWN: [01:15:50] PROBLEM - hsck.lophocmatngu.wiki - reverse DNS on sslhost is UNKNOWN: [01:15:52] PROBLEM - matomo21 HTTPS on matomo21 is UNKNOWN: [01:16:06] PROBLEM - cp5 HTTP 4xx/5xx ERROR Rate on cp5 is WARNING: WARNING - NGINX Error Rate is 49% [01:17:52] PROBLEM - mon21 Current Load on mon21 is CRITICAL: LOAD CRITICAL - total load average: 20.29, 12.58, 5.55 [01:18:45] PROBLEM - matomo21 HTTPS on matomo21 is CRITICAL: HTTP CRITICAL - Invalid HTTP response received from host on port 443: cURL returned 28 - Connection timed out after 10038 milliseconds [01:19:04] RECOVERY - cp5 HTTP 4xx/5xx ERROR Rate on cp5 is OK: OK - NGINX Error Rate is 36% PROBLEM - phorge21 HTTPS on phorge21 is UNKNOWN: [01:20:08] PROBLEM - www.farthestfrontier.wiki - reverse DNS on sslhost is UNKNOWN: PROBLEM - puppet21 HTTPS on puppet21 is UNKNOWN: PROBLEM - mw21 HTTPS on mw21 is UNKNOWN: PROBLEM - data.lophocmatngu.wiki - reverse DNS on sslhost is UNKNOWN: 15 (Terminated).> PROBLEM - cp2 HTTPS on cp2 is UNKNOWN: [01:20:21] PROBLEM - wiki.burning.art - reverse DNS on sslhost is UNKNOWN: [01:20:22] PROBLEM - mon21 grafana.wikitide.net HTTPS on mon21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:21:19] PROBLEM - mason.sagan4.org - reverse DNS on sslhost is UNKNOWN: [01:23:18] PROBLEM - matomo21 HTTPS on matomo21 is UNKNOWN: [01:44:35] PROBLEM - mw2 MediaWiki Rendering on mw2 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:47:20] PROBLEM - mon21 PowerDNS Recursor on mon21 is CRITICAL: CRITICAL - Plugin timed out while executing system call PROBLEM - mem21 Current Load on mem21 is WARNING: LOAD WARNING - total load average: 3.62, 3.57, 2.89 [01:47:21] PROBLEM - mon21 HTTPS on mon21 is UNKNOWN: PROBLEM - mon21 HTTPS on mon21 is UNKNOWN: [01:47:21] PROBLEM - os21 Current Load on os21 is UNKNOWN: [01:47:34] [02WikiTideOrg/puppet] 07Universal-Omega pushed 031 commit to 03master [+0/-0/±1] 13https://github.com/WikiTideOrg/puppet/compare/a8723ef1fd8c...6d8a06934bd0 [01:47:36] [02WikiTideOrg/puppet] 07Universal-Omega 036d8a069 - Update servers [01:48:35] PROBLEM - projects.dmvpetridish.com - reverse DNS on sslhost is UNKNOWN: PROBLEM - mon21 monitoring.wikitide.net HTTPS on mon21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:48:37] PROBLEM - mon21 monitoring.wikitide.net HTTPS on mon21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:48:38] PROBLEM - cp3 Puppet on cp3 is CRITICAL: CHECK_NRPE STATE CRITICAL: Socket timeout after 60 seconds. [01:48:58] WikiTideOrg/puppet - Universal-Omega the build has errored. [01:49:36] PROBLEM - mem21 Current Load on mem21 is CRITICAL: LOAD CRITICAL - total load average: 5.10, 3.84, 3.48 [01:50:05] PROBLEM - farthestfrontier.wiki - reverse DNS on sslhost is UNKNOWN: PROBLEM - jobrunner21 HTTPS on jobrunner21 is UNKNOWN: PROBLEM - cp6 HTTPS on cp6 is UNKNOWN: PROBLEM - mail21 HTTPS on mail21 is UNKNOWN: PROBLEM - mlrpgspeedruns.com [01:50:05] sslhost is UNKNOWN: [01:50:49] PROBLEM - jobrunner1 MediaWiki Rendering on jobrunner1 is UNKNOWN: HTTP UNKNOWN: Failed to unchunk message body [01:50:57] PROBLEM - mw2 MediaWiki Rendering on mw2 is WARNING: HTTP WARNING: HTTP/1.1 404 Not Found - 8191 bytes in 7.028 second response time [01:51:14] PROBLEM - mw1 MediaWiki Rendering on mw1 is UNKNOWN: HTTP UNKNOWN: Failed to unchunk message body [01:51:23] PROBLEM - mon21 Check correctness of the icinga configuration on mon21 is CRITICAL: CHECK_NRPE STATE CRITICAL: Socket timeout after 60 seconds. [01:51:23] PROBLEM - wiki.chevrine.com - reverse DNS on sslhost is UNKNOWN: [01:51:23] PROBLEM - newcascadia.net - reverse DNS on sslhost is UNKNOWN: [01:51:23] PROBLEM - mon21 Check correctness of the icinga configuration on mon21 is CRITICAL: CHECK_NRPE STATE CRITICAL: Socket timeout after 60 seconds. [01:51:23] PROBLEM - cp4 HTTPS on cp4 is UNKNOWN: [01:51:24] PROBLEM - wiki.myehs.eu - reverse DNS on sslhost is UNKNOWN: [01:51:24] PROBLEM - www.lgbtqia.wiki - reverse DNS on sslhost is UNKNOWN: [01:51:25] PROBLEM - nexttide.org - reverse DNS on sslhost is UNKNOWN: [01:51:25] PROBLEM - wikitide.com - reverse DNS on sslhost is UNKNOWN: [01:51:26] PROBLEM - wiki.colleimadcat.com - reverse DNS on sslhost is UNKNOWN: [01:51:26] PROBLEM - alpha.sagan4.org - reverse DNS on sslhost is UNKNOWN: [01:51:27] PROBLEM - cp2 HTTPS on cp2 is CRITICAL: HTTP CRITICAL - Invalid HTTP response received from host on port 443: cURL returned 7 - Failed to connect to cp2.wikitide.net port 443 after 10 ms: Couldn't connect to server [01:51:27] RECOVERY - mon21 HTTPS on mon21 is OK: HTTP OK: HTTP/2 302 - 374 bytes in 0.159 second response time [01:51:28] PROBLEM - os21 Current Load on os21 is CRITICAL: LOAD CRITICAL - total load average: 4.78, 4.16, 3.65 [01:51:28] RECOVERY - mon21 monitoring.wikitide.net HTTPS on mon21 is OK: HTTP OK: HTTP/1.1 302 Found - 298 bytes in 0.259 second response time [01:51:29] RECOVERY - mon21 grafana.wikitide.net HTTPS on mon21 is OK: HTTP OK: HTTP/1.1 302 Found - 429 bytes in 0.362 second response time [01:51:29] PROBLEM - cp3 HTTPS on cp3 is CRITICAL: HTTP CRITICAL: HTTP/2 503 - 2628 bytes in 0.328 second response time [01:51:30] PROBLEM - cp5 HTTPS on cp5 is CRITICAL: HTTP CRITICAL: HTTP/2 503 - 2628 bytes in 0.754 second response time [01:51:30] PROBLEM - cp5 HTTP 4xx/5xx ERROR Rate on cp5 is CRITICAL: CRITICAL - NGINX Error Rate is 85% [01:52:14] PROBLEM - mw21 HTTPS on mw21 is CRITICAL: HTTP CRITICAL - Invalid HTTP response received from host on port 443: cURL returned 28 - Connection timed out after 10083 milliseconds PROBLEM - puppet21 HTTPS on puppet21 is CRITICAL: HTTP CRITICAL - Invalid HTTP response received from host on port 443: cURL returned 28 - Connection timed out after 10072 milliseconds [01:52:26] RECOVERY - mon21 PowerDNS Recursor on mon21 is OK: DNS OK: 0.094 seconds response time. wikitide.org returns 2001:41d0:801:2000::4089,51.75.170.66 [01:52:26] RECOVERY - jobrunner21 HTTPS on jobrunner21 is OK: HTTP OK: HTTP/2 200 - 370 bytes in 0.029 second response time [01:52:30] RECOVERY - cp6 HTTPS on cp6 is OK: HTTP OK: HTTP/2 200 - 2995 bytes in 1.616 second response time PROBLEM - phorge21 HTTPS on phorge21 is CRITICAL: HTTP CRITICAL: HTTP/2 500 - 666 bytes in 2.156 second response time [01:52:56] PROBLEM - mw2 MediaWiki Rendering on mw2 is UNKNOWN: HTTP UNKNOWN: Failed to unchunk message body [01:53:35] PROBLEM - matomo21 HTTPS on matomo21 is CRITICAL: HTTP CRITICAL - Invalid HTTP response received from host on port 443: cURL returned 28 - Connection timed out after 10043 milliseconds [01:53:35] PROBLEM - mail21 HTTPS on mail21 is CRITICAL: HTTP CRITICAL - Invalid HTTP response received from host on port 443: cURL returned 28 - Connection timed out after 10014 milliseconds [01:53:37] RECOVERY - cp4 HTTPS on cp4 is OK: HTTP OK: HTTP/2 200 - 2996 bytes in 0.702 second response time [01:53:38] RECOVERY - cp5 HTTP 4xx/5xx ERROR Rate on cp5 is OK: OK - NGINX Error Rate is 16% [01:53:41] RECOVERY - cp5 HTTPS on cp5 is OK: HTTP OK: HTTP/2 200 - 2974 bytes in 5.089 second response time [01:54:20] PROBLEM - mem21 Current Load on mem21 is WARNING: LOAD WARNING - total load average: 3.43, 3.95, 3.66 [01:54:23] PROBLEM - cp5 Puppet on cp5 is CRITICAL: CRITICAL: Puppet last ran 12 hours ago [01:55:34] PROBLEM - rosettacode.org - LetsEncrypt on sslhost is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:55:34] PROBLEM - rosettacode.org - LetsEncrypt on sslhost is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:55:35] PROBLEM - lophocmatngu.wiki - reverse DNS on sslhost is UNKNOWN: [01:55:35] PROBLEM - wikitide.org - LetsEncrypt on sslhost is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:55:35] PROBLEM - polandballwiki.com - reverse DNS on sslhost is UNKNOWN: [01:55:35] PROBLEM - wikitide.net - reverse DNS on sslhost is UNKNOWN: [01:55:35] PROBLEM - hkrail.wiki - reverse DNS on sslhost is UNKNOWN: [01:55:36] PROBLEM - www.rosettacode.org - reverse DNS on sslhost is UNKNOWN: PROBLEM - www.wikitide.org - reverse DNS on sslhost is UNKNOWN: [01:55:36] RECOVERY - cp3 HTTPS on cp3 is OK: HTTP OK: HTTP/2 200 - 2974 bytes in 0.565 second response time [01:55:43] RECOVERY - mon21 Check correctness of the icinga configuration on mon21 is OK: Icinga configuration is correct [01:55:49] PROBLEM - wikitide.org - reverse DNS on sslhost is UNKNOWN: [01:55:49] PROBLEM - greatamerica.wiki - reverse DNS on sslhost is UNKNOWN: PROBLEM - sdiy.info - reverse DNS on sslhost is UNKNOWN: [01:55:49] PROBLEM - rosettacode.org - reverse DNS on sslhost is UNKNOWN: [01:55:49] PROBLEM - distrowiki.net - reverse DNS on sslhost is UNKNOWN: [01:57:43] PROBLEM - cp5 HTTP 4xx/5xx ERROR Rate on cp5 is WARNING: WARNING - NGINX Error Rate is 42% [01:59:40] RECOVERY - cp5 HTTP 4xx/5xx ERROR Rate on cp5 is OK: OK - NGINX Error Rate is 20% [02:00:03] RECOVERY - mem21 Current Load on mem21 is OK: LOAD OK - total load average: 2.38, 2.98, 3.35 [02:00:12] PROBLEM - cp3 HTTP 4xx/5xx ERROR Rate on cp3 is CRITICAL: CRITICAL - NGINX Error Rate is 76% [02:07:31] PROBLEM - cp5 HTTP 4xx/5xx ERROR Rate on cp5 is WARNING: WARNING - NGINX Error Rate is 43% [02:09:27] RECOVERY - cp5 HTTP 4xx/5xx ERROR Rate on cp5 is OK: OK - NGINX Error Rate is 32% [02:10:16] PROBLEM - phorge1 issue-tracker.wikitide.org HTTPS on phorge1 is CRITICAL: HTTP CRITICAL: HTTP/1.1 500 Internal Server Error - 4248 bytes in 0.032 second response time [02:10:27] PROBLEM - db1 MariaDB on db1 is CRITICAL: Can't connect to server on 'db1.wikitide.net' (115) [02:11:33] PROBLEM - db1 MariaDB Connections on db1 is UNKNOWN: PHP Fatal error: Uncaught mysqli_sql_exception: Connection refused in /usr/lib/nagios/plugins/check_mysql_connections.php:47Stack trace:#0 /usr/lib/nagios/plugins/check_mysql_connections.php(47): mysqli_real_connect(Object(mysqli), 'db1.wikitide.ne...', 'icinga', Object(SensitiveParameterValue), NULL, NULL, NULL, false)#1 {main} thrown in /usr/lib/nagios/plugins/check_mysql_conne [02:11:33] ine 47Fatal error: Uncaught mysqli_sql_exception: Connection refused in /usr/lib/nagios/plugins/check_mysql_connections.php:47Stack trace:#0 /usr/lib/nagios/plugins/check_mysql_connections.php(47): mysqli_real_connect(Object(mysqli), 'db1.wikitide.ne...', 'icinga', Object(SensitiveParameterValue), NULL, NULL, NULL, false)#1 {main} thrown in /usr/lib/nagios/plugins/check_mysql_connections.php on line 47 [02:11:33] PROBLEM - matomo1 HTTPS on matomo1 is CRITICAL: HTTP CRITICAL: HTTP/2 500 - 426 bytes in 0.034 second response time [02:11:48] PROBLEM - phorge1 phorge-static.wikitide.org HTTPS on phorge1 is CRITICAL: HTTP CRITICAL - Invalid HTTP response received from host on port 443: HTTP/1.1 500 Internal Server Error [02:18:14] PROBLEM - projects.dmvpetridish.com - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['ns0.wixdns.net.', 'ns1.wixdns.net.'], 'CNAME': 'mw-lb.wikitide.org.'} [02:18:54] RECOVERY - farthestfrontier.wiki - reverse DNS on sslhost is OK: SSL OK - farthestfrontier.wiki reverse DNS resolves to cp5.wikitide.net - CNAME FLAT [02:19:04] PROBLEM - wiki.chevrine.com - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['ns1.digitalocean.com.', 'ns2.digitalocean.com.', 'ns3.digitalocean.com.'], 'CNAME': 'mw-lb.wikitide.org.'} [02:19:08] PROBLEM - newcascadia.net - reverse DNS on sslhost is CRITICAL: rDNS CRITICAL - newcascadia.net All nameservers failed to answer the query. [02:19:34] PROBLEM - mlrpgspeedruns.com - reverse DNS on sslhost is CRITICAL: rDNS CRITICAL - mlrpgspeedruns.com All nameservers failed to answer the query. [02:19:56] PROBLEM - www.lgbtqia.wiki - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['grannbo.ns.cloudflare.com.', 'rodrigo.ns.cloudflare.com.'], 'CNAME': 'cf-lb.wikitide.org.'} [02:19:57] PROBLEM - wiki.colleimadcat.com - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['dns29.hichina.com.', 'dns30.hichina.com.'], 'CNAME': 'cf-lb.wikitide.org.'} [02:20:03] PROBLEM - meta.sagan4.org - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['ns59.domaincontrol.com.', 'ns60.domaincontrol.com.'], 'CNAME': 'cf-lb.wikitide.org.'} [02:20:03] PROBLEM - www.greatamerica.wiki - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['john.ns.cloudflare.com.', 'vida.ns.cloudflare.com.'], 'CNAME': 'mw-lb.wikitide.org.'} [02:20:25] PROBLEM - wikitide.com - reverse DNS on sslhost is WARNING: rDNS WARNING - reverse DNS entry for wikitide.com could not be found [02:20:26] PROBLEM - beta.sagan4.org - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['ns59.domaincontrol.com.', 'ns60.domaincontrol.com.'], 'CNAME': 'cf-lb.wikitide.org.'} [02:20:27] PROBLEM - www.polandballwiki.com - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['john.ns.cloudflare.com.', 'vida.ns.cloudflare.com.'], 'CNAME': 'mw-lb.wikitide.org.'} [02:20:29] PROBLEM - alpha.sagan4.org - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['ns59.domaincontrol.com.', 'ns60.domaincontrol.com.'], 'CNAME': 'cf-lb.wikitide.org.'} [02:20:38] PROBLEM - hsck.lophocmatngu.wiki - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['lia.ns.cloudflare.com.', 'gerald.ns.cloudflare.com.'], 'CNAME': 'cf-lb.wikitide.org.'} [02:20:47] PROBLEM - wiki.myehs.eu - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['fred.ns.cloudflare.com.', 'ivy.ns.cloudflare.com.'], 'CNAME': 'cf-lb.wikitide.org.'} [02:21:02] PROBLEM - data.lophocmatngu.wiki - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['lia.ns.cloudflare.com.', 'gerald.ns.cloudflare.com.'], 'CNAME': 'cf-lb.wikitide.org.'} [02:21:11] PROBLEM - nexttide.org - reverse DNS on sslhost is CRITICAL: rDNS CRITICAL - nexttide.org All nameservers failed to answer the query. [02:21:23] PROBLEM - www.farthestfrontier.wiki - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['cleo.ns.cloudflare.com.', 'may.ns.cloudflare.com.'], 'CNAME': 'cf-lb.wikitide.org.'} [02:21:39] RECOVERY - wiki.burning.art - reverse DNS on sslhost is OK: SSL OK - wiki.burning.art reverse DNS resolves to cp5.wikitide.net - CNAME OK [02:23:00] PROBLEM - mon21 Current Load on mon21 is WARNING: LOAD WARNING - total load average: 0.06, 0.31, 5.99 [02:23:35] PROBLEM - mason.sagan4.org - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['ns59.domaincontrol.com.', 'ns60.domaincontrol.com.'], 'CNAME': 'cf-lb.wikitide.org.'} [02:24:13] RECOVERY - polandballwiki.com - reverse DNS on sslhost is OK: SSL OK - polandballwiki.com reverse DNS resolves to cp5.wikitide.net - CNAME FLAT [02:24:18] RECOVERY - lophocmatngu.wiki - reverse DNS on sslhost is OK: SSL OK - lophocmatngu.wiki reverse DNS resolves to cp5.wikitide.net - CNAME FLAT [02:24:19] RECOVERY - hkrail.wiki - reverse DNS on sslhost is OK: SSL OK - hkrail.wiki reverse DNS resolves to cp5.wikitide.net - CNAME FLAT [02:24:22] RECOVERY - rosettacode.org - reverse DNS on sslhost is OK: SSL OK - rosettacode.org reverse DNS resolves to cp5.wikitide.net - NS RECORDS OK [02:24:27] RECOVERY - www.rosettacode.org - reverse DNS on sslhost is OK: SSL OK - www.rosettacode.org reverse DNS resolves to cp5.wikitide.net - NS RECORDS OK [02:24:34] RECOVERY - www.wikitide.org - reverse DNS on sslhost is OK: SSL OK - www.wikitide.org reverse DNS resolves to cp3.wikitide.net - NS RECORDS OK [02:24:35] RECOVERY - rosettacode.org - LetsEncrypt on sslhost is OK: OK - Certificate 'rosettacode.org' will expire on Sat 20 Jan 2024 06:52:33 AM GMT +0000. [02:24:40] RECOVERY - wikitide.org - LetsEncrypt on sslhost is OK: OK - Certificate 'wikitide.org' will expire on Tue 20 Feb 2024 12:20:10 PM GMT +0000. [02:24:55] RECOVERY - wikitide.org - reverse DNS on sslhost is OK: SSL OK - wikitide.org reverse DNS resolves to cp3.wikitide.net - NS RECORDS OK [02:25:03] RECOVERY - wikitide.net - reverse DNS on sslhost is OK: SSL OK - wikitide.net reverse DNS resolves to cp3.wikitide.net - NS RECORDS OK [02:25:36] RECOVERY - greatamerica.wiki - reverse DNS on sslhost is OK: SSL OK - greatamerica.wiki reverse DNS resolves to cp5.wikitide.net - CNAME FLAT [02:25:37] RECOVERY - distrowiki.net - reverse DNS on sslhost is OK: SSL OK - distrowiki.net reverse DNS resolves to cp3.wikitide.net - CNAME FLAT [02:25:37] RECOVERY - sdiy.info - reverse DNS on sslhost is OK: SSL OK - sdiy.info reverse DNS resolves to cp5.wikitide.net - NS RECORDS OK [02:27:00] RECOVERY - mon21 Current Load on mon21 is OK: LOAD OK - total load average: 0.06, 0.18, 4.64 [02:28:10] [02WikiTideOrg/puppet] 07Universal-Omega pushed 031 commit to 03master [+0/-0/±1] 13https://github.com/WikiTideOrg/puppet/compare/6d8a06934bd0...1db17177b6bf [02:28:12] PROBLEM - cp5 HTTP 4xx/5xx ERROR Rate on cp5 is WARNING: WARNING - NGINX Error Rate is 43% [02:28:12] [02WikiTideOrg/puppet] 07Universal-Omega 031db1717 - Spacing [02:32:04] PROBLEM - cp5 HTTP 4xx/5xx ERROR Rate on cp5 is CRITICAL: CRITICAL - NGINX Error Rate is 61% [02:34:03] PROBLEM - cp5 HTTP 4xx/5xx ERROR Rate on cp5 is WARNING: WARNING - NGINX Error Rate is 49% [02:36:01] PROBLEM - cp5 HTTP 4xx/5xx ERROR Rate on cp5 is CRITICAL: CRITICAL - NGINX Error Rate is 73% [02:37:57] PROBLEM - cp5 HTTP 4xx/5xx ERROR Rate on cp5 is WARNING: WARNING - NGINX Error Rate is 42% [02:43:46] RECOVERY - cp5 HTTP 4xx/5xx ERROR Rate on cp5 is OK: OK - NGINX Error Rate is 38% [02:49:00] PROBLEM - cp2 HTTPS on cp2 is UNKNOWN: [02:59:33] RECOVERY - db1 MariaDB Connections on db1 is OK: OK connection usage: 0.2%Current connections: 1 [03:00:27] RECOVERY - db1 MariaDB on db1 is OK: Uptime: 68 Threads: 2 Questions: 177 Slow queries: 36 Opens: 34 Open tables: 12 Queries per second avg: 2.602 [03:03:33] RECOVERY - matomo1 HTTPS on matomo1 is OK: HTTP OK: HTTP/2 200 - 224 bytes in 0.208 second response time [03:15:38] PROBLEM - mem21 memcached on mem21 is UNKNOWN: [03:15:38] PROBLEM - bast21 Puppet on bast21 is UNKNOWN: [03:15:41] PROBLEM - puppet21 HTTPS on puppet21 is UNKNOWN: [03:17:01] PROBLEM - wiki.colleimadcat.com - reverse DNS on sslhost is UNKNOWN: PROBLEM - newcascadia.net - reverse DNS on sslhost is UNKNOWN: [03:17:01] PROBLEM - data.lophocmatngu.wiki - reverse DNS on sslhost is UNKNOWN: PROBLEM - wiki.chevrine.com - reverse DNS on sslhost is UNKNOWN: PROBLEM - www.lgbtqia.wiki - reverse DNS on sslhost is UNKNOWN: PROBLEM - alpha.sagan4.org - reverse DNS on sslhost is UN [03:17:01] xceeded.> PROBLEM - projects.dmvpetridish.com - reverse DNS on sslhost is UNKNOWN: [03:17:01] PROBLEM - hsck.lophocmatngu.wiki - reverse DNS on sslhost is UNKNOWN: [03:17:01] PROBLEM - meta.sagan4.org - reverse DNS on sslhost is UNKNOWN: [03:17:02] PROBLEM - www.greatamerica.wiki - reverse DNS on sslhost is UNKNOWN: [03:17:02] PROBLEM - www.polandballwiki.com - reverse DNS on sslhost is UNKNOWN: [03:17:03] PROBLEM - www.farthestfrontier.wiki - reverse DNS on sslhost is UNKNOWN: [03:17:03] PROBLEM - mlrpgspeedruns.com - reverse DNS on sslhost is UNKNOWN: [03:17:04] PROBLEM - wiki.myehs.eu - reverse DNS on sslhost is UNKNOWN: PROBLEM - wikitide.com - reverse DNS on sslhost is UNKNOWN: [03:17:04] PROBLEM - beta.sagan4.org - reverse DNS on sslhost is UNKNOWN: [03:17:05] PROBLEM - cp2 HTTPS on cp2 is CRITICAL: HTTP CRITICAL - Invalid HTTP response received from host on port 443: cURL returned 7 - Failed to connect to cp2.wikitide.net port 443 after 31 ms: Couldn't connect to server [03:17:25] PROBLEM - cp3 HTTP 4xx/5xx ERROR Rate on cp3 is WARNING: WARNING - NGINX Error Rate is 40% [03:18:04] PROBLEM - nexttide.org - reverse DNS on sslhost is UNKNOWN: [03:18:04] PROBLEM - mason.sagan4.org - reverse DNS on sslhost is UNKNOWN: [03:18:05] PROBLEM - bast21 Puppet on bast21 is CRITICAL: CRITICAL: Failed to apply catalog, zero resources tracked by Puppet. It might be a dependency cycle. [03:18:13] PROBLEM - mem21 memcached on mem21 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [03:18:15] PROBLEM - puppet21 HTTPS on puppet21 is CRITICAL: HTTP CRITICAL - Invalid HTTP response received from host on port 443: cURL returned 28 - Connection timed out after 10004 milliseconds [03:18:57] RECOVERY - cp3 HTTP 4xx/5xx ERROR Rate on cp3 is OK: OK - NGINX Error Rate is 22% [03:19:01] PROBLEM - mon21 Current Load on mon21 is CRITICAL: LOAD CRITICAL - total load average: 20.15, 34.36, 30.60 [03:19:07] PROBLEM - mem21 Current Load on mem21 is WARNING: LOAD WARNING - total load average: 3.15, 3.53, 3.37 [03:21:02] RECOVERY - mem21 Current Load on mem21 is OK: LOAD OK - total load average: 2.61, 3.24, 3.28 [03:25:31] RECOVERY - phorge1 issue-tracker.wikitide.org HTTPS on phorge1 is OK: HTTP OK: HTTP/1.1 200 OK - 115674 bytes in 0.321 second response time [03:25:48] RECOVERY - phorge1 phorge-static.wikitide.org HTTPS on phorge1 is OK: HTTP OK: Status line output matched "HTTP/1.1 200" - 58195 bytes in 0.376 second response time [03:26:49] PROBLEM - jobrunner1 MediaWiki Rendering on jobrunner1 is WARNING: HTTP WARNING: HTTP/1.1 404 Not Found - 8191 bytes in 0.188 second response time [03:27:42] PROBLEM - mw2 MediaWiki Rendering on mw2 is WARNING: HTTP WARNING: HTTP/1.1 404 Not Found - 8191 bytes in 0.155 second response time [03:35:47] PROBLEM - mw2 MediaWiki Rendering on mw2 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [03:37:45] PROBLEM - mw2 MediaWiki Rendering on mw2 is WARNING: HTTP WARNING: HTTP/1.1 404 Not Found - 8191 bytes in 0.116 second response time [03:44:26] [02WikiTideOrg/ssl] 07WikiTideSSL pushed 031 commit to 03master [+0/-0/±2] 13https://github.com/WikiTideOrg/ssl/compare/1491bd65da1c...d869af88bf75 [03:44:26] PROBLEM - jobrunner1 Disk Space on jobrunner1 is CRITICAL: CHECK_NRPE: Error - Could not connect to 10.0.0.104: Connection reset by peer [03:44:28] [02WikiTideOrg/ssl] 07WikiTideSSL 03d869af8 - Bot: Adding SSL cert for wiki.myehs.eu Certificate committed by universalomega [03:44:30] PROBLEM - jobrunner1 php-fpm on jobrunner1 is CRITICAL: CHECK_NRPE: Error - Could not connect to 10.0.0.104: Connection reset by peer [03:44:33] PROBLEM - jobrunner1 PowerDNS Recursor on jobrunner1 is CRITICAL: CHECK_NRPE: Error - Could not connect to 10.0.0.104: Connection reset by peer [03:44:42] PROBLEM - jobrunner1 NTP time on jobrunner1 is CRITICAL: CHECK_NRPE: Error - Could not connect to 10.0.0.104: Connection reset by peer [03:44:47] PROBLEM - jobrunner1 ferm_active on jobrunner1 is CRITICAL: CHECK_NRPE: Error - Could not connect to 10.0.0.104: Connection reset by peer [03:44:56] PROBLEM - jobrunner1 JobRunner Service on jobrunner1 is CRITICAL: CHECK_NRPE: Error - Could not connect to 10.0.0.104: Connection reset by peer [03:45:00] PROBLEM - mon21 Current Load on mon21 is WARNING: LOAD WARNING - total load average: 0.32, 0.31, 5.75 [03:45:06] PROBLEM - jobrunner1 Current Load on jobrunner1 is CRITICAL: CHECK_NRPE: Error - Could not connect to 10.0.0.104: Connection reset by peer [03:45:13] [02WikiTideOrg/ssl] 07Universal-Omega pushed 031 commit to 03master [+0/-0/±1] 13https://github.com/WikiTideOrg/ssl/compare/d869af88bf75...92c03967bf16 [03:45:15] [02WikiTideOrg/ssl] 07Universal-Omega 0392c0396 - Remove duplicate [03:45:16] PROBLEM - jobrunner1 conntrack_table_size on jobrunner1 is CRITICAL: CHECK_NRPE: Error - Could not connect to 10.0.0.104: Connection reset by peer [03:45:25] PROBLEM - cp5 HTTP 4xx/5xx ERROR Rate on cp5 is CRITICAL: CRITICAL - NGINX Error Rate is 60% [03:45:35] PROBLEM - www.lgbtqia.wiki - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['grannbo.ns.cloudflare.com.', 'rodrigo.ns.cloudflare.com.'], 'CNAME': 'cf-lb.wikitide.org.'} [03:45:36] PROBLEM - wiki.colleimadcat.com - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['dns29.hichina.com.', 'dns30.hichina.com.'], 'CNAME': 'cf-lb.wikitide.org.'} [03:45:39] PROBLEM - data.lophocmatngu.wiki - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['lia.ns.cloudflare.com.', 'gerald.ns.cloudflare.com.'], 'CNAME': 'cf-lb.wikitide.org.'} [03:45:41] PROBLEM - www.greatamerica.wiki - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['john.ns.cloudflare.com.', 'vida.ns.cloudflare.com.'], 'CNAME': 'mw-lb.wikitide.org.'} [03:45:41] PROBLEM - meta.sagan4.org - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['ns59.domaincontrol.com.', 'ns60.domaincontrol.com.'], 'CNAME': 'cf-lb.wikitide.org.'} [03:45:45] PROBLEM - jobrunner1 nutcracker process on jobrunner1 is CRITICAL: CHECK_NRPE: Error - Could not connect to 10.0.0.104: Connection reset by peer [03:45:46] PROBLEM - jobrunner1 nutcracker port on jobrunner1 is CRITICAL: CHECK_NRPE: Error - Could not connect to 10.0.0.104: Connection reset by peer [03:46:01] PROBLEM - www.farthestfrontier.wiki - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['cleo.ns.cloudflare.com.', 'may.ns.cloudflare.com.'], 'CNAME': 'cf-lb.wikitide.org.'} [03:46:01] PROBLEM - wiki.chevrine.com - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['ns1.digitalocean.com.', 'ns2.digitalocean.com.', 'ns3.digitalocean.com.'], 'CNAME': 'mw-lb.wikitide.org.'} [03:46:04] PROBLEM - wikitide.com - reverse DNS on sslhost is WARNING: rDNS WARNING - reverse DNS entry for wikitide.com could not be found [03:46:05] PROBLEM - newcascadia.net - reverse DNS on sslhost is CRITICAL: rDNS CRITICAL - newcascadia.net All nameservers failed to answer the query. [03:46:06] PROBLEM - www.polandballwiki.com - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['john.ns.cloudflare.com.', 'vida.ns.cloudflare.com.'], 'CNAME': 'mw-lb.wikitide.org.'} [03:46:06] PROBLEM - beta.sagan4.org - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['ns59.domaincontrol.com.', 'ns60.domaincontrol.com.'], 'CNAME': 'cf-lb.wikitide.org.'} [03:46:08] PROBLEM - alpha.sagan4.org - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['ns59.domaincontrol.com.', 'ns60.domaincontrol.com.'], 'CNAME': 'cf-lb.wikitide.org.'} [03:46:09] PROBLEM - cp3 ferm_active on cp3 is CRITICAL: ERROR ferm input drop default policy not set, ferm might not have been started correctly [03:46:16] PROBLEM - hsck.lophocmatngu.wiki - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['lia.ns.cloudflare.com.', 'gerald.ns.cloudflare.com.'], 'CNAME': 'cf-lb.wikitide.org.'} [03:46:26] PROBLEM - wiki.myehs.eu - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['fred.ns.cloudflare.com.', 'ivy.ns.cloudflare.com.'], 'CNAME': 'cf-lb.wikitide.org.'} [03:46:40] PROBLEM - projects.dmvpetridish.com - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['ns1.wixdns.net.', 'ns0.wixdns.net.'], 'CNAME': 'mw-lb.wikitide.org.'} [03:46:40] PROBLEM - mlrpgspeedruns.com - reverse DNS on sslhost is CRITICAL: rDNS CRITICAL - mlrpgspeedruns.com All nameservers failed to answer the query. [03:47:00] RECOVERY - mon21 Current Load on mon21 is OK: LOAD OK - total load average: 0.15, 0.25, 5.07 [03:47:14] PROBLEM - mw1 MediaWiki Rendering on mw1 is CRITICAL: connect to address 10.0.0.101 and port 443: Connection refusedHTTP CRITICAL - Unable to open TCP socket [03:47:51] PROBLEM - nexttide.org - reverse DNS on sslhost is CRITICAL: rDNS CRITICAL - nexttide.org All nameservers failed to answer the query. [03:48:04] PROBLEM - mason.sagan4.org - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['ns59.domaincontrol.com.', 'ns60.domaincontrol.com.'], 'CNAME': 'cf-lb.wikitide.org.'} [03:49:14] PROBLEM - mw1 MediaWiki Rendering on mw1 is UNKNOWN: HTTP UNKNOWN: Failed to unchunk message body [03:50:09] RECOVERY - cp3 ferm_active on cp3 is OK: OK ferm input default policy is set [03:51:16] PROBLEM - cp5 HTTP 4xx/5xx ERROR Rate on cp5 is WARNING: WARNING - NGINX Error Rate is 57% [03:51:27] RECOVERY - cp2 Varnish Backends on cp2 is OK: All 7 backends are healthy [03:53:13] PROBLEM - cp3 HTTP 4xx/5xx ERROR Rate on cp3 is WARNING: WARNING - NGINX Error Rate is 42% [03:55:08] RECOVERY - cp3 HTTP 4xx/5xx ERROR Rate on cp3 is OK: OK - NGINX Error Rate is 31% [03:55:10] PROBLEM - cp5 HTTP 4xx/5xx ERROR Rate on cp5 is CRITICAL: CRITICAL - NGINX Error Rate is 67% [03:55:31] RECOVERY - cp5 Varnish Backends on cp5 is OK: All 7 backends are healthy [03:56:05] RECOVERY - cp3 Varnish Backends on cp3 is OK: All 7 backends are healthy [03:57:14] PROBLEM - mw1 MediaWiki Rendering on mw1 is WARNING: HTTP WARNING: HTTP/1.1 404 Not Found - 8191 bytes in 0.109 second response time [03:59:03] PROBLEM - cp5 HTTP 4xx/5xx ERROR Rate on cp5 is WARNING: WARNING - NGINX Error Rate is 56% [03:59:27] PROBLEM - cp2 Varnish Backends on cp2 is CRITICAL: 3 backends are down. mw1 mw2 mediawiki [04:12:46] RECOVERY - cp5 HTTP 4xx/5xx ERROR Rate on cp5 is OK: OK - NGINX Error Rate is 28% [04:54:19] [02WikiTideOrg/mw-config] 07Universal-Omega pushed 031 commit to 03main [+0/-0/±1] 13https://github.com/WikiTideOrg/mw-config/compare/96768d80a999...554e87a91171 [04:54:20] [02WikiTideOrg/mw-config] 07Universal-Omega 03554e87a - Use bots1 [04:55:12] WikiTideOrg/mw-config - Universal-Omega the build passed. [04:59:30] [02WikiTideOrg/mw-config] 07Universal-Omega pushed 031 commit to 03main [+0/-0/±1] 13https://github.com/WikiTideOrg/mw-config/compare/554e87a91171...908d0cbaa72d [04:59:32] [02WikiTideOrg/mw-config] 07Universal-Omega 03908d0cb - Bump $wgAuthenticationTokenVersion [05:00:21] WikiTideOrg/mw-config - Universal-Omega the build passed. [05:01:25] PROBLEM - test1 MediaWiki Rendering on test1 is UNKNOWN: HTTP UNKNOWN: Failed to unchunk message body [05:25:48] PROBLEM - os21 Current Load on os21 is WARNING: LOAD WARNING - total load average: 1.31, 1.81, 1.99 [05:28:42] PROBLEM - ldap21 Current Load on ldap21 is WARNING: LOAD WARNING - total load average: 1.58, 1.85, 1.98 [05:28:52] PROBLEM - jobchron21 Current Load on jobchron21 is WARNING: LOAD WARNING - total load average: 1.78, 1.88, 1.99 [05:30:42] PROBLEM - ldap21 Current Load on ldap21 is CRITICAL: LOAD CRITICAL - total load average: 2.03, 1.92, 1.99 [05:32:52] PROBLEM - jobchron21 Current Load on jobchron21 is CRITICAL: LOAD CRITICAL - total load average: 2.32, 2.01, 2.01 [05:33:47] PROBLEM - os21 Current Load on os21 is CRITICAL: LOAD CRITICAL - total load average: 2.03, 1.98, 2.00 [05:34:42] PROBLEM - ldap21 Current Load on ldap21 is WARNING: LOAD WARNING - total load average: 1.76, 1.94, 1.98 [05:36:52] PROBLEM - jobchron21 Current Load on jobchron21 is WARNING: LOAD WARNING - total load average: 1.78, 1.95, 1.98 [05:38:42] PROBLEM - ldap21 Current Load on ldap21 is CRITICAL: LOAD CRITICAL - total load average: 2.07, 2.00, 2.00 [05:39:48] PROBLEM - os21 Current Load on os21 is WARNING: LOAD WARNING - total load average: 1.98, 1.95, 1.98 [05:40:52] PROBLEM - jobchron21 Current Load on jobchron21 is CRITICAL: LOAD CRITICAL - total load average: 2.22, 2.05, 2.01 [05:44:44] PROBLEM - meta.sagan4.org - reverse DNS on sslhost is UNKNOWN: [05:44:44] PROBLEM - www.greatamerica.wiki - reverse DNS on sslhost is UNKNOWN: [05:44:44] PROBLEM - os21 Current Load on os21 is CRITICAL: LOAD CRITICAL - total load average: 3.11, 2.41, 2.15 [05:45:54] PROBLEM - wiki.myehs.eu - reverse DNS on sslhost is UNKNOWN: PROBLEM - www.polandballwiki.com - reverse DNS on sslhost is UNKNOWN: PROBLEM - puppet21 HTTPS on puppet21 is UNKNOWN: PROBLEM - hsck.lophocmatngu.wiki - reverse DNS on sslhost is UNKNOWN: minated by signal 15 (Terminated).> [05:45:55] PROBLEM - beta.sagan4.org - reverse DNS on sslhost is UNKNOWN: [05:47:35] PROBLEM - projects.dmvpetridish.com - reverse DNS on sslhost is UNKNOWN: PROBLEM - mw21 HTTPS on mw21 is UNKNOWN: PROBLEM - mail21 HTTPS on mail21 is UNKNOWN: [05:47:46] PROBLEM - mlrpgspeedruns.com - reverse DNS on sslhost is UNKNOWN: [05:49:17] PROBLEM - cp2 HTTPS on cp2 is UNKNOWN: PROBLEM - matomo21 HTTPS on matomo21 is UNKNOWN: PROBLEM - nexttide.org - reverse DNS on sslhost is UNKNOWN: [06:23:14] PROBLEM - mon21 APT on mon21 is UNKNOWN: [06:24:16] PROBLEM - mason.sagan4.org - reverse DNS on sslhost is UNKNOWN: [06:24:21] PROBLEM - mon21 Current Load on mon21 is CRITICAL: LOAD CRITICAL - total load average: 33.22, 32.32, 28.50 [06:24:27] PROBLEM - cp2 HTTPS on cp2 is CRITICAL: HTTP CRITICAL - Invalid HTTP response received from host on port 443: cURL returned 7 - Failed to connect to cp2.wikitide.net port 443 after 18 ms: Couldn't connect to server [06:25:54] PROBLEM - mail21 HTTPS on mail21 is CRITICAL: HTTP CRITICAL - Invalid HTTP response received from host on port 443: cURL returned 28 - Failed to connect to mail21.wikitide.net port 443 after 12382 ms: Timeout was reached [06:25:54] PROBLEM - realitylongterms.studio - reverse DNS on sslhost is UNKNOWN: [06:25:54] PROBLEM - mw21 HTTPS on mw21 is CRITICAL: HTTP CRITICAL - Invalid HTTP response received from host on port 443: cURL returned 28 - Failed to connect to mw21.wikitide.net port 443 after 12383 ms: Timeout was reached [06:25:54] PROBLEM - matomo21 HTTPS on matomo21 is CRITICAL: HTTP CRITICAL - Invalid HTTP response received from host on port 443: cURL returned 28 - Connection timed out after 17542 milliseconds [06:25:54] PROBLEM - www.lgbtqia.wiki - reverse DNS on sslhost is UNKNOWN: [06:25:55] PROBLEM - newcascadia.net - reverse DNS on sslhost is UNKNOWN: [06:25:55] PROBLEM - wikitide.com - reverse DNS on sslhost is UNKNOWN: [06:25:56] PROBLEM - data.lophocmatngu.wiki - reverse DNS on sslhost is UNKNOWN: [06:25:56] PROBLEM - www.farthestfrontier.wiki - reverse DNS on sslhost is UNKNOWN: [06:25:57] PROBLEM - www.wikitide.org - reverse DNS on sslhost is UNKNOWN: [06:25:57] PROBLEM - wiki.chevrine.com - reverse DNS on sslhost is UNKNOWN: [06:25:58] PROBLEM - wiki.colleimadcat.com - reverse DNS on sslhost is UNKNOWN: [06:25:58] PROBLEM - mem21 Current Load on mem21 is CRITICAL: LOAD CRITICAL - total load average: 5.85, 6.33, 5.18 [06:27:02] PROBLEM - mon21 APT on mon21 is CRITICAL: CHECK_NRPE STATE CRITICAL: Socket timeout after 60 seconds. [06:27:03] PROBLEM - alpha.sagan4.org - reverse DNS on sslhost is UNKNOWN: [06:29:09] PROBLEM - puppet21 HTTPS on puppet21 is CRITICAL: HTTP CRITICAL - Invalid HTTP response received from host on port 443: cURL returned 28 - Connection timed out after 10004 milliseconds [06:35:26] PROBLEM - mem21 Current Load on mem21 is WARNING: LOAD WARNING - total load average: 2.46, 3.01, 3.97 [06:41:09] RECOVERY - mem21 Current Load on mem21 is OK: LOAD OK - total load average: 2.27, 2.41, 3.40 [06:54:16] PROBLEM - mason.sagan4.org - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['ns59.domaincontrol.com.', 'ns60.domaincontrol.com.'], 'CNAME': 'cf-lb.wikitide.org.'} [06:54:28] PROBLEM - www.lgbtqia.wiki - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['grannbo.ns.cloudflare.com.', 'rodrigo.ns.cloudflare.com.'], 'CNAME': 'cf-lb.wikitide.org.'} [06:54:29] PROBLEM - wiki.colleimadcat.com - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['dns29.hichina.com.', 'dns30.hichina.com.'], 'CNAME': 'cf-lb.wikitide.org.'} [06:54:32] PROBLEM - data.lophocmatngu.wiki - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['lia.ns.cloudflare.com.', 'gerald.ns.cloudflare.com.'], 'CNAME': 'cf-lb.wikitide.org.'} [06:54:34] RECOVERY - realitylongterms.studio - reverse DNS on sslhost is OK: SSL OK - realitylongterms.studio reverse DNS resolves to cp5.wikitide.net - CNAME FLAT [06:54:53] PROBLEM - www.farthestfrontier.wiki - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['cleo.ns.cloudflare.com.', 'may.ns.cloudflare.com.'], 'CNAME': 'cf-lb.wikitide.org.'} [06:54:54] PROBLEM - wiki.chevrine.com - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['ns1.digitalocean.com.', 'ns2.digitalocean.com.', 'ns3.digitalocean.com.'], 'CNAME': 'mw-lb.wikitide.org.'} [06:54:54] RECOVERY - www.wikitide.org - reverse DNS on sslhost is OK: SSL OK - www.wikitide.org reverse DNS resolves to cp3.wikitide.net - NS RECORDS OK [06:54:55] PROBLEM - newcascadia.net - reverse DNS on sslhost is CRITICAL: rDNS CRITICAL - newcascadia.net All nameservers failed to answer the query. [06:54:57] PROBLEM - wikitide.com - reverse DNS on sslhost is WARNING: rDNS WARNING - reverse DNS entry for wikitide.com could not be found [06:55:00] PROBLEM - mon21 Current Load on mon21 is WARNING: LOAD WARNING - total load average: 0.25, 0.25, 5.31 [06:55:42] PROBLEM - meta.sagan4.org - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['ns59.domaincontrol.com.', 'ns60.domaincontrol.com.'], 'CNAME': 'cf-lb.wikitide.org.'} [06:55:42] PROBLEM - www.greatamerica.wiki - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['john.ns.cloudflare.com.', 'vida.ns.cloudflare.com.'], 'CNAME': 'mw-lb.wikitide.org.'} [06:56:06] PROBLEM - beta.sagan4.org - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['ns59.domaincontrol.com.', 'ns60.domaincontrol.com.'], 'CNAME': 'cf-lb.wikitide.org.'} [06:56:06] PROBLEM - www.polandballwiki.com - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['john.ns.cloudflare.com.', 'vida.ns.cloudflare.com.'], 'CNAME': 'mw-lb.wikitide.org.'} [06:56:08] PROBLEM - alpha.sagan4.org - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['ns59.domaincontrol.com.', 'ns60.domaincontrol.com.'], 'CNAME': 'cf-lb.wikitide.org.'} [06:56:17] PROBLEM - hsck.lophocmatngu.wiki - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['lia.ns.cloudflare.com.', 'gerald.ns.cloudflare.com.'], 'CNAME': 'cf-lb.wikitide.org.'} [06:56:26] PROBLEM - wiki.myehs.eu - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['fred.ns.cloudflare.com.', 'ivy.ns.cloudflare.com.'], 'CNAME': 'cf-lb.wikitide.org.'} [06:56:41] PROBLEM - projects.dmvpetridish.com - reverse DNS on sslhost is WARNING: SSL WARNING - rDNS OK but records conflict. {'NS': ['ns1.wixdns.net.', 'ns0.wixdns.net.'], 'CNAME': 'mw-lb.wikitide.org.'} [06:56:45] PROBLEM - mlrpgspeedruns.com - reverse DNS on sslhost is CRITICAL: rDNS CRITICAL - mlrpgspeedruns.com All nameservers failed to answer the query. [06:56:50] PROBLEM - nexttide.org - reverse DNS on sslhost is CRITICAL: rDNS CRITICAL - nexttide.org All nameservers failed to answer the query. [06:57:00] RECOVERY - mon21 Current Load on mon21 is OK: LOAD OK - total load average: 0.21, 0.24, 4.68 [07:00:47] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is WARNING: WARNING - NGINX Error Rate is 41% [07:02:46] RECOVERY - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is OK: OK - NGINX Error Rate is 39% [07:23:59] PROBLEM - newcascadia.net - reverse DNS on sslhost is WARNING: LifetimeTimeout: The resolution lifetime expired after 5.405 seconds: Server 1.1.1.1 UDP port 53 answered The DNS operation timed out.; Server 1.1.1.1 UDP port 53 answered The DNS operation timed out.; Server 1.1.1.1 UDP port 53 answered The DNS operation timed out. [07:28:35] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is WARNING: WARNING - NGINX Error Rate is 40% [07:30:35] RECOVERY - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is OK: OK - NGINX Error Rate is 39% [07:36:47] RECOVERY - cp6 Varnish Backends on cp6 is OK: All 7 backends are healthy [07:38:51] RECOVERY - cp6 Puppet on cp6 is OK: OK: Puppet is currently enabled, last run 1 minute ago with 0 failures [07:41:28] PROBLEM - cp6 conntrack_table_size on cp6 is CRITICAL: CHECK_NRPE: Error - Could not connect to 139.99.236.151: Connection reset by peer [07:41:32] PROBLEM - cp6 Nginx Backend for test1 on cp6 is CRITICAL: CHECK_NRPE: Error - Could not connect to 139.99.236.151: Connection reset by peer [07:41:44] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is CRITICAL: CHECK_NRPE: Error - Could not connect to 139.99.236.151: Connection reset by peer [07:41:52] PROBLEM - cp6 Nginx Backend for phorge1 on cp6 is CRITICAL: CHECK_NRPE: Error - Could not connect to 139.99.236.151: Connection reset by peer [07:41:56] PROBLEM - cp6 PowerDNS Recursor on cp6 is CRITICAL: CHECK_NRPE: Error - Could not connect to 139.99.236.151: Connection reset by peer [07:41:58] PROBLEM - cp6 NTP time on cp6 is CRITICAL: CHECK_NRPE: Error - Could not connect to 139.99.236.151: Connection reset by peer [07:42:06] PROBLEM - cp6 Nginx Backend for mw1 on cp6 is CRITICAL: CHECK_NRPE: Error - Could not connect to 139.99.236.151: Connection reset by peer [07:42:08] PROBLEM - cp6 Disk Space on cp6 is CRITICAL: CHECK_NRPE: Error - Could not connect to 139.99.236.151: Connection reset by peer [07:42:10] PROBLEM - cp6 ferm_active on cp6 is CRITICAL: CHECK_NRPE: Error - Could not connect to 139.99.236.151: Connection reset by peer [07:42:11] PROBLEM - cp6 Nginx Backend for puppet1 on cp6 is CRITICAL: CHECK_NRPE: Error - Could not connect to 139.99.236.151: Connection reset by peer [07:42:11] PROBLEM - cp6 Varnish Backends on cp6 is CRITICAL: CHECK_NRPE: Error - Could not connect to 139.99.236.151: Connection reset by peer [07:42:14] PROBLEM - cp6 APT on cp6 is CRITICAL: CHECK_NRPE: Error - Could not connect to 139.99.236.151: Connection reset by peer [07:42:16] PROBLEM - cp6 Nginx Backend for mw2 on cp6 is CRITICAL: CHECK_NRPE: Error - Could not connect to 139.99.236.151: Connection reset by peer [07:42:16] PROBLEM - cp6 Puppet on cp6 is CRITICAL: CHECK_NRPE: Error - Could not connect to 139.99.236.151: Connection reset by peer [07:42:18] PROBLEM - cp6 Current Load on cp6 is CRITICAL: CHECK_NRPE: Error - Could not connect to 139.99.236.151: Connection reset by peer [07:42:19] PROBLEM - cp6 Nginx Backend for mail1 on cp6 is CRITICAL: CHECK_NRPE: Error - Could not connect to 139.99.236.151: Connection reset by peer [07:42:22] PROBLEM - cp6 Nginx Backend for matomo1 on cp6 is CRITICAL: CHECK_NRPE: Error - Could not connect to 139.99.236.151: Connection reset by peer [07:45:11] RECOVERY - cp3 Puppet on cp3 is OK: OK: Puppet is currently enabled, last run 49 seconds ago with 0 failures [07:50:22] PROBLEM - cp2 HTTPS on cp2 is UNKNOWN: [07:51:48] PROBLEM - mail21 HTTPS on mail21 is UNKNOWN: [07:52:02] RECOVERY - cp5 Puppet on cp5 is OK: OK: Puppet is currently enabled, last run 7 seconds ago with 0 failures [07:53:04] RECOVERY - cp4 Varnish Backends on cp4 is OK: All 7 backends are healthy [07:53:44] PROBLEM - matomo21 HTTPS on matomo21 is UNKNOWN: [07:54:00] RECOVERY - cp4 Puppet on cp4 is OK: OK: Puppet is currently enabled, last run 3 minutes ago with 0 failures [07:54:39] PROBLEM - puppet21 HTTPS on puppet21 is UNKNOWN: [07:57:51] PROBLEM - data.lophocmatngu.wiki - reverse DNS on sslhost is UNKNOWN: PROBLEM - wiki.colleimadcat.com - reverse DNS on sslhost is UNKNOWN: PROBLEM - www.lgbtqia.wiki - reverse DNS on sslhost is UNKNOWN: [07:59:09] PROBLEM - meta.sagan4.org - reverse DNS on sslhost is UNKNOWN: [07:59:17] PROBLEM - www.greatamerica.wiki - reverse DNS on sslhost is UNKNOWN: [08:00:49] PROBLEM - www.farthestfrontier.wiki - reverse DNS on sslhost is UNKNOWN: [08:11:36] PROBLEM - cp4 Nginx Backend for puppet1 on cp4 is CRITICAL: CHECK_NRPE: Error - Could not connect to 146.59.44.171: Connection reset by peer [08:11:38] PROBLEM - cp4 Nginx Backend for phorge1 on cp4 is CRITICAL: CHECK_NRPE: Error - Could not connect to 146.59.44.171: Connection reset by peer [08:11:40] PROBLEM - cp5 APT on cp5 is CRITICAL: CHECK_NRPE: Error - Could not connect to 15.235.167.159: Connection reset by peer [08:11:41] PROBLEM - cp5 Nginx Backend for matomo1 on cp5 is CRITICAL: CHECK_NRPE: Error - Could not connect to 15.235.167.159: Connection reset by peer [08:11:41] PROBLEM - cp5 Puppet on cp5 is CRITICAL: CHECK_NRPE: Error - Could not connect to 15.235.167.159: Connection reset by peer [08:11:42] PROBLEM - cp4 PowerDNS Recursor on cp4 is CRITICAL: CHECK_NRPE: Error - Could not connect to 146.59.44.171: Connection reset by peer [08:11:43] PROBLEM - cp3 Nginx Backend for mail1 on cp3 is CRITICAL: CHECK_NRPE: Error - Could not connect to 51.75.170.66: Connection reset by peer [08:11:43] PROBLEM - cp3 Nginx Backend for mw1 on cp3 is CRITICAL: CHECK_NRPE: Error - Could not connect to 51.75.170.66: Connection reset by peer [08:11:43] PROBLEM - cp5 conntrack_table_size on cp5 is CRITICAL: CHECK_NRPE: Error - Could not connect to 15.235.167.159: Connection reset by peer [08:11:44] PROBLEM - cp4 APT on cp4 is CRITICAL: CHECK_NRPE: Error - Could not connect to 146.59.44.171: Connection reset by peer [08:11:44] PROBLEM - cp4 Nginx Backend for mail1 on cp4 is CRITICAL: CHECK_NRPE: Error - Could not connect to 146.59.44.171: Connection reset by peer [08:11:46] PROBLEM - cp5 Nginx Backend for phorge1 on cp5 is CRITICAL: CHECK_NRPE: Error - Could not connect to 15.235.167.159: Connection reset by peer [08:11:46] PROBLEM - cp4 Nginx Backend for matomo1 on cp4 is CRITICAL: CHECK_NRPE: Error - Could not connect to 146.59.44.171: Connection reset by peer [08:11:48] PROBLEM - cp5 Nginx Backend for mw2 on cp5 is CRITICAL: CHECK_NRPE: Error - Could not connect to 15.235.167.159: Connection reset by peer [08:11:48] PROBLEM - cp5 Nginx Backend for test1 on cp5 is CRITICAL: CHECK_NRPE: Error - Could not connect to 15.235.167.159: Connection reset by peer [08:11:49] PROBLEM - cp4 Current Load on cp4 is CRITICAL: CHECK_NRPE: Error - Could not connect to 146.59.44.171: Connection reset by peer [08:11:50] PROBLEM - cp5 ferm_active on cp5 is CRITICAL: CHECK_NRPE: Error - Could not connect to 15.235.167.159: Connection reset by peer [08:11:51] PROBLEM - cp3 Nginx Backend for matomo1 on cp3 is CRITICAL: CHECK_NRPE: Error - Could not connect to 51.75.170.66: Connection reset by peer [08:11:51] PROBLEM - cp3 HTTP 4xx/5xx ERROR Rate on cp3 is CRITICAL: CHECK_NRPE: Error - Could not connect to 51.75.170.66: Connection reset by peer [08:11:51] PROBLEM - cp4 Varnish Backends on cp4 is CRITICAL: CHECK_NRPE: Error - Could not connect to 146.59.44.171: Connection reset by peer [08:11:52] PROBLEM - cp3 conntrack_table_size on cp3 is CRITICAL: CHECK_NRPE: Error - Could not connect to 51.75.170.66: Connection reset by peer [08:11:52] PROBLEM - cp4 Nginx Backend for mw1 on cp4 is CRITICAL: CHECK_NRPE: Error - Could not connect to 146.59.44.171: Connection reset by peer [08:11:54] PROBLEM - cp5 Disk Space on cp5 is CRITICAL: CHECK_NRPE: Error - Could not connect to 15.235.167.159: Connection reset by peer [08:11:54] PROBLEM - cp5 Nginx Backend for puppet1 on cp5 is CRITICAL: CHECK_NRPE: Error - Could not connect to 15.235.167.159: Connection reset by peer [08:11:55] PROBLEM - cp4 Disk Space on cp4 is CRITICAL: CHECK_NRPE: Error - Could not connect to 146.59.44.171: Connection reset by peer [08:11:55] PROBLEM - cp4 conntrack_table_size on cp4 is CRITICAL: CHECK_NRPE: Error - Could not connect to 146.59.44.171: Connection reset by peer [08:11:57] PROBLEM - cp3 Nginx Backend for phorge1 on cp3 is CRITICAL: CHECK_NRPE: Error - Could not connect to 51.75.170.66: Connection reset by peer [08:12:01] PROBLEM - cp5 PowerDNS Recursor on cp5 is CRITICAL: CHECK_NRPE: Error - Could not connect to 15.235.167.159: Connection reset by peer [08:12:05] PROBLEM - cp3 Nginx Backend for puppet1 on cp3 is CRITICAL: CHECK_NRPE: Error - Could not connect to 51.75.170.66: Connection reset by peer [08:12:06] PROBLEM - cp3 Varnish Backends on cp3 is CRITICAL: CHECK_NRPE: Error - Could not connect to 51.75.170.66: Connection reset by peer [08:12:07] PROBLEM - cp3 Nginx Backend for test1 on cp3 is CRITICAL: CHECK_NRPE: Error - Could not connect to 51.75.170.66: Connection reset by peer [08:12:08] PROBLEM - cp3 APT on cp3 is CRITICAL: CHECK_NRPE: Error - Could not connect to 51.75.170.66: Connection reset by peer [08:12:12] PROBLEM - cp5 Current Load on cp5 is CRITICAL: CHECK_NRPE: Error - Could not connect to 15.235.167.159: Connection reset by peer [08:12:17] PROBLEM - cp5 HTTP 4xx/5xx ERROR Rate on cp5 is CRITICAL: CHECK_NRPE: Error - Could not connect to 15.235.167.159: Connection reset by peer [08:12:20] PROBLEM - cp3 Puppet on cp3 is CRITICAL: CHECK_NRPE: Error - Could not connect to 51.75.170.66: Connection reset by peer [08:12:20] PROBLEM - cp5 Varnish Backends on cp5 is CRITICAL: CHECK_NRPE: Error - Could not connect to 15.235.167.159: Connection reset by peer [08:12:22] PROBLEM - cp4 Puppet on cp4 is CRITICAL: CHECK_NRPE: Error - Could not connect to 146.59.44.171: Connection reset by peer [08:12:23] PROBLEM - cp4 ferm_active on cp4 is CRITICAL: CHECK_NRPE: Error - Could not connect to 146.59.44.171: Connection reset by peer [08:12:24] PROBLEM - cp3 GDNSD Datacenters on cp3 is CRITICAL: CHECK_NRPE: Error - Could not connect to 51.75.170.66: Connection reset by peer [08:12:25] PROBLEM - cp3 Disk Space on cp3 is CRITICAL: CHECK_NRPE: Error - Could not connect to 51.75.170.66: Connection reset by peer [08:12:25] PROBLEM - cp3 Current Load on cp3 is CRITICAL: CHECK_NRPE: Error - Could not connect to 51.75.170.66: Connection reset by peer [08:12:26] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is CRITICAL: CHECK_NRPE: Error - Could not connect to 146.59.44.171: Connection reset by peer [08:12:26] PROBLEM - cp5 NTP time on cp5 is CRITICAL: CHECK_NRPE: Error - Could not connect to 15.235.167.159: Connection reset by peer [08:12:27] PROBLEM - cp3 ferm_active on cp3 is CRITICAL: CHECK_NRPE: Error - Could not connect to 51.75.170.66: Connection reset by peer [08:12:29] PROBLEM - cp3 NTP time on cp3 is CRITICAL: CHECK_NRPE: Error - Could not connect to 51.75.170.66: Connection reset by peer [08:12:30] PROBLEM - cp5 Nginx Backend for mail1 on cp5 is CRITICAL: CHECK_NRPE: Error - Could not connect to 15.235.167.159: Connection reset by peer [08:12:30] PROBLEM - cp3 Nginx Backend for mw2 on cp3 is CRITICAL: CHECK_NRPE: Error - Could not connect to 51.75.170.66: Connection reset by peer [08:12:31] PROBLEM - cp4 Nginx Backend for test1 on cp4 is CRITICAL: CHECK_NRPE: Error - Could not connect to 146.59.44.171: Connection reset by peer [08:12:31] PROBLEM - cp5 Nginx Backend for mw1 on cp5 is CRITICAL: CHECK_NRPE: Error - Could not connect to 15.235.167.159: Connection reset by peer [08:12:32] PROBLEM - cp4 NTP time on cp4 is CRITICAL: CHECK_NRPE: Error - Could not connect to 146.59.44.171: Connection reset by peer [08:12:36] PROBLEM - cp4 Nginx Backend for mw2 on cp4 is CRITICAL: CHECK_NRPE: Error - Could not connect to 146.59.44.171: Connection reset by peer [08:14:56] [02WikiTideOrg/puppet] 07Universal-Omega pushed 031 commit to 03master [+0/-0/±1] 13https://github.com/WikiTideOrg/puppet/compare/1db17177b6bf...4edfb960d53f [08:14:58] [02WikiTideOrg/puppet] 07Universal-Omega 034edfb96 - Update allowed_hosts [08:15:40] [02WikiTideOrg/puppet] 07Universal-Omega pushed 031 commit to 03master [+0/-0/±1] 13https://github.com/WikiTideOrg/puppet/compare/4edfb960d53f...3cafc5f69073 [08:15:43] [02WikiTideOrg/puppet] 07Universal-Omega 033cafc5f - Use cloud1 [08:17:38] [02WikiTideOrg/puppet] 07Universal-Omega pushed 0316 commits to 03mwdeploy-private-rsync [+48/-46/±62] 13https://github.com/WikiTideOrg/puppet/compare/efed1bd784e2...45f53be2adaa [08:17:39] [02WikiTideOrg/puppet] 07Universal-Omega 0345f53be - Merge branch 'master' into mwdeploy-private-rsync [08:17:42] [02puppet] 07Universal-Omega synchronize pull request 03#45: mwdeploy: support using -private in rsync command - 13https://github.com/WikiTideOrg/puppet/pull/45 [08:19:09] WikiTideOrg/puppet - Universal-Omega the build has errored. [08:19:22] RECOVERY - puppet1 NTP time on puppet1 is OK: NTP OK: Offset 0.004167079926 secs [08:19:24] RECOVERY - bots1 IRC Log Server Bot on bots1 is OK: PROCS OK: 1 process with args 'irclogserverbot.py' [08:19:27] RECOVERY - puppet1 Backups SSLKeys on puppet1 is OK: FILE_AGE OK: /var/log/sslkeys-backup.log is 8363 seconds old and 52 bytes [08:19:32] RECOVERY - bots1 IRC RC Bot on bots1 is OK: PROCS OK: 1 process with args 'ircrcbot.py' [08:19:49] RECOVERY - bots1 NTP time on bots1 is OK: NTP OK: Offset -0.0008465349674 secs [08:19:56] RECOVERY - bots1 ferm_active on bots1 is OK: OK ferm input default policy is set [08:19:57] RECOVERY - matomo1 NTP time on matomo1 is OK: NTP OK: Offset 0.002069294453 secs [08:19:59] [02WikiTideOrg/puppet] 07Universal-Omega pushed 031 commit to 03master [+0/-0/±1] 13https://github.com/WikiTideOrg/puppet/compare/3cafc5f69073...6a8478fced64 [08:20:01] [02WikiTideOrg/puppet] 07Universal-Omega 036a8478f - Update servers in tests [08:20:06] RECOVERY - puppet1 PowerDNS Recursor on puppet1 is OK: DNS OK: 0.451 seconds response time. wikitide.org returns 2001:41d0:801:2000::4089,51.75.170.66 [08:20:06] RECOVERY - cp3 NTP time on cp3 is OK: NTP OK: Offset -0.001378625631 secs [08:20:09] RECOVERY - cp3 Current Load on cp3 is OK: LOAD OK - total load average: 0.56, 0.29, 0.24 [08:20:09] RECOVERY - bots1 IRC-Discord Relay Bot on bots1 is OK: PROCS OK: 2 processes with args 'relaybot' [08:20:13] RECOVERY - puppet1 ferm_active on puppet1 is OK: OK ferm input default policy is set [08:20:15] RECOVERY - cp3 Disk Space on cp3 is OK: DISK OK - free space: / 5841MiB (30% inode=94%); [08:20:15] [02WikiTideOrg/puppet] 07Universal-Omega pushed 032 commits to 03mwdeploy-private-rsync [+0/-0/±2] 13https://github.com/WikiTideOrg/puppet/compare/45f53be2adaa...cd959a59cebe [08:20:16] RECOVERY - cp3 ferm_active on cp3 is OK: OK ferm input default policy is set [08:20:16] [02WikiTideOrg/puppet] 07Universal-Omega 03cd959a5 - Merge branch 'master' into mwdeploy-private-rsync [08:20:17] RECOVERY - bots1 Disk Space on bots1 is OK: DISK OK - free space: / 12699MiB (74% inode=91%); [08:20:17] RECOVERY - matomo1 ferm_active on matomo1 is OK: OK ferm input default policy is set [08:20:19] [02puppet] 07Universal-Omega synchronize pull request 03#45: mwdeploy: support using -private in rsync command - 13https://github.com/WikiTideOrg/puppet/pull/45 [08:20:24] RECOVERY - bots1 PowerDNS Recursor on bots1 is OK: DNS OK: 0.309 seconds response time. wikitide.org returns 2001:41d0:801:2000::4089,51.75.170.66 [08:20:25] RECOVERY - bots1 conntrack_table_size on bots1 is OK: OK: nf_conntrack is 0 % full [08:20:26] RECOVERY - cp3 Nginx Backend for mw2 on cp3 is OK: TCP OK - 0.001 second response time on localhost port 8101 [08:20:28] RECOVERY - bots1 CVT Bot on bots1 is OK: PROCS OK: 2 processes with args 'cvtbot' [08:20:38] RECOVERY - matomo1 PowerDNS Recursor on matomo1 is OK: DNS OK: 0.263 seconds response time. wikitide.org returns 2001:41d0:801:2000::4089,51.75.170.66 [08:20:38] RECOVERY - puppet1 Current Load on puppet1 is OK: LOAD OK - total load average: 0.83, 0.42, 0.30 [08:20:47] RECOVERY - bots1 IRC Log Bot on bots1 is OK: PROCS OK: 1 process with args 'adminlogbot.py' [08:20:53] RECOVERY - puppet1 conntrack_table_size on puppet1 is OK: OK: nf_conntrack is 0 % full [08:20:54] RECOVERY - bots1 IRC Limnoria Bot on bots1 is OK: PROCS OK: 1 process with args 'limnoria' [08:20:59] RECOVERY - mon1 NTP time on mon1 is OK: NTP OK: Offset -0.004622936249 secs [08:21:01] RECOVERY - puppet1 Puppet on puppet1 is OK: OK: Puppet is currently enabled, last run 1 minute ago with 0 failures [08:21:02] RECOVERY - mon1 Puppet on mon1 is OK: OK: Puppet is currently enabled, last run 9 seconds ago with 0 failures [08:21:05] RECOVERY - bots1 Puppet on bots1 is OK: OK: Puppet is currently enabled, last run 1 minute ago with 0 failures [08:21:06] RECOVERY - mon1 Disk Space on mon1 is OK: DISK OK - free space: / 13412MiB (78% inode=91%); [08:21:07] RECOVERY - bots1 Current Load on bots1 is OK: LOAD OK - total load average: 0.08, 0.05, 0.03 [08:21:12] RECOVERY - puppet1 Backups Private on puppet1 is OK: FILE_AGE OK: /var/log/private-backup.log is 19148 seconds old and 93 bytes [08:21:16] RECOVERY - mon1 IRCEcho on mon1 is OK: PROCS OK: 1 process with args '/usr/local/bin/ircecho' [08:21:21] RECOVERY - cp5 Nginx Backend for test1 on cp5 is OK: TCP OK - 0.000 second response time on localhost port 8180 [08:21:21] RECOVERY - cp5 Nginx Backend for matomo1 on cp5 is OK: TCP OK - 0.000 second response time on localhost port 8204 [08:21:22] RECOVERY - cp5 Nginx Backend for phorge1 on cp5 is OK: TCP OK - 0.000 second response time on localhost port 8200 [08:21:22] RECOVERY - mon1 Check correctness of the icinga configuration on mon1 is OK: Icinga configuration is correct [08:21:27] RECOVERY - cp5 Nginx Backend for mw2 on cp5 is OK: TCP OK - 0.000 second response time on localhost port 8101 [08:21:27] RECOVERY - matomo1 conntrack_table_size on matomo1 is OK: OK: nf_conntrack is 0 % full [08:21:30] RECOVERY - cp2 ferm_active on cp2 is OK: OK ferm input default policy is set [08:21:31] RECOVERY - cp3 conntrack_table_size on cp3 is OK: OK: nf_conntrack is 1 % full [08:21:31] RECOVERY - matomo1 php-fpm on matomo1 is OK: PROCS OK: 9 processes with command name 'php-fpm8.2' [08:21:31] RECOVERY - cp5 Puppet on cp5 is OK: OK: Puppet is currently enabled, last run 22 seconds ago with 0 failures [08:21:32] RECOVERY - cp5 conntrack_table_size on cp5 is OK: OK: nf_conntrack is 1 % full [08:21:32] RECOVERY - cp5 Disk Space on cp5 is OK: DISK OK - free space: / 6038MiB (31% inode=94%); [08:21:32] RECOVERY - cp3 Nginx Backend for mw1 on cp3 is OK: TCP OK - 0.005 second response time on localhost port 8100 [08:21:33] RECOVERY - matomo1 Current Load on matomo1 is OK: LOAD OK - total load average: 0.23, 0.19, 0.16 [08:21:34] RECOVERY - cp2 Nginx Backend for puppet1 on cp2 is OK: TCP OK - 0.000 second response time on localhost port 8201 [08:21:36] RECOVERY - cp5 Nginx Backend for puppet1 on cp5 is OK: TCP OK - 0.000 second response time on localhost port 8201 [08:21:38] RECOVERY - cp3 Nginx Backend for mail1 on cp3 is OK: TCP OK - 0.000 second response time on localhost port 8202 [08:21:38] RECOVERY - matomo1 Disk Space on matomo1 is OK: DISK OK - free space: / 13422MiB (78% inode=91%); [08:21:38] RECOVERY - mon1 PowerDNS Recursor on mon1 is OK: DNS OK: 0.034 seconds response time. wikitide.org returns 2001:41d0:801:2000::4089,51.75.170.66 [08:21:39] RECOVERY - cp3 Nginx Backend for phorge1 on cp3 is OK: TCP OK - 0.000 second response time on localhost port 8200 [08:21:40] WikiTideOrg/puppet - Universal-Omega the build passed. [08:21:41] RECOVERY - matomo1 Puppet on matomo1 is OK: OK: Puppet is currently enabled, last run 1 minute ago with 0 failures [08:21:43] RECOVERY - cp5 PowerDNS Recursor on cp5 is OK: DNS OK: 0.787 seconds response time. wikitide.org returns 15.235.167.159,2402:1f00:8000:800::1b3d [08:21:43] RECOVERY - cp3 Varnish Backends on cp3 is OK: All 7 backends are healthy [08:21:46] RECOVERY - cp3 HTTP 4xx/5xx ERROR Rate on cp3 is OK: OK - NGINX Error Rate is 5% [08:21:46] RECOVERY - cp3 Nginx Backend for matomo1 on cp3 is OK: TCP OK - 0.000 second response time on localhost port 8204 [08:21:46] RECOVERY - cp5 ferm_active on cp5 is OK: OK ferm input default policy is set [08:21:47] RECOVERY - cp3 Nginx Backend for test1 on cp3 is OK: TCP OK - 0.000 second response time on localhost port 8180 [08:21:50] RECOVERY - mon1 Backups Grafana on mon1 is OK: FILE_AGE OK: /var/log/grafana-backup.log is 19309 seconds old and 347 bytes [08:21:51] RECOVERY - mon1 ferm_active on mon1 is OK: OK ferm input default policy is set [08:21:51] RECOVERY - cp2 Disk Space on cp2 is OK: DISK OK - free space: / 6474MiB (33% inode=94%); [08:21:52] RECOVERY - mon1 php-fpm on mon1 is OK: PROCS OK: 16 processes with command name 'php-fpm8.2' [08:21:54] RECOVERY - cp3 Nginx Backend for puppet1 on cp3 is OK: TCP OK - 0.000 second response time on localhost port 8201 [08:21:55] RECOVERY - cp5 Current Load on cp5 is OK: LOAD OK - total load average: 0.39, 0.19, 0.11 [08:21:57] RECOVERY - cp3 Puppet on cp3 is OK: OK: Puppet is currently enabled, last run 1 minute ago with 0 failures [08:21:59] RECOVERY - cp2 Nginx Backend for phorge1 on cp2 is OK: TCP OK - 0.000 second response time on localhost port 8200 [08:22:01] WikiTideOrg/puppet - Universal-Omega the build passed. [08:22:03] RECOVERY - cp2 APT on cp2 is OK: APT OK: 0 packages available for upgrade (0 critical updates). [08:22:04] RECOVERY - cp2 PowerDNS Recursor on cp2 is OK: DNS OK: 0.271 seconds response time. wikitide.org returns 2001:41d0:801:2000::4089,51.75.170.66 [08:22:06] RECOVERY - cp2 NTP time on cp2 is OK: NTP OK: Offset 0.001832544804 secs [08:22:10] RECOVERY - cp5 Nginx Backend for mw1 on cp5 is OK: TCP OK - 0.000 second response time on localhost port 8100 [08:22:11] RECOVERY - cp2 Nginx Backend for mw2 on cp2 is OK: TCP OK - 0.001 second response time on localhost port 8101 [08:22:14] RECOVERY - cp2 conntrack_table_size on cp2 is OK: OK: nf_conntrack is 0 % full [08:22:15] RECOVERY - cp5 Varnish Backends on cp5 is OK: All 7 backends are healthy [08:22:15] RECOVERY - mon1 conntrack_table_size on mon1 is OK: OK: nf_conntrack is 0 % full [08:22:17] RECOVERY - cp5 HTTP 4xx/5xx ERROR Rate on cp5 is OK: OK - NGINX Error Rate is 13% [08:22:19] RECOVERY - ns1 NTP time on ns1 is OK: NTP OK: Offset 0.002223879099 secs [08:22:22] RECOVERY - mon1 Current Load on mon1 is OK: LOAD OK - total load average: 0.32, 0.21, 0.13 [08:22:29] RECOVERY - cp5 Nginx Backend for mail1 on cp5 is OK: TCP OK - 0.000 second response time on localhost port 8202 [08:22:33] RECOVERY - ns1 Current Load on ns1 is OK: LOAD OK - total load average: 0.26, 0.12, 0.04 [08:22:44] RECOVERY - ns1 ferm_active on ns1 is OK: OK ferm input default policy is set [08:22:46] RECOVERY - cp2 HTTP 4xx/5xx ERROR Rate on cp2 is OK: OK - NGINX Error Rate is 0% [08:22:52] RECOVERY - cp2 Nginx Backend for test1 on cp2 is OK: TCP OK - 0.001 second response time on localhost port 8180 [08:22:57] RECOVERY - cp2 Nginx Backend for mw1 on cp2 is OK: TCP OK - 0.001 second response time on localhost port 8100 [08:23:03] RECOVERY - ns1 Puppet on ns1 is OK: OK: Puppet is currently enabled, last run 43 seconds ago with 0 failures [08:23:18] RECOVERY - cp2 Puppet on cp2 is OK: OK: Puppet is currently enabled, last run 1 minute ago with 0 failures [08:23:19] RECOVERY - cp5 NTP time on cp5 is OK: NTP OK: Offset 0.001468658447 secs [08:23:20] RECOVERY - cp2 Current Load on cp2 is OK: LOAD OK - total load average: 0.27, 0.15, 0.09 [08:23:33] RECOVERY - ns1 Disk Space on ns1 is OK: DISK OK - free space: / 5427MiB (67% inode=85%); [08:23:36] RECOVERY - ns1 conntrack_table_size on ns1 is OK: OK: nf_conntrack is 0 % full [08:25:23] RECOVERY - db1 PowerDNS Recursor on db1 is OK: DNS OK: 0.219 seconds response time. wikitide.org returns 2001:41d0:801:2000::4089,51.75.170.66 [08:25:27] RECOVERY - db1 Puppet on db1 is OK: OK: Puppet is currently enabled, last run 10 seconds ago with 0 failures [08:25:27] RECOVERY - db1 ferm_active on db1 is OK: OK ferm input default policy is set [08:25:36] RECOVERY - db1 Backups SQL on db1 is OK: FILE_AGE OK: /var/log/backup-logs/sql-backup.log is 99584 seconds old and 1977 bytes [08:25:38] RECOVERY - db1 Disk Space on db1 is OK: DISK OK - free space: / 121019MiB (64% inode=98%); [08:25:52] RECOVERY - db1 Backups SQL wtglobal on db1 is OK: FILE_AGE OK: /var/log/backup-logs/sql-wtglobal-backup-weekly.log is 12340 seconds old and 89 bytes [08:25:53] RECOVERY - os1 NTP time on os1 is OK: NTP OK: Offset 0.00210699439 secs [08:25:54] PROBLEM - os1 Disk Space on os1 is WARNING: DISK WARNING - free space: / 12533MiB (6% inode=99%); [08:25:59] RECOVERY - bast1 NTP time on bast1 is OK: NTP OK: Offset 0.003804147243 secs [08:26:15] RECOVERY - bast1 Disk Space on bast1 is OK: DISK OK - free space: / 5453MiB (67% inode=86%); [08:26:19] RECOVERY - os1 PowerDNS Recursor on os1 is OK: DNS OK: 0.178 seconds response time. wikitide.org returns 2001:41d0:801:2000::4089,51.75.170.66 [08:26:26] RECOVERY - os1 ferm_active on os1 is OK: OK ferm input default policy is set [08:26:26] RECOVERY - bast1 Puppet on bast1 is OK: OK: Puppet is currently enabled, last run 29 seconds ago with 0 failures [08:26:27] RECOVERY - os1 Current Load on os1 is OK: LOAD OK - total load average: 0.17, 0.07, 0.02 [08:26:33] RECOVERY - bast1 PowerDNS Recursor on bast1 is OK: DNS OK: 0.449 seconds response time. wikitide.org returns 2001:41d0:801:2000::4089,51.75.170.66 [08:26:55] RECOVERY - os1 conntrack_table_size on os1 is OK: OK: nf_conntrack is 0 % full [08:27:00] RECOVERY - db1 conntrack_table_size on db1 is OK: OK: nf_conntrack is 0 % full [08:27:04] RECOVERY - bast1 conntrack_table_size on bast1 is OK: OK: nf_conntrack is 0 % full [08:27:07] RECOVERY - db1 Current Load on db1 is OK: LOAD OK - total load average: 0.27, 0.34, 0.37 [08:27:13] RECOVERY - db1 NTP time on db1 is OK: NTP OK: Offset -0.00532540679 secs [08:27:16] RECOVERY - bast1 ferm_active on bast1 is OK: OK ferm input default policy is set [08:27:45] RECOVERY - bast1 Current Load on bast1 is OK: LOAD OK - total load average: 0.12, 0.13, 0.05 [08:29:09] RECOVERY - swiftproxy1 conntrack_table_size on swiftproxy1 is OK: OK: nf_conntrack is 0 % full [08:29:16] RECOVERY - swiftproxy1 Current Load on swiftproxy1 is OK: LOAD OK - total load average: 0.12, 0.09, 0.09 [08:29:32] RECOVERY - swiftproxy1 PowerDNS Recursor on swiftproxy1 is OK: DNS OK: 0.223 seconds response time. wikitide.org returns 2001:41d0:801:2000::4089,51.75.170.66 [08:29:36] RECOVERY - swiftproxy1 NTP time on swiftproxy1 is OK: NTP OK: Offset 0.0009894967079 secs [08:29:36] RECOVERY - swiftproxy1 Disk Space on swiftproxy1 is OK: DISK OK - free space: / 23449MiB (87% inode=95%); [08:29:48] RECOVERY - swiftproxy1 Puppet on swiftproxy1 is OK: OK: Puppet is currently enabled, last run 47 seconds ago with 0 failures [08:30:02] RECOVERY - swiftproxy1 ferm_active on swiftproxy1 is OK: OK ferm input default policy is set [08:30:38] RECOVERY - swiftac1 Disk Space on swiftac1 is OK: DISK OK - free space: / 90140MiB (96% inode=98%); [08:31:10] RECOVERY - swiftac1 ferm_active on swiftac1 is OK: OK ferm input default policy is set [08:31:26] RECOVERY - swiftac1 PowerDNS Recursor on swiftac1 is OK: DNS OK: 0.340 seconds response time. wikitide.org returns 2001:41d0:801:2000::4089,51.75.170.66 [08:31:27] RECOVERY - swiftac1 NTP time on swiftac1 is OK: NTP OK: Offset 0.001703441143 secs [08:31:28] RECOVERY - test1 Disk Space on test1 is OK: DISK OK - free space: / 15992MiB (40% inode=62%); [08:31:31] RECOVERY - test1 PowerDNS Recursor on test1 is OK: DNS OK: 0.293 seconds response time. wikitide.org returns 2001:41d0:801:2000::4089,51.75.170.66 [08:31:37] RECOVERY - cp4 PowerDNS Recursor on cp4 is OK: DNS OK: 0.330 seconds response time. wikitide.org returns 146.59.44.171,2001:41d0:601:1100::4062 [08:31:41] RECOVERY - cp4 Varnish Backends on cp4 is OK: All 7 backends are healthy [08:31:58] RECOVERY - test1 ferm_active on test1 is OK: OK ferm input default policy is set [08:32:04] RECOVERY - swiftac1 Current Load on swiftac1 is OK: LOAD OK - total load average: 1.46, 1.57, 1.51 [08:32:06] RECOVERY - cp4 Nginx Backend for test1 on cp4 is OK: TCP OK - 0.000 second response time on localhost port 8180 [08:32:12] RECOVERY - swiftac1 conntrack_table_size on swiftac1 is OK: OK: nf_conntrack is 6 % full [08:32:13] RECOVERY - mw2 conntrack_table_size on mw2 is OK: OK: nf_conntrack is 1 % full [08:32:14] RECOVERY - cp4 Puppet on cp4 is OK: OK: Puppet is currently enabled, last run 38 seconds ago with 0 failures [08:32:16] RECOVERY - cp4 Nginx Backend for mw2 on cp4 is OK: TCP OK - 0.001 second response time on localhost port 8101 [08:32:21] RECOVERY - mw2 NTP time on mw2 is OK: NTP OK: Offset 0.001953452826 secs [08:32:24] RECOVERY - test1 nutcracker process on test1 is OK: PROCS OK: 1 process with UID = 110 (nutcracker), command name 'nutcracker' [08:32:26] RECOVERY - cp4 NTP time on cp4 is OK: NTP OK: Offset 0.0009343624115 secs [08:32:28] RECOVERY - swiftac1 Puppet on swiftac1 is OK: OK: Puppet is currently enabled, last run 1 minute ago with 0 failures [08:32:32] RECOVERY - mw2 nutcracker port on mw2 is OK: TCP OK - 0.000 second response time on 127.0.0.1 port 11212 [08:32:35] RECOVERY - mw2 Puppet on mw2 is OK: OK: Puppet is currently enabled, last run 24 seconds ago with 0 failures [08:32:37] RECOVERY - test1 php-fpm on test1 is OK: PROCS OK: 8 processes with command name 'php-fpm8.2' [08:32:40] RECOVERY - test1 NTP time on test1 is OK: NTP OK: Offset 0.009962886572 secs [08:32:45] RECOVERY - test1 JobChron Service on test1 is OK: PROCS OK: 1 process with args 'redisJobChronService' [08:32:48] RECOVERY - test1 nutcracker port on test1 is OK: TCP OK - 0.000 second response time on 127.0.0.1 port 11212 [08:32:49] RECOVERY - cp4 Current Load on cp4 is OK: LOAD OK - total load average: 0.05, 0.05, 0.01 [08:32:49] RECOVERY - test1 poolcounter process on test1 is OK: PROCS OK: 1 process with UID = 996 (poolcounter), command name 'poolcounterd' [08:33:06] RECOVERY - cp4 conntrack_table_size on cp4 is OK: OK: nf_conntrack is 0 % full [08:33:09] RECOVERY - cp4 Nginx Backend for phorge1 on cp4 is OK: TCP OK - 0.000 second response time on localhost port 8200 [08:33:10] RECOVERY - mw2 PowerDNS Recursor on mw2 is OK: DNS OK: 0.380 seconds response time. wikitide.org returns 2001:41d0:801:2000::4089,51.75.170.66 [08:33:11] RECOVERY - test1 conntrack_table_size on test1 is OK: OK: nf_conntrack is 0 % full [08:33:12] RECOVERY - test1 Current Load on test1 is OK: LOAD OK - total load average: 0.32, 0.26, 0.16 [08:33:13] RECOVERY - cp4 Nginx Backend for mw1 on cp4 is OK: TCP OK - 0.000 second response time on localhost port 8100 [08:33:17] RECOVERY - cp4 Nginx Backend for mail1 on cp4 is OK: TCP OK - 0.000 second response time on localhost port 8202 [08:33:18] RECOVERY - cp4 ferm_active on cp4 is OK: OK ferm input default policy is set [08:33:23] RECOVERY - cp4 Nginx Backend for matomo1 on cp4 is OK: TCP OK - 0.000 second response time on localhost port 8204 [08:33:25] RECOVERY - cp4 Nginx Backend for puppet1 on cp4 is OK: TCP OK - 0.000 second response time on localhost port 8201 [08:33:27] RECOVERY - mw2 php-fpm on mw2 is OK: PROCS OK: 38 processes with command name 'php-fpm8.2' [08:33:29] RECOVERY - cp4 Disk Space on cp4 is OK: DISK OK - free space: / 6214MiB (32% inode=94%); [08:33:32] RECOVERY - mw2 Disk Space on mw2 is OK: DISK OK - free space: / 24731MiB (60% inode=86%); [08:33:34] RECOVERY - mw2 nutcracker process on mw2 is OK: PROCS OK: 1 process with UID = 110 (nutcracker), command name 'nutcracker' [08:33:48] RECOVERY - mw2 Current Load on mw2 is OK: LOAD OK - total load average: 1.60, 1.16, 1.08 [08:33:58] RECOVERY - mw2 ferm_active on mw2 is OK: OK ferm input default policy is set [08:34:57] RECOVERY - mem1 ferm_active on mem1 is OK: OK ferm input default policy is set [08:35:06] RECOVERY - graylog1 conntrack_table_size on graylog1 is OK: OK: nf_conntrack is 0 % full [08:35:15] RECOVERY - graylog1 Puppet on graylog1 is OK: OK: Puppet is currently enabled, last run 22 seconds ago with 0 failures [08:35:25] RECOVERY - mem1 conntrack_table_size on mem1 is OK: OK: nf_conntrack is 0 % full [08:35:25] RECOVERY - graylog1 Current Load on graylog1 is OK: LOAD OK - total load average: 0.13, 0.04, 0.01 [08:35:26] RECOVERY - mem1 Disk Space on mem1 is OK: DISK OK - free space: / 5089MiB (66% inode=85%); [08:35:32] RECOVERY - services1 ferm_active on services1 is OK: OK ferm input default policy is set [08:35:43] RECOVERY - services1 Disk Space on services1 is OK: DISK OK - free space: / 2146MiB (28% inode=71%); [08:35:55] RECOVERY - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is OK: OK - NGINX Error Rate is 33% [08:35:57] RECOVERY - graylog1 APT on graylog1 is OK: APT OK: 1 packages available for upgrade (0 critical updates). [08:35:58] RECOVERY - graylog1 PowerDNS Recursor on graylog1 is OK: DNS OK: 0.258 seconds response time. wikitide.org returns 2001:41d0:801:2000::4089,51.75.170.66 [08:36:01] RECOVERY - services1 NTP time on services1 is OK: NTP OK: Offset 0.001912355423 secs [08:36:08] RECOVERY - mail1 Current Load on mail1 is OK: LOAD OK - total load average: 0.10, 0.04, 0.01 [08:36:08] RECOVERY - mem1 Puppet on mem1 is OK: OK: Puppet is currently enabled, last run 1 minute ago with 0 failures [08:36:12] RECOVERY - services1 conntrack_table_size on services1 is OK: OK: nf_conntrack is 0 % full [08:36:12] RECOVERY - services1 Puppet on services1 is OK: OK: Puppet is currently enabled, last run 40 seconds ago with 0 failures [08:36:13] RECOVERY - mail1 NTP time on mail1 is OK: NTP OK: Offset -0.004034280777 secs [08:36:14] RECOVERY - cp6 Nginx Backend for mail1 on cp6 is OK: TCP OK - 0.000 second response time on localhost port 8202 [08:36:17] RECOVERY - cp6 Nginx Backend for test1 on cp6 is OK: TCP OK - 0.000 second response time on localhost port 8180 [08:36:19] RECOVERY - mem1 NTP time on mem1 is OK: NTP OK: Offset -0.003176420927 secs [08:36:22] RECOVERY - cp6 Puppet on cp6 is OK: OK: Puppet is currently enabled, last run 14 seconds ago with 0 failures [08:36:22] RECOVERY - cp6 Nginx Backend for matomo1 on cp6 is OK: TCP OK - 0.000 second response time on localhost port 8204 [08:36:25] RECOVERY - cp6 Current Load on cp6 is OK: LOAD OK - total load average: 0.49, 0.14, 0.05 [08:36:28] RECOVERY - mem1 PowerDNS Recursor on mem1 is OK: DNS OK: 0.272 seconds response time. wikitide.org returns 2001:41d0:801:2000::4089,51.75.170.66 [08:36:31] RECOVERY - cp6 Disk Space on cp6 is OK: DISK OK - free space: / 6294MiB (32% inode=94%); [08:36:31] RECOVERY - graylog1 NTP time on graylog1 is OK: NTP OK: Offset -0.007977992296 secs [08:36:36] RECOVERY - mail1 conntrack_table_size on mail1 is OK: OK: nf_conntrack is 0 % full [08:36:43] RECOVERY - graylog1 ferm_active on graylog1 is OK: OK ferm input default policy is set [08:36:45] RECOVERY - mem1 Current Load on mem1 is OK: LOAD OK - total load average: 0.09, 0.12, 0.06 [08:36:45] RECOVERY - services1 PowerDNS Recursor on services1 is OK: DNS OK: 0.325 seconds response time. wikitide.org returns 2001:41d0:801:2000::4089,51.75.170.66 [08:36:49] RECOVERY - mail1 Disk Space on mail1 is OK: DISK OK - free space: / 25233MiB (88% inode=93%); [08:36:52] RECOVERY - mail1 PowerDNS Recursor on mail1 is OK: DNS OK: 0.171 seconds response time. wikitide.org returns 2001:41d0:801:2000::4089,51.75.170.66 [08:36:56] RECOVERY - mail1 ferm_active on mail1 is OK: OK ferm input default policy is set [08:37:08] RECOVERY - cp6 Varnish Backends on cp6 is OK: All 7 backends are healthy [08:37:09] RECOVERY - cp6 Nginx Backend for puppet1 on cp6 is OK: TCP OK - 0.000 second response time on localhost port 8201 [08:37:10] RECOVERY - cp6 ferm_active on cp6 is OK: OK ferm input default policy is set [08:37:11] RECOVERY - mail1 php-fpm on mail1 is OK: PROCS OK: 9 processes with command name 'php-fpm8.2' [08:37:19] RECOVERY - cp6 Nginx Backend for phorge1 on cp6 is OK: TCP OK - 0.000 second response time on localhost port 8200 [08:37:19] RECOVERY - cp6 Nginx Backend for mw1 on cp6 is OK: TCP OK - 0.000 second response time on localhost port 8100 [08:37:20] RECOVERY - cp6 conntrack_table_size on cp6 is OK: OK: nf_conntrack is 0 % full [08:37:20] RECOVERY - cp6 PowerDNS Recursor on cp6 is OK: DNS OK: 1.153 second response time. wikitide.org returns 139.99.236.151,2402:1f00:8100:400::1d23 [08:37:24] RECOVERY - services1 Current Load on services1 is OK: LOAD OK - total load average: 1.02, 0.82, 0.66 [08:37:25] RECOVERY - cp6 Nginx Backend for mw2 on cp6 is OK: TCP OK - 0.000 second response time on localhost port 8101 [08:37:26] RECOVERY - mw1 nutcracker process on mw1 is OK: PROCS OK: 1 process with UID = 110 (nutcracker), command name 'nutcracker' [08:37:27] RECOVERY - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is OK: OK - NGINX Error Rate is 21% [08:37:27] RECOVERY - cp6 NTP time on cp6 is OK: NTP OK: Offset -0.01748153567 secs [08:38:24] RECOVERY - mw1 Disk Space on mw1 is OK: DISK OK - free space: / 21464MiB (52% inode=86%); [08:38:32] RECOVERY - mw1 ferm_active on mw1 is OK: OK ferm input default policy is set [08:38:45] RECOVERY - mw1 Current Load on mw1 is OK: LOAD OK - total load average: 1.13, 0.97, 1.03 [08:38:49] RECOVERY - mw1 conntrack_table_size on mw1 is OK: OK: nf_conntrack is 1 % full [08:38:50] RECOVERY - mw1 php-fpm on mw1 is OK: PROCS OK: 38 processes with command name 'php-fpm8.2' [08:38:54] RECOVERY - mw1 PowerDNS Recursor on mw1 is OK: DNS OK: 0.281 seconds response time. wikitide.org returns 2001:41d0:801:2000::4089,51.75.170.66 [08:38:56] RECOVERY - mw1 nutcracker port on mw1 is OK: TCP OK - 0.000 second response time on 127.0.0.1 port 11212 [08:39:12] RECOVERY - mw1 Puppet on mw1 is OK: OK: Puppet is currently enabled, last run 1 minute ago with 0 failures [08:39:12] RECOVERY - mw1 NTP time on mw1 is OK: NTP OK: Offset 0.00181299448 secs [08:39:36] RECOVERY - jobrunner1 conntrack_table_size on jobrunner1 is OK: OK: nf_conntrack is 0 % full [08:39:37] RECOVERY - jobrunner1 JobRunner Service on jobrunner1 is OK: PROCS OK: 1 process with args 'redisJobRunnerService' [08:39:37] RECOVERY - jobrunner1 Current Load on jobrunner1 is OK: LOAD OK - total load average: 5.84, 4.02, 3.80 [08:39:49] RECOVERY - jobrunner1 nutcracker port on jobrunner1 is OK: TCP OK - 0.000 second response time on 127.0.0.1 port 11212 [08:39:59] RECOVERY - jobrunner1 php-fpm on jobrunner1 is OK: PROCS OK: 6 processes with command name 'php-fpm8.2' [08:40:05] RECOVERY - jobrunner1 NTP time on jobrunner1 is OK: NTP OK: Offset 0.004541784525 secs [08:40:09] RECOVERY - jobrunner1 ferm_active on jobrunner1 is OK: OK ferm input default policy is set [08:40:14] RECOVERY - prometheus1 Disk Space on prometheus1 is OK: DISK OK - free space: / 168883MiB (89% inode=99%); [08:40:25] RECOVERY - prometheus1 Puppet on prometheus1 is OK: OK: Puppet is currently enabled, last run 16 seconds ago with 0 failures [08:40:30] RECOVERY - prometheus1 PowerDNS Recursor on prometheus1 is OK: DNS OK: 0.168 seconds response time. wikitide.org returns 2001:41d0:801:2000::4089,51.75.170.66 [08:40:36] RECOVERY - jobrunner1 Puppet on jobrunner1 is OK: OK: Puppet is currently enabled, last run 55 seconds ago with 0 failures [08:40:41] RECOVERY - jobrunner1 PowerDNS Recursor on jobrunner1 is OK: DNS OK: 0.731 seconds response time. wikitide.org returns 2001:41d0:801:2000::4089,51.75.170.66 [08:40:56] RECOVERY - prometheus1 Current Load on prometheus1 is OK: LOAD OK - total load average: 0.30, 0.15, 0.05 [08:40:56] RECOVERY - jobrunner1 Disk Space on jobrunner1 is OK: DISK OK - free space: / 7909MiB (15% inode=77%); [08:41:01] RECOVERY - jobrunner1 nutcracker process on jobrunner1 is OK: PROCS OK: 1 process with UID = 110 (nutcracker), command name 'nutcracker' [08:41:13] RECOVERY - prometheus1 conntrack_table_size on prometheus1 is OK: OK: nf_conntrack is 0 % full [08:41:15] RECOVERY - prometheus1 NTP time on prometheus1 is OK: NTP OK: Offset 0.004730403423 secs [08:41:22] RECOVERY - prometheus1 ferm_active on prometheus1 is OK: OK ferm input default policy is set [08:41:31] RECOVERY - phorge1 phd on phorge1 is OK: PROCS OK: 1 process with args 'phd' [08:41:33] RECOVERY - jobchron1 Puppet on jobchron1 is OK: OK: Puppet is currently enabled, last run 6 seconds ago with 0 failures [08:41:37] RECOVERY - phorge1 Disk Space on phorge1 is OK: DISK OK - free space: / 13111MiB (76% inode=91%); [08:41:48] RECOVERY - phorge1 Puppet on phorge1 is OK: OK: Puppet is currently enabled, last run 41 seconds ago with 0 failures [08:41:54] RECOVERY - phorge1 conntrack_table_size on phorge1 is OK: OK: nf_conntrack is 0 % full [08:41:59] RECOVERY - jobchron1 conntrack_table_size on jobchron1 is OK: OK: nf_conntrack is 0 % full [08:42:07] RECOVERY - ldap1 Puppet on ldap1 is OK: OK: Puppet is currently enabled, last run 3 seconds ago with 0 failures [08:42:09] RECOVERY - phorge1 PowerDNS Recursor on phorge1 is OK: DNS OK: 0.409 seconds response time. wikitide.org returns 2001:41d0:801:2000::4089,51.75.170.66 [08:42:10] RECOVERY - ldap1 PowerDNS Recursor on ldap1 is OK: DNS OK: 0.276 seconds response time. wikitide.org returns 2001:41d0:801:2000::4089,51.75.170.66 [08:42:13] RECOVERY - phorge1 Current Load on phorge1 is OK: LOAD OK - total load average: 0.18, 0.14, 0.10 [08:42:14] RECOVERY - jobchron1 ferm_active on jobchron1 is OK: OK ferm input default policy is set [08:42:14] RECOVERY - ldap1 ferm_active on ldap1 is OK: OK ferm input default policy is set [08:42:20] RECOVERY - jobchron1 poolcounter process on jobchron1 is OK: PROCS OK: 1 process with UID = 999 (poolcounter), command name 'poolcounterd' [08:42:24] RECOVERY - jobchron1 Current Load on jobchron1 is OK: LOAD OK - total load average: 0.19, 0.11, 0.03 [08:42:26] RECOVERY - swiftobject1 PowerDNS Recursor on swiftobject1 is OK: DNS OK: 0.350 seconds response time. wikitide.org returns 2001:41d0:801:2000::4089,51.75.170.66 [08:42:26] RECOVERY - phorge1 ferm_active on phorge1 is OK: OK ferm input default policy is set [08:42:28] RECOVERY - swiftobject1 Current Load on swiftobject1 is OK: LOAD OK - total load average: 0.48, 0.29, 0.27 [08:42:43] RECOVERY - jobchron1 JobChron Service on jobchron1 is OK: PROCS OK: 1 process with args 'redisJobChronService' [08:42:56] RECOVERY - phorge1 NTP time on phorge1 is OK: NTP OK: Offset -0.002169370651 secs [08:43:01] RECOVERY - jobchron1 Disk Space on jobchron1 is OK: DISK OK - free space: / 2969MiB (52% inode=79%); [08:43:03] RECOVERY - ldap1 Disk Space on ldap1 is OK: DISK OK - free space: / 5093MiB (67% inode=85%); [08:43:04] RECOVERY - phorge1 php-fpm on phorge1 is OK: PROCS OK: 9 processes with command name 'php-fpm8.2' [08:43:08] RECOVERY - ldap1 NTP time on ldap1 is OK: NTP OK: Offset 0.00614798069 secs [08:43:19] RECOVERY - jobchron1 PowerDNS Recursor on jobchron1 is OK: DNS OK: 0.326 seconds response time. wikitide.org returns 2001:41d0:801:2000::4089,51.75.170.66 [08:43:20] RECOVERY - jobchron1 NTP time on jobchron1 is OK: NTP OK: Offset 4.690885544e-05 secs [08:43:20] RECOVERY - swiftobject1 Puppet on swiftobject1 is OK: OK: Puppet is currently enabled, last run 56 seconds ago with 0 failures [08:43:26] RECOVERY - swiftobject1 conntrack_table_size on swiftobject1 is OK: OK: nf_conntrack is 1 % full [08:43:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 42% [08:43:29] RECOVERY - ldap1 conntrack_table_size on ldap1 is OK: OK: nf_conntrack is 0 % full [08:43:44] RECOVERY - ldap1 Current Load on ldap1 is OK: LOAD OK - total load average: 0.10, 0.06, 0.01 [08:43:55] RECOVERY - swiftobject1 ferm_active on swiftobject1 is OK: OK ferm input default policy is set [08:43:57] RECOVERY - swiftobject1 NTP time on swiftobject1 is OK: NTP OK: Offset -0.007766991854 secs [08:44:16] RECOVERY - swiftobject1 Disk Space on swiftobject1 is OK: DISK OK - free space: / 9423MiB (76% inode=88%); [08:46:08] RECOVERY - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is OK: OK - NGINX Error Rate is 39% [08:47:12] PROBLEM - mw2 MediaWiki Rendering on mw2 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [08:47:57] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is WARNING: WARNING - NGINX Error Rate is 46% [08:49:09] PROBLEM - mw2 MediaWiki Rendering on mw2 is WARNING: HTTP WARNING: HTTP/1.1 404 Not Found - 8191 bytes in 0.214 second response time [08:49:58] RECOVERY - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is OK: OK - NGINX Error Rate is 17% [08:50:20] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 48% [08:54:14] RECOVERY - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is OK: OK - NGINX Error Rate is 16% [09:05:58] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 55% [09:07:55] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is CRITICAL: CRITICAL - NGINX Error Rate is 61% [09:09:52] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 51% [09:11:02] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is WARNING: WARNING - NGINX Error Rate is 45% [09:11:22] PROBLEM - mw1 MediaWiki Rendering on mw1 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [09:13:02] RECOVERY - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is OK: OK - NGINX Error Rate is 36% [09:13:16] PROBLEM - mw1 MediaWiki Rendering on mw1 is WARNING: HTTP WARNING: HTTP/1.1 404 Not Found - 8191 bytes in 0.150 second response time [09:15:43] RECOVERY - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is OK: OK - NGINX Error Rate is 34% [09:19:40] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 51% [09:21:37] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is CRITICAL: CRITICAL - NGINX Error Rate is 73% [09:23:34] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 58% [09:31:27] RECOVERY - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is OK: OK - NGINX Error Rate is 23% [09:43:08] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is CRITICAL: CRITICAL - NGINX Error Rate is 69% [09:47:08] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is WARNING: WARNING - NGINX Error Rate is 45% [09:49:09] RECOVERY - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is OK: OK - NGINX Error Rate is 28% [10:03:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 51% [10:05:27] RECOVERY - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is OK: OK - NGINX Error Rate is 37% [10:09:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is CRITICAL: CRITICAL - NGINX Error Rate is 76% [10:13:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 53% [10:15:27] RECOVERY - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is OK: OK - NGINX Error Rate is 39% [10:31:22] PROBLEM - mw1 MediaWiki Rendering on mw1 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [10:33:16] PROBLEM - mw1 MediaWiki Rendering on mw1 is WARNING: HTTP WARNING: HTTP/1.1 404 Not Found - 8191 bytes in 0.141 second response time [10:41:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is CRITICAL: CRITICAL - NGINX Error Rate is 73% [10:45:20] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is WARNING: WARNING - NGINX Error Rate is 44% [10:45:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 50% [10:49:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is CRITICAL: CRITICAL - NGINX Error Rate is 77% [10:51:21] RECOVERY - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is OK: OK - NGINX Error Rate is 27% [10:53:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 45% [10:59:27] RECOVERY - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is OK: OK - NGINX Error Rate is 37% [11:03:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 43% [11:05:12] PROBLEM - mw2 MediaWiki Rendering on mw2 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [11:05:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is CRITICAL: CRITICAL - NGINX Error Rate is 61% [11:07:09] PROBLEM - mw2 MediaWiki Rendering on mw2 is WARNING: HTTP WARNING: HTTP/1.1 404 Not Found - 8191 bytes in 0.142 second response time [11:07:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 50% [11:09:27] RECOVERY - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is OK: OK - NGINX Error Rate is 34% [11:13:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 47% [11:15:27] RECOVERY - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is OK: OK - NGINX Error Rate is 35% [11:19:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 55% [11:21:27] RECOVERY - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is OK: OK - NGINX Error Rate is 36% [11:33:29] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is WARNING: WARNING - NGINX Error Rate is 45% [11:35:30] RECOVERY - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is OK: OK - NGINX Error Rate is 37% [11:39:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is CRITICAL: CRITICAL - NGINX Error Rate is 68% [11:43:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 46% [11:45:27] RECOVERY - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is OK: OK - NGINX Error Rate is 20% [11:57:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 58% [12:01:35] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is CRITICAL: CRITICAL - NGINX Error Rate is 67% [12:03:27] RECOVERY - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is OK: OK - NGINX Error Rate is 33% [12:03:35] RECOVERY - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is OK: OK - NGINX Error Rate is 37% [12:15:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 53% [12:17:27] RECOVERY - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is OK: OK - NGINX Error Rate is 27% [12:17:38] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is CRITICAL: CRITICAL - NGINX Error Rate is 68% [12:19:38] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is WARNING: WARNING - NGINX Error Rate is 47% [12:21:39] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is CRITICAL: CRITICAL - NGINX Error Rate is 61% [12:27:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 54% [12:29:40] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is WARNING: WARNING - NGINX Error Rate is 43% [12:35:41] RECOVERY - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is OK: OK - NGINX Error Rate is 19% [12:37:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is CRITICAL: CRITICAL - NGINX Error Rate is 67% [12:39:42] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is WARNING: WARNING - NGINX Error Rate is 45% [12:41:42] RECOVERY - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is OK: OK - NGINX Error Rate is 17% [12:43:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 55% [12:49:44] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is CRITICAL: CRITICAL - NGINX Error Rate is 75% [12:51:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is CRITICAL: CRITICAL - NGINX Error Rate is 68% [12:53:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 53% [12:53:45] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is WARNING: WARNING - NGINX Error Rate is 40% [13:03:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is CRITICAL: CRITICAL - NGINX Error Rate is 74% [13:05:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 55% [13:07:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is CRITICAL: CRITICAL - NGINX Error Rate is 73% [13:13:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 55% [13:17:27] RECOVERY - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is OK: OK - NGINX Error Rate is 38% [13:21:49] RECOVERY - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is OK: OK - NGINX Error Rate is 38% [13:27:51] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is CRITICAL: CRITICAL - NGINX Error Rate is 72% [13:31:52] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is WARNING: WARNING - NGINX Error Rate is 57% [13:33:52] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is CRITICAL: CRITICAL - NGINX Error Rate is 63% [13:35:52] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is WARNING: WARNING - NGINX Error Rate is 45% [13:39:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 48% [13:41:53] RECOVERY - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is OK: OK - NGINX Error Rate is 33% [13:43:27] RECOVERY - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is OK: OK - NGINX Error Rate is 39% [13:52:21] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 50% [13:55:56] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is WARNING: WARNING - NGINX Error Rate is 51% [13:57:56] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is CRITICAL: CRITICAL - NGINX Error Rate is 61% [13:58:11] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is CRITICAL: CRITICAL - NGINX Error Rate is 60% [13:59:57] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is WARNING: WARNING - NGINX Error Rate is 48% [14:00:09] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 46% [14:01:57] RECOVERY - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is OK: OK - NGINX Error Rate is 38% [14:02:05] RECOVERY - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is OK: OK - NGINX Error Rate is 34% [14:06:02] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 45% [14:07:59] RECOVERY - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is OK: OK - NGINX Error Rate is 33% [14:13:53] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is CRITICAL: CRITICAL - NGINX Error Rate is 78% [14:17:47] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 45% [14:19:44] RECOVERY - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is OK: OK - NGINX Error Rate is 29% [14:20:00] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is CRITICAL: CRITICAL - NGINX Error Rate is 73% [14:22:01] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is WARNING: WARNING - NGINX Error Rate is 40% [14:24:01] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is CRITICAL: CRITICAL - NGINX Error Rate is 75% [14:29:32] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 58% [14:32:03] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is WARNING: WARNING - NGINX Error Rate is 47% [14:36:03] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is CRITICAL: CRITICAL - NGINX Error Rate is 60% [14:37:27] RECOVERY - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is OK: OK - NGINX Error Rate is 30% [14:38:04] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is WARNING: WARNING - NGINX Error Rate is 57% [14:49:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 40% [14:51:27] RECOVERY - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is OK: OK - NGINX Error Rate is 37% [14:52:06] RECOVERY - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is OK: OK - NGINX Error Rate is 34% [15:13:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is CRITICAL: CRITICAL - NGINX Error Rate is 64% [15:18:11] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is WARNING: WARNING - NGINX Error Rate is 58% [15:20:12] RECOVERY - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is OK: OK - NGINX Error Rate is 39% [15:21:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 43% [15:22:31] RECOVERY - mlrpgspeedruns.com - reverse DNS on sslhost is OK: SSL OK - mlrpgspeedruns.com reverse DNS resolves to cp3.wikitide.net - NS RECORDS OK [15:23:27] RECOVERY - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is OK: OK - NGINX Error Rate is 38% [15:34:18] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 40% [15:36:15] RECOVERY - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is OK: OK - NGINX Error Rate is 26% [15:57:19] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is CRITICAL: CRITICAL - NGINX Error Rate is 60% [15:59:11] You’re fine [15:59:19] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is WARNING: WARNING - NGINX Error Rate is 43% [16:01:19] RECOVERY - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is OK: OK - NGINX Error Rate is 33% [16:03:35] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is CRITICAL: CRITICAL - NGINX Error Rate is 61% [16:05:32] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 52% [16:07:29] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is CRITICAL: CRITICAL - NGINX Error Rate is 62% [16:11:22] PROBLEM - mw1 MediaWiki Rendering on mw1 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [16:11:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 49% [16:13:16] PROBLEM - mw1 MediaWiki Rendering on mw1 is WARNING: HTTP WARNING: HTTP/1.1 404 Not Found - 8191 bytes in 0.133 second response time [16:15:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is CRITICAL: CRITICAL - NGINX Error Rate is 64% [16:21:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 48% [16:23:27] RECOVERY - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is OK: OK - NGINX Error Rate is 35% [16:31:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is CRITICAL: CRITICAL - NGINX Error Rate is 71% [16:35:27] RECOVERY - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is OK: OK - NGINX Error Rate is 27% [16:52:17] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is CRITICAL: CRITICAL - NGINX Error Rate is 67% [16:52:30] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is CRITICAL: CRITICAL - NGINX Error Rate is 60% [16:54:14] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 57% [16:54:30] RECOVERY - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is OK: OK - NGINX Error Rate is 38% [17:00:05] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is CRITICAL: CRITICAL - NGINX Error Rate is 76% [17:02:32] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is WARNING: WARNING - NGINX Error Rate is 46% [17:04:32] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is CRITICAL: CRITICAL - NGINX Error Rate is 62% [17:06:33] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is WARNING: WARNING - NGINX Error Rate is 49% [17:08:33] RECOVERY - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is OK: OK - NGINX Error Rate is 29% [17:09:50] RECOVERY - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is OK: OK - NGINX Error Rate is 39% [17:14:35] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is WARNING: WARNING - NGINX Error Rate is 55% [17:15:44] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is CRITICAL: CRITICAL - NGINX Error Rate is 63% [17:20:36] RECOVERY - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is OK: OK - NGINX Error Rate is 35% [17:21:35] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 59% [17:25:29] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is CRITICAL: CRITICAL - NGINX Error Rate is 63% [17:26:37] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is WARNING: WARNING - NGINX Error Rate is 45% [17:27:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 57% [17:28:37] RECOVERY - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is OK: OK - NGINX Error Rate is 37% [17:29:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is CRITICAL: CRITICAL - NGINX Error Rate is 60% [17:31:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 59% [17:35:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is CRITICAL: CRITICAL - NGINX Error Rate is 73% [17:37:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 53% [17:41:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is CRITICAL: CRITICAL - NGINX Error Rate is 73% [17:43:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 41% [17:45:22] PROBLEM - mw1 MediaWiki Rendering on mw1 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [17:45:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is CRITICAL: CRITICAL - NGINX Error Rate is 75% [17:47:16] PROBLEM - mw1 MediaWiki Rendering on mw1 is WARNING: HTTP WARNING: HTTP/1.1 404 Not Found - 8191 bytes in 0.140 second response time [17:53:27] RECOVERY - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is OK: OK - NGINX Error Rate is 38% [17:57:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 54% [17:58:43] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is WARNING: WARNING - NGINX Error Rate is 56% [18:00:43] RECOVERY - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is OK: OK - NGINX Error Rate is 35% [18:03:27] RECOVERY - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is OK: OK - NGINX Error Rate is 37% [18:13:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 55% [18:15:27] RECOVERY - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is OK: OK - NGINX Error Rate is 27% [18:15:47] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is WARNING: WARNING - NGINX Error Rate is 55% [18:19:48] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is CRITICAL: CRITICAL - NGINX Error Rate is 63% [18:21:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 46% [18:21:48] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is WARNING: WARNING - NGINX Error Rate is 43% [18:23:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is CRITICAL: CRITICAL - NGINX Error Rate is 64% [18:23:49] PROBLEM - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is CRITICAL: CRITICAL - NGINX Error Rate is 73% [18:25:12] PROBLEM - mw2 MediaWiki Rendering on mw2 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [18:27:09] PROBLEM - mw2 MediaWiki Rendering on mw2 is WARNING: HTTP WARNING: HTTP/1.1 404 Not Found - 8191 bytes in 0.187 second response time [18:27:27] PROBLEM - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is WARNING: WARNING - NGINX Error Rate is 57% [18:27:49] RECOVERY - cp4 HTTP 4xx/5xx ERROR Rate on cp4 is OK: OK - NGINX Error Rate is 0% [18:29:22] PROBLEM - swiftproxy1 ferm_active on swiftproxy1 is CRITICAL: ERROR ferm input drop default policy not set, ferm might not have been started correctly [18:29:27] RECOVERY - cp6 HTTP 4xx/5xx ERROR Rate on cp6 is OK: OK - NGINX Error Rate is 27% [18:30:34] PROBLEM - swiftproxy1 Puppet on swiftproxy1 is WARNING: WARNING: Puppet is currently disabled, message: reason not specified, last run 32 minutes ago with 0 failures [18:42:35] !log [universalomega@puppet1] Upgraded packages curl, libcurl4, and libcurl3-gnutls on mem1 [18:42:41] RECOVERY - mem1 APT on mem1 is OK: APT OK: 27 packages available for upgrade (0 critical updates). [18:42:44] Logged the message at https://meta.wikitide.org/wiki/Tech:Server_admin_log [18:42:56] !log [universalomega@puppet1] Upgraded packages curl, libcurl4, and libcurl3-gnutls on bots1 [18:43:05] Logged the message at https://meta.wikitide.org/wiki/Tech:Server_admin_log [18:43:12] RECOVERY - bots1 APT on bots1 is OK: APT OK: 27 packages available for upgrade (0 critical updates). [18:43:15] !log [universalomega@puppet1] Upgraded packages curl, libcurl4, and libcurl3-gnutls on matomo1 [18:43:23] Logged the message at https://meta.wikitide.org/wiki/Tech:Server_admin_log [18:43:33] !log [universalomega@puppet1] Upgraded packages xserver-xorg-core, xserver-common, curl, libcurl4, ghostscript, libgs10, libgs10-common, libgs-common, libcurl3-gnutls, and xvfb on mw1 [18:43:42] Logged the message at https://meta.wikitide.org/wiki/Tech:Server_admin_log [18:43:46] !log [universalomega@puppet1] Upgraded packages curl, libcurl4, and libcurl3-gnutls on os1 [18:43:54] Logged the message at https://meta.wikitide.org/wiki/Tech:Server_admin_log [18:43:59] !log [universalomega@puppet1] Upgraded packages curl, libcurl4, and libcurl3-gnutls on db1 [18:44:08] Logged the message at https://meta.wikitide.org/wiki/Tech:Server_admin_log [18:44:17] !log [universalomega@puppet1] Upgraded packages xserver-xorg-core, xserver-common, curl, libcurl4, ghostscript, libgs10, libgs10-common, libgs-common, libcurl3-gnutls, and xvfb on mw2 [18:44:25] Logged the message at https://meta.wikitide.org/wiki/Tech:Server_admin_log [18:44:30] !log [universalomega@puppet1] Upgraded packages curl, libcurl4, and libcurl3-gnutls on mon1 [18:44:36] RECOVERY - db1 APT on db1 is OK: APT OK: 26 packages available for upgrade (0 critical updates). [18:44:39] Logged the message at https://meta.wikitide.org/wiki/Tech:Server_admin_log [18:44:47] !log [universalomega@puppet1] Upgraded packages curl, libcurl4, and libcurl3-gnutls on ldap1 [18:44:50] RECOVERY - os1 APT on os1 is OK: APT OK: 28 packages available for upgrade (0 critical updates). [18:44:54] Logged the message at https://meta.wikitide.org/wiki/Tech:Server_admin_log [18:44:57] RECOVERY - mon1 APT on mon1 is OK: APT OK: 63 packages available for upgrade (0 critical updates). [18:45:09] RECOVERY - matomo1 APT on matomo1 is OK: APT OK: 50 packages available for upgrade (0 critical updates). [18:45:11] !log [universalomega@puppet1] Upgraded packages curl, libcurl4, and libcurl3-gnutls on mail1 [18:45:19] Logged the message at https://meta.wikitide.org/wiki/Tech:Server_admin_log [18:45:26] !log [universalomega@puppet1] Upgraded packages curl, libcurl4, and libcurl3-gnutls on prometheus1 [18:45:31] RECOVERY - mw1 APT on mw1 is OK: APT OK: 56 packages available for upgrade (0 critical updates). [18:45:33] RECOVERY - prometheus1 APT on prometheus1 is OK: APT OK: 27 packages available for upgrade (0 critical updates). [18:45:35] Logged the message at https://meta.wikitide.org/wiki/Tech:Server_admin_log [18:45:40] !log [universalomega@puppet1] Upgraded packages curl, libcurl4, and libcurl3-gnutls on jobchron1 [18:45:48] Logged the message at https://meta.wikitide.org/wiki/Tech:Server_admin_log [18:45:55] !log [universalomega@puppet1] Upgraded packages curl, libcurl4, and libcurl3-gnutls on bast1 [18:45:59] RECOVERY - mw2 APT on mw2 is OK: APT OK: 56 packages available for upgrade (0 critical updates). [18:46:03] Logged the message at https://meta.wikitide.org/wiki/Tech:Server_admin_log [18:46:06] RECOVERY - bast1 APT on bast1 is OK: APT OK: 27 packages available for upgrade (0 critical updates). [18:46:12] !log [universalomega@puppet1] Upgraded packages curl, libcurl4, and libcurl3-gnutls on ns1 [18:46:20] Logged the message at https://meta.wikitide.org/wiki/Tech:Server_admin_log [18:46:27] !log [universalomega@puppet1] Upgraded packages curl, libcurl4, and libcurl3-gnutls on puppet1 [18:46:30] RECOVERY - mail1 APT on mail1 is OK: APT OK: 57 packages available for upgrade (0 critical updates). [18:46:36] Logged the message at https://meta.wikitide.org/wiki/Tech:Server_admin_log [18:46:42] RECOVERY - ldap1 APT on ldap1 is OK: APT OK: 26 packages available for upgrade (0 critical updates). [18:46:45] !log [universalomega@puppet1] Upgraded packages curl, libcurl4, and libcurl3-gnutls on swiftac1 [18:46:49] RECOVERY - ns1 APT on ns1 is OK: APT OK: 27 packages available for upgrade (0 critical updates). [18:46:54] Logged the message at https://meta.wikitide.org/wiki/Tech:Server_admin_log [18:46:57] RECOVERY - puppet1 APT on puppet1 is OK: APT OK: 30 packages available for upgrade (0 critical updates). [18:47:04] !log [universalomega@puppet1] Upgraded packages curl, libcurl4, and libcurl3-gnutls on phorge1 [18:47:11] RECOVERY - jobchron1 APT on jobchron1 is OK: APT OK: 47 packages available for upgrade (0 critical updates). [18:47:13] Logged the message at https://meta.wikitide.org/wiki/Tech:Server_admin_log [18:47:24] RECOVERY - phorge1 APT on phorge1 is OK: APT OK: 42 packages available for upgrade (0 critical updates). [18:47:25] !log [universalomega@puppet1] Upgraded packages xserver-xorg-core, xserver-common, curl, libcurl4, ghostscript, libgs10, libgs10-common, libgs-common, libcurl3-gnutls, and xvfb on jobrunner1 [18:47:33] Logged the message at https://meta.wikitide.org/wiki/Tech:Server_admin_log [18:48:03] !log [universalomega@puppet1] Upgraded packages chromium, chromium-common, chromium-sandbox, curl, libcurl4, and libcurl3-gnutls on services1 [18:48:05] RECOVERY - swiftac1 APT on swiftac1 is OK: APT OK: 27 packages available for upgrade (0 critical updates). [18:48:13] Logged the message at https://meta.wikitide.org/wiki/Tech:Server_admin_log [18:48:20] !log [universalomega@puppet1] Upgraded packages curl, libcurl4, and libcurl3-gnutls on swiftproxy1 [18:48:30] Logged the message at https://meta.wikitide.org/wiki/Tech:Server_admin_log [18:48:31] RECOVERY - services1 APT on services1 is OK: APT OK: 28 packages available for upgrade (0 critical updates). [18:48:38] RECOVERY - jobrunner1 APT on jobrunner1 is OK: APT OK: 56 packages available for upgrade (0 critical updates). [18:48:39] !log [universalomega@puppet1] Upgraded packages xserver-xorg-core, xserver-common, curl, libcurl4, ghostscript, libgs10, libgs10-common, libgs-common, libcurl3-gnutls, and xvfb on test1 [18:48:47] Logged the message at https://meta.wikitide.org/wiki/Tech:Server_admin_log [18:48:58] !log [universalomega@puppet1] Upgraded packages curl, libcurl4, and libcurl3-gnutls on swiftobject1 [18:49:07] Logged the message at https://meta.wikitide.org/wiki/Tech:Server_admin_log [18:49:10] RECOVERY - swiftobject1 APT on swiftobject1 is OK: APT OK: 26 packages available for upgrade (0 critical updates). [18:49:20] RECOVERY - test1 APT on test1 is OK: APT OK: 58 packages available for upgrade (0 critical updates). [18:49:22] RECOVERY - swiftproxy1 APT on swiftproxy1 is OK: APT OK: 26 packages available for upgrade (0 critical updates). [18:51:28] !log [universalomega@puppet1] Upgraded packages curl, libcurl4, and libcurl3-gnutls on cp4 [18:51:33] RECOVERY - cp4 APT on cp4 is OK: APT OK: 29 packages available for upgrade (0 critical updates). [18:51:36] Logged the message at https://meta.wikitide.org/wiki/Tech:Server_admin_log [18:57:10] !log [universalomega@puppet1] Upgraded packages curl, libcurl4, and libcurl3-gnutls on cp6 [18:57:18] Logged the message at https://meta.wikitide.org/wiki/Tech:Server_admin_log [18:57:57] RECOVERY - cp6 APT on cp6 is OK: APT OK: 29 packages available for upgrade (0 critical updates). [18:59:47] !log [universalomega@puppet1] Upgraded packages curl, libcurl4, and libcurl3-gnutls on cp3 [18:59:55] Logged the message at https://meta.wikitide.org/wiki/Tech:Server_admin_log [19:01:34] RECOVERY - cp3 APT on cp3 is OK: APT OK: 0 packages available for upgrade (0 critical updates). [19:02:22] !log [universalomega@puppet1] Upgraded packages curl, libcurl4, and libcurl3-gnutls on cp5 [19:02:30] Logged the message at https://meta.wikitide.org/wiki/Tech:Server_admin_log [19:03:17] RECOVERY - cp5 APT on cp5 is OK: APT OK: 28 packages available for upgrade (0 critical updates). [19:19:18] [02WikiTideOrg/puppet] 07Universal-Omega pushed 031 commit to 03master [+0/-0/±1] 13https://github.com/WikiTideOrg/puppet/compare/6a8478fced64...d8fa954bf12a [19:19:21] [02WikiTideOrg/puppet] 07Universal-Omega 03d8fa954 - Update IRC-Discord-Relay repo [20:11:24] PROBLEM - cp1 Varnish Backends on cp1 is CRITICAL: 3 backends are down. mw1 mw2 mediawiki [20:11:33] PROBLEM - cp1 ferm_active on cp1 is CRITICAL: ERROR ferm input drop default policy not set, ferm might not have been started correctly [20:12:02] PROBLEM - cp1 HTTPS on cp1 is CRITICAL: HTTP CRITICAL: HTTP/2 503 - 2620 bytes in 0.012 second response time [20:12:04] PROBLEM - cp1 HTTP 4xx/5xx ERROR Rate on cp1 is CRITICAL: CRITICAL - NGINX Error Rate is 100% [20:12:07] PROBLEM - cp1 APT on cp1 is CRITICAL: APT CRITICAL: 29 packages available for upgrade (3 critical updates). [20:19:23] RECOVERY - cp1 ferm_active on cp1 is OK: OK ferm input default policy is set [20:22:52] [02WikiTideOrg/dns] 07Universal-Omega pushed 031 commit to 03master [+0/-0/±1] 13https://github.com/WikiTideOrg/dns/compare/4cd0df2f5cc4...dc685b8af7e6 [20:22:54] [02WikiTideOrg/dns] 07Universal-Omega 03dc685b8 - Update cp1 IP [20:27:28] RECOVERY - cp1 APT on cp1 is OK: APT OK: 0 packages available for upgrade (0 critical updates). [20:28:56] RECOVERY - cp1 Varnish Backends on cp1 is OK: All 7 backends are healthy [20:30:18] [02WikiTideOrg/dns] 07Universal-Omega pushed 031 commit to 03master [+0/-0/±1] 13https://github.com/WikiTideOrg/dns/compare/dc685b8af7e6...048c6284be10 [20:30:20] [02WikiTideOrg/dns] 07Universal-Omega 03048c628 - Don't mark cp1 as down [20:34:03] [02WikiTideOrg/dns] 07Universal-Omega pushed 031 commit to 03master [+0/-0/±1] 13https://github.com/WikiTideOrg/dns/compare/048c6284be10...709537a99264 [20:34:06] [02WikiTideOrg/dns] 07Universal-Omega 03709537a - Remove cp1 ipv6 and thumb service [20:38:56] PROBLEM - cp1 ferm_active on cp1 is CRITICAL: ERROR ferm input drop default policy not set, ferm might not have been started correctly [20:40:54] RECOVERY - cp1 ferm_active on cp1 is OK: OK ferm input default policy is set [20:51:53] [02WikiTideOrg/dns] 07Universal-Omega pushed 031 commit to 03master [+0/-0/±1] 13https://github.com/WikiTideOrg/dns/compare/709537a99264...bb17577fa274 [20:51:54] [02WikiTideOrg/dns] 07Universal-Omega 03bb17577 - Use private IP for swift-lb [21:09:04] RECOVERY - swiftproxy1 ferm_active on swiftproxy1 is OK: OK ferm input default policy is set [21:10:15] PROBLEM - cp1 Varnish Backends on cp1 is CRITICAL: 3 backends are down. mw1 mw2 mediawiki [21:13:02] PROBLEM - swiftproxy1 ferm_active on swiftproxy1 is CRITICAL: ERROR ferm input drop default policy not set, ferm might not have been started correctly [21:31:16] PROBLEM - cp1 ferm_active on cp1 is CRITICAL: ERROR ferm input drop default policy not set, ferm might not have been started correctly [21:33:14] RECOVERY - cp1 ferm_active on cp1 is OK: OK ferm input default policy is set [21:34:25] PROBLEM - mw2 Puppet on mw2 is CRITICAL: CRITICAL: Failed to apply catalog, zero resources tracked by Puppet. It might be a dependency cycle. [21:38:54] RECOVERY - swiftproxy1 ferm_active on swiftproxy1 is OK: OK ferm input default policy is set [21:42:54] PROBLEM - swiftproxy1 ferm_active on swiftproxy1 is CRITICAL: ERROR ferm input drop default policy not set, ferm might not have been started correctly [21:44:54] RECOVERY - swiftproxy1 ferm_active on swiftproxy1 is OK: OK ferm input default policy is set [21:46:15] RECOVERY - swiftproxy1 Puppet on swiftproxy1 is OK: OK: Puppet is currently enabled, last run 1 minute ago with 0 failures [22:04:25] RECOVERY - mw2 Puppet on mw2 is OK: OK: Puppet is currently enabled, last run 1 minute ago with 0 failures