[00:00:16] PROBLEM - cp33 Current Load on cp33 is CRITICAL: CRITICAL - load average: 6.03, 4.04, 3.52 [00:00:32] [Grafana] !sre FIRING: The mediawiki job queue has more than 2500 unclaimed jobs https://grafana.miraheze.org/d/GtxbP1Xnk?orgId=1 [00:01:07] PROBLEM - cp23 Current Load on cp23 is CRITICAL: CRITICAL - load average: 6.78, 7.05, 4.22 [00:02:22] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 70% [00:02:26] RECOVERY - cp32 Current Load on cp32 is OK: OK - load average: 3.02, 3.08, 2.75 [00:03:40] RECOVERY - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is OK: OK - NGINX Error Rate is 35% [00:04:16] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 41% [00:06:14] PROBLEM - cp33 Current Load on cp33 is WARNING: WARNING - load average: 2.97, 3.88, 3.75 [00:06:26] PROBLEM - cp32 Current Load on cp32 is CRITICAL: CRITICAL - load average: 8.56, 5.34, 3.67 [00:07:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 45% [00:08:15] PROBLEM - cp33 Current Load on cp33 is CRITICAL: CRITICAL - load average: 5.16, 4.27, 3.89 [00:08:18] PROBLEM - es141 Current Load on es141 is WARNING: WARNING - load average: 3.53, 3.01, 2.25 [00:09:40] RECOVERY - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is OK: OK - NGINX Error Rate is 39% [00:10:19] PROBLEM - es141 Current Load on es141 is CRITICAL: CRITICAL - load average: 4.59, 3.35, 2.45 [00:12:18] PROBLEM - db142 Current Load on db142 is CRITICAL: CRITICAL - load average: 15.75, 9.99, 5.58 [00:12:19] PROBLEM - es141 Current Load on es141 is WARNING: WARNING - load average: 4.00, 3.50, 2.62 [00:12:31] PROBLEM - mw121 Current Load on mw121 is CRITICAL: CRITICAL - load average: 14.62, 12.30, 9.40 [00:12:58] PROBLEM - cp23 Current Load on cp23 is WARNING: WARNING - load average: 3.36, 3.52, 3.94 [00:13:41] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 44% [00:13:59] RECOVERY - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is OK: OK - NGINX Error Rate is 31% [00:14:30] PROBLEM - mw121 Current Load on mw121 is WARNING: WARNING - load average: 10.83, 11.84, 9.59 [00:14:56] PROBLEM - cp23 Current Load on cp23 is CRITICAL: CRITICAL - load average: 14.15, 9.05, 6.01 [00:17:43] PROBLEM - cp23 APT on cp23 is CRITICAL: CHECK_NRPE STATE CRITICAL: Socket timeout after 60 seconds. [00:18:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 73% [00:19:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 56% [00:19:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 72% [00:19:52] RECOVERY - cp23 APT on cp23 is OK: APT OK: 1 packages available for upgrade (0 critical updates). [00:20:02] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 57% [00:20:17] PROBLEM - db142 Current Load on db142 is WARNING: WARNING - load average: 5.22, 7.46, 6.35 [00:20:26] RECOVERY - mw121 Current Load on mw121 is OK: OK - load average: 9.16, 10.14, 9.50 [00:21:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 40% [00:21:50] PROBLEM - en.religiononfire.mar.in.ua - reverse DNS on sslhost is WARNING: NoNameservers: All nameservers failed to answer the query en.religiononfire.mar.in.ua. IN CNAME: Server 2606:4700:4700::1111 UDP port 53 answered SERVFAIL [00:22:17] PROBLEM - db142 Current Load on db142 is CRITICAL: CRITICAL - load average: 8.24, 7.99, 6.69 [00:22:19] PROBLEM - es141 Current Load on es141 is CRITICAL: CRITICAL - load average: 4.51, 3.79, 3.14 [00:23:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 65% [00:23:54] RECOVERY - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is OK: OK - NGINX Error Rate is 36% [00:24:17] PROBLEM - db142 Current Load on db142 is WARNING: WARNING - load average: 5.12, 7.15, 6.56 [00:24:22] PROBLEM - es141 Current Load on es141 is WARNING: WARNING - load average: 3.93, 3.75, 3.20 [00:26:18] PROBLEM - db142 Current Load on db142 is CRITICAL: CRITICAL - load average: 10.16, 7.82, 6.84 [00:28:20] PROBLEM - mw121 Current Load on mw121 is CRITICAL: CRITICAL - load average: 13.88, 11.78, 10.38 [00:29:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 52% [00:29:35] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 57% [00:31:32] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 81% [00:31:39] RECOVERY - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is OK: OK - NGINX Error Rate is 39% [00:32:17] PROBLEM - mw121 Current Load on mw121 is WARNING: WARNING - load average: 10.08, 11.67, 10.75 [00:32:31] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 47% [00:33:28] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 44% [00:34:17] PROBLEM - db142 Current Load on db142 is WARNING: WARNING - load average: 5.26, 7.16, 7.08 [00:35:23] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 74% [00:35:31] PROBLEM - mw122 Current Load on mw122 is CRITICAL: CRITICAL - load average: 12.56, 11.52, 9.75 [00:36:42] PROBLEM - cp23 Current Load on cp23 is WARNING: WARNING - load average: 0.70, 1.83, 3.66 [00:37:31] PROBLEM - mw122 Current Load on mw122 is WARNING: WARNING - load average: 11.06, 11.94, 10.15 [00:37:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 67% [00:37:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 62% [00:38:13] RECOVERY - mw121 Current Load on mw121 is OK: OK - load average: 10.17, 9.52, 10.11 [00:38:14] RECOVERY - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is OK: OK - NGINX Error Rate is 26% [00:40:17] PROBLEM - db142 Current Load on db142 is CRITICAL: CRITICAL - load average: 8.26, 7.50, 7.22 [00:42:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 43% [00:42:09] PROBLEM - mw121 Current Load on mw121 is WARNING: WARNING - load average: 10.64, 10.33, 10.32 [00:42:39] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 0.93, 2.09, 3.31 [00:43:31] RECOVERY - mw122 Current Load on mw122 is OK: OK - load average: 7.56, 9.38, 9.60 [00:44:17] PROBLEM - db142 Current Load on db142 is WARNING: WARNING - load average: 6.88, 7.82, 7.43 [00:45:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 42% [00:45:40] RECOVERY - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is OK: OK - NGINX Error Rate is 37% [00:45:55] RECOVERY - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is OK: OK - NGINX Error Rate is 34% [00:47:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 76% [00:48:05] RECOVERY - mw121 Current Load on mw121 is OK: OK - load average: 9.77, 9.92, 10.15 [00:48:17] RECOVERY - db142 Current Load on db142 is OK: OK - load average: 4.79, 5.53, 6.55 [00:48:19] PROBLEM - es141 Current Load on es141 is CRITICAL: CRITICAL - load average: 4.06, 3.54, 3.48 [00:49:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 40% [00:50:21] PROBLEM - es141 Current Load on es141 is WARNING: WARNING - load average: 3.66, 3.56, 3.50 [00:50:40] RECOVERY - en.religiononfire.mar.in.ua - reverse DNS on sslhost is OK: SSL OK - en.religiononfire.mar.in.ua reverse DNS resolves to cp23.miraheze.org - CNAME OK [00:51:43] PROBLEM - cp32 Varnish Backends on cp32 is CRITICAL: 3 backends are down. mw121 mw131 mw132 [00:54:21] RECOVERY - es141 Current Load on es141 is OK: OK - load average: 2.89, 3.12, 3.33 [00:55:28] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 74% [00:55:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 69% [00:55:43] RECOVERY - cp32 Varnish Backends on cp32 is OK: All 14 backends are healthy [00:56:16] PROBLEM - cp33 NTP time on cp33 is WARNING: NTP WARNING: Offset 0.2506902516 secs [00:57:14] PROBLEM - db142 Current Load on db142 is WARNING: WARNING - load average: 7.53, 6.43, 6.42 [00:57:40] RECOVERY - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is OK: OK - NGINX Error Rate is 31% [00:58:16] RECOVERY - cp33 NTP time on cp33 is OK: NTP OK: Offset -0.03221806884 secs [00:58:18] PROBLEM - es141 Current Load on es141 is CRITICAL: CRITICAL - load average: 4.13, 3.93, 3.63 [00:58:24] PROBLEM - cp23 Current Load on cp23 is CRITICAL: CRITICAL - load average: 4.82, 3.32, 3.02 [00:59:16] RECOVERY - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is OK: OK - NGINX Error Rate is 34% [01:00:22] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 2.50, 3.36, 3.11 [01:01:13] PROBLEM - db142 Current Load on db142 is CRITICAL: CRITICAL - load average: 10.41, 8.27, 7.12 [01:03:08] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 47% [01:03:12] RECOVERY - db142 Current Load on db142 is OK: OK - load average: 5.00, 6.73, 6.69 [01:03:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 40% [01:05:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 75% [01:05:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 54% [01:05:40] RECOVERY - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is OK: OK - NGINX Error Rate is 31% [01:07:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 50% [01:07:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 60% [01:08:49] PROBLEM - mw121 Current Load on mw121 is CRITICAL: CRITICAL - load average: 16.03, 11.86, 10.05 [01:09:08] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 74% [01:10:48] PROBLEM - mw121 Current Load on mw121 is WARNING: WARNING - load average: 11.54, 11.91, 10.32 [01:11:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 44% [01:12:21] PROBLEM - es141 Current Load on es141 is WARNING: WARNING - load average: 3.59, 3.77, 3.95 [01:13:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 59% [01:15:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 70% [01:15:39] RECOVERY - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is OK: OK - NGINX Error Rate is 28% [01:16:18] PROBLEM - es141 Current Load on es141 is CRITICAL: CRITICAL - load average: 4.20, 3.62, 3.82 [01:16:44] RECOVERY - mw121 Current Load on mw121 is OK: OK - load average: 8.85, 10.00, 9.96 [01:18:19] PROBLEM - es141 Current Load on es141 is WARNING: WARNING - load average: 3.98, 3.74, 3.84 [01:19:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 52% [01:20:17] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 56% [01:21:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 63% [01:22:46] PROBLEM - uk.religiononfire.mar.in.ua - reverse DNS on sslhost is CRITICAL: rDNS CRITICAL - uk.religiononfire.mar.in.ua All nameservers failed to answer the query. [01:23:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 55% [01:24:06] RECOVERY - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is OK: OK - NGINX Error Rate is 35% [01:25:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 52% [01:25:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 65% [01:25:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 63% [01:27:08] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 72% [01:27:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 54% [01:27:55] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 69% [01:29:50] RECOVERY - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is OK: OK - NGINX Error Rate is 34% [01:32:16] PROBLEM - cp33 NTP time on cp33 is WARNING: NTP WARNING: Offset 0.1399203837 secs [01:32:19] PROBLEM - es141 Current Load on es141 is CRITICAL: CRITICAL - load average: 4.58, 3.77, 3.68 [01:33:40] RECOVERY - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is OK: OK - NGINX Error Rate is 39% [01:35:44] PROBLEM - cp32 Varnish Backends on cp32 is CRITICAL: 5 backends are down. mw121 mw131 mw132 mw141 mw142 [01:37:31] PROBLEM - mw122 Current Load on mw122 is CRITICAL: CRITICAL - load average: 14.13, 10.41, 8.56 [01:37:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 68% [01:38:32] PROBLEM - mw122 MediaWiki Rendering on mw122 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [01:39:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 55% [01:39:24] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 46% [01:39:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 40% [01:40:22] PROBLEM - es141 Current Load on es141 is WARNING: WARNING - load average: 2.52, 3.47, 3.66 [01:40:28] RECOVERY - mw122 MediaWiki Rendering on mw122 is OK: HTTP OK: HTTP/1.1 200 OK - 8191 bytes in 0.311 second response time [01:41:18] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 62% [01:41:31] RECOVERY - mw122 Current Load on mw122 is OK: OK - load average: 5.30, 9.08, 8.61 [01:41:39] RECOVERY - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is OK: OK - NGINX Error Rate is 36% [01:41:43] RECOVERY - cp32 Varnish Backends on cp32 is OK: All 14 backends are healthy [01:43:08] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 75% [01:43:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 57% [01:44:16] RECOVERY - cp33 NTP time on cp33 is OK: NTP OK: Offset 0.06392276287 secs [01:45:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 41% [01:45:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 76% [01:47:07] RECOVERY - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is OK: OK - NGINX Error Rate is 35% [01:48:18] RECOVERY - es141 Current Load on es141 is OK: OK - load average: 2.24, 2.78, 3.29 [01:49:10] PROBLEM - cp33 Varnish Backends on cp33 is CRITICAL: 7 backends are down. mw121 mw122 mw131 mw132 mw141 mw142 mediawiki [01:50:18] PROBLEM - graylog121 Current Load on graylog121 is WARNING: WARNING - load average: 2.25, 3.27, 3.89 [01:51:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 54% [01:51:09] RECOVERY - cp33 Varnish Backends on cp33 is OK: All 14 backends are healthy [01:52:21] RECOVERY - uk.religiononfire.mar.in.ua - reverse DNS on sslhost is OK: SSL OK - uk.religiononfire.mar.in.ua reverse DNS resolves to cp22.miraheze.org - CNAME OK [01:53:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 60% [01:53:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 47% [01:53:26] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 55% [01:56:18] PROBLEM - graylog121 Current Load on graylog121 is CRITICAL: CRITICAL - load average: 5.48, 4.01, 3.98 [01:56:18] PROBLEM - es141 Current Load on es141 is CRITICAL: CRITICAL - load average: 4.66, 3.78, 3.51 [01:57:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 44% [01:57:21] RECOVERY - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is OK: OK - NGINX Error Rate is 37% [01:58:18] PROBLEM - es141 Current Load on es141 is WARNING: WARNING - load average: 2.99, 3.50, 3.45 [01:59:09] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 99% [02:00:16] PROBLEM - cp33 NTP time on cp33 is WARNING: NTP WARNING: Offset -0.1191035509 secs [02:00:19] PROBLEM - es141 Current Load on es141 is CRITICAL: CRITICAL - load average: 4.06, 3.56, 3.47 [02:01:06] RECOVERY - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is OK: OK - NGINX Error Rate is 27% [02:01:13] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 51% [02:01:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 56% [02:02:16] RECOVERY - cp33 NTP time on cp33 is OK: NTP OK: Offset -0.08380943537 secs [02:03:11] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 61% [02:03:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 84% [02:06:18] PROBLEM - es141 Current Load on es141 is WARNING: WARNING - load average: 3.27, 3.89, 3.71 [02:07:05] RECOVERY - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is OK: OK - NGINX Error Rate is 38% [02:11:13] PROBLEM - es141 PowerDNS Recursor on es141 is CRITICAL: CRITICAL - Plugin timed out while executing system call [02:12:20] PROBLEM - es141 Current Load on es141 is CRITICAL: CRITICAL - load average: 4.24, 4.08, 3.81 [02:12:43] PROBLEM - cp32 Varnish Backends on cp32 is CRITICAL: 4 backends are down. mw122 mw131 mw132 mw141 [02:12:56] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 58% [02:13:56] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 59% [02:15:03] PROBLEM - mw121 Current Load on mw121 is CRITICAL: CRITICAL - load average: 12.96, 11.14, 8.97 [02:15:15] RECOVERY - es141 PowerDNS Recursor on es141 is OK: DNS OK: 1.028 second response time. miraheze.org returns 109.228.51.216,217.174.247.33,2a00:da00:1800:326::1,2a00:da00:1800:328::1 [02:15:50] RECOVERY - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is OK: OK - NGINX Error Rate is 35% [02:16:19] PROBLEM - es141 Current Load on es141 is WARNING: WARNING - load average: 1.96, 3.41, 3.64 [02:16:33] RECOVERY - cp32 Varnish Backends on cp32 is OK: All 14 backends are healthy [02:17:02] PROBLEM - mw121 Current Load on mw121 is WARNING: WARNING - load average: 10.23, 11.12, 9.24 [02:17:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 59% [02:19:01] RECOVERY - mw121 Current Load on mw121 is OK: OK - load average: 8.86, 10.07, 9.07 [02:19:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 53% [02:19:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 67% [02:19:39] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 46% [02:20:16] PROBLEM - cp33 NTP time on cp33 is WARNING: NTP WARNING: Offset 0.374317795 secs [02:20:19] RECOVERY - es141 Current Load on es141 is OK: OK - load average: 2.13, 2.68, 3.28 [02:20:45] RECOVERY - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is OK: OK - NGINX Error Rate is 34% [02:21:34] RECOVERY - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is OK: OK - NGINX Error Rate is 37% [02:23:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 61% [02:23:57] PROBLEM - cp23 NTP time on cp23 is WARNING: NTP WARNING: Offset 0.1121527851 secs [02:24:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 56% [02:25:23] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 60% [02:25:57] RECOVERY - cp23 NTP time on cp23 is OK: NTP OK: Offset 0.09654131532 secs [02:26:37] RECOVERY - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is OK: OK - NGINX Error Rate is 38% [02:27:17] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 57% [02:29:12] RECOVERY - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is OK: OK - NGINX Error Rate is 31% [02:30:29] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 60% [02:34:23] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 59% [02:35:57] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 43% [02:36:15] RECOVERY - cp33 NTP time on cp33 is OK: NTP OK: Offset -0.02847996354 secs [02:37:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 40% [02:39:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 59% [02:39:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 63% [02:40:15] RECOVERY - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is OK: OK - NGINX Error Rate is 34% [02:41:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 79% [02:41:40] RECOVERY - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is OK: OK - NGINX Error Rate is 37% [02:42:43] PROBLEM - mw121 Current Load on mw121 is CRITICAL: CRITICAL - load average: 18.29, 12.73, 10.14 [02:43:09] PROBLEM - cp33 Varnish Backends on cp33 is CRITICAL: 1 backends are down. mw132 [02:43:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 48% [02:44:42] PROBLEM - mw121 Current Load on mw121 is WARNING: WARNING - load average: 10.28, 11.55, 10.03 [02:45:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 59% [02:45:09] RECOVERY - cp33 Varnish Backends on cp33 is OK: All 14 backends are healthy [02:45:30] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 46% [02:45:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 74% [02:48:14] PROBLEM - cp33 Current Load on cp33 is WARNING: WARNING - load average: 1.40, 2.42, 3.78 [02:48:40] RECOVERY - mw121 Current Load on mw121 is OK: OK - load average: 6.52, 9.33, 9.50 [02:49:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 63% [02:49:18] RECOVERY - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is OK: OK - NGINX Error Rate is 33% [02:49:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 57% [02:49:57] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 62% [02:51:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 51% [02:51:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 68% [02:51:50] PROBLEM - en.religiononfire.mar.in.ua - reverse DNS on sslhost is CRITICAL: rDNS CRITICAL - en.religiononfire.mar.in.ua All nameservers failed to answer the query. [02:51:54] RECOVERY - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is OK: OK - NGINX Error Rate is 33% [02:52:14] PROBLEM - cp33 Current Load on cp33 is CRITICAL: CRITICAL - load average: 8.86, 5.63, 4.70 [02:53:08] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 49% [02:55:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 74% [02:55:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 73% [02:55:46] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 48% [02:57:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 56% [02:57:43] RECOVERY - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is OK: OK - NGINX Error Rate is 28% [02:59:07] RECOVERY - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is OK: OK - NGINX Error Rate is 33% [03:00:38] RECOVERY - db101 Backups SQL on db101 is OK: FILE_AGE OK: /var/log/sql-backup.log is 35 seconds old and 275 bytes [03:01:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 67% [03:01:39] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 47% [03:03:06] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 56% [03:03:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 58% [03:05:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 82% [03:06:32] PROBLEM - cp32 HTTPS on cp32 is CRITICAL: HTTP CRITICAL: HTTP/1.1 503 Backend fetch failed - 8191 bytes in 0.429 second response time [03:07:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 56% [03:09:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 72% [03:09:43] PROBLEM - cp32 Varnish Backends on cp32 is CRITICAL: 7 backends are down. mw121 mw122 mw131 mw132 mw141 mw142 mediawiki [03:09:54] Jeez icinga is freaking out [03:11:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 77% [03:13:06] RECOVERY - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is OK: OK - NGINX Error Rate is 37% [03:13:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 63% [03:13:39] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 52% [03:14:26] PROBLEM - cp32 Current Load on cp32 is WARNING: WARNING - load average: 0.66, 1.44, 3.91 [03:15:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 60% [03:17:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 62% [03:17:09] PROBLEM - cp32 Disk Space on cp32 is WARNING: DISK WARNING - free space: / 8123 MB (10% inode=98%); [03:17:39] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 48% [03:18:14] PROBLEM - cp33 Current Load on cp33 is WARNING: WARNING - load average: 1.44, 2.68, 3.89 [03:18:26] RECOVERY - cp32 Current Load on cp32 is OK: OK - load average: 0.14, 0.75, 3.06 [03:19:06] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 57% [03:19:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 55% [03:19:39] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 82% [03:20:18] PROBLEM - graylog121 Current Load on graylog121 is WARNING: WARNING - load average: 1.98, 3.17, 3.89 [03:20:40] PROBLEM - en.religiononfire.mar.in.ua - reverse DNS on sslhost is WARNING: NoNameservers: All nameservers failed to answer the query mar.in.ua. IN NS: Server 2606:4700:4700::1111 UDP port 53 answered SERVFAIL [03:21:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 65% [03:21:39] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 57% [03:22:14] RECOVERY - cp33 Current Load on cp33 is OK: OK - load average: 1.72, 1.97, 3.30 [03:23:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 63% [03:23:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 52% [03:23:39] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 60% [03:25:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 49% [03:25:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 82% [03:25:39] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 48% [03:27:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 67% [03:27:39] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 85% [03:27:43] RECOVERY - cp32 Varnish Backends on cp32 is OK: All 14 backends are healthy [03:29:06] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 48% [03:31:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 49% [03:31:32] PROBLEM - knowledgebase.clientmanager.co.za - reverse DNS on sslhost is WARNING: Timeout: The DNS operation timed out after 5.407189846038818 seconds [03:33:07] RECOVERY - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is OK: OK - NGINX Error Rate is 36% [03:33:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 78% [03:33:10] [02mw-config] 07Reno-Rex opened pull request 03#5191: Modifying $wgAvailableRights and $wgRestrictionLevels - 13https://github.com/miraheze/mw-config/pull/5191 [03:33:39] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 58% [03:33:43] PROBLEM - cp32 Varnish Backends on cp32 is CRITICAL: 7 backends are down. mw121 mw122 mw131 mw132 mw141 mw142 mediawiki [03:34:07] [02mw-config] 07Reno-Rex edited pull request 03#5191: Modifying $wgAvailableRights and $wgRestrictionLevels - 13https://github.com/miraheze/mw-config/pull/5191 [03:34:13] miraheze/mw-config - Reno-Rex the build passed. [03:35:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 65% [03:36:18] PROBLEM - graylog121 Current Load on graylog121 is CRITICAL: CRITICAL - load average: 4.36, 3.68, 3.64 [03:37:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 49% [03:38:18] PROBLEM - graylog121 Current Load on graylog121 is WARNING: WARNING - load average: 2.84, 3.41, 3.55 [03:39:07] RECOVERY - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is OK: OK - NGINX Error Rate is 39% [03:41:52] [02mw-config] 07Reno-Rex edited pull request 03#5191: Modifying $wgAvailableRights and $wgRestrictionLevels - 13https://github.com/miraheze/mw-config/pull/5191 [03:42:07] PROBLEM - cp23 APT on cp23 is CRITICAL: CHECK_NRPE STATE CRITICAL: Socket timeout after 60 seconds. [03:43:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 75% [03:43:08] PROBLEM - cp32 Disk Space on cp32 is CRITICAL: connect to address 2607:f1c0:1800:8100::1 port 5666: Connection refusedconnect to host 2607:f1c0:1800:8100::1 port 5666: Connection refused [03:44:17] RECOVERY - cp23 APT on cp23 is OK: APT OK: 1 packages available for upgrade (0 critical updates). [03:44:18] PROBLEM - graylog121 Current Load on graylog121 is CRITICAL: CRITICAL - load average: 4.14, 3.57, 3.55 [03:44:34] PROBLEM - cp32 APT on cp32 is CRITICAL: connect to address 2607:f1c0:1800:8100::1 port 5666: Connection refusedconnect to host 2607:f1c0:1800:8100::1 port 5666: Connection refused [03:44:41] PROBLEM - cp32 conntrack_table_size on cp32 is CRITICAL: connect to address 2607:f1c0:1800:8100::1 port 5666: Connection refusedconnect to host 2607:f1c0:1800:8100::1 port 5666: Connection refused [03:44:43] PROBLEM - cp32 Stunnel for mw121 on cp32 is CRITICAL: connect to address 2607:f1c0:1800:8100::1 port 5666: Connection refusedconnect to host 2607:f1c0:1800:8100::1 port 5666: Connection refused [03:44:54] PROBLEM - cp32 Stunnel for mw132 on cp32 is CRITICAL: connect to address 2607:f1c0:1800:8100::1 port 5666: Connection refusedconnect to host 2607:f1c0:1800:8100::1 port 5666: Connection refused [03:45:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 46% [03:45:18] PROBLEM - cp32 SSH on cp32 is CRITICAL: connect to address 2607:f1c0:1800:8100::1 and port 22: Connection refused [03:45:19] PROBLEM - cp32 Stunnel for phab121 on cp32 is CRITICAL: connect to address 2607:f1c0:1800:8100::1 port 5666: Connection refusedconnect to host 2607:f1c0:1800:8100::1 port 5666: Connection refused [03:45:27] PROBLEM - cp32 Stunnel for test131 on cp32 is CRITICAL: connect to address 2607:f1c0:1800:8100::1 port 5666: Connection refusedconnect to host 2607:f1c0:1800:8100::1 port 5666: Connection refused [03:45:39] PROBLEM - cp32 PowerDNS Recursor on cp32 is CRITICAL: connect to address 2607:f1c0:1800:8100::1 port 5666: Connection refusedconnect to host 2607:f1c0:1800:8100::1 port 5666: Connection refused [03:45:42] PROBLEM - ns2 GDNSD Datacenters on ns2 is CRITICAL: CRITICAL - 2 datacenters are down: 108.175.15.182/cpweb, 2607:f1c0:1800:8100::1/cpweb [03:45:42] PROBLEM - cp32 Stunnel for mwtask141 on cp32 is CRITICAL: connect to address 2607:f1c0:1800:8100::1 port 5666: Connection refusedconnect to host 2607:f1c0:1800:8100::1 port 5666: Connection refused [03:45:45] PROBLEM - cp32 Stunnel for reports121 on cp32 is CRITICAL: connect to address 2607:f1c0:1800:8100::1 port 5666: Connection refusedconnect to host 2607:f1c0:1800:8100::1 port 5666: Connection refused [03:45:49] PROBLEM - cp32 Stunnel for mw122 on cp32 is CRITICAL: connect to address 2607:f1c0:1800:8100::1 port 5666: Connection refusedconnect to host 2607:f1c0:1800:8100::1 port 5666: Connection refused [03:45:52] PROBLEM - cp32 Puppet on cp32 is CRITICAL: connect to address 2607:f1c0:1800:8100::1 port 5666: Connection refusedconnect to host 2607:f1c0:1800:8100::1 port 5666: Connection refused [03:45:58] PROBLEM - cp32 Stunnel for matomo131 on cp32 is CRITICAL: connect to address 2607:f1c0:1800:8100::1 port 5666: Connection refusedconnect to host 2607:f1c0:1800:8100::1 port 5666: Connection refused [03:46:02] PROBLEM - cp32 Stunnel for mw131 on cp32 is CRITICAL: connect to address 2607:f1c0:1800:8100::1 port 5666: Connection refusedconnect to host 2607:f1c0:1800:8100::1 port 5666: Connection refused [03:46:14] PROBLEM - cp32 NTP time on cp32 is CRITICAL: connect to address 2607:f1c0:1800:8100::1 port 5666: Connection refusedconnect to host 2607:f1c0:1800:8100::1 port 5666: Connection refused [03:46:14] PROBLEM - cp33 Current Load on cp33 is CRITICAL: CRITICAL - load average: 5.13, 3.78, 2.81 [03:46:15] PROBLEM - cp32 ferm_active on cp32 is CRITICAL: connect to address 2607:f1c0:1800:8100::1 port 5666: Connection refusedconnect to host 2607:f1c0:1800:8100::1 port 5666: Connection refused [03:46:16] PROBLEM - cp32 Stunnel for mail121 on cp32 is CRITICAL: connect to address 2607:f1c0:1800:8100::1 port 5666: Connection refusedconnect to host 2607:f1c0:1800:8100::1 port 5666: Connection refused [03:46:18] PROBLEM - cp32 Stunnel for puppet141 on cp32 is CRITICAL: connect to address 2607:f1c0:1800:8100::1 port 5666: Connection refusedconnect to host 2607:f1c0:1800:8100::1 port 5666: Connection refused [03:46:25] PROBLEM - ns1 GDNSD Datacenters on ns1 is CRITICAL: CRITICAL - 2 datacenters are down: 108.175.15.182/cpweb, 2607:f1c0:1800:8100::1/cpweb [03:46:25] PROBLEM - cp32 Current Load on cp32 is CRITICAL: connect to address 2607:f1c0:1800:8100::1 port 5666: Connection refusedconnect to host 2607:f1c0:1800:8100::1 port 5666: Connection refused [03:46:29] PROBLEM - cp32 Stunnel for mw141 on cp32 is CRITICAL: connect to address 2607:f1c0:1800:8100::1 port 5666: Connection refusedconnect to host 2607:f1c0:1800:8100::1 port 5666: Connection refused [03:46:31] PROBLEM - cp32 Stunnel for mon141 on cp32 is CRITICAL: connect to address 2607:f1c0:1800:8100::1 port 5666: Connection refusedconnect to host 2607:f1c0:1800:8100::1 port 5666: Connection refused [03:46:31] PROBLEM - cp32 Stunnel for mw142 on cp32 is CRITICAL: connect to address 2607:f1c0:1800:8100::1 port 5666: Connection refusedconnect to host 2607:f1c0:1800:8100::1 port 5666: Connection refused [03:47:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 71% [03:47:10] PROBLEM - cp33 Varnish Backends on cp33 is CRITICAL: 5 backends are down. mw121 mw122 mw131 mw132 mw141 [03:47:39] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 54% [03:48:14] PROBLEM - cp33 Current Load on cp33 is WARNING: WARNING - load average: 3.84, 4.00, 3.03 [03:48:18] PROBLEM - graylog121 Current Load on graylog121 is WARNING: WARNING - load average: 3.06, 3.44, 3.51 [03:49:10] RECOVERY - cp33 Varnish Backends on cp33 is OK: All 14 backends are healthy [03:49:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 60% [03:49:50] RECOVERY - en.religiononfire.mar.in.ua - reverse DNS on sslhost is OK: SSL OK - en.religiononfire.mar.in.ua reverse DNS resolves to cp22.miraheze.org - CNAME OK [03:50:15] PROBLEM - cp33 Current Load on cp33 is CRITICAL: CRITICAL - load average: 38.43, 12.96, 6.17 [03:50:18] PROBLEM - graylog121 Current Load on graylog121 is CRITICAL: CRITICAL - load average: 4.24, 3.60, 3.54 [03:51:08] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 55% [03:52:18] PROBLEM - graylog121 Current Load on graylog121 is WARNING: WARNING - load average: 3.01, 3.40, 3.48 [03:53:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 90% [03:53:10] RECOVERY - cp32 Stunnel for mw132 on cp32 is OK: TCP OK - 0.001 second response time on localhost port 8107 [03:53:11] PROBLEM - cp33 Varnish Backends on cp33 is CRITICAL: 7 backends are down. mw121 mw122 mw131 mw132 mw141 mw142 mediawiki [03:53:11] RECOVERY - cp32 conntrack_table_size on cp32 is OK: OK: nf_conntrack is 0 % full [03:53:15] RECOVERY - cp32 Stunnel for mw121 on cp32 is OK: TCP OK - 0.000 second response time on localhost port 8104 [03:53:18] RECOVERY - cp32 SSH on cp32 is OK: SSH OK - OpenSSH_8.4p1 Debian-5+deb11u1 (protocol 2.0) [03:53:19] RECOVERY - cp32 Stunnel for phab121 on cp32 is OK: TCP OK - 0.001 second response time on localhost port 8202 [03:53:28] RECOVERY - cp32 Stunnel for test131 on cp32 is OK: TCP OK - 0.000 second response time on localhost port 8180 [03:53:40] RECOVERY - cp32 PowerDNS Recursor on cp32 is OK: DNS OK: 0.077 seconds response time. miraheze.org returns 108.175.15.182,2607:f1c0:1800:26f::1,2607:f1c0:1800:8100::1,74.208.203.152 [03:53:42] RECOVERY - ns2 GDNSD Datacenters on ns2 is OK: OK - all datacenters are online [03:53:43] RECOVERY - cp32 Stunnel for mwtask141 on cp32 is OK: TCP OK - 0.000 second response time on localhost port 8150 [03:53:43] RECOVERY - cp32 Varnish Backends on cp32 is OK: All 14 backends are healthy [03:53:46] RECOVERY - cp32 Stunnel for reports121 on cp32 is OK: TCP OK - 0.000 second response time on localhost port 8205 [03:53:47] RECOVERY - cp32 Stunnel for mw122 on cp32 is OK: TCP OK - 0.000 second response time on localhost port 8105 [03:53:53] RECOVERY - cp32 Puppet on cp32 is OK: OK: Puppet is currently enabled, last run 16 minutes ago with 0 failures [03:53:57] RECOVERY - cp32 Stunnel for matomo131 on cp32 is OK: TCP OK - 0.000 second response time on localhost port 8203 [03:54:02] RECOVERY - cp32 Stunnel for mw131 on cp32 is OK: TCP OK - 0.000 second response time on localhost port 8106 [03:54:15] RECOVERY - cp32 NTP time on cp32 is OK: NTP OK: Offset -0.002507358789 secs [03:54:15] RECOVERY - cp32 ferm_active on cp32 is OK: OK ferm input default policy is set [03:54:16] RECOVERY - cp32 Stunnel for mail121 on cp32 is OK: TCP OK - 0.000 second response time on localhost port 8200 [03:54:18] RECOVERY - graylog121 Current Load on graylog121 is OK: OK - load average: 2.41, 2.95, 3.30 [03:54:19] RECOVERY - cp32 Stunnel for puppet141 on cp32 is OK: TCP OK - 0.000 second response time on localhost port 8204 [03:54:25] RECOVERY - ns1 GDNSD Datacenters on ns1 is OK: OK - all datacenters are online [03:54:26] RECOVERY - cp32 Current Load on cp32 is OK: OK - load average: 0.52, 0.19, 0.07 [03:54:28] RECOVERY - cp32 HTTPS on cp32 is OK: HTTP OK: HTTP/1.1 301 Moved Permanently - 3713 bytes in 0.530 second response time [03:54:29] RECOVERY - cp32 APT on cp32 is OK: APT OK: 1 packages available for upgrade (0 critical updates). [03:54:30] RECOVERY - cp32 Stunnel for mw141 on cp32 is OK: TCP OK - 0.000 second response time on localhost port 8108 [03:54:31] RECOVERY - cp32 Stunnel for mon141 on cp32 is OK: TCP OK - 0.000 second response time on localhost port 8201 [03:54:32] RECOVERY - cp32 Stunnel for mw142 on cp32 is OK: TCP OK - 0.000 second response time on localhost port 8109 [03:55:09] PROBLEM - cp32 Disk Space on cp32 is WARNING: DISK WARNING - free space: / 7304 MB (9% inode=98%); [03:55:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 57% [03:56:07] PROBLEM - cp33 Disk Space on cp33 is WARNING: DISK WARNING - free space: / 8129 MB (10% inode=98%); [03:57:10] RECOVERY - cp33 Varnish Backends on cp33 is OK: All 14 backends are healthy [03:57:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 81% [03:57:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 55% [03:58:18] PROBLEM - graylog121 Current Load on graylog121 is CRITICAL: CRITICAL - load average: 4.31, 3.61, 3.46 [03:59:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 54% [03:59:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 69% [04:00:32] [Grafana] !sre FIRING: The mediawiki job queue has more than 2500 unclaimed jobs https://grafana.miraheze.org/d/GtxbP1Xnk?orgId=1 [04:03:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 78% [04:03:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 55% [04:05:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 52% [04:05:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 62% [04:07:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 52% [04:09:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 64% [04:09:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 68% [04:11:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 58% [04:11:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 59% [04:13:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 67% [04:15:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 49% [04:17:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 65% [04:19:03] PROBLEM - wiki.nj.cn.eu.org - reverse DNS on sslhost is WARNING: NoNameservers: All nameservers failed to answer the query cn.eu.org. IN NS: Server 2606:4700:4700::1111 UDP port 53 answered SERVFAIL [04:19:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 58% [04:19:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 56% [04:19:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 63% [04:20:53] PROBLEM - cp32 Current Load on cp32 is WARNING: WARNING - load average: 4.00, 2.96, 2.10 [04:21:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 75% [04:21:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 76% [04:21:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 59% [04:21:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 58% [04:22:48] PROBLEM - cp32 Current Load on cp32 is CRITICAL: CRITICAL - load average: 4.74, 3.10, 2.22 [04:23:07] RECOVERY - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is OK: OK - NGINX Error Rate is 39% [04:23:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 76% [04:23:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 67% [04:24:44] PROBLEM - cp32 Current Load on cp32 is WARNING: WARNING - load average: 3.77, 3.52, 2.49 [04:25:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 55% [04:25:43] PROBLEM - cp32 Varnish Backends on cp32 is CRITICAL: 7 backends are down. mw121 mw122 mw131 mw132 mw141 mw142 mediawiki [04:26:23] Hmm I can load meta but am not getting any styles loaded [04:26:40] RECOVERY - cp32 Current Load on cp32 is OK: OK - load average: 1.27, 2.67, 2.29 [04:27:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 82% [04:29:06] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 46% [04:29:40] RECOVERY - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is OK: OK - NGINX Error Rate is 35% [04:29:43] RECOVERY - cp32 Varnish Backends on cp32 is OK: All 14 backends are healthy [04:31:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 63% [04:32:14] PROBLEM - cp33 Current Load on cp33 is WARNING: WARNING - load average: 2.82, 2.87, 3.93 [04:33:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 45% [04:33:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 55% [04:35:06] RECOVERY - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is OK: OK - NGINX Error Rate is 38% [04:35:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 59% [04:35:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 67% [04:36:14] RECOVERY - cp33 Current Load on cp33 is OK: OK - load average: 0.72, 1.92, 3.31 [04:39:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 63% [04:39:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 65% [04:39:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 42% [04:41:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 46% [04:41:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 53% [04:41:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 73% [04:41:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 62% [04:41:58] PROBLEM - cp23 NTP time on cp23 is WARNING: NTP WARNING: Offset -0.1003983319 secs [04:43:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 57% [04:43:57] RECOVERY - cp23 NTP time on cp23 is OK: NTP OK: Offset -0.08292248845 secs [04:44:16] PROBLEM - cp33 NTP time on cp33 is WARNING: NTP WARNING: Offset 0.1668264866 secs [04:45:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 72% [04:45:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 74% [04:45:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 85% [04:47:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 49% [04:47:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 56% [04:48:07] PROBLEM - wiki.nj.cn.eu.org - reverse DNS on sslhost is CRITICAL: rDNS CRITICAL - wiki.nj.cn.eu.org All nameservers failed to answer the query. [04:49:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 66% [04:49:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 64% [04:51:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 46% [04:51:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 56% [04:52:14] PROBLEM - cp33 Current Load on cp33 is WARNING: WARNING - load average: 2.57, 3.75, 3.47 [04:53:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 77% [04:53:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 89% [04:53:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 47% [04:53:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 53% [04:54:14] RECOVERY - cp33 Current Load on cp33 is OK: OK - load average: 1.46, 3.02, 3.24 [04:55:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 46% [04:55:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 66% [04:55:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 68% [04:57:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 60% [04:57:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 57% [04:57:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 57% [04:59:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 40% [04:59:24] RECOVERY - knowledgebase.clientmanager.co.za - reverse DNS on sslhost is OK: SSL OK - knowledgebase.clientmanager.co.za reverse DNS resolves to cp23.miraheze.org - CNAME OK [04:59:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 85% [05:01:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 68% [05:01:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 59% [05:01:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 52% [05:02:14] PROBLEM - cp33 Current Load on cp33 is CRITICAL: CRITICAL - load average: 14.62, 7.22, 4.62 [05:03:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 68% [05:05:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 46% [05:05:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 68% [05:06:15] RECOVERY - cp33 NTP time on cp33 is OK: NTP OK: Offset 0.06084731221 secs [05:07:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 59% [05:07:39] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 61% [05:09:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 73% [05:09:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 56% [05:11:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 80% [05:11:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 42% [05:13:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 46% [05:13:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 74% [05:13:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 61% [05:15:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 64% [05:16:14] PROBLEM - cp33 Current Load on cp33 is WARNING: WARNING - load average: 1.46, 3.02, 3.81 [05:17:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 48% [05:17:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 47% [05:17:11] PROBLEM - wiki.nj.cn.eu.org - reverse DNS on sslhost is WARNING: NoNameservers: All nameservers failed to answer the query cn.eu.org. IN NS: Server 2606:4700:4700::1111 UDP port 53 answered SERVFAIL [05:19:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 65% [05:19:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 65% [05:19:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 53% [05:21:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 55% [05:23:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 55% [05:24:14] RECOVERY - cp33 Current Load on cp33 is OK: OK - load average: 1.32, 2.40, 3.28 [05:25:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 47% [05:25:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 81% [05:25:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 87% [05:27:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 67% [05:27:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 52% [05:29:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 57% [05:29:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 61% [05:31:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 64% [05:31:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 59% [05:31:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 51% [05:33:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 62% [05:35:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 65% [05:37:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 60% [05:39:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 56% [05:41:08] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 52% [05:42:07] [02mw-config] 07Reception123 closed pull request 03#5191: Modifying $wgAvailableRights and $wgRestrictionLevels - 13https://github.com/miraheze/mw-config/pull/5191 [05:42:10] [02miraheze/mw-config] 07Reception123 pushed 031 commit to 03master [+0/-0/±1] 13https://github.com/miraheze/mw-config/compare/e846fcec3e89...335bb3b63da9 [05:42:13] [02miraheze/mw-config] 07Reno-Rex 03335bb3b - Modifying $wgAvailableRights and $wgRestrictionLevels (#5191) [05:42:19] !log [reception@mwtask141] starting deploy of {'pull': 'config', 'config': True} to all [05:42:23] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [05:42:30] !log [reception@mwtask141] finished deploy of {'pull': 'config', 'config': True} to all - SUCCESS in 10s [05:42:33] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [05:43:05] miraheze/mw-config - Reception123 the build passed. [05:43:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 60% [05:43:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 54% [05:45:40] RECOVERY - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is OK: OK - NGINX Error Rate is 28% [05:46:04] PROBLEM - cp23 Current Load on cp23 is CRITICAL: CRITICAL - load average: 32.78, 10.13, 4.57 [05:49:06] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 77% [05:49:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 55% [05:49:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 56% [05:51:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 85% [05:51:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 62% [05:51:58] PROBLEM - cp23 Current Load on cp23 is WARNING: WARNING - load average: 0.60, 3.46, 3.28 [05:53:56] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 0.58, 2.57, 2.97 [05:55:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 70% [05:57:09] !log [@test131] starting deploy of {'config': True} to all [05:57:10] !log [@test131] finished deploy of {'config': True} to all - SUCCESS in 0s [05:57:13] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [05:57:17] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [05:57:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 57% [05:57:39] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 42% [05:59:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 74% [05:59:39] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 60% [06:01:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 52% [06:03:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 69% [06:03:39] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 52% [06:05:39] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 60% [06:07:31] RECOVERY - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is OK: OK - NGINX Error Rate is 39% [06:07:39] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 55% [06:08:26] PROBLEM - cp32 Current Load on cp32 is CRITICAL: CRITICAL - load average: 99.74, 61.08, 24.61 [06:11:09] PROBLEM - cp32 Disk Space on cp32 is CRITICAL: DISK CRITICAL - free space: / 4482 MB (5% inode=98%); [06:11:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 59% [06:13:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 55% [06:13:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 88% [06:13:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 80% [06:15:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 56% [06:17:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 57% [06:17:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 63% [06:17:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 64% [06:19:06] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 66% [06:21:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 52% [06:21:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 48% [06:23:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 82% [06:25:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 59% [06:25:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 46% [06:27:06] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 73% [06:27:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 76% [06:28:52] PROBLEM - cp22 Current Load on cp22 is CRITICAL: CRITICAL - load average: 4.49, 9.24, 5.29 [06:29:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 58% [06:29:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 46% [06:29:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 72% [06:31:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 63% [06:31:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 69% [06:31:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 49% [06:33:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 51% [06:35:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 75% [06:35:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 73% [06:40:52] PROBLEM - cp22 Current Load on cp22 is WARNING: WARNING - load average: 0.79, 2.63, 3.83 [06:41:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 58% [06:43:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 47% [06:43:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 60% [06:44:26] PROBLEM - cp32 Current Load on cp32 is WARNING: WARNING - load average: 0.38, 0.90, 3.71 [06:44:52] RECOVERY - cp22 Current Load on cp22 is OK: OK - load average: 0.20, 1.36, 3.04 [06:45:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 59% [06:45:31] RECOVERY - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is OK: OK - NGINX Error Rate is 38% [06:46:26] RECOVERY - cp32 Current Load on cp32 is OK: OK - load average: 0.83, 0.86, 3.34 [06:47:02] PROBLEM - wiki.nj.cn.eu.org - reverse DNS on sslhost is CRITICAL: rDNS CRITICAL - wiki.nj.cn.eu.org All nameservers failed to answer the query. [06:47:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 74% [06:49:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 76% [06:49:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 62% [06:49:39] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 58% [06:51:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 53% [06:53:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 62% [06:57:39] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 80% [06:59:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 52% [06:59:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 51% [07:01:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 76% [07:11:06] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 49% [07:13:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 64% [07:13:07] RECOVERY - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is OK: OK - NGINX Error Rate is 37% [07:13:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 45% [07:15:31] RECOVERY - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is OK: OK - NGINX Error Rate is 31% [07:17:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 63% [07:17:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 59% [07:18:07] PROBLEM - cp33 Disk Space on cp33 is CRITICAL: DISK CRITICAL - free space: / 4488 MB (5% inode=98%); [07:19:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 70% [07:21:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 58% [07:21:39] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 68% [07:23:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 67% [07:23:39] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 58% [07:24:37] PROBLEM - uk.religiononfire.mar.in.ua - reverse DNS on sslhost is WARNING: NoNameservers: All nameservers failed to answer the query mar.in.ua. IN NS: Server 2606:4700:4700::1111 UDP port 53 answered SERVFAIL [07:25:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 55% [07:25:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 52% [07:25:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 51% [07:25:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 64% [07:27:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 62% [07:27:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 70% [07:29:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 65% [07:31:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 45% [07:33:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 46% [07:35:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 69% [07:37:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 67% [07:39:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 59% [07:39:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 55% [07:39:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 42% [07:41:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 71% [07:43:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 73% [07:43:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 63% [07:45:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 54% [07:45:07] RECOVERY - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is OK: OK - NGINX Error Rate is 37% [07:47:04] PROBLEM - wiki.nj.cn.eu.org - reverse DNS on sslhost is WARNING: NoNameservers: All nameservers failed to answer the query cn.eu.org. IN NS: Server 2606:4700:4700::1111 UDP port 53 answered SERVFAIL [07:47:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 63% [07:47:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 51% [07:49:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 65% [07:53:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 59% [07:55:07] RECOVERY - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is OK: OK - NGINX Error Rate is 39% [07:55:07] RECOVERY - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is OK: OK - NGINX Error Rate is 39% [07:55:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 67% [07:57:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 54% [07:57:39] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 57% [07:59:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 59% [07:59:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 82% [08:00:32] [Grafana] !sre FIRING: The mediawiki job queue has more than 2500 unclaimed jobs https://grafana.miraheze.org/d/GtxbP1Xnk?orgId=1 [08:01:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 64% [08:01:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 62% [08:03:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 67% [08:03:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 46% [08:05:23] PROBLEM - cp23 Current Load on cp23 is CRITICAL: CRITICAL - load average: 3.43, 4.88, 3.35 [08:05:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 50% [08:07:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 65% [08:07:22] PROBLEM - cp23 Current Load on cp23 is WARNING: WARNING - load average: 1.05, 3.60, 3.08 [08:09:20] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 2.90, 3.27, 3.00 [08:11:31] RECOVERY - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is OK: OK - NGINX Error Rate is 35% [08:13:04] PROBLEM - cloud10 Puppet on cloud10 is CRITICAL: CRITICAL: Puppet has 1 failures. Last run 3 minutes ago with 1 failures. Failed resources (up to 3 shown): Service[ulogd2] [08:13:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 41% [08:13:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 45% [08:15:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 74% [08:15:39] RECOVERY - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is OK: OK - NGINX Error Rate is 18% [08:17:02] RECOVERY - wiki.nj.cn.eu.org - reverse DNS on sslhost is OK: SSL OK - wiki.nj.cn.eu.org reverse DNS resolves to cp23.miraheze.org - CNAME OK [08:17:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 55% [08:17:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 67% [08:17:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 52% [08:18:29] I guess the errors are kinda to be expected as images are cleared from the cache. If every skin shows at least the site logo and there are pages like the AVID main page that shows a lot of images, then I guess this is because of that [08:19:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 69% [08:19:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 49% [08:19:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 93% [08:20:59] but also, on another note, bluepageswiki has again 10000+ refreshLinks jobs [08:21:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 67% [08:23:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 50% [08:23:46] RECOVERY - uk.religiononfire.mar.in.ua - reverse DNS on sslhost is OK: SSL OK - uk.religiononfire.mar.in.ua reverse DNS resolves to cp22.miraheze.org - CNAME OK [08:24:02] the rise in refreshLinks job seems to coincide with htmlCacheUpdate jobs this time: https://grafana.miraheze.org/d/GtxbP1Xnk/mediawiki?orgId=1&from=now-24h&to=now&var-node=jobchron121&var-job=htmlCacheUpdate&var-job=LocalGlobalUserPageCacheUpdateJob [08:24:29] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 51% [08:25:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 52% [08:25:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 73% [08:25:31] RECOVERY - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is OK: OK - NGINX Error Rate is 39% [08:26:26] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 87% [08:27:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 56% [08:29:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 61% [08:29:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 46% [08:31:06] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 48% [08:33:06] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 66% [08:35:06] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 44% [08:37:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 65% [08:38:06] PROBLEM - cp22 Current Load on cp22 is WARNING: WARNING - load average: 3.82, 3.57, 2.61 [08:39:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 75% [08:39:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 82% [08:40:04] RECOVERY - cp22 Current Load on cp22 is OK: OK - load average: 0.73, 2.48, 2.32 [08:41:04] RECOVERY - cloud10 Puppet on cloud10 is OK: OK: Puppet is currently enabled, last run 1 minute ago with 0 failures [08:41:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 42% [08:43:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 59% [08:45:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 67% [08:45:08] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 97% [08:45:46] PROBLEM - cp23 Current Load on cp23 is CRITICAL: CRITICAL - load average: 3.88, 4.64, 3.15 [08:47:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 50% [08:47:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 40% [08:47:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 51% [08:47:46] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 1.37, 3.36, 2.86 [08:49:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 60% [08:49:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 65% [08:49:31] RECOVERY - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is OK: OK - NGINX Error Rate is 23% [08:51:46] PROBLEM - cp23 Current Load on cp23 is WARNING: WARNING - load average: 1.64, 3.44, 3.09 [08:53:46] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 0.84, 2.53, 2.79 [08:55:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 54% [08:55:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 41% [08:59:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 89% [09:00:20] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 53% [09:01:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 83% [09:02:14] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 75% [09:04:08] RECOVERY - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is OK: OK - NGINX Error Rate is 39% [09:07:58] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 70% [09:09:08] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 48% [09:11:47] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 58% [09:13:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 50% [09:13:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 69% [09:13:41] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 90% [09:15:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 74% [09:19:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 59% [09:21:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 52% [09:23:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 82% [09:26:52] PROBLEM - cp22 Current Load on cp22 is CRITICAL: CRITICAL - load average: 84.90, 176.04, 86.00 [09:29:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 74% [09:33:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 54% [09:33:55] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 56% [09:35:52] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 80% [09:37:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 53% [09:37:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 70% [09:37:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 47% [09:39:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 70% [09:39:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 47% [09:41:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 62% [09:41:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 72% [09:43:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 53% [09:44:27] PROBLEM - cp23 Current Load on cp23 is WARNING: WARNING - load average: 3.39, 3.60, 2.84 [09:46:25] PROBLEM - cp23 Current Load on cp23 is CRITICAL: CRITICAL - load average: 4.22, 3.63, 2.93 [09:47:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 49% [09:48:24] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 2.58, 3.12, 2.82 [09:49:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 66% [09:57:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 64% [09:58:14] PROBLEM - cp23 Current Load on cp23 is WARNING: WARNING - load average: 2.81, 3.41, 3.09 [09:59:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 54% [10:00:12] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 1.82, 2.84, 2.92 [10:01:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 80% [10:03:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 53% [10:03:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 59% [10:05:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 52% [10:07:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 61% [10:08:49] PROBLEM - cp22 Disk Space on cp22 is WARNING: DISK WARNING - free space: / 8185 MB (10% inode=98%); [10:09:06] PROBLEM - cp23 Disk Space on cp23 is WARNING: DISK WARNING - free space: / 8165 MB (10% inode=98%); [10:09:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 49% [10:11:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 76% [10:11:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 64% [10:13:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 59% [10:17:06] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 58% [10:17:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 75% [10:17:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 64% [10:19:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 79% [10:19:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 53% [10:19:39] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 55% [10:21:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 74% [10:21:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 56% [10:23:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 73% [10:27:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 79% [10:28:52] PROBLEM - cp22 Current Load on cp22 is WARNING: WARNING - load average: 3.98, 2.25, 3.92 [10:29:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 42% [10:31:07] RECOVERY - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is OK: OK - NGINX Error Rate is 35% [10:32:52] PROBLEM - cp22 Current Load on cp22 is CRITICAL: CRITICAL - load average: 5.37, 3.69, 4.09 [10:35:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 51% [10:35:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 52% [10:36:52] PROBLEM - cp22 Current Load on cp22 is WARNING: WARNING - load average: 1.45, 2.80, 3.68 [10:37:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 67% [10:38:52] RECOVERY - cp22 Current Load on cp22 is OK: OK - load average: 1.30, 2.17, 3.34 [10:39:31] RECOVERY - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is OK: OK - NGINX Error Rate is 34% [10:39:39] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 52% [10:41:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 66% [10:43:06] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 50% [10:43:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 57% [10:43:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 51% [10:45:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 72% [10:47:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 86% [10:47:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 55% [10:49:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 54% [10:49:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 78% [10:49:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 79% [10:51:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 80% [10:53:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 41% [10:53:46] PROBLEM - cp23 Current Load on cp23 is WARNING: WARNING - load average: 2.27, 3.61, 2.71 [10:55:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 70% [10:55:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 59% [10:55:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 52% [10:55:46] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 1.09, 2.67, 2.47 [10:57:07] RECOVERY - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is OK: OK - NGINX Error Rate is 32% [10:57:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 67% [10:59:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 59% [11:01:31] RECOVERY - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is OK: OK - NGINX Error Rate is 35% [11:01:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 75% [11:03:58] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 50% [11:05:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 44% [11:05:32] [Grafana] !sre RESOLVED: High Job Queue Backlog https://grafana.miraheze.org/d/GtxbP1Xnk?orgId=1 [11:05:53] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 71% [11:07:48] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 55% [11:09:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 80% [11:09:57] PROBLEM - cp23 NTP time on cp23 is WARNING: NTP WARNING: Offset 0.2415628433 secs [11:11:38] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 67% [11:13:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 54% [11:13:33] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 42% [11:13:39] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 59% [11:15:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 71% [11:15:39] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 73% [11:17:24] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 88% [11:19:18] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 48% [11:21:14] RECOVERY - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is OK: OK - NGINX Error Rate is 28% [11:21:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 49% [11:21:50] PROBLEM - en.religiononfire.mar.in.ua - reverse DNS on sslhost is WARNING: NoNameservers: All nameservers failed to answer the query en.religiononfire.mar.in.ua. IN CNAME: Server 2606:4700:4700::1111 UDP port 53 answered SERVFAIL [11:25:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 75% [11:25:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 64% [11:25:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 58% [11:27:32] PROBLEM - cp22 Current Load on cp22 is CRITICAL: CRITICAL - load average: 4.22, 3.74, 2.63 [11:27:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 68% [11:29:30] RECOVERY - cp22 Current Load on cp22 is OK: OK - load average: 2.73, 3.23, 2.56 [11:30:36] PROBLEM - cp23 Current Load on cp23 is CRITICAL: CRITICAL - load average: 6.16, 4.27, 3.14 [11:31:57] RECOVERY - cp23 NTP time on cp23 is OK: NTP OK: Offset -0.01240333915 secs [11:35:19] PROBLEM - gs.sidem.wiki - reverse DNS on sslhost is WARNING: Timeout: The DNS operation timed out after 5.407392978668213 seconds [11:35:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 52% [11:36:31] PROBLEM - cp23 Current Load on cp23 is WARNING: WARNING - load average: 1.72, 3.50, 3.26 [11:38:29] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 1.59, 2.81, 3.03 [11:39:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 72% [11:39:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 47% [11:41:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 44% [11:42:34] PROBLEM - db112 Disk Space on db112 is CRITICAL: DISK CRITICAL - free space: / 7963 MB (5% inode=99%); [11:43:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 63% [11:43:39] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 76% [11:44:20] [02mw-config] 07ugochimobi opened pull request 03#5192: T10722: configure wgImportSources for brolandiawiki - 13https://github.com/miraheze/mw-config/pull/5192 [11:45:17] miraheze/mw-config - ugochimobi the build passed. [11:45:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 55% [11:45:40] RECOVERY - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is OK: OK - NGINX Error Rate is 38% [11:46:34] PROBLEM - db112 Disk Space on db112 is WARNING: DISK WARNING - free space: / 8972 MB (6% inode=99%); [11:47:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 66% [11:49:02] PROBLEM - wiki.nj.cn.eu.org - reverse DNS on sslhost is WARNING: NoNameservers: All nameservers failed to answer the query cn.eu.org. IN NS: Server 2606:4700:4700::1111 UDP port 53 answered SERVFAIL [11:49:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 58% [11:50:40] RECOVERY - en.religiononfire.mar.in.ua - reverse DNS on sslhost is OK: SSL OK - en.religiononfire.mar.in.ua reverse DNS resolves to cp22.miraheze.org - CNAME OK [11:51:39] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 67% [11:57:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 45% [11:59:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 96% [11:59:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 46% [12:00:25] PROBLEM - cp23 Varnish Backends on cp23 is CRITICAL: 7 backends are down. mw121 mw122 mw131 mw132 mw141 mw142 mediawiki [12:01:07] RECOVERY - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is OK: OK - NGINX Error Rate is 39% [12:01:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 67% [12:01:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 57% [12:02:25] RECOVERY - cp23 Varnish Backends on cp23 is OK: All 14 backends are healthy [12:05:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 65% [12:05:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 43% [12:05:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 92% [12:07:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 57% [12:07:07] RECOVERY - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is OK: OK - NGINX Error Rate is 38% [12:08:14] PROBLEM - cp33 Current Load on cp33 is CRITICAL: CRITICAL - load average: 7.55, 4.57, 2.77 [12:09:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 65% [12:11:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 56% [12:11:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 57% [12:12:14] PROBLEM - cp33 Current Load on cp33 is WARNING: WARNING - load average: 2.10, 3.58, 2.81 [12:12:19] PROBLEM - es141 Current Load on es141 is WARNING: WARNING - load average: 3.81, 3.06, 2.32 [12:13:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 91% [12:13:08] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 81% [12:14:14] RECOVERY - cp33 Current Load on cp33 is OK: OK - load average: 2.10, 3.23, 2.78 [12:14:18] RECOVERY - es141 Current Load on es141 is OK: OK - load average: 2.59, 2.95, 2.37 [12:15:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 58% [12:17:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 64% [12:18:05] PROBLEM - wiki.nj.cn.eu.org - reverse DNS on sslhost is CRITICAL: rDNS CRITICAL - wiki.nj.cn.eu.org All nameservers failed to answer the query. [12:19:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 48% [12:21:06] PROBLEM - cp23 Disk Space on cp23 is CRITICAL: DISK CRITICAL - free space: / 4470 MB (5% inode=98%); [12:21:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 58% [12:21:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 74% [12:23:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 81% [12:25:07] RECOVERY - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is OK: OK - NGINX Error Rate is 35% [12:27:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 50% [12:28:49] PROBLEM - cp22 Disk Space on cp22 is CRITICAL: DISK CRITICAL - free space: / 4442 MB (5% inode=98%); [12:29:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is WARNING: WARNING - NGINX Error Rate is 47% [12:29:39] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 86% [12:33:07] RECOVERY - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is OK: OK - NGINX Error Rate is 39% [12:35:02] PROBLEM - cp22 Current Load on cp22 is CRITICAL: CRITICAL - load average: 4.33, 3.89, 3.02 [12:36:25] PROBLEM - ns1 GDNSD Datacenters on ns1 is CRITICAL: CRITICAL - 2 datacenters are down: 108.175.15.182/cpweb, 2607:f1c0:1800:8100::1/cpweb [12:36:28] PROBLEM - cp32 HTTPS on cp32 is CRITICAL: HTTP CRITICAL: HTTP/1.1 502 Bad Gateway - 328 bytes in 0.462 second response time [12:37:00] PROBLEM - cp22 Current Load on cp22 is WARNING: WARNING - load average: 2.35, 3.45, 2.97 [12:37:07] PROBLEM - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is CRITICAL: CRITICAL - NGINX Error Rate is 79% [12:37:42] PROBLEM - ns2 GDNSD Datacenters on ns2 is CRITICAL: CRITICAL - 2 datacenters are down: 108.175.15.182/cpweb, 2607:f1c0:1800:8100::1/cpweb [12:37:44] PROBLEM - cp32 Varnish Backends on cp32 is WARNING: No backends detected. If this is an error, see readme.txt [12:38:57] RECOVERY - cp22 Current Load on cp22 is OK: OK - load average: 2.18, 3.14, 2.91 [12:39:57] PROBLEM - cp32 Puppet on cp32 is CRITICAL: CRITICAL: Puppet has 1 failures. Last run 2 minutes ago with 1 failures. Failed resources (up to 3 shown): Service[varnish] [12:41:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 70% [12:41:28] PROBLEM - cp33 Current Load on cp33 is CRITICAL: CRITICAL - load average: 3.88, 4.58, 3.52 [12:43:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 43% [12:43:09] PROBLEM - cp33 Varnish Backends on cp33 is CRITICAL: 1 backends are down. mw141 [12:45:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 89% [12:46:52] PROBLEM - cp22 Current Load on cp22 is WARNING: WARNING - load average: 3.14, 3.51, 3.10 [12:47:07] RECOVERY - wiki.nj.cn.eu.org - reverse DNS on sslhost is OK: SSL OK - wiki.nj.cn.eu.org reverse DNS resolves to cp22.miraheze.org - CNAME OK [12:47:09] RECOVERY - cp33 Varnish Backends on cp33 is OK: All 14 backends are healthy [12:48:52] RECOVERY - cp22 Current Load on cp22 is OK: OK - load average: 2.42, 3.13, 3.01 [12:49:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 47% [12:51:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 65% [12:53:13] PROBLEM - cp33 Varnish Backends on cp33 is CRITICAL: 1 backends are down. mw141 [12:54:52] PROBLEM - cp22 Current Load on cp22 is CRITICAL: CRITICAL - load average: 7.18, 4.16, 3.39 [12:55:10] PROBLEM - cp33 Varnish Backends on cp33 is WARNING: No backends detected. If this is an error, see readme.txt [12:55:17] PROBLEM - cp23 Current Load on cp23 is WARNING: WARNING - load average: 3.45, 2.83, 2.44 [12:56:19] PROBLEM - cp33 HTTPS on cp33 is CRITICAL: HTTP CRITICAL: HTTP/1.1 502 Bad Gateway - 328 bytes in 0.423 second response time [12:56:45] PROBLEM - cp22 Varnish Backends on cp22 is CRITICAL: 4 backends are down. mw121 mw122 mw131 mw142 [12:56:52] PROBLEM - cp22 Current Load on cp22 is WARNING: WARNING - load average: 2.76, 3.59, 3.28 [12:56:57] PROBLEM - cp33 Current Load on cp33 is WARNING: WARNING - load average: 0.45, 2.87, 3.82 [12:57:16] PROBLEM - cp23 Current Load on cp23 is CRITICAL: CRITICAL - load average: 9.12, 5.70, 3.55 [12:58:28] PROBLEM - cp33 Puppet on cp33 is CRITICAL: CRITICAL: Puppet has 1 failures. Last run 2 minutes ago with 1 failures. Failed resources (up to 3 shown): Service[varnish] [12:58:45] RECOVERY - cp22 Varnish Backends on cp22 is OK: All 14 backends are healthy [12:58:52] PROBLEM - cp22 Current Load on cp22 is CRITICAL: CRITICAL - load average: 31.78, 17.30, 8.45 [12:59:42] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 56% [13:00:00] [02miraheze/mw-config] 07Reception123 pushed 031 commit to 03master [+0/-0/±1] 13https://github.com/miraheze/mw-config/compare/335bb3b63da9...7698e02b9b07 [13:00:01] [02miraheze/mw-config] 07Reception123 037698e02 - disable file uploads [13:00:13] PROBLEM - Host swiftproxy131 is DOWN: CRITICAL - Destination Unreachable (2a10:6740::6:315) [13:00:49] RECOVERY - cp33 Current Load on cp33 is OK: OK - load average: 0.31, 1.52, 3.06 [13:01:08] miraheze/mw-config - Reception123 the build has errored. [13:01:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 70% [13:03:09] RECOVERY - gs.sidem.wiki - reverse DNS on sslhost is OK: SSL OK - gs.sidem.wiki reverse DNS resolves to cp22.miraheze.org - CNAME OK [13:03:24] !log [reception@mwtask141] starting deploy of {'pull': 'config', 'config': True} to all [13:03:28] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [13:03:33] !log [reception@mwtask141] finished deploy of {'pull': 'config', 'config': True} to all - SUCCESS in 8s [13:03:37] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [13:10:29] [02miraheze/mw-config] 07Reception123 pushed 031 commit to 03master [+0/-0/±1] 13https://github.com/miraheze/mw-config/compare/7698e02b9b07...da2c70aa5781 [13:10:31] [02miraheze/mw-config] 07Reception123 03da2c70a - restrict file uploads due to swift/cloud11 issues [13:10:31] !log [reception@mwtask141] starting deploy of {'pull': 'config', 'config': True} to all [13:10:35] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [13:10:40] !log [reception@mwtask141] finished deploy of {'pull': 'config', 'config': True} to all - SUCCESS in 8s [13:10:44] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [13:11:35] miraheze/mw-config - Reception123 the build has errored. [13:16:07] [02miraheze/mw-config] 07Reception123 pushed 032 commits to 03master [+0/-0/±2] 13https://github.com/miraheze/mw-config/compare/da2c70aa5781...f47b88b4659d [13:16:08] [02miraheze/mw-config] 07Reception123 032835798 - Revert "restrict file uploads due to swift/cloud11 issues" [13:16:10] [02miraheze/mw-config] 07Reception123 03f47b88b - Revert "disable file uploads" [13:17:14] miraheze/mw-config - Reception123 the build passed. [13:17:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 59% [13:18:52] PROBLEM - cp22 Current Load on cp22 is WARNING: WARNING - load average: 0.87, 1.92, 3.93 [13:19:39] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 66% [13:22:53] PROBLEM - cp23 Current Load on cp23 is WARNING: WARNING - load average: 1.16, 1.90, 3.66 [13:24:52] RECOVERY - cp22 Current Load on cp22 is OK: OK - load average: 2.09, 2.15, 3.39 [13:28:47] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 1.14, 1.91, 3.17 [13:31:38] PROBLEM - test131 MediaWiki Rendering on test131 is CRITICAL: HTTP CRITICAL: HTTP/1.1 500 Internal Server Error - 700 bytes in 0.677 second response time [13:39:48] !log [@mwtask141] starting deploy of {'config': True} to all [13:40:01] !log [@mwtask141] finished deploy of {'config': True} to all - SUCCESS in 13s [13:47:33] [02miraheze/mw-config] 07Reception123 pushed 031 commit to 03master [+0/-0/±1] 13https://github.com/miraheze/mw-config/compare/f47b88b4659d...eedcead5953b [13:47:35] [02miraheze/mw-config] 07Reception123 03eedcead - make file sitenotice display everywhere [13:47:47] PROBLEM - wiki.yuanpi.eu.org - LetsEncrypt on sslhost is WARNING: WARNING - Certificate 'wiki.yuanpi.eu.org' expires in 15 day(s) (Fri 28 Apr 2023 13:20:24 GMT +0000). [13:48:14] [02miraheze/mw-config] 07Reception123 pushed 031 commit to 03master [+0/-0/±1] 13https://github.com/miraheze/mw-config/compare/eedcead5953b...b16a81f37261 [13:48:16] !log [reception@mwtask141] starting deploy of {'pull': 'config', 'config': True} to all [13:48:17] [02miraheze/mw-config] 07Reception123 03b16a81f - slight wording modification for sitenotice [13:48:24] !log [reception@mwtask141] finished deploy of {'pull': 'config', 'config': True} to all - SUCCESS in 7s [13:48:30] miraheze/mw-config - Reception123 the build passed. [13:48:51] MacFan4000: hmm any idea why that doesn't work? [13:49:13] miraheze/mw-config - Reception123 the build passed. [13:49:31] ah nevermind it does now [13:49:33] [02miraheze/ssl] 07MirahezeSSLBot pushed 031 commit to 03master [+0/-0/±1] 13https://github.com/miraheze/ssl/compare/5fe2077258b1...fa4a9306433e [13:49:34] [02miraheze/ssl] 07MirahezeSSLBot 03fa4a930 - Bot: Update SSL cert for wiki.yuanpi.eu.org [13:52:05] RECOVERY - cp33 HTTPS on cp33 is OK: HTTP OK: HTTP/1.1 301 Moved Permanently - 3713 bytes in 0.534 second response time [13:52:33] [02puppet] 07redbluegreenhat opened pull request 03#3186: Return Varnish's default 200 response on all requests to static - 13https://github.com/miraheze/puppet/pull/3186 [13:53:09] RECOVERY - cp33 Varnish Backends on cp33 is OK: All 14 backends are healthy [13:54:37] [02puppet] 07paladox closed pull request 03#3186: Return Varnish's default 200 response on all requests to static - 13https://github.com/miraheze/puppet/pull/3186 [13:54:39] [02miraheze/puppet] 07paladox pushed 031 commit to 03master [+0/-0/±1] 13https://github.com/miraheze/puppet/compare/e0d4bbbcd78a...fecdaf3c39c8 [13:54:41] [02miraheze/puppet] 07redbluegreenhat 03fecdaf3 - Return Varnish's default 200 response on all requests to static (#3186) [13:55:13] PROBLEM - zhacg.wiki - LetsEncrypt on sslhost is WARNING: WARNING - Certificate 'zhacg.wiki' expires in 15 day(s) (Fri 28 Apr 2023 13:32:33 GMT +0000). [13:56:13] [02miraheze/ssl] 07MirahezeSSLBot pushed 031 commit to 03master [+0/-0/±1] 13https://github.com/miraheze/ssl/compare/fa4a9306433e...95c3122acf7c [13:56:16] [02miraheze/ssl] 07MirahezeSSLBot 0395c3122 - Bot: Update SSL cert for zhacg.wiki [13:56:27] RECOVERY - cp33 Puppet on cp33 is OK: OK: Puppet is currently enabled, last run 1 minute ago with 0 failures [13:57:55] !log [@test131] starting deploy of {'config': True} to all [13:57:56] !log [@test131] finished deploy of {'config': True} to all - SUCCESS in 0s [13:59:36] RECOVERY - test131 MediaWiki Rendering on test131 is OK: HTTP OK: HTTP/1.1 200 OK - 8191 bytes in 0.821 second response time [14:00:08] PROBLEM - cp23 Puppet on cp23 is CRITICAL: CRITICAL: Puppet has 1 failures. Last run 2 minutes ago with 1 failures. Failed resources (up to 3 shown): Exec[load-new-vcl-file] [14:01:26] PROBLEM - cp22 Puppet on cp22 is CRITICAL: CRITICAL: Puppet has 1 failures. Last run 3 minutes ago with 1 failures. Failed resources (up to 3 shown): Exec[load-new-vcl-file] [14:02:07] RECOVERY - cp33 Disk Space on cp33 is OK: DISK OK - free space: / 11608 MB (15% inode=98%); [14:02:28] RECOVERY - cp32 HTTPS on cp32 is OK: HTTP OK: HTTP/1.1 301 Moved Permanently - 3713 bytes in 0.535 second response time [14:02:30] PROBLEM - cp33 Puppet on cp33 is CRITICAL: CRITICAL: Puppet has 1 failures. Last run 2 minutes ago with 1 failures. Failed resources (up to 3 shown): Exec[load-new-vcl-file] [14:03:07] RECOVERY - cp33 HTTP 4xx/5xx ERROR Rate on cp33 is OK: OK - NGINX Error Rate is 1% [14:03:09] RECOVERY - cp32 Disk Space on cp32 is OK: DISK OK - free space: / 11545 MB (15% inode=98%); [14:03:31] RECOVERY - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is OK: OK - NGINX Error Rate is 1% [14:03:42] RECOVERY - ns2 GDNSD Datacenters on ns2 is OK: OK - all datacenters are online [14:03:43] RECOVERY - cp32 Varnish Backends on cp32 is OK: All 14 backends are healthy [14:03:43] well would you look at that [14:03:55] 1% error rate [14:04:01] Seems like puppet is failing on that file though [14:04:12] at least one proxy managed to update at least [14:04:25] RECOVERY - ns1 GDNSD Datacenters on ns1 is OK: OK - all datacenters are online [14:04:36] cp ran out of space [14:04:50] RECOVERY - cp22 Disk Space on cp22 is OK: DISK OK - free space: / 12262 MB (16% inode=98%); [14:05:06] RECOVERY - cp23 Disk Space on cp23 is OK: DISK OK - free space: / 12206 MB (16% inode=98%); [14:06:07] RECOVERY - cp23 Puppet on cp23 is OK: OK: Puppet is currently enabled, last run 21 seconds ago with 0 failures [14:06:27] RECOVERY - cp33 Puppet on cp33 is OK: OK: Puppet is currently enabled, last run 17 seconds ago with 0 failures [14:07:07] RECOVERY - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is OK: OK - NGINX Error Rate is 2% [14:07:26] RECOVERY - cp22 Puppet on cp22 is OK: OK: Puppet is currently enabled, last run 1 minute ago with 0 failures [14:07:40] RECOVERY - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is OK: OK - NGINX Error Rate is 0% [14:07:55] RECOVERY - cp32 Puppet on cp32 is OK: OK: Puppet is currently enabled, last run 1 minute ago with 0 failures [14:08:18] PROBLEM - graylog121 Current Load on graylog121 is WARNING: WARNING - load average: 1.77, 2.86, 3.92 [14:08:20] not suspiring given all the errors [14:08:27] thanks Orange_Star for the temporary fix! [14:08:31] no problem [14:12:18] RECOVERY - graylog121 Current Load on graylog121 is OK: OK - load average: 1.20, 2.02, 3.33 [14:17:03] RECOVERY - wiki.yuanpi.eu.org - LetsEncrypt on sslhost is OK: OK - Certificate 'wiki.yuanpi.eu.org' will expire on Tue 11 Jul 2023 12:49:24 GMT +0000. [14:24:26] RECOVERY - zhacg.wiki - LetsEncrypt on sslhost is OK: OK - Certificate 'zhacg.wiki' will expire on Tue 11 Jul 2023 12:56:06 GMT +0000. [14:31:03] PROBLEM - test131 Puppet on test131 is CRITICAL: CRITICAL: Puppet has 1 failures. Last run 3 minutes ago with 1 failures. Failed resources (up to 3 shown): Service[nginx] [14:59:01] RECOVERY - test131 Puppet on test131 is OK: OK: Puppet is currently enabled, last run 1 minute ago with 0 failures [15:10:03] PROBLEM - www.sidem.wiki - reverse DNS on sslhost is WARNING: NoNameservers: All nameservers failed to answer the query sidem.wiki. IN NS: Server 2606:4700:4700::1111 UDP port 53 answered SERVFAIL [15:26:34] PROBLEM - db112 Disk Space on db112 is CRITICAL: DISK CRITICAL - free space: / 7969 MB (5% inode=99%); [15:39:02] RECOVERY - www.sidem.wiki - reverse DNS on sslhost is OK: SSL OK - www.sidem.wiki reverse DNS resolves to cp23.miraheze.org - CNAME OK [15:48:34] PROBLEM - db112 Disk Space on db112 is WARNING: DISK WARNING - free space: / 9914 MB (7% inode=99%); [15:56:59] [02miraheze/mw-config] 07MacFan4000 pushed 031 commit to 03master [+0/-0/±1] 13https://github.com/miraheze/mw-config/compare/b16a81f37261...891b2a0d928a [15:57:00] [02miraheze/mw-config] 07MacFan4000 03891b2a0 - disable importdump requests and file uploads [15:57:02] !log [@test131] starting deploy of {'config': True} to all [15:57:03] !log [@test131] finished deploy of {'config': True} to all - SUCCESS in 0s [15:57:06] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [15:57:09] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [15:57:30] Reception123: ^ found a solution - revoke the upload permission [15:57:38] (tested on beta) [15:58:02] miraheze/mw-config - MacFan4000 the build passed. [15:59:36] Will that mess things up when it's re-enabled? [15:59:48] If not seems fine to do. I don't have access now though [15:59:54] !log [@mwtask141] starting deploy of {'config': True} to all [15:59:58] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [16:00:01] !log [@mwtask141] finished deploy of {'config': True} to all - SUCCESS in 6s [16:00:05] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [16:00:12] no, $wgRevokePermissions was never in ManageWiki which is why it works [16:11:56] also for SRE and stewards it will still look like it works as global upload would override a local revocation [16:12:45] but I can confirm with with my testing non-privileged account that uploads for everybody else are effectively disabled [16:24:17] PROBLEM - db142 Current Load on db142 is WARNING: WARNING - load average: 7.77, 5.42, 2.70 [16:26:18] RECOVERY - db142 Current Load on db142 is OK: OK - load average: 2.92, 4.80, 2.83 [18:04:17] PROBLEM - cp22 NTP time on cp22 is WARNING: NTP WARNING: Offset 0.3187492788 secs [18:06:34] PROBLEM - db112 Disk Space on db112 is CRITICAL: DISK CRITICAL - free space: / 7962 MB (5% inode=99%); [18:12:26] PROBLEM - db112 Current Load on db112 is CRITICAL: CRITICAL - load average: 11.65, 10.09, 5.69 [18:13:12] !log [salt-user@mwtask141] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=tsbwiki (END - exit=65280) [18:13:14] !log [salt-user@mw141] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=tsbwiki (END - exit=0) [18:13:15] !log [salt-user@mw142] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=tsbwiki (END - exit=0) [18:13:16] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [18:13:16] !log [salt-user@mw131] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=tsbwiki (END - exit=0) [18:13:17] !log [salt-user@mw121] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=tsbwiki (END - exit=0) [18:13:18] !log [salt-user@mw132] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=tsbwiki (END - exit=0) [18:13:19] !log [salt-user@mw122] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=tsbwiki (END - exit=0) [18:13:20] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [18:13:24] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [18:13:29] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [18:13:32] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [18:13:37] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [18:13:41] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [18:13:58] !log [salt-user@mw131] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=sagan4alphawiki (END - exit=65280) [18:13:59] !log [salt-user@mwtask141] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=sagan4alphawiki (END - exit=65280) [18:14:00] !log [salt-user@mw141] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=sagan4alphawiki (END - exit=65280) [18:14:02] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [18:14:06] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [18:14:07] !log [salt-user@mw121] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=sagan4alphawiki (END - exit=256) [18:14:10] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [18:14:14] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [18:14:26] !log [salt-user@mw142] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=sagan4alphawiki (END - exit=256) [18:14:31] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [18:15:03] !log [salt-user@mw132] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=sagan4alphawiki (END - exit=256) [18:15:07] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [18:15:14] !log [salt-user@mw122] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=sagan4alphawiki (END - exit=0) [18:15:17] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [18:15:23] !log [salt-user@mwtask141] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=sagan4betawiki (END - exit=65280) [18:15:27] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [18:15:39] !log [salt-user@mw142] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=sagan4betawiki (END - exit=0) [18:15:41] !log [salt-user@mw132] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=sagan4betawiki (END - exit=0) [18:15:43] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [18:15:43] !log [salt-user@mw141] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=sagan4betawiki (END - exit=0) [18:15:44] !log [salt-user@mw131] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=sagan4betawiki (END - exit=0) [18:15:45] !log [salt-user@mw122] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=sagan4betawiki (END - exit=0) [18:15:46] !log [salt-user@mw121] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=sagan4betawiki (END - exit=0) [18:15:46] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [18:15:50] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [18:15:55] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [18:15:58] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [18:16:03] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [18:16:26] PROBLEM - db112 Current Load on db112 is WARNING: WARNING - load average: 3.77, 7.52, 5.70 [18:17:54] !log [salt-user@mwtask141] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=twistedwonderlandwiki (END - exit=65280) [18:19:08] PROBLEM - wiki.nj.cn.eu.org - reverse DNS on sslhost is CRITICAL: rDNS CRITICAL - wiki.nj.cn.eu.org All nameservers failed to answer the query. [18:19:21] PROBLEM - mw141 MediaWiki Rendering on mw141 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [18:20:06] PROBLEM - mw121 MediaWiki Rendering on mw121 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [18:20:09] PROBLEM - mw122 MediaWiki Rendering on mw122 is CRITICAL: HTTP CRITICAL: HTTP/1.1 502 Bad Gateway - 5903 bytes in 0.013 second response time [18:20:26] RECOVERY - db112 Current Load on db112 is OK: OK - load average: 3.57, 5.74, 5.41 [18:20:47] !log [salt-user@mw131] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=twistedwonderlandwiki (END - exit=0) [18:20:54] PROBLEM - mw132 MediaWiki Rendering on mw132 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [18:20:58] PROBLEM - mw131 MediaWiki Rendering on mw131 is CRITICAL: HTTP CRITICAL: HTTP/1.1 502 Bad Gateway - 5903 bytes in 0.022 second response time [18:20:59] PROBLEM - mw142 MediaWiki Rendering on mw142 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [18:21:08] !log [salt-user@mw121] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=twistedwonderlandwiki (END - exit=0) [18:21:09] !log [salt-user@mw122] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=twistedwonderlandwiki (END - exit=0) [18:21:10] !log [salt-user@mw142] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=twistedwonderlandwiki (END - exit=0) [18:21:11] !log [salt-user@mw141] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=twistedwonderlandwiki (END - exit=0) [18:21:12] !log [salt-user@mw132] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=twistedwonderlandwiki (END - exit=0) [18:21:48] !log [salt-user@mw121] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=projectsekaiwiki (END - exit=0) [18:21:50] PROBLEM - en.religiononfire.mar.in.ua - reverse DNS on sslhost is CRITICAL: rDNS CRITICAL - en.religiononfire.mar.in.ua All nameservers failed to answer the query. [18:21:52] !log [salt-user@mw141] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=projectsekaiwiki (END - exit=0) [18:22:03] !log [salt-user@mw132] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=projectsekaiwiki (END - exit=0) [18:22:15] !log [salt-user@mw122] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=projectsekaiwiki (END - exit=0) [18:22:23] !log [salt-user@mw131] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=projectsekaiwiki (END - exit=0) [18:22:25] PROBLEM - cp23 Varnish Backends on cp23 is CRITICAL: 5 backends are down. mw121 mw131 mw132 mw141 mw142 [18:22:31] !log [salt-user@mwtask141] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=projectsekaiwiki (END - exit=0) [18:22:45] PROBLEM - cp22 Varnish Backends on cp22 is CRITICAL: 7 backends are down. mw121 mw122 mw131 mw132 mw141 mw142 mediawiki [18:22:54] RECOVERY - mw132 MediaWiki Rendering on mw132 is OK: HTTP OK: HTTP/1.1 200 OK - 8191 bytes in 0.576 second response time [18:22:56] RECOVERY - mw131 MediaWiki Rendering on mw131 is OK: HTTP OK: HTTP/1.1 200 OK - 8191 bytes in 0.508 second response time [18:22:57] RECOVERY - mw142 MediaWiki Rendering on mw142 is OK: HTTP OK: HTTP/1.1 200 OK - 8191 bytes in 1.979 second response time [18:23:09] PROBLEM - cp33 Varnish Backends on cp33 is CRITICAL: 7 backends are down. mw121 mw122 mw131 mw132 mw141 mw142 mediawiki [18:23:11] RECOVERY - mw141 MediaWiki Rendering on mw141 is OK: HTTP OK: HTTP/1.1 200 OK - 8191 bytes in 0.760 second response time [18:23:14] !log [salt-user@mw142] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=projectsekaiwiki (END - exit=0) [18:23:43] PROBLEM - cp32 Varnish Backends on cp32 is CRITICAL: 3 backends are down. mw122 mw131 mw132 [18:24:53] !log [salt-user@mw122] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=projectsekaiwiki (END - exit=0) [18:24:56] !log [salt-user@mwtask141] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=projectsekaiwiki (END - exit=0) [18:24:57] !log [salt-user@mw142] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=projectsekaiwiki (END - exit=0) [18:24:58] !log [salt-user@mw121] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=projectsekaiwiki (END - exit=0) [18:24:59] !log [salt-user@mw132] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=projectsekaiwiki (END - exit=0) [18:25:00] !log [salt-user@mw131] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=projectsekaiwiki (END - exit=0) [18:25:01] !log [salt-user@mw141] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=projectsekaiwiki (END - exit=0) [18:26:13] RECOVERY - mw121 MediaWiki Rendering on mw121 is OK: HTTP OK: HTTP/1.1 200 OK - 8191 bytes in 0.309 second response time [18:26:16] RECOVERY - mw122 MediaWiki Rendering on mw122 is OK: HTTP OK: HTTP/1.1 200 OK - 8191 bytes in 0.349 second response time [18:26:18] RECOVERY - cp22 NTP time on cp22 is OK: NTP OK: Offset -0.02493900061 secs [18:27:43] RECOVERY - cp32 Varnish Backends on cp32 is OK: All 14 backends are healthy [18:28:25] RECOVERY - cp23 Varnish Backends on cp23 is OK: All 14 backends are healthy [18:28:45] RECOVERY - cp22 Varnish Backends on cp22 is OK: All 14 backends are healthy [18:29:09] RECOVERY - cp33 Varnish Backends on cp33 is OK: All 14 backends are healthy [18:29:20] hmmm, probably SMW shenanigans? idk what happened here [18:29:44] !log [salt-user@mw142] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=sagan4alphawiki (END - exit=256) [18:29:45] !log [salt-user@mwtask141] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=sagan4alphawiki (END - exit=256) [18:29:46] !log [salt-user@mw141] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=sagan4alphawiki (END - exit=256) [18:29:47] !log [salt-user@mw131] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=sagan4alphawiki (END - exit=256) [18:29:48] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [18:29:48] !log [salt-user@mw121] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=sagan4alphawiki (END - exit=256) [18:29:49] !log [salt-user@mw122] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=sagan4alphawiki (END - exit=256) [18:29:52] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [18:29:56] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [18:29:59] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [18:30:05] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [18:30:09] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [18:32:24] !log [salt-user@mw132] sudo -u www-data php /srv/mediawiki/w/extensions/SemanticMediaWiki/maintenance/setupStore.php --wiki=sagan4alphawiki (END - exit=0) [18:32:28] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [18:33:10] PROBLEM - db112 Current Load on db112 is CRITICAL: CRITICAL - load average: 7.10, 8.72, 6.92 [18:36:48] PROBLEM - mw131 MediaWiki Rendering on mw131 is CRITICAL: HTTP CRITICAL: HTTP/1.1 502 Bad Gateway - 5903 bytes in 0.015 second response time [18:36:59] PROBLEM - mw142 MediaWiki Rendering on mw142 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [18:37:01] PROBLEM - mw141 MediaWiki Rendering on mw141 is CRITICAL: HTTP CRITICAL: HTTP/1.1 502 Bad Gateway - 5903 bytes in 0.006 second response time [18:37:08] I'm just going to say, everything was working until the SMW install was started, I don't know about you guys [18:37:10] PROBLEM - mw132 MediaWiki Rendering on mw132 is CRITICAL: HTTP CRITICAL: HTTP/1.1 502 Bad Gateway - 5903 bytes in 0.008 second response time [18:38:02] PROBLEM - mw121 MediaWiki Rendering on mw121 is CRITICAL: HTTP CRITICAL: HTTP/1.1 502 Bad Gateway - 5903 bytes in 0.012 second response time [18:38:12] PROBLEM - mw122 MediaWiki Rendering on mw122 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [18:38:25] PROBLEM - cp23 Varnish Backends on cp23 is CRITICAL: 5 backends are down. mw121 mw122 mw132 mw141 mw142 [18:38:39] Orange_Star: that is true... [18:38:45] PROBLEM - cp22 Varnish Backends on cp22 is CRITICAL: 7 backends are down. mw121 mw122 mw131 mw132 mw141 mw142 mediawiki [18:39:06] maybe we should kill SMW here /s but also kinda not sarcasm [18:39:09] PROBLEM - cp33 Varnish Backends on cp33 is CRITICAL: 7 backends are down. mw121 mw122 mw131 mw132 mw141 mw142 mediawiki [18:39:10] RECOVERY - mw132 MediaWiki Rendering on mw132 is OK: HTTP OK: HTTP/1.1 200 OK - 8191 bytes in 0.440 second response time [18:39:43] PROBLEM - cp32 Varnish Backends on cp32 is CRITICAL: 7 backends are down. mw121 mw122 mw131 mw132 mw141 mw142 mediawiki [18:39:58] RECOVERY - mw121 MediaWiki Rendering on mw121 is OK: HTTP OK: HTTP/1.1 200 OK - 8191 bytes in 0.395 second response time [18:40:08] RECOVERY - mw122 MediaWiki Rendering on mw122 is OK: HTTP OK: HTTP/1.1 200 OK - 8191 bytes in 0.362 second response time [18:40:43] RECOVERY - mw131 MediaWiki Rendering on mw131 is OK: HTTP OK: HTTP/1.1 200 OK - 8191 bytes in 0.391 second response time [18:41:01] RECOVERY - mw141 MediaWiki Rendering on mw141 is OK: HTTP OK: HTTP/1.1 200 OK - 8191 bytes in 0.435 second response time [18:41:02] RECOVERY - mw142 MediaWiki Rendering on mw142 is OK: HTTP OK: HTTP/1.1 200 OK - 8191 bytes in 0.261 second response time [18:43:09] RECOVERY - cp33 Varnish Backends on cp33 is OK: All 14 backends are healthy [18:43:43] RECOVERY - cp32 Varnish Backends on cp32 is OK: All 14 backends are healthy [18:44:25] RECOVERY - cp23 Varnish Backends on cp23 is OK: All 14 backends are healthy [18:44:45] RECOVERY - cp22 Varnish Backends on cp22 is OK: All 14 backends are healthy [18:50:40] RECOVERY - en.religiononfire.mar.in.ua - reverse DNS on sslhost is OK: SSL OK - en.religiononfire.mar.in.ua reverse DNS resolves to cp23.miraheze.org - CNAME OK [18:53:56] [02WikiDiscover] 07Reception123 closed pull request 03#92: Add a config whether or not to list private wikis - 13https://github.com/miraheze/WikiDiscover/pull/92 [18:53:57] [02miraheze/WikiDiscover] 07Reception123 pushed 033 commits to 03master [+0/-0/±4] 13https://github.com/miraheze/WikiDiscover/compare/acf37a220436...8c58bcda5781 [18:53:58] [02miraheze/WikiDiscover] 07Universal-Omega 03512849c - Add a config whether or not to list private wikis [18:53:59] [02miraheze/WikiDiscover] 07Universal-Omega 03b57f051 - Add config [18:54:02] [02miraheze/WikiDiscover] 07Reception123 038c58bcd - Merge pull request #92 from Universal-Omega/patch-3 [18:54:34] PROBLEM - db112 Disk Space on db112 is WARNING: DISK WARNING - free space: / 9192 MB (6% inode=99%); [18:56:26] PROBLEM - db112 Current Load on db112 is WARNING: WARNING - load average: 1.20, 5.68, 7.94 [18:58:10] miraheze/WikiDiscover - Reception123 the build passed. [19:00:26] RECOVERY - db112 Current Load on db112 is OK: OK - load average: 1.77, 3.76, 6.64 [19:17:13] RECOVERY - wiki.nj.cn.eu.org - reverse DNS on sslhost is OK: SSL OK - wiki.nj.cn.eu.org reverse DNS resolves to cp22.miraheze.org - CNAME OK [19:28:50] [02miraheze/mw-config] 07Reception123 pushed 031 commit to 03master [+0/-0/±1] 13https://github.com/miraheze/mw-config/compare/891b2a0d928a...c54c8d11a412 [19:28:52] [02miraheze/mw-config] 07Reception123 03c54c8d1 - mention uploads in sitenotice [19:29:08] !log [reception@mwtask141] starting deploy of {'pull': 'config', 'config': True} to all [19:29:13] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [19:29:16] !log [reception@mwtask141] finished deploy of {'pull': 'config', 'config': True} to all - SUCCESS in 8s [19:29:21] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [19:29:58] miraheze/mw-config - Reception123 the build passed. [19:56:57] !log [@test131] starting deploy of {'config': True} to all [19:56:58] !log [@test131] finished deploy of {'config': True} to all - SUCCESS in 0s [19:57:01] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [19:57:05] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [21:03:57] PROBLEM - cp23 NTP time on cp23 is WARNING: NTP WARNING: Offset 0.1066554487 secs [21:05:16] PROBLEM - gs.sidem.wiki - reverse DNS on sslhost is CRITICAL: rDNS CRITICAL - gs.sidem.wiki All nameservers failed to answer the query. [21:10:11] PROBLEM - www.sidem.wiki - reverse DNS on sslhost is WARNING: Timeout: The DNS operation timed out after 5.403818368911743 seconds [21:17:57] RECOVERY - cp23 NTP time on cp23 is OK: NTP OK: Offset 0.09980547428 secs [21:22:40] PROBLEM - en.religiononfire.mar.in.ua - reverse DNS on sslhost is WARNING: NoNameservers: All nameservers failed to answer the query en.religiononfire.mar.in.ua. IN CNAME: Server 2606:4700:4700::1111 UDP port 53 answered SERVFAIL [21:28:19] [02mw-config] 07Naleksuh opened pull request 03#5193: disable wiki creations - 13https://github.com/miraheze/mw-config/pull/5193 [21:29:21] miraheze/mw-config - Naleksuh the build passed. [21:33:56] RECOVERY - gs.sidem.wiki - reverse DNS on sslhost is OK: SSL OK - gs.sidem.wiki reverse DNS resolves to cp22.miraheze.org - CNAME OK [21:39:10] RECOVERY - www.sidem.wiki - reverse DNS on sslhost is OK: SSL OK - www.sidem.wiki reverse DNS resolves to cp22.miraheze.org - CNAME OK [21:51:30] PROBLEM - en.religiononfire.mar.in.ua - reverse DNS on sslhost is CRITICAL: rDNS CRITICAL - en.religiononfire.mar.in.ua All nameservers failed to answer the query. [22:00:15] PROBLEM - cp32 NTP time on cp32 is WARNING: NTP WARNING: Offset 0.1303185821 secs [22:07:11] PROBLEM - wiki.andreijiroh.uk.eu.org - reverse DNS on sslhost is CRITICAL: rDNS CRITICAL - wiki.andreijiroh.uk.eu.org All nameservers failed to answer the query. [22:08:15] RECOVERY - cp32 NTP time on cp32 is OK: NTP OK: Offset 0.0877994597 secs [22:11:25] PROBLEM - cp23 Varnish Backends on cp23 is CRITICAL: 7 backends are down. mw121 mw122 mw131 mw132 mw141 mw142 mediawiki [22:15:15] RECOVERY - cp23 Varnish Backends on cp23 is OK: All 14 backends are healthy [22:15:46] [02mw-config] 07MacFan4000 commented on pull request 03#5193: disable wiki creations - 13https://github.com/miraheze/mw-config/pull/5193#issuecomment-1506036237 [22:15:49] [02mw-config] 07MacFan4000 closed pull request 03#5193: disable wiki creations - 13https://github.com/miraheze/mw-config/pull/5193 [22:20:20] PROBLEM - en.religiononfire.mar.in.ua - reverse DNS on sslhost is WARNING: NoNameservers: All nameservers failed to answer the query en.religiononfire.mar.in.ua. IN CNAME: Server 2606:4700:4700::1111 UDP port 53 answered SERVFAIL [22:22:57] [02mw-config] 07Naleksuh commented on pull request 03#5193: disable wiki creations - 13https://github.com/miraheze/mw-config/pull/5193#issuecomment-1506043267 [22:23:42] [02mw-config] 07Naleksuh deleted a comment on pull request 03#5193: disable wiki creations - 13https://github.com/miraheze/mw-config/pull/5193#issuecomment-1506043267 [22:24:16] [02mw-config] 07Naleksuh commented on pull request 03#5193: disable wiki creations - 13https://github.com/miraheze/mw-config/pull/5193#issuecomment-1506044189 [22:26:34] PROBLEM - db112 Disk Space on db112 is CRITICAL: DISK CRITICAL - free space: / 7967 MB (5% inode=99%); [22:26:53] PROBLEM - cp23 Varnish Backends on cp23 is CRITICAL: 1 backends are down. mw142 [22:28:48] RECOVERY - cp23 Varnish Backends on cp23 is OK: All 14 backends are healthy [22:36:24] RECOVERY - wiki.andreijiroh.uk.eu.org - reverse DNS on sslhost is OK: SSL OK - wiki.andreijiroh.uk.eu.org reverse DNS resolves to cp22.miraheze.org - CNAME OK [22:49:51] RECOVERY - en.religiononfire.mar.in.ua - reverse DNS on sslhost is OK: SSL OK - en.religiononfire.mar.in.ua reverse DNS resolves to cp23.miraheze.org - CNAME OK [23:35:40] PROBLEM - uk.religiononfire.mar.in.ua - reverse DNS on sslhost is CRITICAL: rDNS CRITICAL - uk.religiononfire.mar.in.ua All nameservers failed to answer the query.