[00:00:33] RECOVERY - en.religiononfire.mar.in.ua - LetsEncrypt on sslhost is OK: OK - Certificate 'en.religiononfire.mar.in.ua' will expire on Thu 01 Jun 2023 18:42:11 GMT +0000. [00:01:53] PROBLEM - cp23 Current Load on cp23 is CRITICAL: CRITICAL - load average: 12.90, 7.17, 3.31 [00:02:14] PROBLEM - cp23 APT on cp23 is CRITICAL: CHECK_NRPE STATE CRITICAL: Socket timeout after 60 seconds. [00:02:44] PROBLEM - cp32 Current Load on cp32 is CRITICAL: CRITICAL - load average: 7.41, 3.98, 2.96 [00:08:31] PROBLEM - cp32 Current Load on cp32 is WARNING: WARNING - load average: 2.71, 3.82, 3.33 [00:10:20] RECOVERY - cp23 APT on cp23 is OK: APT OK: 2 packages available for upgrade (0 critical updates). [00:10:27] PROBLEM - cp32 Current Load on cp32 is CRITICAL: CRITICAL - load average: 4.07, 4.15, 3.52 [00:13:26] PROBLEM - db142 Current Load on db142 is CRITICAL: CRITICAL - load average: 10.60, 6.48, 4.09 [00:13:46] PROBLEM - cp23 Current Load on cp23 is WARNING: WARNING - load average: 0.45, 3.26, 3.77 [00:13:48] PROBLEM - atnarsia.com - LetsEncrypt on sslhost is CRITICAL: connect to address atnarsia.com and port 443: Network is unreachableHTTP CRITICAL - Unable to open TCP socket [00:14:26] PROBLEM - cp32 Current Load on cp32 is WARNING: WARNING - load average: 3.49, 3.77, 3.51 [00:15:46] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 0.73, 2.43, 3.40 [00:16:26] PROBLEM - cp32 Current Load on cp32 is CRITICAL: CRITICAL - load average: 4.02, 4.01, 3.64 [00:17:25] PROBLEM - db142 Current Load on db142 is WARNING: WARNING - load average: 5.82, 7.11, 4.98 [00:18:26] PROBLEM - cp32 Current Load on cp32 is WARNING: WARNING - load average: 2.21, 3.59, 3.55 [00:19:24] RECOVERY - db142 Current Load on db142 is OK: OK - load average: 5.24, 6.58, 5.05 [00:20:26] PROBLEM - cp32 Current Load on cp32 is CRITICAL: CRITICAL - load average: 6.33, 4.40, 3.82 [00:22:26] PROBLEM - cp32 Current Load on cp32 is WARNING: WARNING - load average: 2.74, 3.89, 3.72 [00:26:26] PROBLEM - cp32 Current Load on cp32 is CRITICAL: CRITICAL - load average: 7.99, 4.80, 4.02 [00:30:18] PROBLEM - es141 Current Load on es141 is CRITICAL: CRITICAL - load average: 4.97, 3.71, 2.69 [00:31:19] PROBLEM - db142 Current Load on db142 is CRITICAL: CRITICAL - load average: 11.67, 8.30, 6.06 [00:32:26] PROBLEM - cp32 Current Load on cp32 is WARNING: WARNING - load average: 2.76, 3.96, 3.96 [00:34:26] PROBLEM - cp32 Current Load on cp32 is CRITICAL: CRITICAL - load average: 7.75, 5.31, 4.43 [00:42:19] PROBLEM - es141 Current Load on es141 is WARNING: WARNING - load average: 2.95, 3.73, 3.51 [00:42:26] PROBLEM - cp32 Current Load on cp32 is WARNING: WARNING - load average: 2.07, 3.42, 3.93 [00:44:19] RECOVERY - es141 Current Load on es141 is OK: OK - load average: 2.37, 3.31, 3.38 [00:46:26] RECOVERY - cp32 Current Load on cp32 is OK: OK - load average: 1.63, 2.27, 3.33 [00:46:57] PROBLEM - uk.religiononfire.mar.in.ua - reverse DNS on sslhost is WARNING: NoNameservers: All nameservers failed to answer the query uk.religiononfire.mar.in.ua. IN CNAME: Server 2606:4700:4700::1111 UDP port 53 answered SERVFAIL [00:47:13] PROBLEM - db142 Current Load on db142 is WARNING: WARNING - load average: 4.95, 7.04, 7.73 [00:53:28] !log [void@puppet141] updated grants for cargouser on all db servers [00:53:33] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [00:59:08] RECOVERY - db142 Current Load on db142 is OK: OK - load average: 3.45, 4.98, 6.45 [01:03:06] PROBLEM - db142 Current Load on db142 is CRITICAL: CRITICAL - load average: 8.49, 7.42, 7.10 [01:05:05] RECOVERY - db142 Current Load on db142 is OK: OK - load average: 3.12, 5.64, 6.48 [01:07:58] PROBLEM - cp32 Current Load on cp32 is CRITICAL: CRITICAL - load average: 2.68, 4.25, 3.62 [01:11:02] PROBLEM - db142 Current Load on db142 is CRITICAL: CRITICAL - load average: 11.04, 8.27, 7.23 [01:14:14] PROBLEM - cp33 Current Load on cp33 is WARNING: WARNING - load average: 3.85, 3.08, 2.53 [01:15:01] PROBLEM - db142 Current Load on db142 is WARNING: WARNING - load average: 4.43, 7.30, 7.16 [01:16:14] RECOVERY - cp33 Current Load on cp33 is OK: OK - load average: 1.95, 2.61, 2.42 [01:17:00] RECOVERY - db142 Current Load on db142 is OK: OK - load average: 0.82, 4.97, 6.32 [01:17:36] PROBLEM - cp32 Current Load on cp32 is WARNING: WARNING - load average: 2.62, 3.67, 3.87 [01:21:27] PROBLEM - cp32 Current Load on cp32 is CRITICAL: CRITICAL - load average: 4.63, 4.14, 3.99 [01:23:22] PROBLEM - cp32 Current Load on cp32 is WARNING: WARNING - load average: 2.86, 3.75, 3.87 [01:29:10] RECOVERY - cp32 Current Load on cp32 is OK: OK - load average: 2.99, 2.83, 3.36 [01:30:42] [02miraheze/ManageWiki] 07The-Voidwalker pushed 031 commit to 03patch-jobqueue [+0/-0/±2] 13https://github.com/miraheze/ManageWiki/commit/919b8f0dd6f2 [01:30:44] [02miraheze/ManageWiki] 07The-Voidwalker 03919b8f0 - use MediaWikiServices for jobqueue [01:30:46] [02ManageWiki] 07The-Voidwalker created branch 03patch-jobqueue - 13https://github.com/miraheze/ManageWiki [01:30:51] [02ManageWiki] 07The-Voidwalker opened pull request 03#391: use MediaWikiServices for jobqueue - 13https://github.com/miraheze/ManageWiki/pull/391 [01:32:07] [02miraheze/ManageWiki] 07The-Voidwalker pushed 031 commit to 03patch-jobqueue [+0/-0/±1] 13https://github.com/miraheze/ManageWiki/compare/919b8f0dd6f2...beab5cfa2662 [01:32:10] [02miraheze/ManageWiki] 07The-Voidwalker 03beab5cf - bump mediawiki version requirement [01:32:12] [02ManageWiki] 07The-Voidwalker synchronize pull request 03#391: use MediaWikiServices for jobqueue - 13https://github.com/miraheze/ManageWiki/pull/391 [01:36:12] miraheze/ManageWiki - The-Voidwalker the build passed. [01:36:25] [02ManageWiki] 07The-Voidwalker closed pull request 03#391: use MediaWikiServices for jobqueue - 13https://github.com/miraheze/ManageWiki/pull/391 [01:36:28] [02miraheze/ManageWiki] 07The-Voidwalker pushed 031 commit to 03master [+0/-0/±3] 13https://github.com/miraheze/ManageWiki/compare/7e2f14bfb7ed...724602d892a5 [01:36:29] [02miraheze/ManageWiki] 07The-Voidwalker 03724602d - use MediaWikiServices for jobqueue (#391) [01:39:47] miraheze/ManageWiki - The-Voidwalker the build passed. [01:44:53] [02miraheze/mediawiki] 07The-Voidwalker pushed 031 commit to 03REL1_40 [+0/-0/±1] 13https://github.com/miraheze/mediawiki/compare/81ffc3983f27...c60e98a38845 [01:44:56] [02miraheze/mediawiki] 07The-Voidwalker 03c60e98a - update ManageWiki [01:46:06] PROBLEM - uk.religiononfire.mar.in.ua - reverse DNS on sslhost is CRITICAL: rDNS CRITICAL - uk.religiononfire.mar.in.ua All nameservers failed to answer the query. [01:50:00] !log [void@test131] starting deploy of {'world': True} to all [01:50:06] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [01:51:50] PROBLEM - en.religiononfire.mar.in.ua - reverse DNS on sslhost is CRITICAL: rDNS CRITICAL - en.religiononfire.mar.in.ua All nameservers failed to answer the query. [01:53:59] !log [void@test131] DEPLOY ABORTED: Non-Zero Exit Code in prep, see output. [01:54:03] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [01:56:33] [02mediawiki] 07The-Voidwalker closed pull request 03#14460: Remove GitInfo hack - 13https://github.com/miraheze/mediawiki/pull/14460 [01:56:34] [02miraheze/mediawiki] 07The-Voidwalker pushed 031 commit to 03REL1_40 [+0/-0/±1] 13https://github.com/miraheze/mediawiki/compare/c60e98a38845...ab6c8f5ae91d [01:56:35] [02miraheze/mediawiki] 07Universal-Omega 03ab6c8f5 - Remove GitInfo hack (#14460) [01:57:02] !log [void@test131] starting deploy of {'world': True} to all [01:57:05] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [02:00:27] !log [void@test131] DEPLOY ABORTED: Non-Zero Exit Code in prep, see output. [02:00:31] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [02:01:03] PROBLEM - cp32 Current Load on cp32 is CRITICAL: CRITICAL - load average: 4.27, 4.43, 3.52 [02:02:59] PROBLEM - cp32 Current Load on cp32 is WARNING: WARNING - load average: 2.77, 3.77, 3.38 [02:04:52] [02miraheze/mediawiki] 07The-Voidwalker pushed 032 commits to 03REL1_40 [+0/-0/±2] 13https://github.com/miraheze/mediawiki/compare/ab6c8f5ae91d...0d54af068258 [02:04:54] [02miraheze/mediawiki] 07The-Voidwalker 03f8e9f54 - update MirahezeMagic [02:04:55] RECOVERY - cp32 Current Load on cp32 is OK: OK - load average: 2.00, 3.19, 3.21 [02:04:57] [02miraheze/mediawiki] 07The-Voidwalker 030d54af0 - Merge branch 'REL1_40' of https://github.com/miraheze/mediawiki into REL1_40 [02:05:19] !log [void@test131] starting deploy of {'world': True} to all [02:05:23] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [02:07:08] [02mediawiki] 07MacFan4000 commented on pull request 03#13790: Bump extensions/ManageWiki from `e16d36e` to `9d7f0fa` - 13https://github.com/miraheze/mediawiki/pull/13790#issuecomment-1502586093 [02:07:11] [02mediawiki] 07dependabot[bot] edited pull request 03#13790: Bump extensions/ManageWiki from `e16d36e` to `9d7f0fa` - 13https://github.com/miraheze/mediawiki/pull/13790 [02:07:30] [02mediawiki] 07MacFan4000 commented on pull request 03#14274: bump extensions/ManageWiki from `c7a0f84` to `7e2f14b` - 13https://github.com/miraheze/mediawiki/pull/14274#issuecomment-1502586300 [02:07:33] [02mediawiki] 07dependabot[bot] edited pull request 03#14274: bump extensions/ManageWiki from `c7a0f84` to `7e2f14b` - 13https://github.com/miraheze/mediawiki/pull/14274 [02:08:01] [02mediawiki] 07dependabot[bot] edited pull request 03#13790: Bump extensions/ManageWiki from `e16d36e` to `9d7f0fa` - 13https://github.com/miraheze/mediawiki/pull/13790 [02:08:16] [02mediawiki] 07dependabot[bot] opened pull request 03#14461: Bump extensions/ManageWiki from `c7a0f84` to `724602d` - 13https://github.com/miraheze/mediawiki/pull/14461 [02:08:18] [02miraheze/mediawiki] 07dependabot[bot] pushed 031 commit to 03dependabot/submodules/REL1_39/extensions/ManageWiki-724602d [+0/-0/±1] 13https://github.com/miraheze/mediawiki/commit/7ee1968ad2a8 [02:08:20] [02miraheze/mediawiki] 07dependabot[bot] 037ee1968 - Bump extensions/ManageWiki from `c7a0f84` to `724602d` [02:08:21] [02mediawiki] 07dependabot[bot] created branch 03dependabot/submodules/REL1_39/extensions/ManageWiki-724602d - 13https://github.com/miraheze/mediawiki [02:08:23] [02mediawiki] 07dependabot[bot] commented on pull request 03#14274: bump extensions/ManageWiki from `c7a0f84` to `7e2f14b` - 13https://github.com/miraheze/mediawiki/pull/14274#issuecomment-1502586776 [02:08:26] [02mediawiki] 07dependabot[bot] edited pull request 03#14274: bump extensions/ManageWiki from `c7a0f84` to `7e2f14b` - 13https://github.com/miraheze/mediawiki/pull/14274 [02:08:29] [02mediawiki] 07dependabot[bot] closed pull request 03#14274: bump extensions/ManageWiki from `c7a0f84` to `7e2f14b` - 13https://github.com/miraheze/mediawiki/pull/14274 [02:08:29] !log [void@test131] DEPLOY ABORTED: Non-Zero Exit Code in prep, see output. [02:08:31] [02mediawiki] 07dependabot[bot] deleted branch 03dependabot/submodules/REL1_40/extensions/ManageWiki-7e2f14b - 13https://github.com/miraheze/mediawiki [02:08:33] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [02:08:34] [02miraheze/mediawiki] 07dependabot[bot] deleted branch 03dependabot/submodules/REL1_40/extensions/ManageWiki-7e2f14b [02:08:37] [02mediawiki] 07github-actions[bot] labeled pull request 03#14461: Bump extensions/ManageWiki from `c7a0f84` to `724602d` - 13https://github.com/miraheze/mediawiki/pull/14461 [02:08:40] [02mediawiki] 07github-actions[bot] labeled pull request 03#14461: Bump extensions/ManageWiki from `c7a0f84` to `724602d` - 13https://github.com/miraheze/mediawiki/pull/14461 [02:08:42] [02mediawiki] 07github-actions[bot] labeled pull request 03#14461: Bump extensions/ManageWiki from `c7a0f84` to `724602d` - 13https://github.com/miraheze/mediawiki/pull/14461 [02:09:15] [02mediawiki] 07MacFan4000 closed pull request 03#13790: Bump extensions/ManageWiki from `e16d36e` to `9d7f0fa` - 13https://github.com/miraheze/mediawiki/pull/13790 [02:09:17] [02mediawiki] 07dependabot[bot] commented on pull request 03#13790: Bump extensions/ManageWiki from `e16d36e` to `9d7f0fa` - 13https://github.com/miraheze/mediawiki/pull/13790#issuecomment-1502587372 [02:09:23] [02mediawiki] 07dependabot[bot] deleted branch 03dependabot/submodules/REL1_39/extensions/ManageWiki-9d7f0fa - 13https://github.com/miraheze/mediawiki [02:09:24] [02miraheze/mediawiki] 07dependabot[bot] deleted branch 03dependabot/submodules/REL1_39/extensions/ManageWiki-9d7f0fa [02:09:47] [02mediawiki] 07MacFan4000 commented on pull request 03#14067: Bump extensions/ManageWiki from `c7a0f84` to `7e2f14b` - 13https://github.com/miraheze/mediawiki/pull/14067#issuecomment-1502587720 [02:09:50] [02mediawiki] 07dependabot[bot] edited pull request 03#14067: Bump extensions/ManageWiki from `c7a0f84` to `7e2f14b` - 13https://github.com/miraheze/mediawiki/pull/14067 [02:10:35] [02mediawiki] 07dependabot[bot] edited pull request 03#14067: Bump extensions/ManageWiki from `c7a0f84` to `7e2f14b` - 13https://github.com/miraheze/mediawiki/pull/14067 [02:10:38] [02miraheze/mediawiki] 07dependabot[bot] pushed 031 commit to 03dependabot/submodules/REL1_39/extensions/ManageWiki-7e2f14b [+0/-0/±1] 13https://github.com/miraheze/mediawiki/compare/2bed8ca1cece...0a2fdc22683c [02:10:40] [02miraheze/mediawiki] 07dependabot[bot] 030a2fdc2 - Bump extensions/ManageWiki from `c7a0f84` to `7e2f14b` [02:10:43] [02mediawiki] 07dependabot[bot] synchronize pull request 03#14067: Bump extensions/ManageWiki from `c7a0f84` to `7e2f14b` - 13https://github.com/miraheze/mediawiki/pull/14067 [02:11:00] [02mediawiki] 07github-actions[bot] labeled pull request 03#14067: Bump extensions/ManageWiki from `c7a0f84` to `7e2f14b` - 13https://github.com/miraheze/mediawiki/pull/14067 [02:15:41] RECOVERY - uk.religiononfire.mar.in.ua - reverse DNS on sslhost is OK: SSL OK - uk.religiononfire.mar.in.ua reverse DNS resolves to cp23.miraheze.org - CNAME OK [02:20:40] RECOVERY - en.religiononfire.mar.in.ua - reverse DNS on sslhost is OK: SSL OK - en.religiononfire.mar.in.ua reverse DNS resolves to cp23.miraheze.org - CNAME OK [02:27:57] PROBLEM - cp23 Current Load on cp23 is CRITICAL: CRITICAL - load average: 5.81, 3.29, 1.66 [02:29:55] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 2.63, 3.29, 1.89 [02:42:27] PROBLEM - db142 Current Load on db142 is WARNING: WARNING - load average: 7.69, 6.98, 4.16 [02:44:26] RECOVERY - db142 Current Load on db142 is OK: OK - load average: 1.89, 5.07, 3.80 [02:51:01] PROBLEM - test131 Puppet on test131 is WARNING: WARNING: Puppet is currently disabled, message: Testing Cargo --Void, last run 23 minutes ago with 0 failures [02:54:25] !log [void@test131] starting deploy of {'folders': 'config,w/extensions/MirahezeMagic'} to all [02:54:26] !log [void@test131] finished deploy of {'folders': 'config,w/extensions/MirahezeMagic'} to all - SUCCESS in 0s [02:54:29] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [02:54:33] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [03:03:28] PROBLEM - cp33 Current Load on cp33 is CRITICAL: CRITICAL - load average: 5.28, 4.31, 3.29 [03:05:24] PROBLEM - cp33 Current Load on cp33 is WARNING: WARNING - load average: 1.90, 3.47, 3.11 [03:07:20] RECOVERY - cp33 Current Load on cp33 is OK: OK - load average: 3.32, 3.40, 3.12 [03:33:58] PROBLEM - cp23 NTP time on cp23 is WARNING: NTP WARNING: Offset 0.1140026152 secs [03:35:02] PROBLEM - cp32 Current Load on cp32 is CRITICAL: CRITICAL - load average: 5.93, 4.65, 3.26 [03:36:57] PROBLEM - cp32 Current Load on cp32 is WARNING: WARNING - load average: 2.12, 3.68, 3.07 [03:40:48] RECOVERY - cp32 Current Load on cp32 is OK: OK - load average: 2.31, 3.34, 3.08 [03:41:58] RECOVERY - cp23 NTP time on cp23 is OK: NTP OK: Offset 0.07820379734 secs [04:08:18] PROBLEM - cp22 NTP time on cp22 is WARNING: NTP WARNING: Offset 0.1308036447 secs [04:18:21] [02miraheze/puppet] 07The-Voidwalker pushed 032 commits to 03patch-LimitCargoDB [+0/-0/±2] 13https://github.com/miraheze/puppet/compare/bfa1838f53bc^...9bc52e799cb9 [04:18:23] [02miraheze/puppet] 07The-Voidwalker 03bfa1838 - update grants for cargouser [04:18:26] [02miraheze/puppet] 07The-Voidwalker 039bc52e7 - update grants [04:18:29] [02puppet] 07The-Voidwalker created branch 03patch-LimitCargoDB - 13https://github.com/miraheze/puppet [04:18:30] [02puppet] 07The-Voidwalker opened pull request 03#3184: Patch-LimitCargoDB - 13https://github.com/miraheze/puppet/pull/3184 [04:19:07] [02puppet] 07The-Voidwalker closed pull request 03#3184: Patch-LimitCargoDB - 13https://github.com/miraheze/puppet/pull/3184 [04:19:17] [02miraheze/puppet] 07The-Voidwalker deleted branch 03patch-LimitCargoDB [04:19:19] [02puppet] 07The-Voidwalker deleted branch 03patch-LimitCargoDB - 13https://github.com/miraheze/puppet [04:20:17] RECOVERY - cp22 NTP time on cp22 is OK: NTP OK: Offset 0.08603909612 secs [04:20:30] [02miraheze/puppet] 07The-Voidwalker pushed 031 commit to 03master [+0/-0/±1] 13https://github.com/miraheze/puppet/compare/0be907ef7c84...241009f1d77c [04:20:31] [02miraheze/puppet] 07The-Voidwalker 03241009f - update cargo grants [04:23:52] [02miraheze/mw-config] 07The-Voidwalker pushed 031 commit to 03patch-addCargoScript [+0/-0/±1] 13https://github.com/miraheze/mw-config/compare/ceedfd60e1af...58ccff1caff4 [04:23:53] [02miraheze/mw-config] 07The-Voidwalker 0358ccff1 - set $wgCargoDBname [04:23:54] [02mw-config] 07The-Voidwalker synchronize pull request 03#5182: add new createCargoDB.php to cargo install - 13https://github.com/miraheze/mw-config/pull/5182 [04:24:56] miraheze/mw-config - The-Voidwalker the build passed. [04:27:01] RECOVERY - test131 Puppet on test131 is OK: OK: Puppet is currently enabled, last run 1 minute ago with 0 failures [04:50:34] PROBLEM - db112 Disk Space on db112 is WARNING: DISK WARNING - free space: / 14595 MB (10% inode=98%); [04:57:20] !log [@test131] starting deploy of {'config': True} to all [04:57:21] !log [@test131] finished deploy of {'config': True} to all - SUCCESS in 0s [04:57:25] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [04:57:30] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [05:05:37] [Grafana] !sre FIRING: The mediawiki job queue has more than 2500 unclaimed jobs https://grafana.miraheze.org/d/GtxbP1Xnk?orgId=1 [05:35:14] PROBLEM - gs.sidem.wiki - reverse DNS on sslhost is WARNING: NoNameservers: All nameservers failed to answer the query gs.sidem.wiki. IN CNAME: Server 2606:4700:4700::1111 UDP port 53 answered SERVFAIL [05:45:41] PROBLEM - uk.religiononfire.mar.in.ua - reverse DNS on sslhost is CRITICAL: rDNS CRITICAL - uk.religiononfire.mar.in.ua All nameservers failed to answer the query. [06:03:54] PROBLEM - gs.sidem.wiki - reverse DNS on sslhost is CRITICAL: rDNS CRITICAL - gs.sidem.wiki All nameservers failed to answer the query. [06:15:16] RECOVERY - uk.religiononfire.mar.in.ua - reverse DNS on sslhost is OK: SSL OK - uk.religiononfire.mar.in.ua reverse DNS resolves to cp22.miraheze.org - CNAME OK [06:33:09] RECOVERY - gs.sidem.wiki - reverse DNS on sslhost is OK: SSL OK - gs.sidem.wiki reverse DNS resolves to cp23.miraheze.org - CNAME OK [06:58:36] PROBLEM - mem141 NTP time on mem141 is WARNING: NTP WARNING: Offset 0.1079033315 secs [07:08:17] PROBLEM - db142 Current Load on db142 is CRITICAL: CRITICAL - load average: 10.08, 6.32, 3.90 [07:10:17] RECOVERY - db142 Current Load on db142 is OK: OK - load average: 3.24, 5.26, 3.83 [07:17:21] RECOVERY - mem141 NTP time on mem141 is OK: NTP OK: Offset 0.08994072676 secs [07:30:17] PROBLEM - cp22 NTP time on cp22 is WARNING: NTP WARNING: Offset 0.1012299657 secs [07:36:17] RECOVERY - cp22 NTP time on cp22 is OK: NTP OK: Offset 0.09999075532 secs [08:07:58] PROBLEM - cp23 NTP time on cp23 is WARNING: NTP WARNING: Offset 0.1217982173 secs [08:13:11] PROBLEM - mw121 Current Load on mw121 is WARNING: WARNING - load average: 11.61, 10.04, 8.48 [08:13:57] RECOVERY - cp23 NTP time on cp23 is OK: NTP OK: Offset 0.09811675549 secs [08:15:10] RECOVERY - mw121 Current Load on mw121 is OK: OK - load average: 7.93, 9.25, 8.38 [08:29:46] PROBLEM - cp23 Current Load on cp23 is WARNING: WARNING - load average: 3.66, 2.97, 1.85 [08:31:46] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 1.01, 2.20, 1.70 [08:56:39] PROBLEM - mw121 Current Load on mw121 is CRITICAL: CRITICAL - load average: 13.37, 12.15, 9.73 [08:57:31] PROBLEM - mw122 Current Load on mw122 is WARNING: WARNING - load average: 9.98, 10.78, 9.10 [08:58:38] PROBLEM - mw121 Current Load on mw121 is WARNING: WARNING - load average: 10.44, 10.93, 9.53 [08:59:31] RECOVERY - mw122 Current Load on mw122 is OK: OK - load average: 7.66, 9.78, 8.94 [09:00:37] RECOVERY - mw121 Current Load on mw121 is OK: OK - load average: 8.57, 9.99, 9.35 [09:10:35] [Grafana] !sre FIRING: The mediawiki job queue has more than 2500 unclaimed jobs https://grafana.miraheze.org/d/GtxbP1Xnk?orgId=1 [09:28:42] PROBLEM - cp23 Current Load on cp23 is CRITICAL: CRITICAL - load average: 20.05, 8.07, 3.92 [09:34:36] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 0.78, 3.39, 3.17 [09:47:57] PROBLEM - cp23 NTP time on cp23 is WARNING: NTP WARNING: Offset 0.1300979555 secs [09:57:58] RECOVERY - cp23 NTP time on cp23 is OK: NTP OK: Offset 0.07949835062 secs [09:59:11] PROBLEM - cp23 Current Load on cp23 is WARNING: WARNING - load average: 3.45, 2.45, 2.21 [10:01:10] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 2.24, 2.38, 2.22 [10:09:30] PROBLEM - cloud11 IPMI Sensors on cloud11 is UNKNOWN: Cannot access cache directory: /tmp/.freeipmi-root-> Execution of /usr/sbin/ipmi-sel failed with return code 1.-> /usr/sbin/ipmi-sel was executed with the following parameters: sudo /usr/sbin/ipmi-sel --output-event-state --interpret-oem-data --entity-sensor-names --sensor-types=all [10:11:57] PROBLEM - swiftac111 SSH on swiftac111 is CRITICAL: connect to address 2a10:6740::6:202 and port 22: No route to host [10:12:02] PROBLEM - swiftac111 Current Load on swiftac111 is CRITICAL: connect to address 2a10:6740::6:202 port 5666: No route to hostconnect to host 2a10:6740::6:202 port 5666: No route to host [10:12:10] PROBLEM - swiftac111 APT on swiftac111 is CRITICAL: connect to address 2a10:6740::6:202 port 5666: No route to hostconnect to host 2a10:6740::6:202 port 5666: No route to host [10:12:15] PROBLEM - swiftac111 Swift Container Service on swiftac111 is CRITICAL: connect to address 2a10:6740::6:202 and port 6001: No route to host [10:12:24] PROBLEM - swiftac111 conntrack_table_size on swiftac111 is CRITICAL: connect to address 2a10:6740::6:202 port 5666: No route to hostconnect to host 2a10:6740::6:202 port 5666: No route to host [10:12:25] PROBLEM - ping6 on swiftac111 is CRITICAL: CRITICAL - Destination Unreachable (2a10:6740::6:202) [10:12:29] PROBLEM - swiftac111 Puppet on swiftac111 is CRITICAL: connect to address 2a10:6740::6:202 port 5666: No route to hostconnect to host 2a10:6740::6:202 port 5666: No route to host [10:12:40] PROBLEM - swiftproxy111 HTTPS on swiftproxy111 is CRITICAL: HTTP CRITICAL - Invalid HTTP response received from host on port 443: HTTP/1.1 401 Unauthorized [10:12:49] PROBLEM - swiftac111 Swift Account Service on swiftac111 is CRITICAL: connect to address 2a10:6740::6:202 and port 6002: No route to host [10:12:57] PROBLEM - Host swiftac111 is DOWN: CRITICAL - Destination Unreachable (2a10:6740::6:202) [10:12:57] PROBLEM - swiftac111 ferm_active on swiftac111 is CRITICAL: connect to address 2a10:6740::6:202 port 5666: No route to hostconnect to host 2a10:6740::6:202 port 5666: No route to host [10:13:14] PROBLEM - swiftproxy111 HTTP on swiftproxy111 is CRITICAL: HTTP CRITICAL - Invalid HTTP response received from host: HTTP/1.1 401 Unauthorized [10:13:21] PROBLEM - swiftproxy131 HTTP on swiftproxy131 is CRITICAL: HTTP CRITICAL - Invalid HTTP response received from host: HTTP/1.1 401 Unauthorized [10:13:58] PROBLEM - swiftproxy131 HTTPS on swiftproxy131 is CRITICAL: HTTP CRITICAL - Invalid HTTP response received from host on port 443: HTTP/1.1 401 Unauthorized [10:28:46] PROBLEM - cp23 Current Load on cp23 is CRITICAL: CRITICAL - load average: 5.78, 3.82, 2.45 [10:30:45] PROBLEM - cp23 Current Load on cp23 is WARNING: WARNING - load average: 3.79, 3.61, 2.54 [10:32:10] PROBLEM - cp23 APT on cp23 is CRITICAL: CHECK_NRPE STATE CRITICAL: Socket timeout after 60 seconds. [10:32:43] PROBLEM - cp23 Current Load on cp23 is CRITICAL: CRITICAL - load average: 7.82, 5.24, 3.26 [10:37:09] RECOVERY - cp23 APT on cp23 is OK: APT OK: 2 packages available for upgrade (0 critical updates). [10:38:38] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 1.01, 2.97, 2.92 [10:46:00] PROBLEM - swiftproxy111 Puppet on swiftproxy111 is WARNING: WARNING: Puppet last ran 1 hour ago [10:52:23] PROBLEM - cloud11 Puppet on cloud11 is WARNING: WARNING: Puppet last ran 1 hour ago [10:58:17] PROBLEM - cp22 NTP time on cp22 is WARNING: NTP WARNING: Offset 0.1341492534 secs [11:09:57] PROBLEM - cp23 NTP time on cp23 is WARNING: NTP WARNING: Offset 0.1727074981 secs [11:10:17] RECOVERY - cp22 NTP time on cp22 is OK: NTP OK: Offset 0.0862852037 secs [11:26:18] I received a Grafana alert to the email, looks like there are a lot of refreshLinks jobs for bluepageswiki [11:28:24] I can't help but think this has to do with SMW, judging by the smw.changePropagationUpdate jobs when the refreshLinks jobs cascaded in: https://grafana.miraheze.org/d/GtxbP1Xnk/mediawiki?orgId=1&from=now-12h&to=now&var-node=jobchron121&var-job=AssembleUploadChunks&var-job=CentralAuthCreateLocalAccountJob&var-job=CentralAuthUnattachUserJob&var-job=ChangeDeletionNotification&var-job=ChangeNotification&var-job=ChangeVisibilityNotific [11:28:24] ation&var-job=CleanTermsIfUnused&var-job=CreateWikiJob&var-job=DataDumpGenerateJob&var-job=DeleteJob&var-job=DeleteTranslatableBundleJob&var-job=DispatchChangeDeletionNotification&var-job=DispatchChangeVisibilityNotification&var-job=DispatchChanges&var-job=EchoNotificationDeleteJob&var-job=EchoNotificationJob&var-job=EchoPushNotificationRequest&var-job=EntityChangeNotification&var-job=GlobalNewFilesDeleteJob&var-job=GlobalNewFiles [11:28:25] InsertJob&var-job=GlobalNewFilesMoveJob&var-job=GlobalUserPageLocalJobSubmitJob&var-job=InitImageDataJob&var-job=LocalGlobalUserPageCacheUpdateJob&var-job=LocalPageMoveJob&var-job=LocalRenameUserJob&var-job=LoginNotifyChecks&var-job=MDCreatePage&var-job=MDDeletePage&var-job=MWScriptJob&var-job=MassMessageJob&var-job=MassMessageServerSideJob&var-job=MassMessageSubmitJob&var-job=MessageGroupStatesUpdaterJob&var-job=MessageGroupStats [11:28:30] RebuildJob&var-job=MessageIndexRebuildJob&var-job=MessageUpdateJob&var-job=MoveTranslatableBundleJob&var-job=NamespaceMigrationJob&var-job=PageProperties&var-job=PublishStashedFile&var-job=PurgeEntityData&var-job=RecordLintJob&var-job=RemovePIIJob&var-job=RenderTranslationPageJob&var-job=RequestWikiAIJob&var-job=SMWRefreshJob&var-job=SMWUpdateJob&var-job=SMW%5CChangePropagationClassUpdateJob&var-job=SMW%5CChangePropagationDispatch [11:28:35] Job&var-job=SMW%5CChangePropagationUpdateJob&var-job=SMW%5CEntityIdDisposerJob&var-job=SMW%5CFulltextSearchTableRebuildJob&var-job=SMW%5CFulltextSearchTableUpdateJob&var-job=SMW%5CPropertyStatisticsRebuildJob&var-job=SMW%5CRefreshJob&var-job=SMW%5CUpdateDispatcherJob&var-job=SMW%5CUpdateJob&var-job=SetContainersAccessJob&var-job=TTMServerMessageUpdateJob&var-job=ThumbnailRender&var-job=TranslatableBundleDeleteJob&var-job=Translata [11:28:42] bleBundleMoveJob&var-job=TranslateRenderJob&var-job=TranslateSandboxEmailJob&var-job=TranslationNotificationsEmailJob&var-job=TranslationNotificationsSubmitJob&var-job=TranslationsUpdateJob&var-job=UpdateMessageBundle&var-job=UpdateRepoOnDelete&var-job=UpdateRepoOnMove&var-job=UpdateTranslatablePageJob&var-job=UpdateTranslatorActivity&var-job=activityUpdateJob&var-job=cargoPopulateTable&var-job=categoryMembershipChange&var-job=cdn [11:28:47] Purge&var-job=clearUserWatchlist&var-job=clearWatchlistNotifications&var-job=compileArticleMetadata&var-job=constraintsRunCheck&var-job=constraintsTableUpdate&var-job=crosswikiSuppressUser&var-job=deleteLinks&var-job=deletePage&var-job=dtImport&var-job=edReparse&var-job=enotifNotify&var-job=enqueue&var-job=fixDoubleRedirect&var-job=flaggedrevs_CacheUpdate&var-job=globalUsageCachePurge&var-job=htmlCacheUpdate&var-job=menteeOverview [11:28:52] UpdateDataForMentor&var-job=newUserMessageJob&var-job=newcomerTasksCacheRefreshJob&var-job=null&var-job=pageFormsCreatePage&var-job=pageSchemasCreatePage&var-job=reassignMenteesJob&var-job=recentChangesUpdate&var-job=refreshLinksDynamic&var-job=refreshLinksPrioritized&var-job=renameUser&var-job=revertedTagUpdate&var-job=sendMail&var-job=setUserMentorDatabaseJob&var-job=smw.changePropagationClassUpdate&var-job=smw.changePropagation [11:28:57] Dispatch&var-job=smw.changePropagationUpdate&var-job=smw.deferredConstraintCheckUpdateJob&var-job=smw.elasticFileIngest&var-job=smw.elasticIndexerRecovery&var-job=smw.entityIdDisposer&var-job=smw.fulltextSearchTableRebuild&var-job=smw.fulltextSearchTableUpdate&var-job=smw.parserCachePurgeJob&var-job=smw.propertyStatisticsRebuild&var-job=smw.refresh&var-job=smw.update&var-job=smw.updateDispatcher&var-job=updateBetaFeaturesUserCount [11:29:02] s&var-job=userEditCountInit&var-job=userGroupExpiry&var-job=userOptionsUpdate&var-job=watchlistExpiry&var-job=webVideoTranscode&var-job=webVideoTranscodePrioritized&var-job=wikibase-InjectRCRecords&var-job=wikibase-addUsagesForPage [11:29:06] oops, IRC doesn't like that URL much huh [11:29:39] https://grafana.miraheze.org/d/GtxbP1Xnk/mediawiki?orgId=1&from=now-12h&to=now&var-node=jobchron121&var-job=SMW%5CChangePropagationUpdateJob&var-job=smw.changePropagationUpdate [11:29:42] much better [11:30:07] hmm [11:39:57] RECOVERY - cp23 NTP time on cp23 is OK: NTP OK: Offset 0.07882192731 secs [11:44:52] PROBLEM - cp22 Current Load on cp22 is WARNING: WARNING - load average: 3.62, 2.41, 1.62 [11:48:07] PROBLEM - cp22 APT on cp22 is CRITICAL: CHECK_NRPE STATE CRITICAL: Socket timeout after 60 seconds. [11:50:04] RECOVERY - cp22 APT on cp22 is OK: APT OK: 2 packages available for upgrade (0 critical updates). [11:50:52] RECOVERY - cp22 Current Load on cp22 is OK: OK - load average: 1.52, 2.82, 2.19 [11:53:57] PROBLEM - cp23 NTP time on cp23 is WARNING: NTP WARNING: Offset -0.1168517768 secs [11:55:58] RECOVERY - cp23 NTP time on cp23 is OK: NTP OK: Offset -0.0863019526 secs [11:57:05] PROBLEM - graylog121 Current Load on graylog121 is WARNING: WARNING - load average: 3.46, 2.93, 2.57 [11:59:05] RECOVERY - graylog121 Current Load on graylog121 is OK: OK - load average: 1.79, 2.55, 2.48 [11:59:21] PROBLEM - cp23 Current Load on cp23 is WARNING: WARNING - load average: 3.42, 2.99, 2.06 [12:01:19] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 1.41, 2.40, 1.96 [12:24:17] PROBLEM - db142 Current Load on db142 is CRITICAL: CRITICAL - load average: 9.43, 6.17, 3.44 [12:25:58] PROBLEM - cp23 NTP time on cp23 is WARNING: NTP WARNING: Offset 0.1216821373 secs [12:26:09] PROBLEM - wiki.geoparkcorumbatai.com.br - reverse DNS on sslhost is WARNING: NoNameservers: All nameservers failed to answer the query geoparkcorumbatai.com.br. IN NS: Server 2606:4700:4700::1111 UDP port 53 answered SERVFAIL [12:26:17] RECOVERY - db142 Current Load on db142 is OK: OK - load average: 2.57, 4.83, 3.30 [12:29:01] PROBLEM - cp23 Current Load on cp23 is CRITICAL: CRITICAL - load average: 27.66, 11.04, 4.93 [12:30:04] RECOVERY - cp23 NTP time on cp23 is OK: NTP OK: Offset 0.07005962729 secs [12:33:57] PROBLEM - graylog121 Current Load on graylog121 is WARNING: WARNING - load average: 3.76, 3.14, 2.67 [12:35:56] RECOVERY - graylog121 Current Load on graylog121 is OK: OK - load average: 2.73, 2.90, 2.63 [12:42:49] PROBLEM - cp23 Current Load on cp23 is WARNING: WARNING - load average: 1.32, 2.62, 3.84 [12:42:50] Orange_Star: refreshLinks gets stuck sometimes due to memory issues [12:43:06] Just runJobs on that wiki with max memory until it’s clear [12:46:46] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 2.34, 2.23, 3.39 [12:51:50] PROBLEM - en.religiononfire.mar.in.ua - reverse DNS on sslhost is WARNING: NoNameservers: All nameservers failed to answer the query mar.in.ua. IN NS: Server 2606:4700:4700::1111 UDP port 53 answered SERVFAIL [12:54:48] RECOVERY - wiki.geoparkcorumbatai.com.br - reverse DNS on sslhost is OK: SSL OK - wiki.geoparkcorumbatai.com.br reverse DNS resolves to cp22.miraheze.org - CNAME OK [13:00:32] PROBLEM - cp23 Current Load on cp23 is CRITICAL: CRITICAL - load average: 4.80, 3.77, 3.33 [13:02:30] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 1.77, 2.90, 3.05 [13:05:25] [02miraheze/MirahezeMagic] 07Reception123 pushed 031 commit to 03master [+0/-0/±1] 13https://github.com/miraheze/MirahezeMagic/compare/66b9db8d1970...5a59909a890e [13:05:28] [02miraheze/MirahezeMagic] 07Reception123 035a59909 - temporary message regarding cloud11 for uploads [13:08:36] [02miraheze/mw-config] 07Reception123 pushed 031 commit to 03Reception123-patch-3 [+0/-0/±1] 13https://github.com/miraheze/mw-config/commit/997a709c2aa0 [13:08:37] [02miraheze/mw-config] 07Reception123 03997a709 - cloud11 failiures/upload issues sitenotice for File: pages [13:08:39] [02mw-config] 07Reception123 created branch 03Reception123-patch-3 - 13https://github.com/miraheze/mw-config [13:08:43] [02mw-config] 07Reception123 opened pull request 03#5190: cloud11 failiures/upload issues sitenotice for File: pages - 13https://github.com/miraheze/mw-config/pull/5190 [13:09:51] miraheze/mw-config - Reception123 the build passed. [13:10:35] [Grafana] !sre FIRING: The mediawiki job queue has more than 2500 unclaimed jobs https://grafana.miraheze.org/d/GtxbP1Xnk?orgId=1 [13:12:23] PROBLEM - mw121 Current Load on mw121 is CRITICAL: CRITICAL - load average: 13.64, 11.19, 9.38 [13:13:02] miraheze/MirahezeMagic - Reception123 the build passed. [13:14:22] PROBLEM - mw121 Current Load on mw121 is WARNING: WARNING - load average: 11.19, 10.71, 9.40 [13:14:32] looks like the jobs are getting done very slowly, it has slowed down since 13:48 UTC+2 [13:14:48] PROBLEM - graylog121 Current Load on graylog121 is WARNING: WARNING - load average: 3.65, 3.44, 3.00 [13:15:45] yeah refreshLinks can be like that... [13:16:11] [02mw-config] 07Reception123 synchronize pull request 03#5190: cloud11 failiures/upload issues sitenotice for File: pages - 13https://github.com/miraheze/mw-config/pull/5190 [13:16:13] [02miraheze/mw-config] 07Reception123 pushed 031 commit to 03Reception123-patch-3 [+0/-0/±1] 13https://github.com/miraheze/mw-config/compare/997a709c2aa0...0f08cf1a2d9c [13:16:15] [02miraheze/mw-config] 07Reception123 030f08cf1 - Update Sitenotice.php [13:16:20] RECOVERY - mw121 Current Load on mw121 is OK: OK - load average: 8.28, 9.84, 9.24 [13:16:47] PROBLEM - graylog121 Current Load on graylog121 is CRITICAL: CRITICAL - load average: 4.34, 3.67, 3.13 [13:16:49] [02miraheze/mediawiki] 07Reception123 pushed 031 commit to 03REL1_39 [+0/-0/±1] 13https://github.com/miraheze/mediawiki/compare/9d6baedf2911...7e26f1ff0680 [13:16:51] [02miraheze/mediawiki] 07Reception123 037e26f1f - Update MirahezeMagic [13:17:14] miraheze/mw-config - Reception123 the build passed. [13:17:14] !log [reception@mwtask141] starting deploy of {'pull': 'config', 'config': True, 'world': True, 'l10n': True} to all [13:17:19] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [13:17:20] !log [reception@mwtask141] DEPLOY ABORTED: Non-Zero Exit Code in prep, see output. [13:17:26] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [13:17:29] [02mw-config] 07Reception123 closed pull request 03#5190: cloud11 failiures/upload issues sitenotice for File: pages - 13https://github.com/miraheze/mw-config/pull/5190 [13:17:30] [02miraheze/mw-config] 07Reception123 pushed 031 commit to 03master [+0/-0/±1] 13https://github.com/miraheze/mw-config/compare/421e3e47cc14...e846fcec3e89 [13:17:30] !log [reception@mwtask141] starting deploy of {'pull': 'config', 'config': True, 'world': True, 'l10n': True} to all [13:17:31] [02miraheze/mw-config] 07Reception123 03e846fce - cloud11 failiures/upload issues sitenotice for File: pages (#5190) [13:17:34] [02miraheze/mw-config] 07Reception123 deleted branch 03Reception123-patch-3 [13:17:35] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [13:17:36] [02mw-config] 07Reception123 deleted branch 03Reception123-patch-3 - 13https://github.com/miraheze/mw-config [13:18:28] miraheze/mw-config - Reception123 the build passed. [13:18:47] PROBLEM - graylog121 Current Load on graylog121 is WARNING: WARNING - load average: 3.30, 3.52, 3.14 [13:20:40] RECOVERY - en.religiononfire.mar.in.ua - reverse DNS on sslhost is OK: SSL OK - en.religiononfire.mar.in.ua reverse DNS resolves to cp23.miraheze.org - CNAME OK [13:22:14] !log [reception@mwtask141] DEPLOY ABORTED: Non-Zero Exit Code in prep, see output. [13:22:18] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [13:22:46] RECOVERY - graylog121 Current Load on graylog121 is OK: OK - load average: 2.73, 3.30, 3.15 [13:24:40] !log [reception@mwtask141] starting deploy of {'pull': 'config', 'config': True, 'l10n': True, 'folders': 'w/extensions/MirahezeMagic'} to all [13:24:45] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [13:27:12] !log [@test131] starting deploy of {'config': True} to all [13:27:13] !log [@test131] finished deploy of {'config': True} to all - SUCCESS in 0s [13:27:17] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [13:27:21] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [13:28:05] PROBLEM - cp23 Current Load on cp23 is CRITICAL: CRITICAL - load average: 5.16, 3.11, 2.30 [13:31:07] PROBLEM - mw121 Current Load on mw121 is WARNING: WARNING - load average: 11.12, 10.29, 9.73 [13:32:25] PROBLEM - cp23 APT on cp23 is CRITICAL: CHECK_NRPE STATE CRITICAL: Socket timeout after 60 seconds. [13:34:25] RECOVERY - cp23 APT on cp23 is OK: APT OK: 2 packages available for upgrade (0 critical updates). [13:35:05] PROBLEM - mw121 Current Load on mw121 is CRITICAL: CRITICAL - load average: 12.40, 11.05, 10.16 [13:35:59] PROBLEM - cp23 Current Load on cp23 is WARNING: WARNING - load average: 2.02, 3.99, 3.44 [13:36:43] PROBLEM - graylog121 Current Load on graylog121 is CRITICAL: CRITICAL - load average: 4.73, 3.85, 3.48 [13:37:03] PROBLEM - mw121 Current Load on mw121 is WARNING: WARNING - load average: 9.96, 10.96, 10.26 [13:37:57] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 2.47, 3.39, 3.28 [13:38:43] PROBLEM - graylog121 Current Load on graylog121 is WARNING: WARNING - load average: 3.36, 3.54, 3.40 [13:39:02] RECOVERY - mw121 Current Load on mw121 is OK: OK - load average: 8.43, 9.76, 9.88 [13:42:42] RECOVERY - graylog121 Current Load on graylog121 is OK: OK - load average: 3.03, 3.30, 3.35 [13:48:07] PROBLEM - cp22 APT on cp22 is CRITICAL: CHECK_NRPE STATE CRITICAL: Socket timeout after 60 seconds. [13:49:41] !log [reception@mwtask141] starting deploy of {'l10n': True, 'folders': 'w/extensions/MirahezeMagic', 'ignoretime': True} to all [13:49:46] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [13:50:03] RECOVERY - cp22 APT on cp22 is OK: APT OK: 2 packages available for upgrade (0 critical updates). [13:55:24] !log [reception@mwtask141] starting deploy of {'pull': 'config', 'config': True} to all [13:55:30] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [13:55:40] !log [reception@mwtask141] finished deploy of {'pull': 'config', 'config': True} to all - SUCCESS in 15s [13:55:47] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [13:58:46] PROBLEM - mw121 Current Load on mw121 is WARNING: WARNING - load average: 10.95, 11.62, 10.60 [13:59:47] PROBLEM - cp23 Current Load on cp23 is CRITICAL: CRITICAL - load average: 6.22, 3.63, 2.72 [14:02:13] PROBLEM - cp23 APT on cp23 is CRITICAL: CHECK_NRPE STATE CRITICAL: Socket timeout after 60 seconds. [14:02:47] PROBLEM - db142 Current Load on db142 is WARNING: WARNING - load average: 7.96, 6.54, 4.90 [14:03:37] PROBLEM - graylog121 Current Load on graylog121 is WARNING: WARNING - load average: 3.57, 3.71, 3.44 [14:03:46] PROBLEM - cp23 Current Load on cp23 is WARNING: WARNING - load average: 2.92, 3.54, 2.92 [14:04:42] PROBLEM - mw121 Current Load on mw121 is CRITICAL: CRITICAL - load average: 13.42, 11.54, 10.83 [14:04:46] PROBLEM - db142 Current Load on db142 is CRITICAL: CRITICAL - load average: 8.88, 7.28, 5.36 [14:04:52] RECOVERY - cp23 APT on cp23 is OK: APT OK: 2 packages available for upgrade (0 critical updates). [14:05:36] RECOVERY - graylog121 Current Load on graylog121 is OK: OK - load average: 2.89, 3.33, 3.32 [14:05:46] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 2.06, 3.13, 2.85 [14:06:19] PROBLEM - es141 Current Load on es141 is CRITICAL: CRITICAL - load average: 4.16, 3.12, 2.14 [14:06:46] PROBLEM - db142 Current Load on db142 is WARNING: WARNING - load average: 6.79, 7.18, 5.57 [14:07:31] PROBLEM - mw122 Current Load on mw122 is CRITICAL: CRITICAL - load average: 14.02, 11.02, 9.33 [14:08:18] RECOVERY - es141 Current Load on es141 is OK: OK - load average: 2.35, 2.85, 2.16 [14:08:45] RECOVERY - db142 Current Load on db142 is OK: OK - load average: 2.71, 5.67, 5.23 [14:09:31] PROBLEM - mw122 Current Load on mw122 is WARNING: WARNING - load average: 11.94, 11.34, 9.66 [14:11:35] PROBLEM - graylog121 Current Load on graylog121 is WARNING: WARNING - load average: 3.84, 3.82, 3.53 [14:15:31] RECOVERY - mw122 Current Load on mw122 is OK: OK - load average: 9.05, 10.17, 9.70 [14:16:07] PROBLEM - cp22 Current Load on cp22 is CRITICAL: CRITICAL - load average: 18.50, 6.83, 3.40 [14:16:34] PROBLEM - mw121 Current Load on mw121 is WARNING: WARNING - load average: 8.91, 11.23, 11.53 [14:18:15] PROBLEM - cp22 APT on cp22 is CRITICAL: CHECK_NRPE STATE CRITICAL: Socket timeout after 60 seconds. [14:19:34] PROBLEM - graylog121 Current Load on graylog121 is CRITICAL: CRITICAL - load average: 4.33, 3.87, 3.63 [14:20:11] RECOVERY - cp22 APT on cp22 is OK: APT OK: 2 packages available for upgrade (0 critical updates). [14:21:33] PROBLEM - graylog121 Current Load on graylog121 is WARNING: WARNING - load average: 3.82, 3.81, 3.64 [14:21:50] PROBLEM - en.religiononfire.mar.in.ua - reverse DNS on sslhost is WARNING: NoNameservers: All nameservers failed to answer the query mar.in.ua. IN NS: Server 2606:4700:4700::1111 UDP port 53 answered SERVFAIL [14:24:29] RECOVERY - mw121 Current Load on mw121 is OK: OK - load average: 8.98, 9.00, 10.17 [14:24:38] PROBLEM - db142 Current Load on db142 is WARNING: WARNING - load average: 7.75, 7.08, 5.98 [14:25:33] RECOVERY - graylog121 Current Load on graylog121 is OK: OK - load average: 2.30, 3.05, 3.37 [14:26:37] PROBLEM - db142 Current Load on db142 is CRITICAL: CRITICAL - load average: 9.14, 7.25, 6.14 [14:27:47] !log [reception@mwtask141] finished deploy of {'l10n': True, 'folders': 'w/extensions/MirahezeMagic', 'ignoretime': True} to all - SUCCESS in 2280s [14:27:52] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [14:29:47] PROBLEM - cp23 Current Load on cp23 is CRITICAL: CRITICAL - load average: 6.58, 3.80, 2.64 [14:29:53] PROBLEM - cp22 Current Load on cp22 is WARNING: WARNING - load average: 0.64, 2.08, 3.65 [14:30:36] RECOVERY - db142 Current Load on db142 is OK: OK - load average: 2.51, 6.20, 6.10 [14:31:51] RECOVERY - cp22 Current Load on cp22 is OK: OK - load average: 1.24, 1.79, 3.36 [14:32:13] PROBLEM - cp23 APT on cp23 is CRITICAL: CHECK_NRPE STATE CRITICAL: Socket timeout after 60 seconds. [14:32:22] PROBLEM - mw121 Current Load on mw121 is CRITICAL: CRITICAL - load average: 12.68, 10.33, 10.18 [14:34:21] RECOVERY - mw121 Current Load on mw121 is OK: OK - load average: 7.52, 9.25, 9.80 [14:38:30] PROBLEM - graylog121 Current Load on graylog121 is WARNING: WARNING - load average: 3.93, 3.52, 3.36 [14:40:10] RECOVERY - cp23 APT on cp23 is OK: APT OK: 2 packages available for upgrade (0 critical updates). [14:41:46] PROBLEM - cp23 Current Load on cp23 is WARNING: WARNING - load average: 1.90, 3.54, 3.58 [14:43:47] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 1.54, 2.95, 3.36 [14:44:13] PROBLEM - mw121 Current Load on mw121 is WARNING: WARNING - load average: 11.10, 10.35, 9.93 [14:46:12] RECOVERY - mw121 Current Load on mw121 is OK: OK - load average: 9.09, 9.92, 9.84 [14:48:28] PROBLEM - graylog121 Current Load on graylog121 is CRITICAL: CRITICAL - load average: 4.47, 3.92, 3.58 [14:50:27] PROBLEM - graylog121 Current Load on graylog121 is WARNING: WARNING - load average: 3.52, 3.83, 3.59 [14:50:40] RECOVERY - en.religiononfire.mar.in.ua - reverse DNS on sslhost is OK: SSL OK - en.religiononfire.mar.in.ua reverse DNS resolves to cp23.miraheze.org - CNAME OK [14:57:47] PROBLEM - cp23 Current Load on cp23 is CRITICAL: CRITICAL - load average: 10.38, 5.07, 3.51 [15:00:25] PROBLEM - graylog121 Current Load on graylog121 is CRITICAL: CRITICAL - load average: 4.23, 3.83, 3.67 [15:01:18] PROBLEM - cp23 APT on cp23 is CRITICAL: CHECK_NRPE STATE CRITICAL: Socket timeout after 60 seconds. [15:02:25] PROBLEM - graylog121 Current Load on graylog121 is WARNING: WARNING - load average: 3.25, 3.60, 3.60 [15:03:19] RECOVERY - cp23 APT on cp23 is OK: APT OK: 2 packages available for upgrade (0 critical updates). [15:03:46] PROBLEM - cp23 Current Load on cp23 is WARNING: WARNING - load average: 1.95, 3.98, 3.69 [15:05:47] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 1.32, 3.02, 3.37 [15:12:23] PROBLEM - graylog121 Current Load on graylog121 is CRITICAL: CRITICAL - load average: 4.10, 3.69, 3.61 [15:13:52] PROBLEM - mw121 Current Load on mw121 is CRITICAL: CRITICAL - load average: 12.02, 10.98, 9.87 [15:17:49] PROBLEM - mw121 Current Load on mw121 is WARNING: WARNING - load average: 11.31, 11.62, 10.41 [15:18:22] PROBLEM - graylog121 Current Load on graylog121 is WARNING: WARNING - load average: 3.72, 3.86, 3.73 [15:20:22] PROBLEM - graylog121 Current Load on graylog121 is CRITICAL: CRITICAL - load average: 4.20, 3.95, 3.77 [15:21:47] RECOVERY - mw121 Current Load on mw121 is OK: OK - load average: 8.47, 10.09, 10.07 [15:22:21] PROBLEM - graylog121 Current Load on graylog121 is WARNING: WARNING - load average: 3.39, 3.74, 3.72 [15:24:21] PROBLEM - graylog121 Current Load on graylog121 is CRITICAL: CRITICAL - load average: 4.70, 4.17, 3.88 [15:25:43] PROBLEM - mw121 Current Load on mw121 is CRITICAL: CRITICAL - load average: 12.17, 11.16, 10.51 [15:27:41] PROBLEM - mw121 Current Load on mw121 is WARNING: WARNING - load average: 9.31, 10.37, 10.30 [15:28:20] PROBLEM - graylog121 Current Load on graylog121 is WARNING: WARNING - load average: 3.51, 3.89, 3.84 [15:29:52] PROBLEM - cp23 Current Load on cp23 is CRITICAL: CRITICAL - load average: 15.20, 12.91, 6.77 [15:37:35] RECOVERY - mw121 Current Load on mw121 is OK: OK - load average: 9.32, 9.76, 10.14 [15:40:18] PROBLEM - graylog121 Current Load on graylog121 is CRITICAL: CRITICAL - load average: 4.61, 3.74, 3.67 [15:41:35] PROBLEM - mw121 Current Load on mw121 is CRITICAL: CRITICAL - load average: 14.34, 11.69, 10.79 [15:42:18] PROBLEM - graylog121 Current Load on graylog121 is WARNING: WARNING - load average: 3.19, 3.44, 3.56 [15:43:35] PROBLEM - mw121 Current Load on mw121 is WARNING: WARNING - load average: 10.48, 11.18, 10.72 [15:43:47] PROBLEM - cp23 Current Load on cp23 is WARNING: WARNING - load average: 0.58, 1.89, 3.68 [15:45:46] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 0.84, 1.53, 3.32 [15:51:35] RECOVERY - mw121 Current Load on mw121 is OK: OK - load average: 7.65, 9.41, 10.09 [15:52:18] RECOVERY - graylog121 Current Load on graylog121 is OK: OK - load average: 2.13, 2.94, 3.32 [15:57:35] PROBLEM - mw121 Current Load on mw121 is CRITICAL: CRITICAL - load average: 12.11, 10.79, 10.38 [15:59:35] RECOVERY - mw121 Current Load on mw121 is OK: OK - load average: 7.10, 9.31, 9.89 [15:59:47] PROBLEM - cp23 Current Load on cp23 is CRITICAL: CRITICAL - load average: 9.56, 5.60, 3.75 [16:01:14] PROBLEM - cp23 APT on cp23 is CRITICAL: CHECK_NRPE STATE CRITICAL: Socket timeout after 60 seconds. [16:02:18] PROBLEM - graylog121 Current Load on graylog121 is CRITICAL: CRITICAL - load average: 5.19, 3.86, 3.53 [16:03:15] RECOVERY - cp23 APT on cp23 is OK: APT OK: 2 packages available for upgrade (0 critical updates). [16:03:46] PROBLEM - cp23 Current Load on cp23 is WARNING: WARNING - load average: 1.25, 3.51, 3.36 [16:05:46] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 1.29, 2.75, 3.09 [16:07:31] PROBLEM - mw122 Current Load on mw122 is WARNING: WARNING - load average: 10.68, 9.34, 8.18 [16:08:35] !log puppet141: upgrade puppet-agent puppetdb puppetdb-termini puppetserver tzdata [16:08:39] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [16:09:31] RECOVERY - mw122 Current Load on mw122 is OK: OK - load average: 7.92, 8.66, 8.07 [16:13:07] PROBLEM - mw131 Puppet on mw131 is CRITICAL: CRITICAL: Failed to apply catalog, zero resources tracked by Puppet. It might be a dependency cycle. [16:13:21] PROBLEM - mwtask141 Puppet on mwtask141 is CRITICAL: CRITICAL: Puppet has 440 failures. Last run 2 minutes ago with 440 failures. Failed resources (up to 3 shown): File[/etc/ssl/private/iol.wiki.key],File[/etc/ssl/localcerts/roll-of-arms.com.crt],File[/etc/ssl/private/roll-of-arms.com.key],File[/etc/ssl/localcerts/wiki.yahyabd.xyz.crt] [16:13:26] PROBLEM - cp22 Puppet on cp22 is CRITICAL: CRITICAL: Failed to apply catalog, zero resources tracked by Puppet. It might be a dependency cycle. [16:14:18] PROBLEM - graylog121 Current Load on graylog121 is WARNING: WARNING - load average: 3.91, 3.89, 3.86 [16:14:52] PROBLEM - mw121 Puppet on mw121 is CRITICAL: CRITICAL: Puppet has 838 failures. Last run 3 minutes ago with 838 failures. Failed resources (up to 3 shown): File[/etc/ssl/private/n64brew.dev.key],File[/etc/ssl/localcerts/infectowiki.com.crt],File[/etc/ssl/private/infectowiki.com.key],File[/etc/ssl/localcerts/history.sdtef.org.crt] [16:16:42] !log upgrade puppet-agent on all servers except cloud11 + the vms on cloud11 [16:16:49] Logged the message at https://meta.miraheze.org/wiki/Tech:Server_admin_log [16:18:18] PROBLEM - graylog121 Current Load on graylog121 is CRITICAL: CRITICAL - load average: 5.17, 4.32, 4.02 [16:20:16] [02miraheze/puppet] 07paladox pushed 031 commit to 03paladox-patch-3 [+0/-0/±1] 13https://github.com/miraheze/puppet/commit/c2059cef1551 [16:20:18] PROBLEM - graylog121 Current Load on graylog121 is WARNING: WARNING - load average: 3.20, 3.83, 3.87 [16:20:18] PROBLEM - es141 Current Load on es141 is CRITICAL: CRITICAL - load average: 4.01, 2.49, 1.52 [16:20:19] [02miraheze/puppet] 07paladox 03c2059ce - relaybot: also use proxy for https apt repo [16:20:22] [02puppet] 07paladox created branch 03paladox-patch-3 - 13https://github.com/miraheze/puppet [16:20:24] [02puppet] 07paladox opened pull request 03#3185: relaybot: also use proxy for https apt repo - 13https://github.com/miraheze/puppet/pull/3185 [16:20:26] [02puppet] 07paladox closed pull request 03#3185: relaybot: also use proxy for https apt repo - 13https://github.com/miraheze/puppet/pull/3185 [16:20:27] [02miraheze/puppet] 07paladox pushed 031 commit to 03master [+0/-0/±1] 13https://github.com/miraheze/puppet/compare/241009f1d77c...e0d4bbbcd78a [16:20:28] [02miraheze/puppet] 07paladox 03e0d4bbb - relaybot: also use proxy for https apt repo (#3185) [16:20:29] [02puppet] 07paladox deleted branch 03paladox-patch-3 - 13https://github.com/miraheze/puppet [16:20:32] [02miraheze/puppet] 07paladox deleted branch 03paladox-patch-3 [16:21:05] PROBLEM - db142 Current Load on db142 is CRITICAL: CRITICAL - load average: 8.46, 6.00, 3.70 [16:22:19] PROBLEM - es141 Current Load on es141 is WARNING: WARNING - load average: 3.44, 2.85, 1.78 [16:23:36] PROBLEM - swiftobject121 Puppet on swiftobject121 is CRITICAL: CRITICAL: Puppet has 1 failures. Last run 2 minutes ago with 1 failures. Failed resources (up to 3 shown): Package[puppet-agent] [16:23:59] PROBLEM - swiftobject122 Puppet on swiftobject122 is CRITICAL: CRITICAL: Puppet has 1 failures. Last run 2 minutes ago with 1 failures. Failed resources (up to 3 shown): Package[puppet-agent] [16:24:18] PROBLEM - graylog121 Current Load on graylog121 is CRITICAL: CRITICAL - load average: 4.50, 3.91, 3.88 [16:24:18] RECOVERY - es141 Current Load on es141 is OK: OK - load average: 3.07, 2.98, 1.96 [16:27:03] PROBLEM - db142 Current Load on db142 is WARNING: WARNING - load average: 6.30, 7.16, 5.11 [16:27:17] PROBLEM - mw121 Current Load on mw121 is WARNING: WARNING - load average: 11.44, 10.51, 10.06 [16:28:28] PROBLEM - cp23 Current Load on cp23 is CRITICAL: CRITICAL - load average: 6.42, 4.26, 2.95 [16:29:02] RECOVERY - db142 Current Load on db142 is OK: OK - load average: 1.74, 5.15, 4.62 [16:31:14] PROBLEM - mw121 Current Load on mw121 is CRITICAL: CRITICAL - load average: 13.37, 11.36, 10.46 [16:32:14] PROBLEM - cp23 APT on cp23 is CRITICAL: CHECK_NRPE STATE CRITICAL: Socket timeout after 60 seconds. [16:33:13] PROBLEM - mw121 Current Load on mw121 is WARNING: WARNING - load average: 10.90, 10.87, 10.38 [16:35:11] RECOVERY - mw121 Current Load on mw121 is OK: OK - load average: 8.08, 9.97, 10.11 [16:36:18] PROBLEM - graylog121 Current Load on graylog121 is WARNING: WARNING - load average: 2.81, 3.75, 3.96 [16:37:29] RECOVERY - cp23 APT on cp23 is OK: APT OK: 1 packages available for upgrade (0 critical updates). [16:38:19] PROBLEM - cp23 Current Load on cp23 is WARNING: WARNING - load average: 2.07, 3.66, 3.42 [16:40:17] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 1.59, 2.82, 3.13 [16:41:06] PROBLEM - mw121 Current Load on mw121 is CRITICAL: CRITICAL - load average: 14.40, 11.00, 10.30 [16:41:07] RECOVERY - mw131 Puppet on mw131 is OK: OK: Puppet is currently enabled, last run 3 seconds ago with 0 failures [16:41:19] RECOVERY - mwtask141 Puppet on mwtask141 is OK: OK: Puppet is currently enabled, last run 44 seconds ago with 0 failures [16:42:43] RECOVERY - mw121 Puppet on mw121 is OK: OK: Puppet is currently enabled, last run 1 minute ago with 0 failures [16:43:26] RECOVERY - cp22 Puppet on cp22 is OK: OK: Puppet is currently enabled, last run 11 seconds ago with 0 failures [16:44:18] PROBLEM - graylog121 Current Load on graylog121 is CRITICAL: CRITICAL - load average: 4.82, 4.02, 3.93 [16:45:04] PROBLEM - mw121 Current Load on mw121 is WARNING: WARNING - load average: 11.98, 11.56, 10.72 [16:49:01] PROBLEM - mw121 Current Load on mw121 is CRITICAL: CRITICAL - load average: 13.98, 11.82, 10.95 [16:49:45] RECOVERY - swiftobject121 Puppet on swiftobject121 is OK: OK: Puppet is currently enabled, last run 11 seconds ago with 0 failures [16:49:59] RECOVERY - swiftobject122 Puppet on swiftobject122 is OK: OK: Puppet is currently enabled, last run 23 seconds ago with 0 failures [16:52:58] PROBLEM - mw121 Current Load on mw121 is WARNING: WARNING - load average: 9.63, 11.05, 10.85 [17:00:01] PROBLEM - cp23 Current Load on cp23 is CRITICAL: CRITICAL - load average: 10.19, 6.72, 3.95 [17:02:26] PROBLEM - cp23 APT on cp23 is CRITICAL: CHECK_NRPE STATE CRITICAL: Socket timeout after 60 seconds. [17:02:52] RECOVERY - mw121 Current Load on mw121 is OK: OK - load average: 6.53, 9.18, 10.15 [17:04:41] RECOVERY - cp23 APT on cp23 is OK: APT OK: 1 packages available for upgrade (0 critical updates). [17:10:35] [Grafana] !sre FIRING: The mediawiki job queue has more than 2500 unclaimed jobs https://grafana.miraheze.org/d/GtxbP1Xnk?orgId=1 [17:10:45] PROBLEM - mw121 Current Load on mw121 is CRITICAL: CRITICAL - load average: 17.69, 12.49, 11.01 [17:18:56] PROBLEM - swiftproxy131 Current Load on swiftproxy131 is CRITICAL: CRITICAL - load average: 10.91, 7.88, 6.36 [17:21:47] PROBLEM - cp23 Current Load on cp23 is WARNING: WARNING - load average: 1.00, 1.74, 3.66 [17:23:46] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 0.65, 1.35, 3.28 [17:24:36] PROBLEM - mw121 Current Load on mw121 is WARNING: WARNING - load average: 10.76, 11.98, 12.00 [17:24:48] RECOVERY - swiftproxy131 Current Load on swiftproxy131 is OK: OK - load average: 4.83, 6.70, 6.40 [17:25:01] PROBLEM - cp32 Current Load on cp32 is CRITICAL: CRITICAL - load average: 5.23, 5.39, 3.65 [17:28:19] PROBLEM - es141 Current Load on es141 is WARNING: WARNING - load average: 3.79, 2.51, 1.53 [17:28:52] PROBLEM - cp32 Current Load on cp32 is WARNING: WARNING - load average: 1.92, 3.62, 3.31 [17:30:18] RECOVERY - es141 Current Load on es141 is OK: OK - load average: 2.83, 2.52, 1.66 [17:30:48] RECOVERY - cp32 Current Load on cp32 is OK: OK - load average: 1.57, 2.92, 3.09 [17:34:41] PROBLEM - cp32 Current Load on cp32 is WARNING: WARNING - load average: 3.98, 3.63, 3.34 [17:36:37] PROBLEM - cp32 Current Load on cp32 is CRITICAL: CRITICAL - load average: 5.72, 4.07, 3.52 [17:38:32] PROBLEM - cp32 Current Load on cp32 is WARNING: WARNING - load average: 3.40, 3.96, 3.56 [17:40:25] PROBLEM - mw121 Current Load on mw121 is CRITICAL: CRITICAL - load average: 14.14, 11.01, 10.94 [17:40:28] RECOVERY - cp32 Current Load on cp32 is OK: OK - load average: 2.39, 3.35, 3.38 [17:43:25] PROBLEM - swiftproxy131 Current Load on swiftproxy131 is CRITICAL: CRITICAL - load average: 9.79, 8.37, 6.86 [17:45:22] PROBLEM - swiftproxy131 Current Load on swiftproxy131 is WARNING: WARNING - load average: 5.73, 7.39, 6.69 [17:45:25] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 50% [17:46:21] PROBLEM - mw121 Current Load on mw121 is WARNING: WARNING - load average: 9.43, 11.17, 11.15 [17:46:26] PROBLEM - cp32 Current Load on cp32 is WARNING: WARNING - load average: 3.93, 3.65, 3.50 [17:47:22] RECOVERY - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is OK: OK - NGINX Error Rate is 24% [17:48:26] RECOVERY - cp32 Current Load on cp32 is OK: OK - load average: 2.73, 3.27, 3.37 [17:49:01] PROBLEM - swiftproxy111 Current Load on swiftproxy111 is CRITICAL: CRITICAL - load average: 11.13, 8.27, 5.88 [17:49:16] PROBLEM - swiftproxy131 Current Load on swiftproxy131 is CRITICAL: CRITICAL - load average: 7.85, 8.14, 7.17 [17:51:14] PROBLEM - swiftproxy131 Current Load on swiftproxy131 is WARNING: WARNING - load average: 6.26, 7.47, 7.04 [17:52:17] PROBLEM - mw121 Current Load on mw121 is CRITICAL: CRITICAL - load average: 12.10, 10.88, 10.95 [17:53:11] PROBLEM - swiftproxy131 Current Load on swiftproxy131 is CRITICAL: CRITICAL - load average: 9.39, 8.35, 7.42 [17:53:36] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 53% [17:54:16] PROBLEM - mw121 Current Load on mw121 is WARNING: WARNING - load average: 10.29, 10.71, 10.88 [17:55:01] PROBLEM - swiftproxy111 Current Load on swiftproxy111 is WARNING: WARNING - load average: 6.10, 7.39, 6.31 [17:57:01] PROBLEM - swiftproxy111 Current Load on swiftproxy111 is CRITICAL: CRITICAL - load average: 7.72, 8.15, 6.75 [17:59:01] PROBLEM - swiftproxy111 Current Load on swiftproxy111 is WARNING: WARNING - load average: 6.63, 7.50, 6.68 [17:59:03] PROBLEM - swiftproxy131 Current Load on swiftproxy131 is WARNING: WARNING - load average: 4.07, 7.31, 7.42 [17:59:31] RECOVERY - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is OK: OK - NGINX Error Rate is 35% [18:01:01] RECOVERY - swiftproxy111 Current Load on swiftproxy111 is OK: OK - load average: 4.55, 6.42, 6.38 [18:01:25] PROBLEM - knowledgebase.clientmanager.co.za - reverse DNS on sslhost is WARNING: Timeout: The DNS operation timed out after 5.407663583755493 seconds [18:02:57] RECOVERY - swiftproxy131 Current Load on swiftproxy131 is OK: OK - load average: 5.14, 5.85, 6.79 [18:08:07] RECOVERY - mw121 Current Load on mw121 is OK: OK - load average: 8.12, 9.07, 9.93 [18:12:03] PROBLEM - mw121 Current Load on mw121 is CRITICAL: CRITICAL - load average: 13.27, 11.89, 10.88 [18:14:02] PROBLEM - mw121 Current Load on mw121 is WARNING: WARNING - load average: 11.23, 11.68, 10.93 [18:18:04] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 59% [18:19:58] RECOVERY - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is OK: OK - NGINX Error Rate is 27% [18:25:54] RECOVERY - mw121 Current Load on mw121 is OK: OK - load average: 10.12, 9.85, 10.19 [18:27:22] PROBLEM - swiftproxy131 Current Load on swiftproxy131 is CRITICAL: CRITICAL - load average: 14.23, 8.95, 7.28 [18:27:50] PROBLEM - cp23 Current Load on cp23 is WARNING: WARNING - load average: 3.42, 2.77, 1.91 [18:29:01] PROBLEM - swiftproxy111 Current Load on swiftproxy111 is CRITICAL: CRITICAL - load average: 8.12, 7.40, 6.28 [18:29:48] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 2.70, 2.94, 2.09 [18:30:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 52% [18:31:01] PROBLEM - swiftproxy111 Current Load on swiftproxy111 is WARNING: WARNING - load average: 7.69, 7.19, 6.32 [18:33:01] RECOVERY - swiftproxy111 Current Load on swiftproxy111 is OK: OK - load average: 5.78, 6.68, 6.24 [18:34:00] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 47% [18:37:55] RECOVERY - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is OK: OK - NGINX Error Rate is 15% [18:38:09] RECOVERY - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is OK: OK - NGINX Error Rate is 35% [18:39:05] PROBLEM - swiftproxy131 Current Load on swiftproxy131 is WARNING: WARNING - load average: 5.75, 7.83, 7.79 [18:41:02] PROBLEM - swiftproxy131 Current Load on swiftproxy131 is CRITICAL: CRITICAL - load average: 7.52, 8.22, 7.96 [18:41:58] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 45% [18:42:59] PROBLEM - swiftproxy131 Current Load on swiftproxy131 is WARNING: WARNING - load average: 6.41, 7.96, 7.92 [18:45:35] [Grafana] !sre RESOLVED: High Job Queue Backlog https://grafana.miraheze.org/d/GtxbP1Xnk?orgId=1 [18:46:54] PROBLEM - swiftproxy131 Current Load on swiftproxy131 is CRITICAL: CRITICAL - load average: 11.50, 8.53, 8.09 [18:47:41] RECOVERY - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is OK: OK - NGINX Error Rate is 23% [18:49:35] PROBLEM - mw121 Current Load on mw121 is WARNING: WARNING - load average: 10.95, 10.45, 10.04 [18:51:01] PROBLEM - swiftproxy111 Current Load on swiftproxy111 is CRITICAL: CRITICAL - load average: 8.44, 7.20, 6.32 [18:52:21] PROBLEM - uk.religiononfire.mar.in.ua - reverse DNS on sslhost is CRITICAL: rDNS CRITICAL - uk.religiononfire.mar.in.ua All nameservers failed to answer the query. [18:53:01] RECOVERY - swiftproxy111 Current Load on swiftproxy111 is OK: OK - load average: 3.78, 5.90, 5.95 [18:54:42] PROBLEM - swiftproxy131 Current Load on swiftproxy131 is WARNING: WARNING - load average: 4.52, 6.88, 7.71 [18:55:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 52% [18:56:40] PROBLEM - swiftproxy131 Current Load on swiftproxy131 is CRITICAL: CRITICAL - load average: 11.59, 8.73, 8.29 [18:57:35] RECOVERY - mw121 Current Load on mw121 is OK: OK - load average: 7.48, 9.97, 10.12 [18:58:37] PROBLEM - swiftproxy131 Current Load on swiftproxy131 is WARNING: WARNING - load average: 4.61, 7.16, 7.78 [18:59:46] PROBLEM - cp23 Current Load on cp23 is CRITICAL: CRITICAL - load average: 5.38, 3.82, 2.45 [19:00:34] PROBLEM - swiftproxy131 Current Load on swiftproxy131 is CRITICAL: CRITICAL - load average: 10.99, 8.81, 8.30 [19:01:35] PROBLEM - mw121 Current Load on mw121 is CRITICAL: CRITICAL - load average: 12.57, 11.10, 10.51 [19:01:47] PROBLEM - cp23 Current Load on cp23 is WARNING: WARNING - load average: 3.86, 3.73, 2.58 [19:03:01] PROBLEM - swiftproxy111 Current Load on swiftproxy111 is CRITICAL: CRITICAL - load average: 8.57, 9.01, 7.24 [19:03:31] RECOVERY - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is OK: OK - NGINX Error Rate is 22% [19:03:35] PROBLEM - mw121 Current Load on mw121 is WARNING: WARNING - load average: 8.77, 10.29, 10.30 [19:03:46] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 1.73, 2.94, 2.43 [19:05:35] RECOVERY - mw121 Current Load on mw121 is OK: OK - load average: 7.39, 9.34, 9.95 [19:07:01] PROBLEM - swiftproxy111 Current Load on swiftproxy111 is WARNING: WARNING - load average: 4.51, 7.04, 6.89 [19:07:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 46% [19:08:33] PROBLEM - swiftproxy131 Current Load on swiftproxy131 is WARNING: WARNING - load average: 4.54, 7.18, 7.87 [19:09:31] RECOVERY - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is OK: OK - NGINX Error Rate is 32% [19:10:33] PROBLEM - swiftproxy131 Current Load on swiftproxy131 is CRITICAL: CRITICAL - load average: 18.06, 10.98, 9.13 [19:12:31] PROBLEM - mw121 Current Load on mw121 is CRITICAL: CRITICAL - load average: 13.54, 11.61, 10.67 [19:13:18] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 63% [19:13:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 44% [19:14:30] PROBLEM - mw121 Current Load on mw121 is WARNING: WARNING - load average: 9.51, 10.51, 10.36 [19:15:16] RECOVERY - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is OK: OK - NGINX Error Rate is 36% [19:15:31] RECOVERY - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is OK: OK - NGINX Error Rate is 34% [19:16:28] PROBLEM - mw121 Current Load on mw121 is CRITICAL: CRITICAL - load average: 12.28, 11.28, 10.66 [19:17:01] PROBLEM - swiftproxy111 Current Load on swiftproxy111 is CRITICAL: CRITICAL - load average: 8.26, 7.59, 7.17 [19:18:27] PROBLEM - mw121 Current Load on mw121 is WARNING: WARNING - load average: 8.96, 10.69, 10.54 [19:21:01] PROBLEM - swiftproxy111 Current Load on swiftproxy111 is WARNING: WARNING - load average: 5.40, 7.46, 7.31 [19:21:56] RECOVERY - uk.religiononfire.mar.in.ua - reverse DNS on sslhost is OK: SSL OK - uk.religiononfire.mar.in.ua reverse DNS resolves to cp22.miraheze.org - CNAME OK [19:23:01] PROBLEM - swiftproxy111 Current Load on swiftproxy111 is CRITICAL: CRITICAL - load average: 15.16, 9.50, 8.01 [19:24:30] PROBLEM - cp33 Current Load on cp33 is WARNING: WARNING - load average: 3.64, 2.62, 2.02 [19:24:53] PROBLEM - db142 Current Load on db142 is WARNING: WARNING - load average: 6.95, 5.24, 3.29 [19:25:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 44% [19:26:22] RECOVERY - mw121 Current Load on mw121 is OK: OK - load average: 8.02, 9.19, 9.95 [19:26:26] RECOVERY - cp33 Current Load on cp33 is OK: OK - load average: 1.93, 2.31, 1.98 [19:26:53] PROBLEM - db142 Current Load on db142 is CRITICAL: CRITICAL - load average: 10.09, 6.59, 4.00 [19:27:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 61% [19:28:52] RECOVERY - db142 Current Load on db142 is OK: OK - load average: 3.55, 5.36, 3.87 [19:29:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 57% [19:29:47] PROBLEM - cp23 Current Load on cp23 is CRITICAL: CRITICAL - load average: 6.05, 3.83, 2.46 [19:30:48] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 54% [19:31:13] PROBLEM - cp23 APT on cp23 is CRITICAL: CHECK_NRPE STATE CRITICAL: Socket timeout after 60 seconds. [19:32:46] RECOVERY - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is OK: OK - NGINX Error Rate is 32% [19:33:18] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 41% [19:33:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 60% [19:35:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 44% [19:36:19] RECOVERY - cp23 APT on cp23 is OK: APT OK: 1 packages available for upgrade (0 critical updates). [19:37:15] PROBLEM - cp32 Current Load on cp32 is WARNING: WARNING - load average: 3.87, 3.42, 2.69 [19:39:07] RECOVERY - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is OK: OK - NGINX Error Rate is 29% [19:39:11] RECOVERY - cp32 Current Load on cp32 is OK: OK - load average: 2.04, 2.99, 2.63 [19:41:10] PROBLEM - mw121 Current Load on mw121 is CRITICAL: CRITICAL - load average: 13.57, 10.98, 10.00 [19:41:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 62% [19:41:46] PROBLEM - cp23 Current Load on cp23 is WARNING: WARNING - load average: 0.70, 2.92, 3.65 [19:43:09] RECOVERY - mw121 Current Load on mw121 is OK: OK - load average: 8.86, 10.07, 9.78 [19:43:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 40% [19:43:46] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 0.92, 2.29, 3.33 [19:45:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 60% [19:47:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 42% [19:51:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 81% [19:53:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 53% [19:53:41] PROBLEM - cp32 Current Load on cp32 is WARNING: WARNING - load average: 3.73, 3.01, 2.69 [19:55:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 85% [19:55:36] RECOVERY - cp32 Current Load on cp32 is OK: OK - load average: 2.60, 2.92, 2.70 [19:57:31] RECOVERY - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is OK: OK - NGINX Error Rate is 36% [19:59:17] PROBLEM - knowledgebase.clientmanager.co.za - reverse DNS on sslhost is CRITICAL: rDNS CRITICAL - knowledgebase.clientmanager.co.za All nameservers failed to answer the query. [20:01:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 46% [20:01:47] PROBLEM - cp23 Current Load on cp23 is CRITICAL: CRITICAL - load average: 4.64, 4.37, 3.26 [20:03:46] PROBLEM - cp23 Current Load on cp23 is WARNING: WARNING - load average: 1.80, 3.47, 3.07 [20:05:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 68% [20:05:46] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 1.23, 2.68, 2.82 [20:05:48] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 44% [20:07:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 44% [20:07:45] RECOVERY - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is OK: OK - NGINX Error Rate is 28% [20:08:15] PROBLEM - cp32 NTP time on cp32 is CRITICAL: NTP CRITICAL: Offset 0.5525172651 secs [20:08:24] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 50% [20:09:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 79% [20:11:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 53% [20:12:12] PROBLEM - swiftobject111 Disk Space on swiftobject111 is CRITICAL: connect to address 2a10:6740::6:203 port 5666: No route to hostconnect to host 2a10:6740::6:203 port 5666: No route to host [20:12:13] RECOVERY - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is OK: OK - NGINX Error Rate is 37% [20:12:45] PROBLEM - mw121 Current Load on mw121 is WARNING: WARNING - load average: 10.52, 10.74, 9.84 [20:13:14] PROBLEM - swiftobject111 SSH on swiftobject111 is CRITICAL: connect to address 2a10:6740::6:203 and port 22: No route to host [20:13:25] PROBLEM - swiftobject111 PowerDNS Recursor on swiftobject111 is CRITICAL: connect to address 2a10:6740::6:203 port 5666: No route to hostconnect to host 2a10:6740::6:203 port 5666: No route to host [20:13:25] PROBLEM - swiftobject111 ferm_active on swiftobject111 is CRITICAL: connect to address 2a10:6740::6:203 port 5666: No route to hostconnect to host 2a10:6740::6:203 port 5666: No route to host [20:13:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 80% [20:13:31] PROBLEM - swiftobject111 NTP time on swiftobject111 is CRITICAL: connect to address 2a10:6740::6:203 port 5666: No route to hostconnect to host 2a10:6740::6:203 port 5666: No route to host [20:13:39] PROBLEM - swiftobject111 conntrack_table_size on swiftobject111 is CRITICAL: connect to address 2a10:6740::6:203 port 5666: No route to hostconnect to host 2a10:6740::6:203 port 5666: No route to host [20:13:39] PROBLEM - ping6 on swiftobject111 is CRITICAL: CRITICAL - Destination Unreachable (2a10:6740::6:203) [20:13:59] PROBLEM - swiftobject112 Current Load on swiftobject112 is CRITICAL: connect to address 2a10:6740::6:204 port 5666: No route to hostconnect to host 2a10:6740::6:204 port 5666: No route to host [20:14:03] PROBLEM - Host swiftobject111 is DOWN: CRITICAL - Destination Unreachable (2a10:6740::6:203) [20:14:03] PROBLEM - swiftobject111 Swift Object Service on swiftobject111 is CRITICAL: connect to address 2a10:6740::6:203 and port 6000: No route to host [20:14:03] PROBLEM - swiftobject111 Puppet on swiftobject111 is CRITICAL: connect to address 2a10:6740::6:203 port 5666: No route to hostconnect to host 2a10:6740::6:203 port 5666: No route to host [20:14:04] PROBLEM - swiftobject112 conntrack_table_size on swiftobject112 is CRITICAL: connect to address 2a10:6740::6:204 port 5666: No route to hostconnect to host 2a10:6740::6:204 port 5666: No route to host [20:14:07] PROBLEM - swiftobject113 SSH on swiftobject113 is CRITICAL: connect to address 2a10:6740::6:205 and port 22: No route to host [20:14:08] PROBLEM - swiftobject112 Disk Space on swiftobject112 is CRITICAL: connect to address 2a10:6740::6:204 port 5666: No route to hostconnect to host 2a10:6740::6:204 port 5666: No route to host [20:14:10] PROBLEM - swiftobject113 PowerDNS Recursor on swiftobject113 is CRITICAL: connect to address 2a10:6740::6:205 port 5666: No route to hostconnect to host 2a10:6740::6:205 port 5666: No route to host [20:14:21] PROBLEM - swiftobject113 ferm_active on swiftobject113 is CRITICAL: connect to address 2a10:6740::6:205 port 5666: No route to hostconnect to host 2a10:6740::6:205 port 5666: No route to host [20:14:21] PROBLEM - swiftobject113 NTP time on swiftobject113 is CRITICAL: connect to address 2a10:6740::6:205 port 5666: No route to hostconnect to host 2a10:6740::6:205 port 5666: No route to host [20:14:28] PROBLEM - swiftobject113 Current Load on swiftobject113 is CRITICAL: connect to address 2a10:6740::6:205 port 5666: No route to hostconnect to host 2a10:6740::6:205 port 5666: No route to host [20:14:29] PROBLEM - Host swiftobject113 is DOWN: CRITICAL - Destination Unreachable (2a10:6740::6:205) [20:14:33] PROBLEM - swiftobject121 Current Load on swiftobject121 is CRITICAL: connect to address 2a10:6740::6:316 port 5666: No route to hostconnect to host 2a10:6740::6:316 port 5666: No route to host [20:14:36] PROBLEM - ping6 on swiftobject112 is CRITICAL: CRITICAL - Destination Unreachable (2a10:6740::6:204) [20:14:43] RECOVERY - mw121 Current Load on mw121 is OK: OK - load average: 8.89, 9.97, 9.66 [20:14:51] PROBLEM - swiftobject112 Swift Object Service on swiftobject112 is CRITICAL: connect to address 2a10:6740::6:204 and port 6000: No route to host [20:14:51] PROBLEM - swiftobject112 NTP time on swiftobject112 is CRITICAL: connect to address 2a10:6740::6:204 port 5666: No route to hostconnect to host 2a10:6740::6:204 port 5666: No route to host [20:15:00] PROBLEM - swiftobject112 APT on swiftobject112 is CRITICAL: connect to address 2a10:6740::6:204 port 5666: No route to hostconnect to host 2a10:6740::6:204 port 5666: No route to host [20:15:01] PROBLEM - swiftobject121 Swift Object Service on swiftobject121 is CRITICAL: connect to address 2a10:6740::6:316 and port 6000: No route to host [20:15:05] PROBLEM - swiftobject112 SSH on swiftobject112 is CRITICAL: connect to address 2a10:6740::6:204 and port 22: No route to host [20:15:10] PROBLEM - swiftobject112 PowerDNS Recursor on swiftobject112 is CRITICAL: connect to address 2a10:6740::6:204 port 5666: No route to hostconnect to host 2a10:6740::6:204 port 5666: No route to host [20:15:14] PROBLEM - swiftobject112 Puppet on swiftobject112 is CRITICAL: connect to address 2a10:6740::6:204 port 5666: No route to hostconnect to host 2a10:6740::6:204 port 5666: No route to host [20:15:14] PROBLEM - swiftobject112 ferm_active on swiftobject112 is CRITICAL: connect to address 2a10:6740::6:204 port 5666: No route to hostconnect to host 2a10:6740::6:204 port 5666: No route to host [20:15:17] PROBLEM - swiftobject121 ferm_active on swiftobject121 is CRITICAL: connect to address 2a10:6740::6:316 port 5666: No route to hostconnect to host 2a10:6740::6:316 port 5666: No route to host [20:15:17] PROBLEM - swiftobject121 conntrack_table_size on swiftobject121 is CRITICAL: connect to address 2a10:6740::6:316 port 5666: No route to hostconnect to host 2a10:6740::6:316 port 5666: No route to host [20:15:18] PROBLEM - Host swiftobject112 is DOWN: CRITICAL - Destination Unreachable (2a10:6740::6:204) [20:15:24] PROBLEM - Host swiftobject122 is DOWN: CRITICAL - Destination Unreachable (2a10:6740::6:317) PROBLEM - swiftobject122 NTP time on swiftobject122 is CRITICAL: connect to address 2a10:6740::6:317 port 5666: No route to hostconnect to host 2a10:6740::6:317 port 5666: No route to host [20:15:24] PROBLEM - swiftobject122 SSH on swiftobject122 is CRITICAL: connect to address 2a10:6740::6:317 and port 22: No route to host [20:15:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 46% [20:15:33] PROBLEM - swiftobject121 SSH on swiftobject121 is CRITICAL: connect to address 2a10:6740::6:316 and port 22: No route to host [20:15:33] PROBLEM - swiftobject121 APT on swiftobject121 is CRITICAL: connect to address 2a10:6740::6:316 port 5666: No route to hostconnect to host 2a10:6740::6:316 port 5666: No route to host [20:15:52] PROBLEM - swiftobject121 PowerDNS Recursor on swiftobject121 is CRITICAL: connect to address 2a10:6740::6:316 port 5666: No route to hostconnect to host 2a10:6740::6:316 port 5666: No route to host [20:15:53] PROBLEM - swiftobject121 Puppet on swiftobject121 is CRITICAL: connect to address 2a10:6740::6:316 port 5666: No route to hostconnect to host 2a10:6740::6:316 port 5666: No route to host [20:15:53] PROBLEM - Host swiftobject121 is DOWN: CRITICAL - Destination Unreachable (2a10:6740::6:316) [20:19:33] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 82% [20:20:03] PROBLEM - swiftproxy111 Puppet on swiftproxy111 is CRITICAL: connect to address 2a10:6740::6:201 port 5666: No route to hostconnect to host 2a10:6740::6:201 port 5666: No route to host [20:20:15] RECOVERY - cp32 NTP time on cp32 is OK: NTP OK: Offset -0.000156134367 secs [20:20:39] PROBLEM - swiftproxy111 NTP time on swiftproxy111 is CRITICAL: connect to address 2a10:6740::6:201 port 5666: No route to hostconnect to host 2a10:6740::6:201 port 5666: No route to host [20:20:43] PROBLEM - swiftproxy111 APT on swiftproxy111 is CRITICAL: connect to address 2a10:6740::6:201 port 5666: No route to hostconnect to host 2a10:6740::6:201 port 5666: No route to host [20:20:43] PROBLEM - swiftproxy111 ferm_active on swiftproxy111 is CRITICAL: connect to address 2a10:6740::6:201 port 5666: No route to hostconnect to host 2a10:6740::6:201 port 5666: No route to host [20:20:48] PROBLEM - swiftproxy111 PowerDNS Recursor on swiftproxy111 is CRITICAL: connect to address 2a10:6740::6:201 port 5666: No route to hostconnect to host 2a10:6740::6:201 port 5666: No route to host [20:20:53] PROBLEM - swiftproxy111 SSH on swiftproxy111 is CRITICAL: connect to address 2a10:6740::6:201 and port 22: No route to host [20:20:53] PROBLEM - swiftproxy111 memcached on swiftproxy111 is CRITICAL: connect to address 2a10:6740::6:201 and port 11211: No route to host [20:20:53] PROBLEM - ping6 on swiftproxy111 is CRITICAL: CRITICAL - Destination Unreachable (2a10:6740::6:201) [20:21:00] PROBLEM - swiftproxy111 Disk Space on swiftproxy111 is CRITICAL: connect to address 2a10:6740::6:201 port 5666: No route to hostconnect to host 2a10:6740::6:201 port 5666: No route to host [20:21:00] PROBLEM - cp32 Current Load on cp32 is CRITICAL: CRITICAL - load average: 8.44, 6.61, 4.36 [20:21:38] PROBLEM - Host swiftproxy111 is DOWN: CRITICAL - Destination Unreachable (2a10:6740::6:201) [20:23:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 59% [20:24:14] PROBLEM - cp33 Current Load on cp33 is CRITICAL: CRITICAL - load average: 4.35, 3.82, 2.58 [20:25:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 73% [20:26:14] PROBLEM - cp33 Current Load on cp33 is WARNING: WARNING - load average: 3.05, 3.66, 2.68 [20:27:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 44% [20:28:14] RECOVERY - cp33 Current Load on cp33 is OK: OK - load average: 2.13, 3.08, 2.58 [20:29:18] PROBLEM - knowledgebase.clientmanager.co.za - reverse DNS on sslhost is WARNING: Timeout: The DNS operation timed out after 5.403587341308594 seconds [20:29:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 65% [20:29:46] PROBLEM - cp23 Current Load on cp23 is CRITICAL: CRITICAL - load average: 5.50, 3.60, 2.67 [20:31:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 52% [20:31:46] PROBLEM - cp23 Current Load on cp23 is WARNING: WARNING - load average: 3.62, 3.40, 2.70 [20:33:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 72% [20:33:46] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 0.96, 2.45, 2.43 [20:35:26] PROBLEM - gs.sidem.wiki - reverse DNS on sslhost is CRITICAL: rDNS CRITICAL - gs.sidem.wiki All nameservers failed to answer the query. [20:35:39] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 43% [20:37:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 58% [20:37:40] RECOVERY - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is OK: OK - NGINX Error Rate is 34% [20:43:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 63% [20:43:46] PROBLEM - cp23 Current Load on cp23 is WARNING: WARNING - load average: 3.14, 3.52, 2.84 [20:44:32] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 46% [20:44:43] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 55% [20:45:47] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 1.47, 2.78, 2.65 [20:46:37] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 65% [20:47:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 49% [20:48:32] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 42% [20:50:26] RECOVERY - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is OK: OK - NGINX Error Rate is 36% [20:51:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 70% [20:54:15] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 50% [20:54:19] RECOVERY - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is OK: OK - NGINX Error Rate is 27% [20:55:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 59% [20:56:10] RECOVERY - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is OK: OK - NGINX Error Rate is 33% [20:59:46] PROBLEM - cp23 Current Load on cp23 is CRITICAL: CRITICAL - load average: 7.88, 4.39, 2.95 [21:00:00] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 47% [21:01:46] PROBLEM - cp23 Current Load on cp23 is WARNING: WARNING - load average: 2.98, 3.62, 2.84 [21:03:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 75% [21:03:46] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 1.13, 2.74, 2.61 [21:03:48] RECOVERY - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is OK: OK - NGINX Error Rate is 37% [21:04:13] RECOVERY - gs.sidem.wiki - reverse DNS on sslhost is OK: SSL OK - gs.sidem.wiki reverse DNS resolves to cp22.miraheze.org - CNAME OK [21:07:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 51% [21:07:57] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 49% [21:09:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 74% [21:11:51] RECOVERY - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is OK: OK - NGINX Error Rate is 35% [21:12:23] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 40% [21:13:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 47% [21:14:18] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 62% [21:15:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 64% [21:16:12] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 43% [21:16:26] PROBLEM - cp32 Current Load on cp32 is WARNING: WARNING - load average: 2.02, 2.72, 3.74 [21:18:06] RECOVERY - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is OK: OK - NGINX Error Rate is 26% [21:18:26] PROBLEM - cp32 Current Load on cp32 is CRITICAL: CRITICAL - load average: 4.34, 3.08, 3.72 [21:18:26] PROBLEM - cloud11 Puppet on cloud11 is CRITICAL: connect to address 2a10:6740::6:200 port 5666: No route to hostconnect to host 2a10:6740::6:200 port 5666: No route to host [21:18:30] PROBLEM - cloud11 IPMI Sensors on cloud11 is CRITICAL: CHECK_NRPE STATE CRITICAL: Socket timeout after 60 seconds. [21:18:47] PROBLEM - cloud11 SMART on cloud11 is CRITICAL: connect to address 2a10:6740::6:200 port 5666: No route to hostconnect to host 2a10:6740::6:200 port 5666: No route to host [21:20:15] PROBLEM - cloud11 Current Load on cloud11 is CRITICAL: connect to address 2a10:6740::6:200 port 5666: No route to hostconnect to host 2a10:6740::6:200 port 5666: No route to host [21:20:15] PROBLEM - cloud11 APT on cloud11 is CRITICAL: connect to address 2a10:6740::6:200 port 5666: No route to hostconnect to host 2a10:6740::6:200 port 5666: No route to host [21:20:15] PROBLEM - cloud11 Disk Space on cloud11 is CRITICAL: connect to address 2a10:6740::6:200 port 5666: No route to hostconnect to host 2a10:6740::6:200 port 5666: No route to host [21:20:21] PROBLEM - cloud11 PowerDNS Recursor on cloud11 is CRITICAL: connect to address 2a10:6740::6:200 port 5666: No route to hostconnect to host 2a10:6740::6:200 port 5666: No route to host [21:20:27] PROBLEM - cloud11 conntrack_table_size on cloud11 is CRITICAL: connect to address 2a10:6740::6:200 port 5666: No route to hostconnect to host 2a10:6740::6:200 port 5666: No route to host [21:20:40] PROBLEM - cloud11 ferm_active on cloud11 is CRITICAL: connect to address 2a10:6740::6:200 port 5666: No route to hostconnect to host 2a10:6740::6:200 port 5666: No route to host [21:20:40] PROBLEM - cloud11 NTP time on cloud11 is CRITICAL: connect to address 2a10:6740::6:200 port 5666: No route to hostconnect to host 2a10:6740::6:200 port 5666: No route to host [21:20:59] PROBLEM - cloud11 SSH on cloud11 is CRITICAL: connect to address 2a10:6740::6:200 and port 22: No route to host [21:21:21] PROBLEM - Host cloud11 is DOWN: CRITICAL - Destination Unreachable (2a10:6740::6:200) [21:21:56] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 55% [21:23:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 54% [21:23:50] RECOVERY - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is OK: OK - NGINX Error Rate is 37% [21:27:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 49% [21:27:43] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 41% [21:29:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 64% [21:29:46] PROBLEM - cp23 Current Load on cp23 is CRITICAL: CRITICAL - load average: 6.33, 4.77, 2.89 [21:31:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 57% [21:31:47] PROBLEM - cp23 Current Load on cp23 is WARNING: WARNING - load average: 2.13, 3.67, 2.71 [21:33:27] RECOVERY - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is OK: OK - NGINX Error Rate is 35% [21:33:46] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 1.19, 2.81, 2.51 [21:35:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 86% [21:35:40] RECOVERY - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is OK: OK - NGINX Error Rate is 32% [21:37:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 59% [21:41:33] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 77% [21:47:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 51% [21:49:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 65% [21:53:46] PROBLEM - mw121 Current Load on mw121 is CRITICAL: CRITICAL - load average: 13.68, 11.56, 9.51 [21:54:26] PROBLEM - cp32 Current Load on cp32 is WARNING: WARNING - load average: 1.28, 2.22, 3.83 [21:55:45] PROBLEM - mw121 Current Load on mw121 is WARNING: WARNING - load average: 9.50, 10.71, 9.44 [21:57:43] PROBLEM - mw121 Current Load on mw121 is CRITICAL: CRITICAL - load average: 12.41, 11.23, 9.77 [21:57:48] PROBLEM - cp23 Current Load on cp23 is CRITICAL: CRITICAL - load average: 8.10, 4.29, 3.01 [21:58:02] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 46% [21:59:17] RECOVERY - knowledgebase.clientmanager.co.za - reverse DNS on sslhost is OK: SSL OK - knowledgebase.clientmanager.co.za reverse DNS resolves to cp22.miraheze.org - CNAME OK [21:59:42] PROBLEM - mw121 Current Load on mw121 is WARNING: WARNING - load average: 10.10, 10.90, 9.83 [21:59:59] RECOVERY - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is OK: OK - NGINX Error Rate is 39% [22:00:26] RECOVERY - cp32 Current Load on cp32 is OK: OK - load average: 3.11, 2.41, 3.38 [22:00:36] [02WikiDiscover] 07Universal-Omega opened pull request 03#92: Add a config whether or not to list private wikis - 13https://github.com/miraheze/WikiDiscover/pull/92 [22:01:41] RECOVERY - mw121 Current Load on mw121 is OK: OK - load average: 7.78, 9.76, 9.54 [22:01:51] [02WikiDiscover] 07Universal-Omega synchronize pull request 03#92: Add a config whether or not to list private wikis - 13https://github.com/miraheze/WikiDiscover/pull/92 [22:05:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 56% [22:05:41] miraheze/WikiDiscover - Universal-Omega the build passed. [22:05:46] PROBLEM - cp33 Current Load on cp33 is WARNING: WARNING - load average: 3.74, 2.72, 2.26 [22:05:46] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 1.70, 3.33, 3.39 [22:07:42] RECOVERY - cp33 Current Load on cp33 is OK: OK - load average: 2.08, 2.49, 2.24 [22:07:43] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 44% [22:09:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 71% [22:11:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 59% [22:11:40] RECOVERY - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is OK: OK - NGINX Error Rate is 25% [22:13:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 73% [22:14:48] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 67% [22:15:46] PROBLEM - cp23 Current Load on cp23 is CRITICAL: CRITICAL - load average: 3.59, 4.55, 3.88 [22:16:42] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 50% [22:19:21] PROBLEM - cp33 Current Load on cp33 is CRITICAL: CRITICAL - load average: 5.13, 3.60, 2.68 [22:20:26] PROBLEM - cp23 APT on cp23 is CRITICAL: CHECK_NRPE STATE CRITICAL: Socket timeout after 60 seconds. [22:20:32] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 47% [22:21:17] PROBLEM - cp33 Current Load on cp33 is WARNING: WARNING - load average: 3.83, 3.69, 2.82 [22:21:47] PROBLEM - cp23 Current Load on cp23 is WARNING: WARNING - load average: 2.45, 3.58, 3.74 [22:22:26] RECOVERY - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is OK: OK - NGINX Error Rate is 33% [22:22:30] RECOVERY - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is OK: OK - NGINX Error Rate is 36% [22:22:39] PROBLEM - en.religiononfire.mar.in.ua - reverse DNS on sslhost is CRITICAL: rDNS CRITICAL - en.religiononfire.mar.in.ua All nameservers failed to answer the query. [22:23:13] PROBLEM - cp33 Current Load on cp33 is CRITICAL: CRITICAL - load average: 5.12, 4.39, 3.19 [22:23:17] RECOVERY - cp23 APT on cp23 is OK: APT OK: 1 packages available for upgrade (0 critical updates). [22:25:09] PROBLEM - cp33 Current Load on cp33 is WARNING: WARNING - load average: 2.83, 3.68, 3.06 [22:25:46] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 2.54, 2.80, 3.39 [22:26:15] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 46% [22:27:05] PROBLEM - cp33 Current Load on cp33 is CRITICAL: CRITICAL - load average: 7.44, 4.29, 3.31 [22:29:01] RECOVERY - cp33 Current Load on cp33 is OK: OK - load average: 2.26, 3.35, 3.08 [22:29:34] PROBLEM - cp32 Current Load on cp32 is CRITICAL: CRITICAL - load average: 3.95, 4.16, 3.20 [22:29:46] PROBLEM - cp23 Current Load on cp23 is CRITICAL: CRITICAL - load average: 5.33, 4.88, 4.08 [22:31:25] PROBLEM - knowledgebase.clientmanager.co.za - reverse DNS on sslhost is WARNING: Timeout: The DNS operation timed out after 5.405061960220337 seconds [22:31:30] RECOVERY - cp32 Current Load on cp32 is OK: OK - load average: 1.53, 3.31, 3.01 [22:31:46] PROBLEM - cp23 Current Load on cp23 is WARNING: WARNING - load average: 1.33, 3.52, 3.68 [22:32:01] RECOVERY - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is OK: OK - NGINX Error Rate is 39% [22:33:46] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 0.80, 2.58, 3.31 [22:37:07] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 64% [22:37:18] PROBLEM - cp32 Current Load on cp32 is CRITICAL: CRITICAL - load average: 5.94, 5.00, 3.75 [22:37:44] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 56% [22:39:04] RECOVERY - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is OK: OK - NGINX Error Rate is 39% [22:41:09] PROBLEM - cp32 Current Load on cp32 is WARNING: WARNING - load average: 1.41, 3.73, 3.57 [22:45:01] RECOVERY - cp32 Current Load on cp32 is OK: OK - load average: 1.65, 2.97, 3.35 [22:45:22] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 60% [22:46:48] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 51% [22:47:16] RECOVERY - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is OK: OK - NGINX Error Rate is 36% [22:48:45] RECOVERY - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is OK: OK - NGINX Error Rate is 29% [22:52:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 71% [22:57:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 53% [22:58:31] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 42% [22:59:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 70% [23:00:38] RECOVERY - knowledgebase.clientmanager.co.za - reverse DNS on sslhost is OK: SSL OK - knowledgebase.clientmanager.co.za reverse DNS resolves to cp23.miraheze.org - CNAME OK [23:02:18] PROBLEM - es141 Current Load on es141 is CRITICAL: CRITICAL - load average: 4.74, 3.01, 1.89 [23:06:19] RECOVERY - es141 Current Load on es141 is OK: OK - load average: 1.84, 2.92, 2.17 [23:08:18] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 63% [23:09:19] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 42% [23:10:15] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 42% [23:11:14] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 70% [23:12:12] RECOVERY - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is OK: OK - NGINX Error Rate is 35% [23:13:08] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 54% [23:14:15] PROBLEM - cp32 NTP time on cp32 is WARNING: NTP WARNING: Offset 0.1946879625 secs [23:15:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 68% [23:16:04] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 41% [23:16:14] PROBLEM - cp33 Current Load on cp33 is WARNING: WARNING - load average: 3.41, 3.09, 2.77 [23:17:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 57% [23:18:14] RECOVERY - cp33 Current Load on cp33 is OK: OK - load average: 2.46, 2.76, 2.68 [23:19:59] RECOVERY - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is OK: OK - NGINX Error Rate is 36% [23:20:19] RECOVERY - en.religiononfire.mar.in.ua - reverse DNS on sslhost is OK: SSL OK - en.religiononfire.mar.in.ua reverse DNS resolves to cp23.miraheze.org - CNAME OK [23:21:07] RECOVERY - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is OK: OK - NGINX Error Rate is 35% [23:21:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 58% [23:23:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 69% [23:23:51] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 45% [23:25:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 48% [23:25:48] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 68% [23:29:09] PROBLEM - cp33 Current Load on cp33 is CRITICAL: CRITICAL - load average: 4.53, 3.76, 3.24 [23:29:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 54% [23:29:42] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 46% [23:31:05] RECOVERY - cp33 Current Load on cp33 is OK: OK - load average: 2.75, 3.28, 3.12 [23:31:31] PROBLEM - cp23 Current Load on cp23 is CRITICAL: CRITICAL - load average: 2.22, 4.04, 2.88 [23:31:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 74% [23:33:29] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 0.76, 2.92, 2.61 [23:33:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 47% [23:35:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 65% [23:37:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is CRITICAL: CRITICAL - NGINX Error Rate is 61% [23:39:07] PROBLEM - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is WARNING: WARNING - NGINX Error Rate is 56% [23:39:39] RECOVERY - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is OK: OK - NGINX Error Rate is 36% [23:40:26] PROBLEM - cp32 Current Load on cp32 is CRITICAL: CRITICAL - load average: 4.32, 3.57, 2.29 [23:41:20] PROBLEM - cp23 Current Load on cp23 is CRITICAL: CRITICAL - load average: 6.24, 3.69, 2.82 [23:41:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 58% [23:42:26] RECOVERY - cp32 Current Load on cp32 is OK: OK - load average: 1.94, 3.04, 2.26 [23:43:18] RECOVERY - cp23 Current Load on cp23 is OK: OK - load average: 1.35, 2.71, 2.57 [23:43:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 68% [23:43:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 47% [23:45:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 55% [23:46:15] RECOVERY - cp32 NTP time on cp32 is OK: NTP OK: Offset 0.07370448112 secs [23:47:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 83% [23:49:07] RECOVERY - cp23 HTTP 4xx/5xx ERROR Rate on cp23 is OK: OK - NGINX Error Rate is 34% [23:49:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 71% [23:51:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 47% [23:55:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is WARNING: WARNING - NGINX Error Rate is 48% [23:57:31] PROBLEM - cp32 HTTP 4xx/5xx ERROR Rate on cp32 is CRITICAL: CRITICAL - NGINX Error Rate is 73% [23:57:39] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is CRITICAL: CRITICAL - NGINX Error Rate is 60% [23:58:16] PROBLEM - cp33 Current Load on cp33 is WARNING: WARNING - load average: 2.93, 3.62, 3.36 [23:58:26] PROBLEM - cp32 Current Load on cp32 is WARNING: WARNING - load average: 3.97, 3.17, 2.69 [23:59:40] PROBLEM - cp22 HTTP 4xx/5xx ERROR Rate on cp22 is WARNING: WARNING - NGINX Error Rate is 56%