[08:12:31] hello folks [08:12:49] if you are ok I can proceed with the reimage of the first eqiad VM [08:12:52] (kubernetes1005) [08:20:55] <_joe_> elukey: keep an eye on kask/sessionstore's metrics [08:20:59] <_joe_> but yes go on [08:27:39] _joe_ ack namely https://grafana.wikimedia.org/d/000001590/sessionstore?orgId=1 right? [08:28:09] <_joe_> yes [08:28:28] ack, so I am draining 1005, will wait a little before proceeding [08:34:34] nothing weird in metrics, proceeding [09:03:07] 10serviceops, 10Release-Engineering-Team, 10Scap: Deploy Scap version 4.5.0 - https://phabricator.wikimedia.org/T304134 (10jnuche) [09:05:46] kubernetes1005 ready to be uncordoned, I'd ask for a quick review before doing it though [09:20:07] uncordoned, will stop any reimage since it is enough for a Friday, will resume next week with the other 3 vms [09:28:53] jayme: o/ I am wondering one thing - should we drain kubernetes200[1-4] (one at the time) as prep step for decom? [09:29:04] (now that codfw is on bullseye) [09:29:15] (and expanded) [09:33:04] <_joe_> elukey: I would assume so, but again not on a friday :) [09:33:28] draining codfw nodes? :D [09:34:04] (we have already disabled scheduling a while ago) [09:35:13] ack then, there are still pods running on those nodes, we can wait Monday [09:35:34] yeah, there haven't been much deploy movement [09:38:11] 10serviceops, 10Product-Infrastructure-Team-Backlog, 10SRE, 10Maps (Geoshapes), and 2 others: New Service Request geoshapes - https://phabricator.wikimedia.org/T274388 (10MSantos) @akosiaris the initial geoshapes deployment-charts is created and ready to move forward: https://gerrit.wikimedia.org/r/c/opera... [09:38:27] 10serviceops, 10Product-Infrastructure-Team-Backlog, 10SRE, 10Maps (Geoshapes), and 2 others: New Service Request geoshapes - https://phabricator.wikimedia.org/T274388 (10MSantos) [09:51:48] 10serviceops, 10Prod-Kubernetes, 10Kubernetes, 10Patch-For-Review: setup/install kubernetes10[18-22] - https://phabricator.wikimedia.org/T293728 (10akosiaris) 05Open→03Resolved [09:51:50] 10serviceops, 10Prod-Kubernetes, 10Kubernetes, 10Machine-Learning-Team (Active Tasks), 10Patch-For-Review: Move kubernetes workers to bullseye and docker to overlayfs - https://phabricator.wikimedia.org/T300744 (10akosiaris) [09:51:54] 10serviceops, 10DC-Ops, 10SRE, 10ops-eqiad: Q1:(Need By: TBD) rack/setup/install kubernetes10[18-21] - https://phabricator.wikimedia.org/T290202 (10akosiaris) [09:52:17] 10serviceops, 10decommission-hardware: decommission kubernetes100[1-4] - https://phabricator.wikimedia.org/T303044 (10akosiaris) 05Stalled→03Open [09:52:21] 10serviceops, 10Prod-Kubernetes, 10Kubernetes, 10Patch-For-Review: setup/install kubernetes10[18-22] - https://phabricator.wikimedia.org/T293728 (10akosiaris) [09:52:24] 10serviceops, 10Prod-Kubernetes, 10Kubernetes, 10Machine-Learning-Team (Active Tasks), 10Patch-For-Review: Move kubernetes workers to bullseye and docker to overlayfs - https://phabricator.wikimedia.org/T300744 (10akosiaris) [10:06:34] I have a question for node decom [10:06:54] do we need to kubectl drain + delete or is it sufficient drain? [10:07:47] https://wikitech.wikimedia.org/wiki/Kubernetes/Clusters/Add_or_remove_nodes#Removing_a_node says drain, decom, delete [10:12:31] ahh I didn't find it, must have looked in the wrong place, thanks [10:13:41] at least that worked once in staging :-p [10:50:10] 10serviceops, 10decommission-hardware: decommission kubernetes200[1-4] - https://phabricator.wikimedia.org/T303045 (10akosiaris) 05Stalled→03Open [10:50:13] 10serviceops, 10Prod-Kubernetes, 10Patch-For-Review: setup/install kubernetes20[1(89)|2(012)] - https://phabricator.wikimedia.org/T302208 (10akosiaris) [10:50:16] 10serviceops, 10Prod-Kubernetes, 10Kubernetes, 10Machine-Learning-Team (Active Tasks), 10Patch-For-Review: Move kubernetes workers to bullseye and docker to overlayfs - https://phabricator.wikimedia.org/T300744 (10akosiaris) [11:29:28] 10serviceops, 10DC-Ops, 10SRE, 10ops-eqiad: Q3:(Need By: TBD) rack/setup/install parse100[01-24] - https://phabricator.wikimedia.org/T299573 (10akosiaris) >>! In T299573#7785936, @Cmjohnson wrote: > @akosiaris I can spread the other 3 between B and D if that works better for you? Yeah that sounds fine. T... [11:35:14] 10serviceops, 10DC-Ops, 10SRE, 10ops-eqiad: Q3:(Need By: TBD) rack/setup/install conf100[789] - https://phabricator.wikimedia.org/T301272 (10akosiaris) >>! In T301272#7782148, @cmooney wrote: > FYI I don't believe there is any reason E/F would be ruled out for these, if space/power is tight in the existing...