[02:14:12] !log admin downtiming SMART alerts on dumps server labstore1007 for the weekend because it has been flapping T281045 [02:14:13] T281045: labstore1007 crashed after storage controller errors--replace disk? - https://phabricator.wikimedia.org/T281045 [02:14:15] Logged the message at https://wikitech.wikimedia.org/wiki/Nova_Resource:Admin/SAL [02:14:15] T281045: labstore1007 crashed after storage controller errors--replace disk? - https://phabricator.wikimedia.org/T281045 [02:23:24] Anyone knows what's up with FontCDN on wmflabs? CSS files (https://tools-static.wmflabs.org/fontcdn/css?family=Roboto) are being served properly but font files (e.g. https://tools-static.wmflabs.org/fontcdn/s/roboto/v27/KFOmCnqEu92Fr1Mu72xKOzY.woff2) are 502'ing. [02:24:21] The mess can be visualized on https://fontcdn.toolforge.org/ where everything is just browser fallback font (for me, at least) [02:39:00] chlod: I'll see if I can figure anything out quickly. We may need to open a bug and look more tomorrow (/me is getting tired) [02:39:24] bd808: Got it. Thanks! [02:42:47] "no live upstreams while connecting to upstream" in the nginx error log. Now to check what it's really doing [02:51:54] !log tools Restarted nginx on tools-static-14 to see if that clears up the fontcdn 502 errors [02:51:58] Logged the message at https://wikitech.wikimedia.org/wiki/Nova_Resource:Tools/SAL [02:52:29] chlod: turning it off and on seemed to work. It's like I'm a pro SRE or something :) [02:53:02] thank you so much! :) [04:33:31] Logged the message at https://wikitech.wikimedia.org/wiki/Nova_Resource:Tools.bridgebot/SAL [04:34:56] I wonder what we need to change to make the wm-bot on Libera.Chat relay the dologmsg notices instead of the Freenode bot? [05:16:52] Logged the message at https://wikitech.wikimedia.org/wiki/Nova_Resource:Tools.bridgebot/SAL [14:32:17] !log tools manually remove valhallasw yuvipanda admin certificates and from configmap and restart maintain-kubeusers pod T282725 [14:32:20] Logged the message at https://wikitech.wikimedia.org/wiki/Nova_Resource:Tools/SAL [14:32:21] T282725: maintain-kubeusers: Remove admin access from users removed from tools.admin - https://phabricator.wikimedia.org/T282725 [14:47:31] !log tools manually remove jeh admin certificates and from maintain-kubeusers configmap T282725 [14:47:35] Logged the message at https://wikitech.wikimedia.org/wiki/Nova_Resource:Tools/SAL [14:47:35] T282725: maintain-kubeusers: Remove admin access from users removed from tools.admin - https://phabricator.wikimedia.org/T282725 [16:50:05] Hi, I'm kinda new to the whole Toolforge and Kubernetes stuff. I'd like to setup a Python 3.8 bot to run on toolforge. I already set up the tool itself (https://toolsadmin.wikimedia.org/tools/id/lingua-libre-bot) and I'm ssh'd into the server with the proper tool account. I've been reading through the docs I could find (most notably: [16:50:05] https://wikitech.wikimedia.org/wiki/Help:Toolforge/Kubernetes#Kubernetes_cronjobs), but before I start fiddling with Kubernetes, I thought I'd try to make the bot works at least once with the shell. Which I can't, because it requires Python 3.8 and python3 --version returns 3.5.3 [16:52:51] Poslovitch: There's Python 3.7 on Kubernetes, I'm afraid there is no 3.8 or newer since we usually have what released Debian versions have (Debian 9/Stretch is running on the bastions and the grid with 3.5, Debian 10/Buster has 3.7, and Debian 11/Bullseye which is very close to release but not out yet will have 3.9) [16:54:07] Okay. 3.7 should be fine. Now, how do I get myself a kubernetes installed? The page about Kubernetes seems to be about webservices, which my bot is not, I believe [16:59:11] Kubernetes is a platform for running containers across multiple servers, it's already installed but you just need to tell it what to do (which can be sometimes difficult due to all of its complex features). We have easy-to-use tooling for webservices, but similar for continuous jobs and cronjobs are still in development. You can talk to Kubernetes [16:59:11] directly using the `kubectl` command in the meantime. There's are a few examples on that page, and if you have any specific questions or issues I and others here can help with them. [17:01:06] So, I can create a pod with kubectl and make it use this `image: docker-registry.tools.wmflabs.org/toolforge-python37-sssd-base:latest` as the docker image in which my code will run? Does that mean the bot will run on Python 3.7? [17:03:38] Basically yes. You should create a Deployment object instead of creating a Pod directly, so it makes sure your tool/pod is always running even if that specific pod gets deleted, for example if we have to restart a node for updates. [17:04:28] Okay. Sounds a bit tough to do, but hopefully I'll figure out a way. Thanks for the hints! [17:05:53] bd808: I know what to fix for the wm-bot !log relay, I will get to it in a bit [17:11:49] legoktm: awesome. thank you [18:57:40] I trust: MacFan4000!.*@user/macfan4000 (2admin), .*@user/majavah (2admin), .*@user/bd808 (2admin), [18:57:40] @trusted [18:57:49] want to be trusted too? [18:58:17] Successfully added .*@user/legoktm [18:58:17] @trustadd .*@user/legoktm admin [18:58:19] ty [19:12:55] * legoktm pokes wm-bot [19:12:57] Relay was enabled [19:12:57] @relay-on [19:12:59] test [19:13:06] yesss [19:16:13] I submitted a patch to change the default port, I guess I could just set the hiera [19:18:12] merged, it'll take a bit for puppet to roll it out [19:32:07] puppet is really slow... :/ [19:34:32] !log tools.admin dologmsg now logs to LiberaChat relay! [19:34:34] Logged the message at https://wikitech.wikimedia.org/wiki/Nova_Resource:Tools.admin/SAL [19:34:49] there we go ^.^ [20:03:37] i'm trying to mount a cinder volume but it just doesn't work. could be a volume deleted not long ago be a problem here? [20:03:58] +with the same name [20:18:21] no dice, wth [20:21:23] !log tools.integraality Deploy b52079f [20:21:25] Logged the message at https://wikitech.wikimedia.org/wiki/Nova_Resource:Tools.integraality/SAL [20:24:27] !log tools.integraality Deploy 2f39621 [20:24:28] Logged the message at https://wikitech.wikimedia.org/wiki/Nova_Resource:Tools.integraality/SAL [20:33:38] gifti: hmmm... I haven't done much debugging with cinder volumes yet. andrewbogott or bstorm are either of you about? [20:35:17] gifti: can you be more specific than "doesn't work"? how are you doing that? any errors? [20:40:06] majavah: i had a 100G volume attached (it was mounted properly), i detached it, deleted it, created a new 80G one, attached it, forgot to edit fstab, ran wmcs-prepare-cinder-volume, edited fstab, ran the command again, it was successful, no errors, but it wasn't mounted, i tried it one more time, this time properly, same results, then i tried it with another volume name, same results again [20:47:26] i can mount it on another instance, so i think i broke something on the instance i want to attach it on [20:47:43] what does /etc/fstab look like? [20:48:42] UUID=5eb86d23-f988-42e6-affd-9cc9409a40cb / ext4 defaults 11 [20:48:42] /dev/vd/second-local-disk /srv ext4 defaults 0 0 [20:48:42] UUID=308ee617-48e2-4ac9-8a38-a3821cb61793 /data ext4 nofail,x-systemd.device-timeout=2s 0 2 [20:49:26] uh [20:49:55] does that vm also have a lvm-based extra disk? [20:50:01] yes [20:50:08] hmm [20:50:18] i wanted to migrate some data off of it [20:50:42] it worked two months ago [20:57:39] !log tools.integraality Deploy d170ea1 [20:57:41] Logged the message at https://wikitech.wikimedia.org/wiki/Nova_Resource:Tools.integraality/SAL [20:57:59] !log tools.integraality Deploy 67c42f43 [20:58:00] Logged the message at https://wikitech.wikimedia.org/wiki/Nova_Resource:Tools.integraality/SAL [20:58:07] !log tools.integraality Deploy 258aae9a [20:58:08] Logged the message at https://wikitech.wikimedia.org/wiki/Nova_Resource:Tools.integraality/SAL [20:59:11] gifti: sorry I missed the ping before, I can have a look if you're still here [20:59:52] can you tell me the fqdn of the affected instance? [21:02:18] andrewbogott: it's taxonbota-b.dwl.eqiad1.wikimedia.cloud [21:04:46] gifti: ok, I'm going to start by removing this line from fstab: 'UUID=308ee617-48e2-4ac9-8a38-a3821cb61793 /data ext4 nofail,x-systemd.device-timeout=2s 0 2' that seems like a relic from a previous cinder mount, does that sound right? [21:05:46] yes, that was my last failed attempt [21:06:18] ok [21:06:30] I'm looking at 'lsblk' and I don't see the volume presenting to the OS [21:06:41] it's currently somewhere else [21:06:47] should i reattach it? [21:07:10] this is the 'taxonbot' volume? Horizon shows it as attached [21:07:32] 'attached to: /dev/vdb on taxonbot-b' [21:07:40] it's taxonbota-test [21:08:59] horizon shows that one as '/dev/sdb on taxonbota-c' [21:09:18] that's right [21:09:37] ok, you've lost me then [21:09:54] i want to attach it to taxonbota-b [21:10:51] which i have done now [21:10:57] stop please [21:11:01] ok [21:11:22] do you not see in Horizon that the 'taxonbot' volume is already attached to taxonbot-b? [21:11:39] i do [21:11:46] Is it your intent to have both volumes attached at once? [21:11:46] how is that relevant? [21:11:54] no [21:12:20] it's all named pretty similarly [21:12:40] i want taxonbot on taxonbot-b and taxonbota on taxonbota-b [21:12:44] oh shit taxonbota-b and taxonbot-b are two different things [21:12:46] * andrewbogott sights [21:12:56] ok, so now I see the volume in lsblk, that seems good [21:13:52] go ahead and try running wmcs-prepare-cinder-volume and let's see what it does :) [21:14:37] it asks me where i want to mount it [21:14:48] that seems good [21:15:16] i will say /data [21:15:49] ok [21:15:55] it's ready to prepare and mount and i continue [21:16:12] yep, go ahead [21:16:23] it says it mounts and updates fstab [21:16:27] and is done [21:17:03] there is no mountpoint in lsblk, no corresponding output in mount and nothing in /data [21:17:07] hm, fstab looks right but it didn't mount [21:18:28] yeah [21:18:38] and 'mount /data' returns success but doesn't do anything [21:19:39] did wmcs-prepare-cinder-volume say it was going to format things? [21:20:05] no [21:20:10] it was already formatted [21:20:19] !log tools.lingua-libre-bot Mom, am I really on TV? [21:20:20] Logged the message at https://wikitech.wikimedia.org/wiki/Nova_Resource:Tools.lingua-libre-bot/SAL [21:20:34] ok [21:21:41] I tried mounting it as /waffle instead of /data and it worked [21:21:50] huh [21:21:51] !log tools.lingua-libre-bot Mom, am I really on TV? [21:21:51] Logged the message at https://wikitech.wikimedia.org/wiki/Nova_Resource:Tools.lingua-libre-bot/SAL [21:21:55] so... that's interesting! No idea why it cares unless there's some residual state from something being mounted there before [21:22:20] Have you tried rebooting at any point in all this? (Is this a server that's safe to reboot?) [21:22:42] i did not and it's not safe to reboot it at all [21:23:08] i would have to coordinate with doctaxon for that [21:23:17] ok [21:23:27] then... I guess we treat /data as cursed for now :( [21:23:32] if it works with another path, i'm happy [21:23:45] I'm sorry I don't have an actual explanation [21:23:52] that's ok [21:23:59] thank you for your idea [21:24:31] the server will be scrapped soon hopefully [21:38:03] when I was mounting stuff I had to fiddle with systemd because it wanted to control the mounts [21:38:08] I don't remember what I actually did [21:38:25] mmmm systemd [21:38:29] !log tools.translate-link deployed d829336d71 (“show” mode and other improvements) [21:38:31] Logged the message at https://wikitech.wikimedia.org/wiki/Nova_Resource:Tools.translate-link/SAL [22:04:33] !log tools.ranker pulled efa1cecac0 (README update, no webservice restart) [22:04:35] Logged the message at https://wikitech.wikimedia.org/wiki/Nova_Resource:Tools.ranker/SAL [22:05:27] !log tools.translate-link pulled 11922d2df5 (README update, no webservice restart) [22:05:28] Logged the message at https://wikitech.wikimedia.org/wiki/Nova_Resource:Tools.translate-link/SAL [22:35:10] chicocvenancio: do you think I can get rights to PAWS again if I wanted? [22:35:14] (not that I'm asking for it now...) [22:35:28] but sort of testing willingness of the cloud team :) [23:28:59] bd808: what do you think I need to do to request more rights for the PAWS deployment? [23:30:59] yuvipanda: file a phabricator task and have no-one with similae rights oppose within a week, basically [23:31:02] https://wikitech.wikimedia.org/wiki/Help:Access_policies [23:31:11] makes sense [23:31:13] yuvipanda: open a phab task, tag it with #wmcs-kanban, and let me know. :) [23:31:43] this is a part of IRC you can't do anywhere else [23:31:51] god-damn-sucker: :) [23:32:14] god-damn-sucker: reminded me of https://bash.toolforge.org/quip/AVQ_K14_1EdYBLP_7qWd [23:33:36] hahahahahahahah [23:37:09] bd808: https://phabricator.wikimedia.org/T283443 [23:39:39] no promises on actually doing anything tho [23:53:23] and hopefully https://github.com/toolforge/paws/pull/70 will setup automatic build + push of the user image, making chicocvenancio's life eaiser :) [23:56:26] yuvipanda: if you can trick^Wconvince madhu to come around too I'll give you root everywhere. ;) [23:56:39] * bd808 misses the "old" gang [23:56:41] bd808: I'm 100% convinced that isn't going to happen :) [23:57:18] she's engineering management at hashicorp now, but really she wants to work on making clothes for a living eventually... [23:58:46] clothes are cool. bootstrapping physical manufacturing sounds meh, but I'm sure madhu can do it if she decides to do it