[06:08:59] ryankemper: thanks for the merge! [06:15:43] (good morning :) [06:16:20] wdqs1013 seems to have some issue with blazegraph, afaics it should be safe to depool/restart but lemme know if there is anything ongoing [06:45:33] dcausse: it seems we are down to wdqs1007, wdqs1012 and wdqs1013 [06:48:06] do we have any we can proceed in CEST without MrG? [06:53:06] elukey: nothing going on with the wdqs1013 atn\ [06:53:12] s/atn/atm [07:00:54] zpapierski: o/ so ok to depool/restart blazegraph in there? [07:03:15] sure [07:04:11] and thanks [07:32:22] zpapierski: I'm afraid we'll have to wait for Ryan to continue [07:32:50] elukey: thanks for the patch over the week-end! :) [07:32:54] and sorry for the noise [07:35:29] dcausse: np! Happy to help :) I am restart bz on wdqs1013 [07:35:39] thanks! [07:35:53] (all done :) [07:36:03] thx! [07:36:32] anyway - 3 servers left, we - and by we I mean Ryan - should probably be able to finish today [07:36:46] sounds weird [07:37:19] yes should be done today hopefully [09:14:01] ejoseph: how are things? [09:14:45] let me know when you have time, so we can sync up via meet [09:32:38] To join the video meeting, click this link: https://meet.google.com/ati-hniw-bod [09:32:38] Otherwise, to join by phone, dial +27 10 823 0514 and enter this PIN: 821 382 084# [09:32:38] To view more phone numbers, click this link: https://tel.meet/ati-hniw-bod?hs=5 [09:34:09] wait a sec, I'm having issues with my bt headphones [09:34:25] this way anyone can join that call, you probably want to create a new one, this is a public channel ;) [09:34:28] it seems ubuntu upgrade might've broken something after all [09:34:43] Oh [09:34:49] volans: sure? [09:35:08] I think phone + pin should be enough no? [09:35:29] oh right [09:36:02] ejoseph: post only links next time, any non-wmf person would need to be permitted antry [09:36:07] s/antry/entry [09:36:28] oh ok [10:30:24] lunch [10:41:33] break [13:51:13] hey folks super sorry for the dse k8s meeting, totally spaced out doing maintenance on kafka :( [13:59:18] elukey: np! [14:22:17] The SSH command responded with a non-zero exit status. Vagrant [14:22:17] assumes that this means the command failed. The output for this command [14:22:17] should be in the log above. Please read the output to determine what [14:22:17] went wrong. [14:25:45] the last log i got while setting up cirruss search [14:26:30] ejoseph: while doing "vagrant up" ? [14:27:16] ejoseph: provide the "red" logs from the command [14:27:55] yes [14:30:35] https://usercontent.irccloud-cdn.com/file/NXtta9yU/Screenshot%202021-10-19%20at%2004.21.19.png [14:31:00] not enough, there has to be more issues before that [14:33:15] hm... if the catalog was "Applied" it means that puppet did provision properly... [14:33:45] sorry my bad [15:08:31] ryankemper: are you still having trouble connecting? [15:09:00] Trey314159: yeah, I’m trying restarting [15:09:08] cool [16:20:47] folks qq - given https://grafana.wikimedia.org/d/000000489/wikidata-query-service?orgId=1&viewPanel=8&refresh=1m&from=now-12h&to=now&var-cluster_name=wdqs for wdqs1013, can I repool it? (depooled this morning after restarting blazegraph) [16:21:04] elukey: I repooled it :) [16:21:07] ah! [16:21:12] nevermind then :) [16:27:42] ah also congrats all for the new streaming updater [16:27:48] I know it was a loong project :) [16:28:38] thanks :) [16:59:11] does try-with-resources do anything (useful) with ByteArrayOutputStream? Maybe it's just how i'm using it, but to get the content from the baos i have to close the thing wrapping it, which closes the byte array [16:59:37] i guess not a big deal, but something about this feels awkward and i'm probably doing it wrong :) https://phabricator.wikimedia.org/P17502 [17:03:54] ByteArrayOutputStream no but I think ObjectOutputStream may have some buffers to flush, but in your example you should declare them in the same try( block ) I think [17:04:54] you can have multiple declarations in there [17:07:16] hm this won't work [17:07:25] yea the ObjectOutputStream has to be closed, otherwise it doesn't finish writing [17:08:02] yea can't declare in the same block, or can but then have to flush/close the oos manually. Ok maybe this is fine then :) Something about it just felt off [17:08:22] it definitely feels weird [17:09:06] might just that I rarely used ByteArrayOuputStream with try-with-res..? [17:09:28] that was my other thought, maybe try-with is useful for OutputStream generally, but pointless in the ByteArray case [17:10:05] yes I think so [17:10:45] ok, without the nested try blocks it looks less silly :) [17:18:17] dinner [20:36:41] hey, is this where work on search-loader2001 is happening as well? [20:36:56] just wanted to report this: The following units failed: wmf_auto_restart_mjolnir-kafka-bulk-daemon.service [20:37:08] but maybe it's still WIP ? [20:37:21] mutante: hmm, we have some graphs, lemme check. [20:37:32] This should be used, but we might not notice codfw breaking since users don't read from codfw [20:37:34] ebernhardson: thanks, just going through Icinga web UI [20:38:02] we can ACK it but even nicer would be to change puppet so it only monitors where users read from I guess :) [20:38:53] well, we still care that codfw is broken. I'm just saying we usually find out things are broken because someone tells us they aren't getting the data. They wont notice if they don't read it :) Indeed something is wrong in codfw... [20:39:10] ah, *nod* yea [20:39:14] alright [20:41:19] hmm, curious only codfw failed. It's having trouble with a wiki it doesn't know anything about (shnwiki). It should probably fail more gracefully [20:42:24] mutante: certainly ack for now, i'll see if i can make any easy fix to log instead of fail [20:44:27] ebernhardson: ACK, will ACK :) thanks [22:21:03] to follow-up on above, it looks like the root cause is a change to codfw configuration broke the cross-cluster autodetection. mjolnir finds 3 clusters from search.svc.eqiad.wmnet, but only 1 from search.svc.codfw.wmnet