[08:31:38] there just has been some s2 query spikes: https://grafana.wikimedia.org/d/000000278/mysql-aggregated?orgId=1&refresh=1m&var-site=eqiad&var-group=core&var-shard=s2&var-role=All&from=1639103492641&to=1639125092641 [08:33:34] https://grafana-rw.wikimedia.org/d/000000273/mysql?viewPanel=3&orgId=1&refresh=1m&from=now-24h&to=now&forceLogin=true&var-job=All&var-server=db1162&var-port=9104 [08:35:51] I only see MediaWiki\Revision\RevisionStore::loadSlotRecords at the moment, so "page visits" maybe? [08:36:06] to itwiki? [08:36:13] I am also seeing some API queries to itwiki [08:36:20] scanning 500k rows [08:36:24] could be a bot [08:37:31] yeah, most of the queries are for itwiki from what I can see [08:40:31] A bot called "MediaWikiCrawler-Google/2.0" doing api calls? [08:41:19] pretty much all the queries on db1162 are for itwiki [08:42:13] the ip space seems google owned [08:42:33] although it could be google cloud [08:42:39] should we ask traffic to block it? [08:42:51] let's ask what other know about it [09:59:09] db backups are about to be back into a good state, codfw commons media backups are at 25% [10:18:12] jynus: that means parsing (edit or uncached view), that function needs caching T297147 [10:18:12] T297147: RevisionStore::newRevisionSlots() needs a cache - https://phabricator.wikimedia.org/T297147 [10:18:22] there is already a patch for it :) [10:29:11] I reworked the db error dashboard and added both to logstash home page [10:30:27] Amir1: very cool :) [10:30:41] kormat: ^^ <3 [11:08:52] all backups green now [11:15:07] btw, I'm starting the clean up of old rows of flaggedtemplates. I think this will make it much smaller https://phabricator.wikimedia.org/T296380#7562464 [11:15:25] Running it on alswiki now but it's taking a long time :( [11:16:13] I think this will upwards of 1B rows [11:17:30] *will delete [11:28:26] oh and this might cause alerts in backups ^ [11:31:48] alerts why? you are only doing deletes, right? [11:41:39] ah, I get it now, because of size changes [11:42:04] that is ok, as long as we see it is intended, it will be easy to spot :-) [11:42:39] :D yup, let's see how much it will delete [12:23:05] I will briefly disconnecto to move for lunch. After it, I think I will increase disk size of backup2* hosts. [12:23:59] as backup is like 30% done, but disk usage is at 50%, so better increase it so it doesn't alert during Christmas [14:07:15] I will restart mysql on db1123 on Monday to clear that warning about memory usage [14:14:06] :-( [14:28:35] I can't believe the clean up of flaggedtemplates for alswiki (a super tiny wiki with 80k pages) is taking five hours now and it's not even half-way done. [14:44:57] huh. i think i have a wmfdb PoC replacement for mysql.py implemented. that was both worse and better than i expected. [16:08:42] * Emperor offers to declare it "too OOPy" ;-) [17:32:14] Now s1 with the QPS 100% increase [17:34:01] same query patterns as before [18:05:03] seems to be trending down: https://grafana.wikimedia.org/d/000000273/mysql?viewPanel=16&orgId=1&refresh=1m&var-job=All&var-server=db1163&var-port=9104&from=1639148680805&to=1639159480805 [18:06:23] backup2* maintenance seems to be completed, leaving it here for the week, have a nice weekend