[04:02:33] [1/2] I had a question before i uploaded the .xml file for the new wiki we're making [04:02:34] [2/2] is there a virtual .XML thing i can try to see what it looks like? test it out before I submit it for uploading and importing to a new wiki? [04:04:00] Go to Special:Export on any wiki (Wikipedia, Miraheze Meta, etc.), insert the name of a page that exists, click "Export" and open the file [04:04:47] [1/2] the problem is that the wiki in question no longer exists [04:04:48] [2/2] we're salvaging the .xml and image files from an archive.org backup [04:04:54] and the .xml alone is over 17 gigs [04:05:03] sorry, i should have led with that [04:05:11] ohhh [04:05:24] you're trying to see how a wiki would look like if the file were imported? [04:05:32] we decided to move to miraheze for stability's sake [04:05:49] yeah, and to give me a chance to scrub the accounts of people who don't want to move, scrub stuff that doesn't meet community guidelines, scrub dead links [04:06:18] [1/3] no sense exporting stuff to a new wiki here at miraheze that we immediately have to delete, of course [04:06:19] [2/3] i want a chance to cleanse it [04:06:19] [3/3] but doing that to raw files would tkae approximately one metric week [04:06:20] there's no way to preview what an XML dump will import, sadly [04:06:28] ffffffff [04:06:31] okay, uh [04:06:59] [1/2] is there a way i could import the .xml dump to a wiki that isn't public? [04:06:59] [2/2] Create the wiki on miraheze but don't 'publish' it until i have a chance to scrub it? [04:07:19] there is, yes, using a standard private wiki [04:07:46] [1/2] thanks [04:07:46] [2/2] is there a tutorial on how to set that up? I can do it this weekend [04:08:00] i know about the Import Wiki page on the miraheze website, but i haven't used it yet [04:08:37] [[Special:RequestWiki]] to request a wiki and [[Special:RequestImportDump]] to upload the file [04:08:38] [1/2] [04:08:38] [2/2] [04:08:53] though the limit for Special:RequestImportDump is like 2GB iirc? [04:09:14] [1/2] oh [04:09:14] [2/2] yeah that won't fly, this wiki is 17.3gb of raw .xml [04:10:46] what wiki are you importing, if I may ask? [05:02:12] I mean I guess you could try and install a local MediaWiki version too (not that I have any idea how difficult that is) [05:15:39] [1/2] https://discord.com/channels/407504499280707585/615786602454581249/1201871363103797290 [05:15:39] [2/2] https://discord.com/channels/407504499280707585/615786602454581249/1201962714734538812 [05:16:06] ah, interesting [05:16:20] We have quite a few chunky wikis that are seeking to migrate, as of late [06:34:08] <.labster> You should be able to import just a portion of the wiki, just use/write some programs to just return a portion of the XML. [06:36:42] <.labster> I may have had the worst experience importing of anyone here, since I had to convert every page of a wiki from PmWiki to Mediawiki while at the same time creating categories and rewriting most links to include spaces and propercase instead of all CamelCase. And then of course writing Mediawiki XML from the converted pages. [06:37:04] <.labster> Though recovery from archive.org sounds pretty bad in other ways. [08:36:39] Depends on if Wikiteam dumped it or if it's through Wayback [21:51:58] i am most certainly not a programmer [21:52:01] we're kind of flying blind [21:52:09] the guy who ran the last server alone has dropped off the face of the earth [21:52:42] once I have the .xml files, then i can pursue next steps [22:10:56] Ouch, yeah, scraper tool may be the way to go then. 😦 [22:18:38] we're working on getting it [22:18:50] the problem is, 17.30 gb is a huge file [22:19:00] i'm honestly not sure how we would transfer it [22:19:10] Is that compressed? [22:19:43] MEGA free accounts are 20 Gb now I think [22:20:10] and it's w/ files? [22:21:36] also you can make divided archive like I have to do at work w/ a 70 Gb data base whenever devs need it to debug an error [22:21:53] https://cdn.discordapp.com/attachments/1006789349498699827/1202740636336525312/image.png?ex=65ce8e81&is=65bc1981&hm=7f7207bbcceef24d3202bc8d4c8a34ffad9fe0ddb17c142f266308246c19a750& [22:36:23] divided archive is effective if you don't have a neat host to send the whole blob onto [22:43:38] we'll try to use Mega [22:43:54] then i'll direct download the .XML files and see what the best next step would be [22:44:17] we weren't expecting to have to do this, but the old admin has simply dropped off the face of the earth and the ad hoc leadership committee is kind of scrambling [22:45:11] It happens [23:11:39] https://wormhole.app/ is decent, but it's P2P for larger files like that so it could result in IP address leakages if that's a concern [23:11:51] thanks [23:14:12] I think disroot also has smt [23:14:46] You could also upload to https://archive.org if it's a wiki dump. Helps with ensuring it doesn't get lost too [23:15:06] it's there already actually [23:15:10] Oh nic [23:20:13] <.labster> Definitely compress the file. 7z should be really good if you let it build a big dictionary. [23:39:39] for transmissal to me, yeah [23:39:47] but i can't import it compressed, right?