00:32:37march_happy quits [Ping timeout: 245 seconds]
00:33:03march_happy (march_happy) joins
00:33:04dm4v_ joins
00:35:10dm4v quits [Ping timeout: 265 seconds]
00:35:10dm4v_ is now known as dm4v
00:36:23BlueMaxima joins
00:37:09lennier2 joins
00:39:42lennier1 quits [Ping timeout: 245 seconds]
00:39:51lennier2 is now known as lennier1
00:57:29Arcorann (Arcorann) joins
00:59:53march_happy quits [Remote host closed the connection]
01:01:42march_happy (march_happy) joins
01:21:18Hackerpcs quits [Client Quit]
01:23:28Hackerpcs (Hackerpcs) joins
01:27:12march_happy quits [Ping timeout: 245 seconds]
01:30:59march_happy (march_happy) joins
01:50:48G4te_Keep3r3 quits [Client Quit]
02:26:31Gray_cat joins
02:36:28<Gray_cat>I've heard that Livejournal had recently turned off anonymous comments - as people suggest, "to track all political commenters, but since LJ is small, 'They' are likely to give up on tracking and just shut it down". This sort of talk has happened before, of course, but still - is there an LJ archive going on? o_O
02:48:30<Doranwen>Gray_cat: not that I know of - some fandom people (not archiveteam) developed a method of saving individual LJs and comms but it requires Microsoft Office and Windows and all of that, so me with my Linux isn't able to do that
02:50:19<Doranwen>someone was working on a script method that would work for people who wanted to save their own LJs and whatever they had access to, but I think life took over and it was never finished
02:51:50Doranwen wishes she had the skills to develop it herself but that's not remotely possible in the near future, and probably not ever
02:52:55tzt quits [Ping timeout: 265 seconds]
02:58:52nico_32 quits [Ping timeout: 245 seconds]
03:02:28G4te_Keep3r3 joins
03:05:22<Gray_cat>I was thinking more about "scrape publicly-facing stuff - posts and comments"...
03:06:04<Gray_cat>Though chaps moving to dreamwidth might be interested in a full archive, private posts included (if DW does imports?)
03:13:27march_happy quits [Read error: Connection reset by peer]
03:13:47march_happy (march_happy) joins
03:15:57nico_32 (nico) joins
03:15:57march_happy quits [Read error: Connection reset by peer]
03:16:40march_happy (march_happy) joins
03:18:52march_happy quits [Read error: Connection reset by peer]
03:19:11march_happy (march_happy) joins
03:20:14Gray_cat has half an IT education and would like to see the script to see if it is within that half's abilities
03:20:15march_happy quits [Read error: Connection reset by peer]
03:20:46march_happy (march_happy) joins
03:22:54march_happy quits [Read error: Connection reset by peer]
03:35:00tzt (tzt) joins
03:36:37<Doranwen>Gray_cat: oh yeah, most people here seem only interested in the publicly-facing stuff - which, I agree is very important! - but there's a whole lot of locked stuff that many of us would love to save more efficiently than copy/paste… :/
03:38:07<Doranwen>I'm a member of quite a few communities on there - haven't used my account in several years but I still have access to all of them
03:38:27<Doranwen>as for the script, I believe thuban was working on it at one point, but you'll have to ask him how far it got
03:39:30<Doranwen>(and yes, DW does imports - currently they're still able, intermittently, to import from LJ, but who knows how much longer that will be, and that doesn't help with communities one is not the owner of but may have posted a lot to)
03:41:16<Doranwen>should anything be developed, I *definitely* want to know about it - there are several of us in the fandom world who would be using that script almost non-stop (as much as rate limiting will allow) to save tons of LJs
03:42:32<Gray_cat>Oh, if I were a proper user, I'd also be interested in locked stuff, definitely - but, being a lurker, first thing I remember are the journals I like to read (and one or two that had disappeared... together with the cookies recipes and story leads :/)
03:42:58<Doranwen>yup, plenty to save out there
03:43:55<Doranwen>I'd like to be able to save both, is all - limiting to public-only leaves me with no way to save the large quantities of private stuff I can currently see if I feel like browsing LJ
03:44:25<Doranwen>but there was stuff I only watched public posts for too, so
03:46:05<Gray_cat>No intent to limit to public-only - just a feeling that "public" is the only thing available to me and "private" is something I've never seen, thus, is lost anyway. Silly, I guess?
03:46:35<Doranwen>no, it makes sense, what you see is what you feel for
03:47:29<Doranwen>the trouble with the private stuff is most of it is fandom, and from what I've seen, the intersection of fandom and "tech-savvy enough to build a tool to archive it" is a very slim slice of people
03:48:01<Doranwen>(well, most of the stuff that others could access besides the super private life update stuff that only a user and their circle of friends ever saw)
03:48:56<Doranwen>I think if the intersection were greater, we'd probably have a better tool by now - but the one tech person who could build something only knew Excel macros, hence the tool they came up with requiring MSOffice :P
03:49:56<Doranwen>it's plenty user-friendly for the less techie people, but not so helpful for those of us who are techie enough to run Linux (but not enough to create something to archive with)
03:55:31<Gray_cat>Sheesh. I know, I know, being Turing-complete, most languages can do most things, but an archiver in Excel macros? I'd guess that's art, like the music machine made out of floppy drives, not a real-life "we've done the only thing we knew how!"
03:57:02<Gray_cat>But fandom people can do the most gorgeous art! [in "Fifth Element"-voice] That is definitely worth saving.
04:00:01treora_ quits [Quit: blub blub.]
04:01:13treora joins
04:44:42Gray_cat quits [Client Quit]
04:45:17Gray_cat joins
04:55:47DogsRNice quits [Read error: Connection reset by peer]
05:02:41jacobk joins
06:42:35jodizzle quits [Remote host closed the connection]
06:42:35dm4v quits [Client Quit]
06:42:35summerisle quits [Client Quit]
06:42:35sknebel quits [Client Quit]
06:42:35qwertyasdfuiopghjkl quits [Client Quit]
06:42:40dm4v joins
06:43:16summerisle (summerisle) joins
06:43:41sknebel (sknebel) joins
06:44:11jodizzle (jodizzle) joins
09:09:38DiscantX joins
09:10:24Discant joins
09:12:36BlueMaxima quits [Client Quit]
09:14:17DiscantX quits [Ping timeout: 245 seconds]
09:20:57Discant quits [Ping timeout: 245 seconds]
10:15:34Discant joins
11:12:00Discant quits [Read error: Connection reset by peer]
11:12:16mutantmonkey quits [Remote host closed the connection]
11:12:21Discant joins
11:13:47DiscantX joins
11:17:06Discant quits [Ping timeout: 246 seconds]
11:18:11Discant joins
11:20:57DiscantX quits [Ping timeout: 245 seconds]
11:26:39mutantmonkey (mutantmonkey) joins
12:39:12Discant quits [Ping timeout: 265 seconds]
12:49:05dm4v quits [Client Quit]
12:49:41dm4v joins
13:04:57HP_Archivist (HP_Archivist) joins
13:46:45TheTechRobo quits [Remote host closed the connection]
13:49:20TheTechRobo (TheTechRobo) joins
14:26:51Mateon2 joins
14:28:02Mateon1 quits [Ping timeout: 245 seconds]
14:28:02Mateon2 is now known as Mateon1
14:44:22Dalek quits [Client Quit]
14:47:22Dalek (Dalek) joins
14:48:08Dalek quits [Client Quit]
14:50:50Dalek (Dalek) joins
14:52:47Dalek quits [Client Quit]
14:53:27Arcorann quits [Ping timeout: 245 seconds]
14:54:28Dalek (Dalek) joins
14:55:56HP_Archivist quits [Client Quit]
14:57:47Dalek quits [Client Quit]
15:00:33Dalek (Dalek) joins
15:01:50HP_Archivist (HP_Archivist) joins
16:55:30jtagcat62 quits [Quit: Bye!]
16:59:32jtagcat62 (jtagcat) joins
18:13:36nico_32 quits [Ping timeout: 246 seconds]
19:04:31thetechrobo_ joins
19:04:45TheTechRobo quits [Remote host closed the connection]
19:15:00nico_32 (nico) joins
19:19:17tzt quits [Ping timeout: 245 seconds]
19:58:40qwertyasdfuiopghjkl joins
20:00:38Discant joins
20:50:25HP_Archivist quits [Client Quit]
21:15:24Discant quits [Ping timeout: 265 seconds]
21:32:57rocketdive joins
21:41:58rocketdive quits [Remote host closed the connection]
21:42:45thetechrobo_ is now known as TheTechRobo
21:45:07rocketdive joins
21:46:08rocketdive quits [Remote host closed the connection]
21:46:43rocketdive joins
21:50:40rocketdive is now known as kwjejdjsjs
21:51:31rocketdive (rocketdive) joins
21:53:22kwjejdjsjs quits [Remote host closed the connection]
22:34:51rocketdive quits [Remote host closed the connection]
22:35:23rocketdive (rocketdive) joins
22:44:13flashfire42 quits [Quit: The Lounge - https://thelounge.chat]
22:44:13kiska quits [Quit: The Lounge - https://thelounge.chat]
22:44:13Ryz2 quits [Quit: The Lounge - https://thelounge.chat]
22:44:13s-crypt quits [Quit: The Lounge - https://thelounge.chat]
22:46:52Ryz2 (Ryz) joins
22:46:53s-crypt (s-crypt) joins
22:47:01flashfire42 (flashfire42) joins
22:47:52kiska (kiska) joins
23:28:41<rocketdive>hey, i'm new here, is the data submitted on warrior actually saved to your device/hard drive? like the people on the leaderboard that have 1tb+, is that just what has been submitted or is that data actually being saved to their disk space?
23:29:26<rocketdive>i'm just wondering because i'm currently helping with one of the projects but my laptop is not that big lol
23:30:43<thuban>rocketdive: it's saved temporarily, then deleted once it's uploaded. the warrior won't use more than 60GB at once.
23:30:51<TheTechRobo>rocketdive: The data is sent up to the Internet Archive.
23:30:53<thuban>https://wiki.archiveteam.org/index.php?title=ArchiveTeam_Warrior#How_much_disk_space_will_the_Warrior_use?
23:31:02<TheTechRobo>damn, ninja'd
23:31:26<rocketdive>ah okay, i read that but i wasn't sure if the data remains on your device or not. thanks!
23:35:52<thuban>Gray_cat, Doranwen: the basic post-and-comment scraper is working, but it doesn't fetch most of the actual contents; i put it aside partly because i got busy and partly because lj's cruftiness made the architecture/data organization an ugly mess and i was really unhappy with it.
23:40:17<thuban>i kinda punted because i wanted to talk to J A A (who wrote the snscrape package it's a module for) about it, but he's even busier than i am... having thought about it some more, though, i _might_ have a solution. i'll try and implement it and see how it goes.