| 01:53:46 | | qwertyasdfuiopghjkl quits [Remote host closed the connection] |
| 04:49:05 | | qw3rty_ joins |
| 04:52:37 | | qw3rty__ quits [Ping timeout: 244 seconds] |
| 08:00:47 | | Daloader joins |
| 08:08:26 | | Daloader quits [Ping timeout: 265 seconds] |
| 08:53:31 | | ragu joins |
| 09:42:00 | | qwertyasdfuiopghjkl joins |
| 10:53:12 | | rewby quits [Changing host] |
| 10:53:12 | | rewby (rewby) joins |
| 11:06:44 | | HackMii quits [Ping timeout: 258 seconds] |
| 11:08:56 | | HackMii (hacktheplanet) joins |
| 12:05:38 | | Gereon62 quits [Quit: The Lounge - https://thelounge.chat] |
| 12:06:04 | | Gereon62 (Gereon) joins |
| 18:07:46 | | qwertyasdfuiopghjkl quits [Ping timeout: 244 seconds] |
| 18:56:37 | | appledash joins |
| 18:57:22 | <appledash> | Hey there - I've known about ArchiveTeam for quite some time, but I recently started looking into it a bit more, and I found out about the URLTeam project. Given I run a URL shortener, I had to see if my site was on the list - it is! |
| 18:57:34 | <appledash> | I was wondering how I could provide you guys a CSV file or something of the ~100,000 short URLs on my site |
| 19:05:20 | <appledash> | Alas, it be late for me, so I will likely return any messages tomorrow :) I run https://derpy.me |
| 19:18:36 | | qwertyasdfuiopghjkl joins |
| 19:39:37 | | asdf joins |
| 19:45:10 | | asdf quits [Remote host closed the connection] |
| 19:51:22 | | asdf joins |
| 19:52:26 | | asdf quits [Remote host closed the connection] |
| 20:12:46 | <@JAA> | appledash: Lovely to hear, and that would be great! :-) You could dump it onto https://transfer.archivete.am/ (prior compression appreciated if it turns out to be large, but doesn't sound too big at 100k), and then we'll take care of it from there (uploading to the Internet Archive with the right metadata etc.). Feel free to include any additional columns you have and are comfortable with sharing publicly; |
| 20:12:52 | <@JAA> | I don't know what you're tracking, but link creation datetime and view counts are common examples I've seen in the past. |
| 20:15:29 | <@JAA> | Any unambiguous data format will do. CSV is what this crawling project is currently producing, but if something else is easier for you, that's just as fine. |