| 00:05:38 | | etnguyen03 quits [Client Quit] |
| 00:14:37 | | Zephyr joins |
| 00:14:52 | <Zephyr> | Hello all :) |
| 00:15:35 | <Zephyr> | I have a request. I totally passed the news 10 years ago about my Xfire profile. I need to shut up one of my friends who keeps telling me that I've never played 1,400 hours of Counter Strike Source. I searched the web a bit, and apart from finding my account name in a 211 MB .txt file, I didn't find much. |
| 00:15:35 | <Zephyr> | Do you know of a website that lists data? I saw that some repositories had 7 TB of data, but how can I be sure there will be datas about hours linked to my profile ? |
| 00:18:39 | <nicolas17> | https://wiki.archiveteam.org/index.php/Xfire isn't entirely clear but it suggests most user profile data was lost |
| 00:18:49 | <nicolas17> | "ArchiveTeam successfully downloaded most of the videos, and like 20% of the screenshots. There was, unfortunately, no time left for the user profiles" |
| 00:25:03 | <nicolas17> | I manually looked at 3 random archives (out of 1700) and they were all videos and screenshots |
| 00:40:30 | <Zephyr> | nicolas17 so no possibility to find an xml with hours data ? |
| 00:41:02 | <nicolas17> | do you know where that was on the original website? |
| 00:41:13 | <Zephyr> | yes |
| 00:41:21 | <Zephyr> | http://xfire.com/profile |
| 00:41:35 | <Zephyr> | and for mine it was http://xfire.com/profile/zephyr2090 |
| 00:44:21 | <nicolas17> | downloading all CDXs, 30%... |
| 00:45:51 | <kiska> | I think the format was profile.xfire.com/<slug> and crash.xfire.com/<slug> as well |
| 00:46:01 | | etnguyen03 (etnguyen03) joins |
| 00:49:40 | | anarcat quits [Ping timeout: 256 seconds] |
| 00:51:43 | <kiska> | I am very doubtful that we will find something as I do remember running out of time to get profile data |
| 00:59:55 | <Zephyr> | Maybe you're right kiska |
| 01:00:37 | <Zephyr> | but on internet archive, there is a link like I wrote |
| 01:00:50 | <Zephyr> | but both are possible |
| 01:01:17 | <nicolas17> | https://web.archive.org/web/*/http://xfire.com/profile/zephyr2090 there is not |
| 01:07:06 | <Zephyr> | yes my profile has not been archived |
| 01:07:09 | <Zephyr> | but other yes |
| 01:07:31 | <Zephyr> | that's why I'm pretty sure of my link |
| 01:12:50 | | DogsRNice joins |
| 01:13:32 | <nicolas17> | https://web.archive.org/web/*/http://xfire.com/profile/ze* |
| 01:15:28 | <nicolas17> | I looked at some and they weren't even saved by archiveteam, but by unrelated widespread crawls like https://archive.org/details/alexacrawls?tab=about |
| 01:16:07 | | anarcat (anarcat) joins |
| 01:19:24 | | Wohlstand quits [Quit: Wohlstand] |
| 01:20:14 | <Zephyr> | and in CDXs ? nicolas17 |
| 01:32:14 | | Zephyr quits [Client Quit] |
| 01:39:15 | | Zephyr joins |
| 01:39:19 | <Zephyr> | hi again |
| 01:39:24 | <Zephyr> | sorry I was disconnected |
| 01:39:42 | <Zephyr> | if you write something nicolas17 tell me ^^' |
| 02:01:30 | <Zephyr> | I found in my old 2012 computer, the Xfire folder |
| 02:01:57 | <Zephyr> | is any chance Xfire set in a file, all my data? |
| 02:07:46 | <twiswist> | I have some WARCs obtained by running gallery-dl and wget through warcprox and the data is both in the output folders and in the WARC. I want both, but dislike the duplication and excessive filesize, so I'm considering "shucking" the warc by identifying HTTP responses that are already on disk verbatim, carving them out and replacing them with a reference, to be filled in later if I ever need the actual WARC back. |
| 02:08:03 | <twiswist> | My question: is this unwise? |
| 02:21:49 | <twiswist> | Good alternatives are a WARC FUSE filesystem (doesn't exist yet), just tanking it and buying more hard drive, or deleting the warc (who cares really) |
| 02:43:50 | | cyanbox joins |
| 03:12:11 | <TheTechRobo> | WARC contains the raw HTTP response, including transfer encoding if any which would make that difficult. Even if the server doesn't use transfer-encoding, you would still need to store the headers, and be able to assemble them back into a bit-for-bit version of the record. It'd be hard to get right without losing information. |
| 03:13:44 | | etnguyen03 quits [Client Quit] |
| 03:14:05 | <TheTechRobo> | A WARC fuse filesystem would be a cool idea, actually. Or even just a GUI tool to browse them. You could even add a header to store the filename so it can have a nice directory structure when extracted. (Obviously not asking you to build it, but it would be a cool idea.) |
| 03:21:46 | <pabs> | TheTechRobo: a web search found https://github.com/PsypherPunk/python-webarchive-fuse/ /cc twiswist |
| 03:22:02 | <pabs> | (archived since 2020) |
| 03:22:11 | | qw3rty_ quits [Ping timeout: 272 seconds] |
| 03:22:15 | | etnguyen03 (etnguyen03) joins |
| 03:22:37 | | qw3rty_ joins |
| 03:25:29 | <pabs> | and that links to https://github.com/ukwa/webarchive-fuse which is even more obsolete |
| 03:28:48 | <h2ibot> | PaulWise edited The WARC Ecosystem (+388, add WARC FUSE tools): https://wiki.archiveteam.org/?diff=57702&oldid=56652 |
| 03:32:48 | <h2ibot> | PaulWise edited The WARC Ecosystem (+94, add section for other lists of WARC tools): https://wiki.archiveteam.org/?diff=57703&oldid=57702 |
| 03:35:11 | <twiswist> | The responses I care about are images and have no transfer-encoding, so no problem there. The HTTP headers don't bother me either, all I have to do is identify the HTTP response body inside the http response inside the warc entry and replace that with a reference if it matches any file. I've opted to just keep everything for now |
| 03:35:27 | <twiswist> | Thank you for adding to the wiki article |
| 03:38:49 | <h2ibot> | PaulWise edited The WARC Ecosystem (-2, typos): https://wiki.archiveteam.org/?diff=57704&oldid=57703 |
| 03:54:43 | <pabs> | twiswist: sounds like libarchive might have WARC support and there is a fuse filesystem for libarchive https://github.com/mxmlnkn/ratarmount/issues/128 https://github.com/cybernoid/archivemount |
| 03:55:44 | <pabs> | ah, archivemount is in Debian, checking if it works |
| 03:56:52 | <h2ibot> | PaulWise edited The WARC Ecosystem (+571, add gowarc, java-warc, metawarc): https://wiki.archiveteam.org/?diff=57705&oldid=57704 |
| 04:02:26 | | etnguyen03 quits [Remote host closed the connection] |
| 04:02:51 | <pabs> | twiswist: archivemount just works! |
| 04:03:28 | <pabs> | ah, it includes HTTP headers in each file though... |
| 04:03:58 | <pabs> | should have been put as xattrs |
| 04:11:54 | <h2ibot> | PaulWise edited The WARC Ecosystem (+259, add libarchive/archivemount): https://wiki.archiveteam.org/?diff=57706&oldid=57705 |
| 04:42:27 | <nicolas17> | I downloaded all cdx files in https://archive.org/details/archiveteam_xfire, among the archived URLs there are 0 instances of "profile" or "zephyr" |
| 04:43:01 | <nicolas17> | fun fact, searching for "zephy" finds some false positives in the base32 hashes :D |
| 04:49:06 | | DopefishJustin quits [Remote host closed the connection] |
| 05:11:05 | | Zephyr quits [Quit: Ooops, wrong browser tab.] |
| 05:18:35 | | matoro quits [Quit: https://quassel-irc.org - Chat comfortably. Anywhere.] |
| 05:20:14 | | matoro joins |
| 05:24:54 | | matoro quits [Client Quit] |
| 05:25:05 | | matoro joins |
| 06:01:25 | | Island quits [Read error: Connection reset by peer] |
| 06:17:29 | | DogsRNice quits [Read error: Connection reset by peer] |
| 08:11:40 | | wotd quits [Ping timeout: 256 seconds] |
| 08:51:54 | | TheEnbyperor_ quits [Ping timeout: 256 seconds] |
| 09:02:50 | | TheEnbyperor joins |
| 09:02:56 | | TheEnbyperor_ (TheEnbyperor) joins |
| 09:16:32 | | ducky (ducky) joins |
| 09:21:17 | | ducky quits [Ping timeout: 272 seconds] |
| 09:58:01 | | szczot3k quits [Ping timeout: 272 seconds] |
| 10:00:34 | | ducky (ducky) joins |
| 10:04:48 | | szczot3k (szczot3k) joins |
| 10:05:37 | | ducky quits [Ping timeout: 272 seconds] |
| 10:11:29 | | ducky (ducky) joins |
| 10:16:23 | | ducky quits [Ping timeout: 272 seconds] |
| 10:24:01 | <h2ibot> | Manu edited Discourse/archived (+89, forum.clockworkpi.com has been archived before): https://wiki.archiveteam.org/?diff=57707&oldid=57679 |
| 10:32:51 | | midou quits [Ping timeout: 272 seconds] |
| 10:33:03 | <h2ibot> | Manu edited Discourse/archived (+86, Queued meta.discourse.org): https://wiki.archiveteam.org/?diff=57708&oldid=57707 |
| 10:35:03 | <h2ibot> | Manu edited Discourse/archived (+29, llllllll.co required Firefox UA): https://wiki.archiveteam.org/?diff=57709&oldid=57708 |
| 10:37:25 | | midou joins |
| 11:00:03 | | Bleo182600722719623455222 quits [Quit: The Lounge - https://thelounge.chat] |
| 11:02:47 | | Bleo182600722719623455222 joins |
| 11:02:49 | | NatTheCat6 (NatTheCat) joins |
| 11:05:04 | | NatTheCat quits [Ping timeout: 256 seconds] |
| 11:05:04 | | NatTheCat6 is now known as NatTheCat |
| 11:31:35 | | ducky (ducky) joins |
| 11:36:11 | | ducky quits [Ping timeout: 272 seconds] |
| 12:16:21 | | FiTheArchiver joins |
| 12:20:10 | | FiTheArchiver quits [Client Quit] |
| 12:21:09 | | LddPotato quits [Ping timeout: 272 seconds] |
| 12:38:51 | | croissant_ joins |
| 12:41:15 | | Wohlstand (Wohlstand) joins |
| 12:42:03 | | croissant quits [Ping timeout: 272 seconds] |
| 13:05:20 | | Zephyr joins |
| 13:05:26 | <Zephyr> | Hi nicolas17 ^^ |
| 13:05:33 | <Zephyr> | Sorry I was disconnected again x) |
| 13:05:52 | <Zephyr> | you found "zephy" in hash32 then? |
| 13:06:00 | <Zephyr> | but no zephyr or zephyr2090 |
| 13:25:36 | <h2ibot> | Cruller edited Deathwatch (+369, /* 2025 */ Add goo portal and いまトピ): https://wiki.archiveteam.org/?diff=57710&oldid=57688 |
| 13:29:56 | <kiska> | Here is what I found https://server8.kiska.pw/uploads/a6ffe94b73246591/image.png https://web.archive.org/web/*/http://xfire.com/profile/zep* |
| 13:34:16 | <@arkiver> | asie: hmm, what is your idea behind that proposed channel name? |
| 13:41:39 | <h2ibot> | Cruller edited Goo (+236, /* Sites */ Add News, Ranking, Poratal, and いまトピ): https://wiki.archiveteam.org/?diff=57711&oldid=57441 |
| 13:46:07 | <@arkiver> | for Archiwum Allegro, we're going to scan 17 billion IDs |
| 13:51:14 | | etnguyen03 (etnguyen03) joins |
| 13:52:12 | <Zephyr> | kiska yep I didnt see my account unfortunatly |
| 14:03:12 | <asie> | arkiver: music tempo |
| 14:03:36 | <asie> | "allegro" is fast and bright, "prestissimo" is extremely fast (because you're going to need to grab a lot of data quickly) |
| 14:08:25 | <@arkiver> | asie: aha |
| 14:08:52 | <@arkiver> | alright let's go with that one |
| 14:16:25 | | wickedplayer494 quits [Ping timeout: 272 seconds] |
| 14:17:23 | | wickedplayer494 joins |
| 14:43:10 | | yupela joins |
| 14:44:49 | <yupela> | Hi. Recently came across this community and thought there might be a solution to a problem I was having a while ago. Are there any good tools to capture full page screenshots? I've tried finding existing software and also writing my own but so many pages don't play nice and have odd behaviours when scrolling. |
| 14:46:14 | <yupela> | Most software I found to do it will lock the viewport of the browser, apply some CSS rules to try and hold stuff still, then scroll down a full page at a time taking a screenshot, then stitch them all together |
| 14:47:28 | | yupela quits [Client Quit] |
| 14:48:21 | <twiswist> | Firefox used to (and palemoon still does) have an unusual extra command bar summoned with shift-f2 where you can write screenshot --fullpage --clipboard and it actually does it. I don't remember where firefox moved this functionality |
| 14:50:14 | <justauser|m> | An "unusual command bar" lives at Ctrl+Shift+J, but I think it's a different one. |
| 14:51:15 | | etnguyen03 quits [Client Quit] |
| 14:53:40 | <TheTechRobo> | brozzler does this to do it in one go: https://github.com/internetarchive/brozzler/blob/master/brozzler/browser.py#L798 |
| 14:54:13 | <TheTechRobo> | but that doesn't work well if the page is extremely long. Chrome really doesn't like big screenshots. |
| 14:58:30 | | Larsenv quits [Quit: The Lounge - https://thelounge.chat] |
| 15:00:09 | | wickedplayer494 is now authenticated as wickedplayer494 |
| 15:25:42 | <cruller> | When taking screenshots in Firefox, the maximum length of the longer side is as follows: |
| 15:25:52 | <cruller> | When executing :screenshot --fullpage: 10,000 pixels |
| 15:25:59 | <cruller> | When using “Save full page” from https://support.mozilla.org/en-US/kb/take-screenshots-firefox: 32,700 pixels |
| 15:27:43 | | wotd joins |
| 15:40:19 | <cruller> | Chrome's “Capture full size screenshot” did not work on https://www.tandfonline.com/doi/full/10.1080/15548627.2015.1100356 (extremely long page). Additionally, attempting to “Capture node screenshot” on /html/body simply caused it to crash. |
| 15:48:57 | <h2ibot> | Manu edited Discourse (+47, Active Discourses: Add forum.elivelinux.org): https://wiki.archiveteam.org/?diff=57712&oldid=57681 |
| 16:04:28 | <cruller> | Is it difficult to capture elements whose behavior changes depending on whether they are in the visible area? Simple lazy loading aside. |
| 16:08:18 | <cruller> | When converting dynamic things to static ones, the question of “which state to freeze” always arises... |
| 16:13:27 | <justauser|m> | Lazy loading is commonly handled by scrolldown-scrollup-capture pattern I believe? |
| 16:16:18 | <justauser|m> | For harder things, I'd guess I want the invisible state. |
| 16:16:18 | <justauser|m> | Less likely to be animated, for example. |
| 16:26:04 | <h2ibot> | Manu edited Discourse (+46, Active Discourses: Add discuss.gradle.org): https://wiki.archiveteam.org/?diff=57713&oldid=57712 |
| 16:31:04 | <h2ibot> | Manu edited Discourse (+47, Active Discourses: Add forum.cursor.com): https://wiki.archiveteam.org/?diff=57714&oldid=57713 |
| 16:47:30 | <cruller> | I guess some people prefer to capture every element on the page in the state it appears when inside the viewport. |
| 17:12:52 | | etnguyen03 (etnguyen03) joins |
| 17:29:02 | | NF885 (NF885) joins |
| 17:30:54 | | etnguyen03 quits [Client Quit] |
| 17:33:19 | | Wake quits [Quit: Ping timeout (120 seconds)] |
| 17:40:01 | | etnguyen03 (etnguyen03) joins |
| 18:05:43 | | DogsRNice joins |
| 18:12:40 | | NF885 quits [Client Quit] |
| 18:15:54 | | lennier2_ joins |
| 18:18:34 | | lennier2 quits [Ping timeout: 256 seconds] |
| 18:24:55 | | pseudorizer quits [Quit: ZNC 1.10.1 - https://znc.in] |
| 18:25:37 | | pseudorizer (pseudorizer) joins |
| 18:29:25 | <Zephyr> | any news nicolas17? |
| 18:29:32 | <Zephyr> | sorry to bother you |
| 18:41:40 | | Zachava joins |
| 18:43:41 | | Wake joins |
| 18:50:40 | | Wohlstand quits [Remote host closed the connection] |
| 18:50:56 | | Wohlstand (Wohlstand) joins |
| 18:52:51 | | Zachava is now authenticated as Zachava |
| 19:04:13 | | etnguyen03 quits [Client Quit] |
| 19:13:44 | | Island joins |
| 19:43:10 | | ichdasich quits [Quit: reboot] |
| 19:44:29 | | ichdasich joins |
| 21:08:56 | <h2ibot> | Exorcism edited ArchiveTeam Warrior (+0, v4.1.0 -> v4.2.0): https://wiki.archiveteam.org/?diff=57715&oldid=56980 |
| 21:13:09 | | szczot3k quits [Ping timeout: 272 seconds] |
| 21:23:02 | | vitaly-zdanevich joins |
| 21:24:24 | <vitaly-zdanevich> | This old website looks like will die soon, links redirect to a newer version that also last updated in 2022 https://www.game-over.net/reviews.php?id=452&page=reviews |
| 21:27:59 | <pokechu22> | will run it in #archivebot |
| 21:33:33 | | szczot3k (szczot3k) joins |
| 21:35:25 | | VerifiedJ quits [Quit: The Lounge - https://thelounge.chat] |
| 21:35:57 | | VerifiedJ (VerifiedJ) joins |
| 21:40:30 | | vitaly-zdanevich is now authenticated as vitaly-zdanevich |
| 21:42:50 | | etnguyen03 (etnguyen03) joins |
| 21:59:34 | | etnguyen03 quits [Remote host closed the connection] |
| 22:00:46 | | etnguyen03 (etnguyen03) joins |
| 22:02:53 | <nicolas17> | Zephyr: I found no "zephyr" in any URL |
| 22:02:58 | <nicolas17> | the "zephy" false positives were like |
| 22:03:00 | <nicolas17> | com,xfire,video)/116d03-2.jpg 20150620065120 http://video.xfire.com/116d03-2.jpg image/jpeg 200 PTGCHZEPHYV4MYCAS3CHYFRRMAKXOJQK - - 6773 24968964850 archiveteam_xfire_20150622041426/xfire_20150622041426.megawarc.warc.gz |
| 22:03:02 | <nicolas17> | com,xfire,video)/127219-2.jpg 20150620005731 http://video.xfire.com/127219-2.jpg image/jpeg 200 ZT6FD44IQZNMTTKH6WJTMPNCZEPHY234 - - 7489 30895929300 archiveteam_xfire_20150622041437/xfire_20150622041437.megawarc.warc.gz |
| 22:04:26 | <nicolas17> | it seems there's no profiles archived here at all, the ones that do exist on the wayback machine didn't come from archiveteam's project |
| 22:06:49 | | skyrocket quits [Remote host closed the connection] |
| 22:11:01 | | etnguyen03 quits [Client Quit] |
| 22:22:05 | | skyrocket joins |
| 22:54:26 | | etnguyen03 (etnguyen03) joins |
| 22:57:38 | | etnguyen03 quits [Remote host closed the connection] |
| 23:02:06 | | etnguyen03 (etnguyen03) joins |
| 23:33:29 | | etnguyen03 quits [Client Quit] |