00:05:38etnguyen03 quits [Client Quit]
00:14:37Zephyr joins
00:14:52<Zephyr>Hello all :)
00:15:35<Zephyr>I have a request. I totally passed the news 10 years ago about my Xfire profile. I need to shut up one of my friends who keeps telling me that I've never played 1,400 hours of Counter Strike Source. I searched the web a bit, and apart from finding my account name in a 211 MB .txt file, I didn't find much.
00:15:35<Zephyr>Do you know of a website that lists data? I saw that some repositories had 7 TB of data, but how can I be sure there will be datas about hours linked to my profile ?
00:18:39<nicolas17>https://wiki.archiveteam.org/index.php/Xfire isn't entirely clear but it suggests most user profile data was lost
00:18:49<nicolas17>"ArchiveTeam successfully downloaded most of the videos, and like 20% of the screenshots. There was, unfortunately, no time left for the user profiles"
00:25:03<nicolas17>I manually looked at 3 random archives (out of 1700) and they were all videos and screenshots
00:40:30<Zephyr>nicolas17 so no possibility to find an xml with hours data ?
00:41:02<nicolas17>do you know where that was on the original website?
00:41:13<Zephyr>yes
00:41:21<Zephyr>http://xfire.com/profile
00:41:35<Zephyr>and for mine it was http://xfire.com/profile/zephyr2090
00:44:21<nicolas17>downloading all CDXs, 30%...
00:45:51<kiska>I think the format was profile.xfire.com/<slug> and crash.xfire.com/<slug> as well
00:46:01etnguyen03 (etnguyen03) joins
00:49:40anarcat quits [Ping timeout: 256 seconds]
00:51:43<kiska>I am very doubtful that we will find something as I do remember running out of time to get profile data
00:59:55<Zephyr>Maybe you're right kiska
01:00:37<Zephyr>but on internet archive, there is a link like I wrote
01:00:50<Zephyr>but both are possible
01:01:17<nicolas17>https://web.archive.org/web/*/http://xfire.com/profile/zephyr2090 there is not
01:07:06<Zephyr>yes my profile has not been archived
01:07:09<Zephyr>but other yes
01:07:31<Zephyr>that's why I'm pretty sure of my link
01:12:50DogsRNice joins
01:13:32<nicolas17>https://web.archive.org/web/*/http://xfire.com/profile/ze*
01:15:28<nicolas17>I looked at some and they weren't even saved by archiveteam, but by unrelated widespread crawls like https://archive.org/details/alexacrawls?tab=about
01:16:07anarcat (anarcat) joins
01:19:24Wohlstand quits [Quit: Wohlstand]
01:20:14<Zephyr>and in CDXs ? nicolas17
01:32:14Zephyr quits [Client Quit]
01:39:15Zephyr joins
01:39:19<Zephyr>hi again
01:39:24<Zephyr>sorry I was disconnected
01:39:42<Zephyr>if you write something nicolas17 tell me ^^'
02:01:30<Zephyr>I found in my old 2012 computer, the Xfire folder
02:01:57<Zephyr>is any chance Xfire set in a file, all my data?
02:07:46<twiswist>I have some WARCs obtained by running gallery-dl and wget through warcprox and the data is both in the output folders and in the WARC. I want both, but dislike the duplication and excessive filesize, so I'm considering "shucking" the warc by identifying HTTP responses that are already on disk verbatim, carving them out and replacing them with a reference, to be filled in later if I ever need the actual WARC back.
02:08:03<twiswist>My question: is this unwise?
02:21:49<twiswist>Good alternatives are a WARC FUSE filesystem (doesn't exist yet), just tanking it and buying more hard drive, or deleting the warc (who cares really)
02:43:50cyanbox joins
03:12:11<TheTechRobo>WARC contains the raw HTTP response, including transfer encoding if any which would make that difficult. Even if the server doesn't use transfer-encoding, you would still need to store the headers, and be able to assemble them back into a bit-for-bit version of the record. It'd be hard to get right without losing information.
03:13:44etnguyen03 quits [Client Quit]
03:14:05<TheTechRobo>A WARC fuse filesystem would be a cool idea, actually. Or even just a GUI tool to browse them. You could even add a header to store the filename so it can have a nice directory structure when extracted. (Obviously not asking you to build it, but it would be a cool idea.)
03:21:46<pabs>TheTechRobo: a web search found https://github.com/PsypherPunk/python-webarchive-fuse/ /cc twiswist
03:22:02<pabs>(archived since 2020)
03:22:11qw3rty_ quits [Ping timeout: 272 seconds]
03:22:15etnguyen03 (etnguyen03) joins
03:22:37qw3rty_ joins
03:25:29<pabs>and that links to https://github.com/ukwa/webarchive-fuse which is even more obsolete
03:28:48<h2ibot>PaulWise edited The WARC Ecosystem (+388, add WARC FUSE tools): https://wiki.archiveteam.org/?diff=57702&oldid=56652
03:32:48<h2ibot>PaulWise edited The WARC Ecosystem (+94, add section for other lists of WARC tools): https://wiki.archiveteam.org/?diff=57703&oldid=57702
03:35:11<twiswist>The responses I care about are images and have no transfer-encoding, so no problem there. The HTTP headers don't bother me either, all I have to do is identify the HTTP response body inside the http response inside the warc entry and replace that with a reference if it matches any file. I've opted to just keep everything for now
03:35:27<twiswist>Thank you for adding to the wiki article
03:38:49<h2ibot>PaulWise edited The WARC Ecosystem (-2, typos): https://wiki.archiveteam.org/?diff=57704&oldid=57703
03:54:43<pabs>twiswist: sounds like libarchive might have WARC support and there is a fuse filesystem for libarchive https://github.com/mxmlnkn/ratarmount/issues/128 https://github.com/cybernoid/archivemount
03:55:44<pabs>ah, archivemount is in Debian, checking if it works
03:56:52<h2ibot>PaulWise edited The WARC Ecosystem (+571, add gowarc, java-warc, metawarc): https://wiki.archiveteam.org/?diff=57705&oldid=57704
04:02:26etnguyen03 quits [Remote host closed the connection]
04:02:51<pabs>twiswist: archivemount just works!
04:03:28<pabs>ah, it includes HTTP headers in each file though...
04:03:58<pabs>should have been put as xattrs
04:11:54<h2ibot>PaulWise edited The WARC Ecosystem (+259, add libarchive/archivemount): https://wiki.archiveteam.org/?diff=57706&oldid=57705
04:42:27<nicolas17>I downloaded all cdx files in https://archive.org/details/archiveteam_xfire, among the archived URLs there are 0 instances of "profile" or "zephyr"
04:43:01<nicolas17>fun fact, searching for "zephy" finds some false positives in the base32 hashes :D
04:49:06DopefishJustin quits [Remote host closed the connection]
05:11:05Zephyr quits [Quit: Ooops, wrong browser tab.]
05:18:35matoro quits [Quit: https://quassel-irc.org - Chat comfortably. Anywhere.]
05:20:14matoro joins
05:24:54matoro quits [Client Quit]
05:25:05matoro joins
06:01:25Island quits [Read error: Connection reset by peer]
06:17:29DogsRNice quits [Read error: Connection reset by peer]
08:11:40wotd quits [Ping timeout: 256 seconds]
08:51:54TheEnbyperor_ quits [Ping timeout: 256 seconds]
09:02:50TheEnbyperor joins
09:02:56TheEnbyperor_ (TheEnbyperor) joins
09:16:32ducky (ducky) joins
09:21:17ducky quits [Ping timeout: 272 seconds]
09:58:01szczot3k quits [Ping timeout: 272 seconds]
10:00:34ducky (ducky) joins
10:04:48szczot3k (szczot3k) joins
10:05:37ducky quits [Ping timeout: 272 seconds]
10:11:29ducky (ducky) joins
10:16:23ducky quits [Ping timeout: 272 seconds]
10:24:01<h2ibot>Manu edited Discourse/archived (+89, forum.clockworkpi.com has been archived before): https://wiki.archiveteam.org/?diff=57707&oldid=57679
10:32:51midou quits [Ping timeout: 272 seconds]
10:33:03<h2ibot>Manu edited Discourse/archived (+86, Queued meta.discourse.org): https://wiki.archiveteam.org/?diff=57708&oldid=57707
10:35:03<h2ibot>Manu edited Discourse/archived (+29, llllllll.co required Firefox UA): https://wiki.archiveteam.org/?diff=57709&oldid=57708
10:37:25midou joins
11:00:03Bleo182600722719623455222 quits [Quit: The Lounge - https://thelounge.chat]
11:02:47Bleo182600722719623455222 joins
11:02:49NatTheCat6 (NatTheCat) joins
11:05:04NatTheCat quits [Ping timeout: 256 seconds]
11:05:04NatTheCat6 is now known as NatTheCat
11:31:35ducky (ducky) joins
11:36:11ducky quits [Ping timeout: 272 seconds]
12:16:21FiTheArchiver joins
12:20:10FiTheArchiver quits [Client Quit]
12:21:09LddPotato quits [Ping timeout: 272 seconds]
12:38:51croissant_ joins
12:41:15Wohlstand (Wohlstand) joins
12:42:03croissant quits [Ping timeout: 272 seconds]
13:05:20Zephyr joins
13:05:26<Zephyr>Hi nicolas17 ^^
13:05:33<Zephyr>Sorry I was disconnected again x)
13:05:52<Zephyr>you found "zephy" in hash32 then?
13:06:00<Zephyr>but no zephyr or zephyr2090
13:25:36<h2ibot>Cruller edited Deathwatch (+369, /* 2025 */ Add goo portal and いまトピ): https://wiki.archiveteam.org/?diff=57710&oldid=57688
13:29:56<kiska>Here is what I found https://server8.kiska.pw/uploads/a6ffe94b73246591/image.png https://web.archive.org/web/*/http://xfire.com/profile/zep*
13:34:16<@arkiver>asie: hmm, what is your idea behind that proposed channel name?
13:41:39<h2ibot>Cruller edited Goo (+236, /* Sites */ Add News, Ranking, Poratal, and いまトピ): https://wiki.archiveteam.org/?diff=57711&oldid=57441
13:46:07<@arkiver>for Archiwum Allegro, we're going to scan 17 billion IDs
13:51:14etnguyen03 (etnguyen03) joins
13:52:12<Zephyr>kiska yep I didnt see my account unfortunatly
14:03:12<asie>arkiver: music tempo
14:03:36<asie>"allegro" is fast and bright, "prestissimo" is extremely fast (because you're going to need to grab a lot of data quickly)
14:08:25<@arkiver>asie: aha
14:08:52<@arkiver>alright let's go with that one
14:16:25wickedplayer494 quits [Ping timeout: 272 seconds]
14:17:23wickedplayer494 joins
14:43:10yupela joins
14:44:49<yupela>Hi. Recently came across this community and thought there might be a solution to a problem I was having a while ago. Are there any good tools to capture full page screenshots? I've tried finding existing software and also writing my own but so many pages don't play nice and have odd behaviours when scrolling.
14:46:14<yupela>Most software I found to do it will lock the viewport of the browser, apply some CSS rules to try and hold stuff still, then scroll down a full page at a time taking a screenshot, then stitch them all together
14:47:28yupela quits [Client Quit]
14:48:21<twiswist>Firefox used to (and palemoon still does) have an unusual extra command bar summoned with shift-f2 where you can write screenshot --fullpage --clipboard and it actually does it. I don't remember where firefox moved this functionality
14:50:14<justauser|m>An "unusual command bar" lives at Ctrl+Shift+J, but I think it's a different one.
14:51:15etnguyen03 quits [Client Quit]
14:53:40<TheTechRobo>brozzler does this to do it in one go: https://github.com/internetarchive/brozzler/blob/master/brozzler/browser.py#L798
14:54:13<TheTechRobo>but that doesn't work well if the page is extremely long. Chrome really doesn't like big screenshots.
14:58:30Larsenv quits [Quit: The Lounge - https://thelounge.chat]
15:25:42<cruller>When taking screenshots in Firefox, the maximum length of the longer side is as follows:
15:25:52<cruller>When executing :screenshot --fullpage: 10,000 pixels
15:25:59<cruller>When using “Save full page” from https://support.mozilla.org/en-US/kb/take-screenshots-firefox: 32,700 pixels
15:27:43wotd joins
15:40:19<cruller>Chrome's “Capture full size screenshot” did not work on https://www.tandfonline.com/doi/full/10.1080/15548627.2015.1100356 (extremely long page). Additionally, attempting to “Capture node screenshot” on /html/body simply caused it to crash.
15:48:57<h2ibot>Manu edited Discourse (+47, Active Discourses: Add forum.elivelinux.org): https://wiki.archiveteam.org/?diff=57712&oldid=57681
16:04:28<cruller>Is it difficult to capture elements whose behavior changes depending on whether they are in the visible area? Simple lazy loading aside.
16:08:18<cruller>When converting dynamic things to static ones, the question of “which state to freeze” always arises...
16:13:27<justauser|m>Lazy loading is commonly handled by scrolldown-scrollup-capture pattern I believe?
16:16:18<justauser|m>For harder things, I'd guess I want the invisible state.
16:16:18<justauser|m>Less likely to be animated, for example.
16:26:04<h2ibot>Manu edited Discourse (+46, Active Discourses: Add discuss.gradle.org): https://wiki.archiveteam.org/?diff=57713&oldid=57712
16:31:04<h2ibot>Manu edited Discourse (+47, Active Discourses: Add forum.cursor.com): https://wiki.archiveteam.org/?diff=57714&oldid=57713
16:47:30<cruller>I guess some people prefer to capture every element on the page in the state it appears when inside the viewport.
17:12:52etnguyen03 (etnguyen03) joins
17:29:02NF885 (NF885) joins
17:30:54etnguyen03 quits [Client Quit]
17:33:19Wake quits [Quit: Ping timeout (120 seconds)]
17:40:01etnguyen03 (etnguyen03) joins
18:05:43DogsRNice joins
18:12:40NF885 quits [Client Quit]
18:15:54lennier2_ joins
18:18:34lennier2 quits [Ping timeout: 256 seconds]
18:24:55pseudorizer quits [Quit: ZNC 1.10.1 - https://znc.in]
18:25:37pseudorizer (pseudorizer) joins
18:29:25<Zephyr>any news nicolas17?
18:29:32<Zephyr>sorry to bother you
18:41:40Zachava joins
18:43:41Wake joins
18:50:40Wohlstand quits [Remote host closed the connection]
18:50:56Wohlstand (Wohlstand) joins
19:04:13etnguyen03 quits [Client Quit]
19:13:44Island joins
19:43:10ichdasich quits [Quit: reboot]
19:44:29ichdasich joins
21:08:56<h2ibot>Exorcism edited ArchiveTeam Warrior (+0, v4.1.0 -> v4.2.0): https://wiki.archiveteam.org/?diff=57715&oldid=56980
21:13:09szczot3k quits [Ping timeout: 272 seconds]
21:23:02vitaly-zdanevich joins
21:24:24<vitaly-zdanevich>This old website looks like will die soon, links redirect to a newer version that also last updated in 2022 https://www.game-over.net/reviews.php?id=452&page=reviews
21:27:59<pokechu22>will run it in #archivebot
21:33:33szczot3k (szczot3k) joins
21:35:25VerifiedJ quits [Quit: The Lounge - https://thelounge.chat]
21:35:57VerifiedJ (VerifiedJ) joins
21:42:50etnguyen03 (etnguyen03) joins
21:59:34etnguyen03 quits [Remote host closed the connection]
22:00:46etnguyen03 (etnguyen03) joins
22:02:53<nicolas17>Zephyr: I found no "zephyr" in any URL
22:02:58<nicolas17>the "zephy" false positives were like
22:03:00<nicolas17>com,xfire,video)/116d03-2.jpg 20150620065120 http://video.xfire.com/116d03-2.jpg image/jpeg 200 PTGCHZEPHYV4MYCAS3CHYFRRMAKXOJQK - - 6773 24968964850 archiveteam_xfire_20150622041426/xfire_20150622041426.megawarc.warc.gz
22:03:02<nicolas17>com,xfire,video)/127219-2.jpg 20150620005731 http://video.xfire.com/127219-2.jpg image/jpeg 200 ZT6FD44IQZNMTTKH6WJTMPNCZEPHY234 - - 7489 30895929300 archiveteam_xfire_20150622041437/xfire_20150622041437.megawarc.warc.gz
22:04:26<nicolas17>it seems there's no profiles archived here at all, the ones that do exist on the wayback machine didn't come from archiveteam's project
22:06:49skyrocket quits [Remote host closed the connection]
22:11:01etnguyen03 quits [Client Quit]
22:22:05skyrocket joins
22:54:26etnguyen03 (etnguyen03) joins
22:57:38etnguyen03 quits [Remote host closed the connection]
23:02:06etnguyen03 (etnguyen03) joins
23:33:29etnguyen03 quits [Client Quit]