00:00:03Hackerpcs (Hackerpcs) joins
00:21:36cyan_box joins
00:22:19Arcorann_ joins
00:22:57Mateon1 quits [Ping timeout: 272 seconds]
00:24:18Mateon1 joins
00:25:29cyanbox_ quits [Ping timeout: 272 seconds]
00:26:07Arcorann quits [Ping timeout: 272 seconds]
00:52:20etnguyen03 quits [Client Quit]
01:17:38<pabs>datechnoman: got any URLs for *.in-addr.arpa *.ip6.arpa ? for https://cabforum.org/2025/11/10/ballot-sc-086v3-sunset-the-inclusion-of-ip-reverse-address-domain-names/
01:18:52etnguyen03 (etnguyen03) joins
01:31:16etnguyen03 quits [Client Quit]
01:35:56etnguyen03 (etnguyen03) joins
01:48:27Mateon1 quits [Ping timeout: 272 seconds]
01:49:29Mateon1 joins
02:01:45Mateon1 quits [Ping timeout: 272 seconds]
02:02:45Mateon1 joins
02:23:06Suika quits [Quit: Server is ded]
02:23:55Suika joins
02:25:11mr_sarge quits [Ping timeout: 272 seconds]
02:25:18mr_sarge (sarge) joins
03:31:45Kotomind_ joins
03:40:33Kotomind_ quits [Ping timeout: 272 seconds]
03:41:55PredatorIWD256 joins
03:43:46PredatorIWD25 quits [Ping timeout: 256 seconds]
03:43:46PredatorIWD256 is now known as PredatorIWD25
04:02:02BennyOtt quits [Quit: ZNC 1.10.1 - https://znc.in]
04:04:23BennyOtt (BennyOtt) joins
04:05:21etnguyen03 quits [Client Quit]
04:05:50etnguyen03 (etnguyen03) joins
04:13:18etnguyen03 quits [Remote host closed the connection]
04:33:13nexussfan quits [Read error: Connection reset by peer]
04:33:52Island quits [Read error: Connection reset by peer]
05:04:48n9nes quits [Ping timeout: 256 seconds]
05:05:08n9nes joins
05:06:50SootBector quits [Remote host closed the connection]
05:08:01SootBector (SootBector) joins
06:04:05nexussfan (nexussfan) joins
06:23:57nexussfan quits [Client Quit]
06:54:20Wohlstand (Wohlstand) joins
06:57:55hexagonwin (hexagonwin) joins
07:00:11SootBector quits [Remote host closed the connection]
07:01:20SootBector (SootBector) joins
07:09:05sec^nd quits [Remote host closed the connection]
07:09:39sec^nd (second) joins
07:11:39s-crypt quits [Quit: Ping timeout (120 seconds)]
07:11:51s-crypt (s-crypt) joins
08:22:30Shard114 (Shard) joins
08:22:37Wohlstand quits [Client Quit]
08:23:08Shard11 quits [Ping timeout: 256 seconds]
08:23:08Shard114 is now known as Shard11
08:31:59FireFly quits [Quit: Upgrading...]
08:58:29nine quits [Quit: See ya!]
08:58:42nine joins
08:58:42nine quits [Changing host]
08:58:42nine (nine) joins
09:58:31Kotomind_ joins
10:02:15Kotomind_ quits [Client Quit]
10:03:43gogo joins
10:03:48Webuser064005 joins
10:03:57Webuser064005 quits [Client Quit]
10:12:15<Guest>does urls-grab only get websites on the public internet (not tor)?
10:15:51<@imer>Guest: yep
10:22:55<Guest>then how did the warrior end up downloading CSAM? this isnt something i ran into but someone posted about it happening to them in #archiveteam. like what would have happened if it was a residential connection vs hetzner? i wouldnt want my door busted down for running an AT project D:
10:26:02<@imer>as far as we can tell, it's a honeypot - so maybe a site that was previously csam related. urls does grab the front page of domains it sees (and can follow links down based on some heuristics), so if it was linked somewhere else it'll discover it that way
10:27:21<@imer>this is also why we generally don't recommend people run urls
10:29:24<@imer>as for the legal side of things, I'm not a lawyer. I do think that if something like that would happen, we would be able to conclusively prove it wasn't the person, but the (automated) warrior downloading these things on their network
10:30:06<@imer>I don't think aside from occasional abuse emails (most of which are bogus) we've had any issues
10:32:13gogo quits [Ping timeout: 272 seconds]
10:35:45simon816 quits [Quit: ZNC 1.10.1 - https://znc.in]
10:40:42simon816 (simon816) joins
10:49:59<Guest>got it
11:11:29ats quits [Ping timeout: 272 seconds]
11:15:13ats (ats) joins
11:53:10RadRooster quits [Remote host closed the connection]
12:00:01Bleo182600722719623455222 quits [Quit: The Lounge - https://thelounge.chat]
12:02:50Bleo182600722719623455222 joins
12:25:29Dada joins
12:40:27Gadelhas562873784438 quits [Quit: Ping timeout (120 seconds)]
12:43:50irisfreckles13 joins
13:22:58Dada quits [Remote host closed the connection]
13:24:30Dada joins
13:51:11<Yakov>https://www.apkmirror.com/apk/discord/discord-chat-for-gamers/ Has anyone seen this before? Discord name is redacted in apkmirror because of a DMCA...
13:52:46<Yakov>It's really weird that apkmirror would do that yet still redistribute their apk
14:10:32<TheTechRobo>I do think the URLs project should have a stronger warning in the Warrior web UI than just "WARNING IP BLOCK LISTS" given that things like this can happen
14:11:48<klea>Is there some project to grab tor websites?
14:18:04Webuser178095 joins
14:18:08Webuser178095 quits [Client Quit]
14:27:11sec^nd quits [Remote host closed the connection]
14:27:39sec^nd (second) joins
14:31:37Arcorann_ quits [Ping timeout: 272 seconds]
14:35:20<TheTechRobo>There was, but I don't remember what happened to it; it hasn't been running for awhile
15:25:45gosc joins
15:26:06<gosc>apparently amazon disabled downloading mobile apps on everything but an amazon fire tablet last year
15:26:25<gosc>does anyone own a fire tablet/fire tv?
15:27:00stepney141 quits [Ping timeout: 256 seconds]
15:27:08<gosc>there's a whole bunch of free apps on there I wanted to grab, which are very obscure and still downloadable
15:27:34<schwarzkatz|m>Yakov: I have seen that before, it has been like that for some time I think
15:30:34stepney141 (stepney141) joins
15:31:57<gosc>here's the apps I need to grab, I'm kind of busy so I'll just leave them here: https://www.amazon.com/s?i=mobile-apps&rh=p_4%3AJTWebMan&search-type=ss
15:32:00<gosc>and https://www.amazon.com/aibee-Rose-Guns-Days-Season1/dp/B00ANFT1NE/
15:32:17<gosc>I understand if no one wants to pay for the second one but the first link has entirely free apps
15:36:50<gosc>actually wait, this would make the amazon appstore endangered by archiveteam standards
15:36:52<gosc>let me add that
15:46:07<h2ibot>Calmevening edited Alive... OR ARE THEY (+281): https://wiki.archiveteam.org/?diff=60368&oldid=60359
16:06:16<justauser>blog.archive.today now doesn't resolve in many places and https://archive-is.tumblr.com/ stopped redirecting. Has two recent posts.
16:06:45<anarcat>quick, archive it
16:06:47<anarcat>oh shit
16:07:40<justauser>Already SPN'd the latter.
16:08:03<justauser>Doing something with AT as whole is planned for ages, AFAIK?
16:09:08<justauser>!ao'd the redirect.
16:16:39<Guest>https://transfer.archivete.am/EtrtB/blogat_spn.png
16:16:40<eggdrop>inline (for browser viewing): https://transfer.archivete.am/inline/EtrtB/blogat_spn.png
16:16:49<Guest>(spn fails to resolve domain)
16:18:29DogsRNice joins
16:19:11<justauser>klea: Pair.com? Don't tell me their security@ is a mailing list.
16:19:52<klea>justauser: i sent a email to webmaster@marcdashevsky.com not to pair.com, maybe they handle mail via mailing list?
16:19:56<klea>would be annoying probably.
16:21:02<justauser>The bounce is from pair.com, and I still remember the rush to save their mailing lists. Thus the conclusion.
16:21:14<justauser>Nothing certain, but a good suspicion.
16:24:48<schwarzkatz|m>does the WBM care a lot about query params? Say, I request ?page=2&foo=bar but only ?page=2 is saved. would I get an answer?
16:24:48<schwarzkatz|m>background: a website I am collecting links for pollutes almost all pagination with some random parameters that do not disappear again from the url after visiting another page; I am thinking about just removing them but am not sure if that is okay
16:24:59lucifer_sam quits [Ping timeout: 272 seconds]
16:26:12<justauser>Almost exact match.
16:26:31<justauser>Details on the wiki.
16:26:41<justauser>However, the opposite sort of works.
16:27:28<justauser>If you go to ?page=2, but only ?page=2&foo=bar is saved, it'll suggest to show pages with this prefix and you'll see the saved one.
16:28:12<justauser>Does it mean URLs grow infinitely?
16:29:40<schwarzkatz|m>justauser: hm, and is that also depending on the order? in this case the page params gets added at the end :/
16:30:11<justauser>Meh.
16:30:22<justauser>Another case of user-hostility.
16:30:25<schwarzkatz|m>justauser: luckily no, once they are added they stay there
16:30:39<justauser>That's exactly what I meant.
16:30:56<justauser>If parameters are only added and never removed while you browse...
16:31:06<justauser>..this could explode.
16:32:11<schwarzkatz|m>at least the nav bar on top has hardlinks :D
16:32:11<justauser>arkiver: Two days left for opendiary. Are you still hoping to do something?
16:32:40<schwarzkatz|m>I think I'll just keep them then. thanks! the website in question is https://www.animepro.de/anima/test/dvd-serien in case you are wondering
16:32:49<justauser>I'd say strip.
16:33:09<justauser>Only save a single canonical URL for a page.
16:35:36<schwarzkatz|m>oh? from what you said it sounded like this would break the links :o
16:36:16<justauser>It would, but from what you say they are unpredictable to an outsider anyways.
16:36:44<justauser>If we were doing a crawl, it would make some sense to have every internal link good.
16:37:08<justauser>Unless it results in hundreds of duplicated captures, taht is.
16:37:43<schwarzkatz|m>oh, I'm so sorry, the context is important of course
16:37:43<schwarzkatz|m>am collecting the links for archivebot
16:38:14<justauser>So !ao-ing page list instead of letting it crawl?
16:39:02<schwarzkatz|m>yeah, I always was under the impression that is more accurate, no?
16:42:03<justauser>It depends (tm).
16:42:25<justauser>It's more effort and a risk of making a dumb mistake.
16:42:58<justauser>Has to be used when the site is uncrawlable, but this doesn't seem to be the case here?
16:44:11<justauser>Clicked around a bit, doesn't seem that bad in terms of possible duplicate captures.
16:45:14<justauser>It also creates some links to nothing, but you can't tell without fetching that it's one.
16:45:48<justauser>Or does it?
16:46:05<schwarzkatz|m>I see... in that case an automated crawl of that site would be nice! I don't think there will be any updates to the site in the forseeable future
16:46:20<schwarzkatz|m>justauser: do you have an example?
16:48:10<justauser>I suspected https://www.animepro.de/anima/test/dvd-serien?pimcore_request_source=staticroute&controller=%40AppBundle\Controller\AnimaController&bundle=AppBundle&action=list&filter=Z&page=3 , but it actually works fine.
16:48:47<justauser>Despite pointing to page number 3 out of 1. There is probably some workaround server-side.
16:49:21<schwarzkatz|m>ah, it also keeps the page param when filtering :D
16:50:35gogo joins
17:01:29<schwarzkatz|m>I think this would create a lot of duplicates. should I manually collect the links then? or is blacklisting the &filter urls a good option here?
17:03:23<justauser>I suspect this will cause us to miss things.
17:03:43<justauser>How much duplicates do you expect?
17:04:09<justauser>Perhaps we can start the crawl from an URL that's already saturated with possible parameters?
17:06:43<schwarzkatz|m>justauser: I am not sure how archivebot crawls a page, but if we get a &filter=A to Z for every single page without the filter (for every category of article) it would be quite a lot
17:07:24<justauser>Uhhh...
17:09:32Kotomind joins
17:10:24<schwarzkatz|m>I suspect it to be x25 of useless links
17:11:33<justauser>I ran out of good ideas. Maybe someone with more experience has some.
17:16:17ThreeHM quits [Ping timeout: 272 seconds]
17:18:05ThreeHM (ThreeHeadedMonkey) joins
17:26:58gosc quits [Client Quit]
17:49:55Cuphead2527480 (Cuphead2527480) joins
18:11:55irisfreckles13 quits [Read error: Connection reset by peer]
18:32:29<h2ibot>Justauser edited List of websites excluded from the Wayback Machine/Partial exclusions (+67, Added g2a.com, cryptomus.com - hinted by…): https://wiki.archiveteam.org/?diff=60369&oldid=60345
18:37:08gosc joins
18:39:53Kotomind quits [Ping timeout: 272 seconds]
18:45:42itachi1706 quits [Quit: Bye :P]
18:47:19gosc quits [Client Quit]
18:51:40itachi1706 (itachi1706) joins
18:57:32<h2ibot>Justauser edited List of websites excluded from the Wayback Machine (+28, Added prodcreators.com - hint by archiveis@twitter): https://wiki.archiveteam.org/?diff=60370&oldid=60360
19:02:49<justauser>https://hunch.ly/ - another archival service. Commercial, closed.
19:03:11Wohlstand1 (Wohlstand) joins
19:05:33Wohlstand1 is now known as Wohlstand
19:13:49<@arkiver>justauser: not sure if there's public archives?
19:14:28<@arkiver>justauser: getting it ready now
19:14:39<@arkiver>6 million posts, hopefully enoughtime
19:19:37<schwarzkatz|m>good luck!
19:19:43<@arkiver>yeah :)
19:28:38<h2ibot>KleaBot made 2 bot changes: https://wiki.archiveteam.org/index.php?title=Special:Contributions/KleaBot&offset=20260129192830&limit=2&namespace=2&wpfilters[]=nsInvert&wpfilters[]=associated
19:30:01<@arkiver>imer: when you are around, can we have a target for opendiary?
19:30:02<@arkiver>with
19:30:02gogo quits [Client Quit]
19:30:06<@arkiver>opendiary_
19:30:10<@arkiver>archiveteam_opendiary_
19:30:15<@arkiver>Archive Team Open Diary:
19:30:27<@arkiver>will very likely not be huge at all
19:46:26croissant_ joins
19:49:56croissant quits [Ping timeout: 256 seconds]
19:51:41<h2ibot>Manuel edited Discourse/uncategorized (-53, Remove mozilla discourse, because it is already…): https://wiki.archiveteam.org/?diff=60373&oldid=60196
19:51:42<h2ibot>Ljcool2006 edited KamTape (-3): https://wiki.archiveteam.org/?diff=60374&oldid=58118
19:51:43<h2ibot>Brad edited Deathwatch (+243, added soldaat van oranje): https://wiki.archiveteam.org/?diff=60375&oldid=60308
19:55:41<h2ibot>Arkiver uploaded File:Opendiary-icon.png: https://wiki.archiveteam.org/?title=File%3AOpendiary-icon.png
19:55:54Kabaya joins
19:56:13gogo joins
19:59:03Kabaya3 quits [Ping timeout: 272 seconds]
20:07:39<klea>nulldata: could you add an entry in the deathwatch for trinket?
20:13:44gogo quits [Client Quit]
20:14:29gogo joins
20:16:57Webuser195119 joins
20:17:38Webuser195119 quits [Client Quit]
20:28:21lflare quits [Killed (vindobona.hackint.org (Nickname regained by services))]
20:28:23lflare (lflare) joins
20:39:46Cuphead2527480 quits [Client Quit]
20:47:45<@arkiver>opendiary project started, it's an easy site
20:47:56<@arkiver>whenever there is a target, items will statr finishing
20:48:16<@imer>soon
20:48:33<klea>i wonder why the process to make targets is manual, it should all be automatable
20:48:37<klea>ie, put file in a single monorepo
20:48:44<klea>edit wiki
20:48:45<klea>kaboom
20:49:33<@imer>it is largely automated (thanks to rew.by), just gotta add the naming and title arkiver gives me and run the automation to deploy the target
20:49:56<@arkiver>imer: yep, if you can't do it now, no worries (though the deadline is a bit close)
20:49:59<@imer>there's some tuning if we know the project will be big and server selection which would be harder to automate
20:50:02<@arkiver>but doesn't have to be in this single minute right now :P
20:50:04<@imer>targets up
20:50:06<@arkiver>wooh
20:50:08<@arkiver>:)
20:50:17<@imer>was already doing it when you sent your message
20:51:48<that_lurker>are the targets running on the new colocation server(s) or somewhere else?
20:52:01<@imer>drone poked as well
20:53:21<klea>is there some way for us boring users to check which target server was choosen, or there's no way?
20:54:10Wohlstand quits [Client Quit]
20:54:19<@imer>that_lurker: elsewhere, no real changes there in a while. most small projects are going to end up on n905na which is a hetzner box. the other main target is optane10 which is closer to IA physically, but has small disks
20:54:34<@imer>klea: there isn't unless you run the project and look at the logs
20:54:42<klea>oh
20:54:44<klea>annoying.
20:55:08<klea>i guess it doesn't matter too much :p
20:57:30<@imer>arkiver: we seem to have overloaded the site
20:57:39<@arkiver>imer: yeah
20:57:44<@arkiver>well it's responding but slow
20:58:13<that_lurker>get DPoS'd :-P
21:06:35corentin quits [Quit: Ping timeout (120 seconds)]
21:07:01corentin joins
21:07:19Island joins
21:07:53<klea>wait, did arkiver accidentally switch the automatic warrior project to it, and DPoS them and have it bork, so changed again?
21:09:46BearFortress quits []
21:15:46v01d joins
21:24:44gogo quits [Client Quit]
21:26:16v01d quits [Ping timeout: 256 seconds]
21:31:40<PC>Yakov: i saw the apkmirror thing too, yeah. been a thing for a couple months at least
21:37:10lunik1 quits [Quit: :x]
21:37:35lunik1 joins
21:50:53BearFortress joins
21:59:18Sluggs (Sluggs) joins
21:59:18Sluggs quits [Excess Flood]
22:02:47gogo joins
22:02:59Sluggs (Sluggs) joins
22:04:19<IDK>0 on https://www.opendiary.com/
22:04:31<IDK>banned before I even started? 😭
22:04:32<datechnoman>Website is dead lol
22:04:44<datechnoman>I cant even load it from my home connection
22:04:54<datechnoman>Run on a potato?
22:05:43<IDK>Amazon potato in particular
22:05:57<klea>APotato
22:06:26<IDK>website runs on ec2 instance, we probably sent their aws bills to the 12th digit
22:06:35<klea>smh
22:08:02<IDK>honestly, who knows if their entire backend is just a series of ec2 instances
22:08:37<@imer>should calm down once the warriors finish up the items they started.. eventually..
22:08:53Sk1d joins
22:09:45<@imer>gonna set limit to 0 for a bit
22:09:45<@imer>!remindme 10min
22:09:46<eggdrop>[remind] ok, i'll remind you at 2026-01-29T22:19:45Z
22:13:05<@imer>"Open Diary will cease operations on February 28th, 2026 and the site will no longer be available. At that time, all servers will cease operations. We have extended this deadline to allow additional time for members to download their diaries."
22:13:05<@imer>more time woo
22:14:34<IDK>lol, the singular ec2 server behind the opendiary website accepts ssh with password, no keys required
22:15:00<h2ibot>Imer edited Deathwatch (+48, /* 2026 */): https://wiki.archiveteam.org/?diff=60377&oldid=60375
22:15:23<@imer>did indeed not put what I did, bad imer
22:19:47<eggdrop>[remind] imer:
22:20:12<@imer>!remindme 10min lift opendiary limit?
22:20:12<eggdrop>[remind] ok, i'll remind you at 2026-01-29T22:30:12Z
22:30:12<eggdrop>[remind] imer: lift opendiary limit?
22:30:28<@imer>looking better, but not quite fast yet
22:30:31<@imer>!remindme 10min lift opendiary limit?
22:30:31<eggdrop>[remind] ok, i'll remind you at 2026-01-29T22:40:31Z
22:32:31Radzig quits [Quit: ZNC 1.10.1 - https://znc.in]
22:37:22<@imer>limit to 120/min
22:39:31Radzig joins
22:39:50Arcorann_ joins
22:40:32<eggdrop>[remind] imer: lift opendiary limit?
22:41:08<@imer>-> 300/min
22:42:07<@imer>too much. 180/min
22:45:51<@JAA>Re archive.today blog, https://archive-is.tumblr.com/ looks fine to me now.
22:46:07<@JAA>Both blog.archive.today and blog.archive.ph redirect there.
22:49:46<@imer>down to 120/min again, still slow
22:50:08PredatorIWD25 quits [Quit: Ping timeout (120 seconds)]
22:50:20PredatorIWD25 joins
23:00:21nexussfan (nexussfan) joins
23:03:29sec^nd quits [Remote host closed the connection]
23:04:42<@imer>seems to have completely died again, i did turn it down to 60/min as well. so unsure. i'm heading off for today, ~1s load time of the front page was normal for reference
23:05:12<@imer>ah, lots of warrior returns so maybe a long sleep expired
23:05:15sec^nd (second) joins
23:05:53mrminemeet quits [Ping timeout: 272 seconds]
23:05:57mrminemeet_ joins
23:11:53Kabaya quits [Read error: Connection reset by peer]
23:11:54Kabaya2 joins
23:16:40atphoenix_ (atphoenix) joins
23:18:23<nicolas17>imer: what's the limit now?
23:18:44<nicolas17>I see completion at 200/min
23:19:49atphoenix__ quits [Ping timeout: 272 seconds]
23:20:42<nicolas17>I think items take too long
23:21:33<nicolas17>maybe disabling multiitems would let you better observe the consequences of rate limit changes
23:22:30<nicolas17>still seeing 200/min
23:22:37<nicolas17>despite the limit being at 60/min