02:54:10tzt quits [Read error: Connection reset by peer]
02:55:05tzt (tzt) joins
06:23:13Pedrosso quits [Ping timeout: 272 seconds]
06:23:13balrog quits [Ping timeout: 272 seconds]
06:24:07Pedrosso joins
06:27:15balrog (balrog) joins
06:49:06DogsRNice quits [Read error: Connection reset by peer]
07:31:28Sluggs quits [Ping timeout: 256 seconds]
09:50:46tzt quits [Quit: tzt]
09:50:59tzt (tzt) joins
14:10:53<justauser|m>Proteopedia.org dump almost complete. Lost several images over 500 and regexp errors, but still going to download more than the server thinks it has.
14:12:55<justauser|m>3D images probably not included. I'm not sure how exactly do they work.
14:29:06<justauser|m>Looks like the data is from a huge DB https://www.wwpdb.org and the display software is open-source at https://bioinformatics.org/firstglance/download/.
14:31:01<justauser|m>pabs: Looks like a false positive on part of DWD.
14:31:12<justauser|m>https://www.wiki.balug.org/robots.txt
14:31:40<justauser|m>Only specific pages are disallowed, but the regexp doesn't check for EOL.
14:33:02<justauser|m>I can patch it out, or we may report and wait for a fix.
14:41:42<@arkiver>justauser|m: how do you discover more than the server "thinks it has"?
14:51:54<justauser|m>No idea.
14:52:46<justauser|m>I ran WT3, with some patches. Maybe they caused this, maybe not.
14:53:19<justauser|m>It's possible the statistics page counts the files on disk while listing refers to the DB. I've had some 404s already.
14:53:50<justauser|m>(but only two so far)
15:07:29<justauser|m>Patches were related to the listing, so it's pretty plausible. I'll count actually downloaded images once done.
15:57:52<@arkiver>interesting, thanks
15:58:17<@arkiver>how did you determine in this case how much the server thinks it has?
19:12:47<@JAA>(Please preserve the patches as well.)
21:47:42DogsRNice joins