06:21:00<Vito``>hi, just learned that Fileplanet was being shut down and archived
06:22:00<Vito``>I used to help run Polycount, which used to have all their 3D models hosted there
06:23:00<Vito``>is the best way to find all the files they lost by going through the metadata of the tars?
06:23:00<Vito``>the wiki says all the data on the page is otherwise outdated
08:23:00<godane>good think i backed stillflying.net up: http://fireflyfans.net/mthread.aspx?bid=2&tid=53804
09:25:00<schbirid1>Vito``: #fireplanet :)
09:25:00<schbirid1>Vito``: we have not uploaded much yet
09:25:00<schbirid1>we will have a nice interface some day
09:25:00<schbirid1>but actually not for the polycount stuff (because that is from the older planet* hostisng and people put private files up their spaces
09:25:00<schbirid1>we got ALL the files so we cannot publish that
09:26:00<schbirid1>i am trying to host it so that if you know a path, you can download it (no public index)
09:26:00<schbirid1>that should prevent privacy issues
09:30:00<schbirid1>Vito``: i have the whole planetquake stuff locally on my machine,so if you need a specific file, just shout
09:30:00<schbirid1>i thought the models were mirrored by others already though, eg leileilol
09:51:00<Vito``>schbirid1: if I compiled a list of paths you have them locally
09:51:00<Vito``>?
09:53:00<schbirid1>yeah
11:44:00<hiker1>http://www.familyguyonline.com/ is shutting down Jan. 18. Might be worth grabbing whatever is one the site now, to remember the game.
11:44:00<hiker1>they will probably redirect the domain eventually.
15:05:00<no2pencil>Merry Christmas Archivers!!
15:31:00<hiker1>Is there a tutorial somewhere on how to use wget for different sites?
15:37:00<Nemo_bis>hiker1: there are wget examples on the page of many services on our wiki
15:38:00<hiker1>What do you mean?
15:38:00<hiker1>Can you give me an example?
15:45:00<tef>wget -r -nH np -Amp3 --cut-dirs=1 http://foo.com/~tef/filez
15:46:00<tef>makes a directory 'filez' with all the mp3s it found
15:46:00<tef>-r - recursive, follow links
15:46:00<tef>-nH - don't make a directory for the host (foo.com)
15:46:00<tef>-np - don't go to a parent directory
15:46:00<tef>--cut-dirs=1 strip '~tef' from the path
15:46:00<tef>-Amp3 - only save mp3s
15:48:00<hiker1>That doesn't use warc output.
15:50:00<Deewiant>http://www.archiveteam.org/index.php?title=Wget_with_WARC_output#Usage
15:50:00<tef>oh
15:51:00<hiker1>I was using like rewrite urls
15:51:00<hiker1>and some other commands
15:51:00<hiker1>It varies so much by website
15:54:00<hiker1>and to download all the sites prerequisites
17:54:00<hiker1>Is it possible to append to a warc file?
17:55:00<hiker1>or append to a wget mirror?
17:55:00<hiker1>The site I mirrored apparently uses a subdomain, but I used the --no-parent argument.
17:57:00<hiker1>I also used --convert-links, but it did not convert links to the subdomain.
19:09:00<schbirid1>hiker2: from what i know, no
19:09:00<schbirid1>you can use -c but iirc it does not work too well with -m usualy
19:10:00<hiker2>Someone in here mentioned they grabbed all the urls from a site before actually downloading the site. Is this possible? useful?
19:21:00<schbirid1>depends on the website
19:21:00<schbirid1>you can use --spider
19:21:00<schbirid1>BUT that will download, just not store
19:21:00<hiker2>When would that be useful?
19:21:00<schbirid1>if you have no space and want to find out about the site structure
19:22:00<schbirid1>or if you are just interested in the URLs, not the data
19:22:00<hiker2>It seems that since wget has no way to continue warc downloads, it would be useful to create a program that does.
19:22:00<hiker2>*can
19:23:00<hiker2>wget doesn't seem particularly well-suited to download complete mirrors of websites.
19:32:00<schbirid1>it could be better for sure
19:32:00<schbirid1>also eats memory :(
19:32:00<schbirid1>there is heretix which archive.org uses but i never tried that
19:32:00<hiker2>httrack as well
19:32:00<hiker2>but I don't think it supports WARC
19:33:00<schbirid1>i have had awful results with httrack
19:34:00<hiker2>someone wrote http://code.google.com/p/httrack2arc/
19:34:00<hiker2>which converts httrack to ARC format
19:34:00<hiker2>When I used HTTrack it worked for what I needed.
19:34:00<hiker2>I think it resumes too
20:04:00<ersi>Too bad it's running on a retarded operating system with a crappy file system that's case insensitive
20:06:00<schbirid1>httrack is on linux too
20:07:00<ersi>Huh, didn't know that
20:21:00<SketchCow>MERRY CHRISTMAS ARCHIVE TEAM
20:21:00<SketchCow>JESUS SAVES AND SO DO WE
20:23:00<SmileyG>\o/
20:35:00<ersi>http://i.imgur.com/Jek9D.jpg
21:07:00<rubita>http://www.carolinaherrera.com/212/es/areyouonthelist?share=2zkuHzwOxvy930fvZN7HOVc97XE-GNOL1fzysCqIoynkz4rz3EUUdzs6j6FXsjB4447F-isvxjqkXd4Qey2GHw#teaser
21:14:00<rubita>http://www.carolinaherrera.com/212/es/areyouonthelist?share=XTv1etZcVd-19S-VT5m1-oIXWSwtlJ3dj4ARKTLVwK7kz4rz3EUUdzs6j6FXsjB4447F-isvxjqkXd4Qey2GHw#episodio-1
21:24:00<SketchCow>BUT MY EXPENSIVE UNWANTED THING
21:25:00<chronomex>I like how the first thing to load on that page is a php error
22:44:00<tef>heretrix isn't that good :v
22:59:00<SketchCow>what, in general?
23:01:00<ersi>I guess in the context of the earlier conversation, ie for a random-person-grab-site-expedition
23:10:00<tef>SketchCow: well, it's a million lines of code, kinda interweaved. it sorta does the job though
23:10:00<tef>my impression from picking through it trying to find out the format ideosynchrasies of ARC made me unhappy
23:11:00<tef>at work we use something like phantomjs + mitmproxy to dump warcs.
23:17:00<tef>don't get me wrong, i haven't had to use it in anger, but wget should perform just as well, considering it likely has very similar crawling logic
23:20:00<hiker2>Is there a way to get wget to download external images?
23:20:00<hiker2>like from tinypic.