[prev in list] [next in list] [prev in thread] [next in thread] 

List:       wget
Subject:    Re: Question
From:       Micah Cowan <micah () cowan ! name>
Date:       2007-08-07 18:43:09
Message-ID: 46B8BD3D.4020804 () cowan ! name
[Download RAW message or body]

-----BEGIN PGP SIGNED MESSAGE-----
Hash: SHA256

It seems to me that you can simply start a recursive,
non-parent-traversing fetch (-r -np) of the page with the links, and
you'll end up with the PDF files you want (plus anything else linked to
on that page). If the PDF files are stored in different directories on
the website, they'll be in different directories in the fetch;
otherwise, they won't be, and yeah you'd need to write some script to do
what you need (sorry, no samples available).

- -Micah

Andra Isan wrote:
> I have a paper proceeding and I want to follow a link of that proceeding
> and go to a paper link, then follow the paper link and go to author link
> and then follow author link which leads to all the paper that the author
> has written. I want to place all these pdf files( papers of one author)
> into a directory. So, at the end I have directories of all authors
> containing papers that those authors have written. (one directory for
> each author) 
> I am not sure if I can do it with Wget or not.
> 
> */Micah Cowan <micah@cowan.name>/* wrote:
> 
> I don't think your request is very clear. Certainly you can download PDF
> files with Wget. What do you mean by "organize them in a directory"?
> What sort of organization do you want? Please be as specific as
> possible.
-----BEGIN PGP SIGNATURE-----
Version: GnuPG v1.4.6 (GNU/Linux)
Comment: Using GnuPG with Mozilla - http://enigmail.mozdev.org

iD8DBQFGuL097M8hyUobTrERCKqYAJ9/tC05b5+NI2xacmJfNqzQnzZEdgCfY+m7
UbasnhbVBKEk13w82PcJO6Q=
=TeLr
-----END PGP SIGNATURE-----
[prev in list] [next in list] [prev in thread] [next in thread] 

Configure | About | News | Add a list | Sponsored by KoreLogic