Re: [squid-users] caching websites automatically

From: John Doe <jdmls_at_yahoo.com>
Date: Wed, 11 Feb 2009 01:59:32 -0800 (PST)

From: Leonardo Rodrigues Magalh�es <[email protected]> > lorenor escreveu: > > I'm searching for a method to cache websites automatically with squid. > > The goal is to give squid a list of URLs and the proxy will cache the > > sites. > > I know only one way to cache a site. A client have to make a request. > > But is there another way without client interaction? > No, squid has no mode do that automagically. > But ..... with some linux clients, wget for example, you can easily do that > cd /tmp/garbage > sites.txt should contain URL of the sites you wanna to fetch > www.onesite.com > www.othersite.com > www.anything.com > export http_proxy=http://your.squid.ip.box:3128 > wget -i sites.txt --mirror > that should fetch in mirror style (EVERYTHING) from the informed sites and > save them under the directory you started wget. Depending on the amount of data, > that could take a long time to run. You can probably erase everything after wget > finishes, but it may be intelligent to keep the files and run mirror again some > days after, which will make MUCH less traffic being generated. > and, in the end of the process, squid should have cached everything that is > cacheable according to site configurations and your caching parameters as well. > squid has no automatic mode for doing that, but that can be easily done with > wget. You can use the '--delete-after' option to cleanup automaticaly.. And maybe '--limit-rate' if you don't want to hammer the servers... JD
Received on Wed Feb 11 2009 - 09:59:44 MST

This archive was generated by hypermail 2.2.0 : Wed Feb 11 2009 - 12:00:01 MST