mån 2007-03-26 klockan 11:40 +0200 skrev Guillaume Smet:
> On 3/26/07, Henrik Nordstrom <henrik@henriknordstrom.net> wrote:
> > One way is to set up a separate set of cache_peer for these robots,
> > using the no-cache cache_peer option to avoid having that traffic
> > cached. Then use cache_peer_access with suitable acls to route the robot
> > requests via these peers and deny them from the other normal set of
> > peers.
>
> AFAICS, it won't solve the problem as the robots won't be able to
> access the "global" cache readonly.
It does.
Squid processing sequence is kind of:
1. accept request
2. http_access etc
3. cache lookup, send response if a cache hit.
4. if cache miss, look for a cache_peer
5. cache response if allowed
Regards
Henrik
This archive was generated by hypermail pre-2.1.9 : Sat Mar 31 2007 - 13:00:02 MDT