I have been using squid-cache at home on my firewall for years now, I
use it for the normal standard old stuff of simply caching of where we
surf.
I am writing a new kiosk based software package that has a GUI app as
the front end for the operator and apache is serving up the pages to
the web browser clients. What is being served up are images. The
images that come into the GUI app are full size images, 4 megapixel on
up.
Current what I am doing is after my software copies the full size
images into the computer, it then creates the two different web
images, one is a small thumbnail (120x180) the other is a larger image
for the screen (400x600). When one is doing this to 200 images at one
time, it takes a while, too long in my opinion.
My first though was to have the indexing page detect if the smaller
images where there and create them, page by page, and then save the
smaller image so that next time it was snappy. Then it dawned on me:
Isn't that was things like squid-cache do? Cache these processed
files?
So the question is: Is squid-cache (on Windows) the right tool to
cache these images? I know that apache can be setup as a cache, but I
don't know anything about that. Will I be better off using apache?
The other question I have to ask someone, more myself then anyone...
Am I making this too complicated by adding a proxy along with the web
server?
Thoughts and opinions?
-- The contents of this e-mail are intended for the named addressee only. It contains information that may be confidential. Unless you are the named addressee or an authorized designee, you may not copy or use it, or disclose it to anyone else. If you received it in error please notify us immediately and then destroy it.Received on Wed Mar 14 2007 - 11:36:27 MDT
This archive was generated by hypermail pre-2.1.9 : Sat Mar 31 2007 - 13:00:02 MDT