load balancing
    David Cuthbert 
    dacut at kanga.org
       
    Fri Nov  3 23:37:45 PST 2006
    
    
  
1103073858.GB829 at xxxxxxxxxxxxxxxxx>
In-Reply-To: <20061103073858.GB829 at xxxxxxxxxxxxxxxxx>
Content-Type: text/plain; charset=ISO-8859-1; format=flowed
Content-Transfer-Encoding: 7bit
Lines: 22
Message-ID: <454c434a$0$788$415eb37d at xxxxxxxxxxxxxxxxxxxxxxxxxxxxxx>
NNTP-Posting-Host: 67.182.133.151
X-Trace: 1162625866 crater_reader.dragonflybsd.org 788 67.182.133.151
Xref: crater_reader.dragonflybsd.org dragonfly.users:7942
Joerg Sonnenberger wrote:
> Only if you use a broken web server. With a proper O(1) event
> notification mechanism and async web servers, keep-alive is a huge win,
> if you don't have a single server load high enough to run out of file
> descriptors.
Heh... well:
- We only recently got proper O(1) event notifications,
- We definitely don't have an asynchronous server, and
- During the peak, our onlines run pretty darn close to full capacity.
To expand on the second point a bit: when you make a request to our 
server, that thread is tied up for the duration of your request.  Each 
server has a fixed number of threads, so tying one of these up to handle 
keep-alive would be a big deal.  Our way of gracefully handling a 
request which holds the thread too long is to just kill it (and log tons 
of errors and page a few engineers).  Yeah, ugly.
Even if we were to fix the thread issue, there's enough custom legacy 
crap which assumes a request-per-thread model that lots of stuff would 
break.  I'd like to think we're unique in being this crufty, but the 
anecdotal tales I hear lead me to believe this is about par... :-(
    
    
More information about the Users
mailing list