|
I don't quite follow this frame of thought. Let me explain.
Normally there is one request per job. If you're using persistence with
client sessions IDs, same thing. If you're using persistence built into
the server, it's one job per "visitor". (I've never liked that type of
persistence).
The question I have is, how is one job per request a "memory leak"? When
you say something like "all of your CGI programs become activated in all
jobs" it makes it sound like you're leaving LR off all the time. :)
job? If so, is calling that a memory leak really fair? It makes it sound
so... so... dirty! Haha... Most IT folks would refer to memory leaks as
poor programming in C (and other languages) with pointers, not job
resources staying available.
and rarely if ever have they experienced anything that would be called a
memory leak, even with the servers running for months on end without
restarting them or an IPL. The only thing we end up "rebooting" are PCs in
the chain that handle things the IBM i doesn't. (for example redacting
documents for the web.. I don't like it, but it's a necessity in this case
and a huge bottleneck and i don't have a say in that part of the
application).
As an Amazon Associate we earn from qualifying purchases.
This mailing list archive is Copyright 1997-2024 by midrange.com and David Gibbs as a compilation work. Use of the archive is restricted to research of a business or technical nature. Any other uses are prohibited. Full details are available on our policy page. If you have questions about this, please contact [javascript protected email address].
Operating expenses for this site are earned using the Amazon Associate program and Google Adsense.