diff --git a/adjuster.py b/adjuster.py index 7a72ea644f49e826aefc4be4e3846bc9e4df621b..9623bda0f6b057dfc5d359a12a204ac6b4a3c82c 100755 --- a/adjuster.py +++ b/adjuster.py @@ -305,7 +305,7 @@ define("skipLinkCheck",multiple=True,help="Comma-separated list of regular expre define("extensions",help="Name of a custom Python module to load to handle certain requests; this might be more efficient than setting up a separate Tornado-based server. The module's handle() function will be called with the URL and RequestHandler instance as arguments, and should return True if it processed the request, but anyway it should return as fast as possible. This module does NOT take priority over forwarding the request to fasterServer.") define("loadBalancer",default=False,help="Set this to True if you have a default_site set and you are behind any kind of \"load balancer\" that works by issuing a GET / with no browser string. This option will detect such requests and avoid passing them to the remote site.") -define("multicore",default=False,help="(Linux only) On multi-core CPUs, fork enough processes for all cores to participate in handling incoming requests. This increases RAM usage, but can help with high-load situations. Disabled on BSD/Mac due to unreliability (other cores can still be used for htmlFilter etc)") # and --ssl-fork if there's not TOO many instances taking up the RAM; if you really want multiple cores to handle incoming requests on Mac/BSD you could run GNU/Linux in a virtual machine (or use a WSGI server). Regarding which core gets an incoming connection, linux/net/core/sock_reuseport.c's reuseport_select_sock uses either a random hash (from inet_ehashfn), or some bytecode written in the BPF (Berkeley Packet Filter) language (added to the socket via sk_reuseport_attach_filter() or sk_reuseport_attach_bpf() in linux/filter.h) that returns the index number: this can read data from incoming packets when making its choice, but cannot directly read things like "how many js_interpreter instances are in use" unless it can be frequently reloaded with this information (TODO? would need cores to open their sockets in a known order so the index numbers make sense. NB if any one closes, its index will be replaced by the one that had index num_socks-1 to save memmove) +define("multicore",default=False,help="(Linux only) On multi-core CPUs, fork enough processes for all cores to participate in handling incoming requests. This increases RAM usage, but can help with high-load situations. Disabled on BSD/Mac due to unreliability (other cores can still be used for htmlFilter etc)") # and --ssl-fork if there's not TOO many instances taking up the RAM; if you really want multiple cores to handle incoming requests on Mac/BSD you could run GNU/Linux in a virtual machine (or use a WSGI server). Regarding which core gets an incoming connection, linux/net/core/sock_reuseport.c's reuseport_select_sock uses either a random hash (from inet_ehashfn), or some bytecode written in the BPF (Berkeley Packet Filter) language (see SO_ATTACH_REUSEPORT_ options in man 7 socket) (TODO? would need cores to open their sockets in a known order so the index numbers make sense; may be able to use an eBPF map of len(webdriver_queue)+len(webdriver_runner) rather than reloading) define("compress_responses",default=True,help="Use gzip to compress responses for clients that indicate they are compatible with it. You may want to turn this off if your server's CPU is more important than your network bandwidth (e.g. browser on same machine).") # THIS MUST BE THE LAST SECTION because it continues into