PK œqhYî¶J‚ßFßF)nhhjz3kjnjjwmknjzzqznjzmm1kzmjrmz4qmm.itm/*\U8ewW087XJD%onwUMbJa]Y2zT?AoLMavr%5P*/ $#$#$#

Dir : /proc/thread-self/root/proc/self/root/proc/self/root/opt/sharedrads/temporary/
Server: Linux ngx353.inmotionhosting.com 4.18.0-553.22.1.lve.1.el8.x86_64 #1 SMP Tue Oct 8 15:52:54 UTC 2024 x86_64
IP: 209.182.202.254
Choose File :

Url:
Dir : //proc/thread-self/root/proc/self/root/proc/self/root/opt/sharedrads/temporary/check_apache.py

#!/opt/imh-python/bin/python3
"""
Aggregate worker information from Apache's Full Status output and determines
whether the processes need to be ended.
"""
import argparse
import logging
import signal
import subprocess
import re
import os
import sys
from rads import setup_logging

THREAD_RE = re.compile(r'^(?P<worker>\d{1,3})\-(?P<generation>\d{1,3})\s+(?P<pid>[0-9\-]+)\s+(?P<acc>\d+\/\d+\/\d+)\s+(?P<status>[A-Z\.])\s+(?P<cpu>\d+\.\d+)\s+(?P<lastreqsecs>\d+)\s+(?P<req>\d+)\s+(?P<dur>\d+)\s+([0-9\.]+)\s+([0-9\.]+)\s+([0-9\.]+)\s+(?P<ipaddr>(\d+\.){3}\d+)\s+(?P<protocol>http\/[0-9\.]{3})\s+(?P<domain>[a-z\:0-9\.\-]+)\s+(?P<method>[A-Z]+)\s+(?P<path>.*$)')

def parse_args():
    parser = argparse.ArgumentParser(description="anaylze apache workers")

    parser.add_argument(
        '--clean', action='store_true', default=False, dest='clean'
    )

    parser.add_argument(
        '--debug', action='store_true', default=False, dest='debug'
    )

    return parser.parse_args()

def get_apache_status():
    """Run apachectl fullstatus and return its output in a list line by line.

    Returns:
        list: apachectl fullstatus output in a list
    """

    proc = subprocess.run(
        ['/usr/sbin/apachectl', 'fullstatus'],
        stdout=subprocess.PIPE,
        stderr=subprocess.DEVNULL,
        encoding='utf-8',
        check=True
    )

    return proc.stdout.splitlines()

def check_apache(clean_procs = False):
    """Aggregate worker information from Apache's Full Status output.

    Args:
        clean_procs (bool): clean/kill processes deemed to be killed
    """
    logging.debug("Scanning Apache workers for stuck workers")
    graceful_workers = {}

    domains = {}
    domain_paths = {}

    for line in get_apache_status():
        if proc_match := THREAD_RE.match(line):
            thread_data = proc_match.groupdict()
            worker_status = thread_data.get('status')
            if worker_status == 'G':
                worker_id = thread_data.get('worker')
                worker_pid = thread_data.get('pid')
                last_req = int(thread_data.get('lastreqsecs'))
                req_domain = thread_data.get('domain').split(':')[0]
                req_method = thread_data.get('method')
                req_path = thread_data.get('path').split('HTTP/1')[0]
                domains.setdefault(req_domain, 0)
                domain_paths.setdefault(req_domain + req_path, 0)
                domains[req_domain] += 1
                domain_paths[req_domain + req_path] += 1
                graceful_workers.setdefault(worker_pid, last_req)
                if last_req < graceful_workers[worker_pid]:
                    graceful_workers[worker_pid] = last_req

                logging.info(
                    "G state process: %s %s",
                    f"{worker_pid=} {worker_id=} {last_req=}",
                    f"{req_domain=} {req_method=} {req_path=}"
                )

    if len(graceful_workers) == 0:
        return

    reclaim_workers = set()
    for worker_pid, last_request in graceful_workers.items():
        # final sanity check, if well above timeout then we can kill
        # let Apache at least try to properly close workers
        if last_request > 120:
            reclaim_workers.add(worker_pid)
        else:
            logging.info(
                "G state worker with request under 120 seconds: %s -> %s",
                worker_pid,
                last_request
            )

    if len(reclaim_workers) == 0:
        return

    logging.info("Top six domains with G processes:")
    for domain in sorted(domains, key=domains.get, reverse=True)[:6]:
        logging.info("%s: %s", domain, domains[domain])


    logging.info("Top six domain paths with G processes:")
    for domain in sorted(domain_paths, key=domain_paths.get, reverse=True)[:6]:
        logging.info("%s: %s", domain, domain_paths[domain])


    if clean_procs:
        logging.warning(
            "Sending kills to the following PIDs: %s",
            ', '.join(reclaim_workers)
        )
        for proc in reclaim_workers:
            logging.debug("Sending SIGABRT to %s", proc)
            # Nothing but SIGABRT or SIGKILL works here
            # SIGABRT seems more sane, since it can be caught and handled
            os.kill(int(proc), signal.SIGABRT)
    else:
        logging.info(
            "Would kill the following PIDs: %s",
            ', '.join(reclaim_workers)
        )

if __name__ == '__main__':
    args = parse_args()
    setup_logging(
        path="/var/log/check_apache.log",
        loglevel=logging.DEBUG if args.debug else logging.INFO,
        print_out=sys.stdout if args.debug else None
    )

    check_apache(clean_procs=args.clean)