mirror of https://github.com/jumpserver/jumpserver
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
538 lines
14 KiB
538 lines
14 KiB
#!/usr/bin/env python3 |
|
# coding: utf-8 |
|
|
|
import os |
|
import subprocess |
|
import threading |
|
import datetime |
|
import logging |
|
import logging.handlers |
|
import time |
|
import argparse |
|
import sys |
|
import shutil |
|
import signal |
|
from collections import defaultdict |
|
import daemon |
|
from daemon import pidfile |
|
|
|
BASE_DIR = os.path.dirname(os.path.abspath(__file__)) |
|
sys.path.insert(0, BASE_DIR) |
|
|
|
try: |
|
from apps.jumpserver import const |
|
__version__ = const.VERSION |
|
except ImportError as e: |
|
logging.info("Not found __version__: {}".format(e)) |
|
logging.info("Sys path: {}".format(sys.path)) |
|
logging.info("Python is: ") |
|
logging.info(subprocess.call('which python', shell=True)) |
|
__version__ = 'Unknown' |
|
try: |
|
import apps |
|
logging.info("List apps: {}".format(os.listdir('apps'))) |
|
logging.info('apps is: {}'.format(apps)) |
|
except: |
|
pass |
|
|
|
try: |
|
from apps.jumpserver.conf import load_user_config |
|
CONFIG = load_user_config() |
|
except ImportError as e: |
|
logging.info("Import error: {}".format(e)) |
|
logging.info("Could not find config file, `cp config_example.yml config.yml`") |
|
sys.exit(1) |
|
|
|
os.environ["PYTHONIOENCODING"] = "UTF-8" |
|
APPS_DIR = os.path.join(BASE_DIR, 'apps') |
|
LOG_DIR = os.path.join(BASE_DIR, 'logs') |
|
TMP_DIR = os.path.join(BASE_DIR, 'tmp') |
|
HTTP_HOST = CONFIG.HTTP_BIND_HOST or '127.0.0.1' |
|
HTTP_PORT = CONFIG.HTTP_LISTEN_PORT or 8080 |
|
WS_PORT = CONFIG.WS_LISTEN_PORT or 8082 |
|
DEBUG = CONFIG.DEBUG or False |
|
LOG_LEVEL = CONFIG.LOG_LEVEL or 'INFO' |
|
|
|
START_TIMEOUT = 40 |
|
WORKERS = 4 |
|
DAEMON = False |
|
LOG_KEEP_DAYS = 7 |
|
logging.basicConfig( |
|
format='%(asctime)s %(message)s', level=logging.INFO, |
|
datefmt='%Y-%m-%d %H:%M:%S' |
|
) |
|
|
|
EXIT_EVENT = threading.Event() |
|
LOCK = threading.Lock() |
|
files_preserve = [] |
|
STOP_TIMEOUT = 10 |
|
|
|
logger = logging.getLogger() |
|
|
|
try: |
|
os.makedirs(os.path.join(BASE_DIR, "data", "static")) |
|
os.makedirs(os.path.join(BASE_DIR, "data", "media")) |
|
except: |
|
pass |
|
|
|
|
|
def check_database_connection(): |
|
os.chdir(os.path.join(BASE_DIR, 'apps')) |
|
for i in range(60): |
|
logging.info("Check database connection ...") |
|
code = subprocess.call("python manage.py showmigrations users ", shell=True) |
|
if code == 0: |
|
logging.info("Database connect success") |
|
return |
|
time.sleep(1) |
|
logging.info("Connection database failed, exist") |
|
sys.exit(10) |
|
|
|
|
|
def make_migrations(): |
|
logging.info("Check database structure change ...") |
|
os.chdir(os.path.join(BASE_DIR, 'apps')) |
|
logging.info("Migrate model change to database ...") |
|
subprocess.call('python3 manage.py migrate', shell=True) |
|
|
|
|
|
def collect_static(): |
|
logging.info("Collect static files") |
|
os.chdir(os.path.join(BASE_DIR, 'apps')) |
|
command = 'python3 manage.py collectstatic --no-input -c &> /dev/null ' |
|
subprocess.call(command, shell=True) |
|
logging.info("Collect static files done") |
|
|
|
|
|
def prepare(): |
|
check_database_connection() |
|
make_migrations() |
|
collect_static() |
|
|
|
|
|
def check_pid(pid): |
|
""" Check For the existence of a unix pid. """ |
|
try: |
|
os.kill(pid, 0) |
|
except OSError: |
|
return False |
|
else: |
|
return True |
|
|
|
|
|
def get_pid_file_path(s): |
|
return os.path.join(TMP_DIR, '{}.pid'.format(s)) |
|
|
|
|
|
def get_log_file_path(s): |
|
return os.path.join(LOG_DIR, '{}.log'.format(s)) |
|
|
|
|
|
def get_pid_from_file(path): |
|
if os.path.isfile(path): |
|
with open(path) as f: |
|
try: |
|
return int(f.read().strip()) |
|
except ValueError: |
|
return 0 |
|
return 0 |
|
|
|
|
|
def get_pid(s): |
|
pid_file = get_pid_file_path(s) |
|
return get_pid_from_file(pid_file) |
|
|
|
|
|
def is_running(s, unlink=True): |
|
pid_file = get_pid_file_path(s) |
|
|
|
if os.path.isfile(pid_file): |
|
pid = get_pid(s) |
|
if pid == 0: |
|
return False |
|
elif check_pid(pid): |
|
return True |
|
|
|
if unlink: |
|
os.unlink(pid_file) |
|
return False |
|
|
|
|
|
def parse_service(s): |
|
all_services = [ |
|
'gunicorn', 'celery_ansible', 'celery_default', |
|
'beat', 'flower', 'daphne', |
|
] |
|
if s == 'all': |
|
return all_services |
|
elif s == "web": |
|
return ['gunicorn', 'flower', 'daphne'] |
|
elif s == 'ws': |
|
return ['daphne'] |
|
elif s == "task": |
|
return ["celery_ansible", "celery_default", "beat"] |
|
elif s == "celery": |
|
return ["celery_ansible", "celery_default"] |
|
elif "," in s: |
|
services = set() |
|
for i in s.split(','): |
|
services.update(parse_service(i)) |
|
return services |
|
else: |
|
return [s] |
|
|
|
|
|
def get_start_gunicorn_kwargs(): |
|
print("\n- Start Gunicorn WSGI HTTP Server") |
|
prepare() |
|
bind = '{}:{}'.format(HTTP_HOST, HTTP_PORT) |
|
log_format = '%(h)s %(t)s "%(r)s" %(s)s %(b)s ' |
|
|
|
cmd = [ |
|
'gunicorn', 'jumpserver.wsgi', |
|
'-b', bind, |
|
'-k', 'gthread', |
|
'--threads', '10', |
|
'-w', str(WORKERS), |
|
'--max-requests', '4096', |
|
'--access-logformat', log_format, |
|
'--access-logfile', '-' |
|
] |
|
|
|
if DEBUG: |
|
cmd.append('--reload') |
|
return {'cmd': cmd, 'cwd': APPS_DIR} |
|
|
|
|
|
def get_start_daphne_kwargs(): |
|
print("\n- Start Daphne ASGI WS Server") |
|
cmd = [ |
|
'daphne', 'jumpserver.asgi:application', |
|
'-b', HTTP_HOST, |
|
'-p', str(WS_PORT), |
|
] |
|
return {'cmd': cmd, 'cwd': APPS_DIR} |
|
|
|
|
|
def get_start_celery_ansible_kwargs(): |
|
print("\n- Start Celery as Distributed Task Queue: Ansible") |
|
return get_start_worker_kwargs('ansible', 4) |
|
|
|
|
|
def get_start_celery_default_kwargs(): |
|
print("\n- Start Celery as Distributed Task Queue: Celery") |
|
return get_start_worker_kwargs('celery', 2) |
|
|
|
|
|
def get_start_worker_kwargs(queue, num): |
|
# Todo: Must set this environment, otherwise not no ansible result return |
|
os.environ.setdefault('PYTHONOPTIMIZE', '1') |
|
os.environ.setdefault('ANSIBLE_FORCE_COLOR', 'True') |
|
|
|
if os.getuid() == 0: |
|
os.environ.setdefault('C_FORCE_ROOT', '1') |
|
|
|
cmd = [ |
|
'celery', 'worker', |
|
'-A', 'ops', |
|
'-l', 'INFO', |
|
'-c', str(num), |
|
'-Q', queue, |
|
'-n', '{}@%h'.format(queue) |
|
] |
|
return {"cmd": cmd, "cwd": APPS_DIR} |
|
|
|
|
|
def get_start_flower_kwargs(): |
|
print("\n- Start Flower as Task Monitor") |
|
if os.getuid() == 0: |
|
os.environ.setdefault('C_FORCE_ROOT', '1') |
|
|
|
cmd = [ |
|
'celery', 'flower', |
|
'-A', 'ops', |
|
'-l', 'INFO', |
|
'--url_prefix=flower', |
|
'--auto_refresh=False', |
|
'--max_tasks=1000', |
|
'--tasks_columns=uuid,name,args,state,received,started,runtime,worker' |
|
] |
|
return {"cmd": cmd, "cwd": APPS_DIR} |
|
|
|
|
|
def get_start_beat_kwargs(): |
|
print("\n- Start Beat as Periodic Task Scheduler") |
|
os.environ.setdefault('PYTHONOPTIMIZE', '1') |
|
if os.getuid() == 0: |
|
os.environ.setdefault('C_FORCE_ROOT', '1') |
|
|
|
scheduler = "django_celery_beat.schedulers:DatabaseScheduler" |
|
cmd = [ |
|
'celery', 'beat', |
|
'-A', 'ops', |
|
'-l', 'INFO', |
|
'--scheduler', scheduler, |
|
'--max-interval', '60' |
|
] |
|
return {"cmd": cmd, 'cwd': APPS_DIR} |
|
|
|
|
|
processes = {} |
|
|
|
|
|
def watch_services(): |
|
max_retry = 3 |
|
services_retry = defaultdict(int) |
|
stopped_services = {} |
|
|
|
def check_services(): |
|
now = datetime.datetime.now().strftime('%Y-%m-%d %H:%M:%S') |
|
for s, p in processes.items(): |
|
print("{} Check service status: {} -> ".format(now, s), end='') |
|
try: |
|
p.wait(timeout=1) |
|
except subprocess.TimeoutExpired: |
|
pass |
|
ok = is_running(s) |
|
if not ok: |
|
stopped_services[s] = '' |
|
print("stopped with code: {}({})".format(p.returncode, p.pid)) |
|
else: |
|
print("running at {}".format(p.pid)) |
|
stopped_services.pop(s, None) |
|
services_retry.pop(s, None) |
|
|
|
def retry_start_stopped_services(): |
|
for s in stopped_services: |
|
if services_retry[s] > max_retry: |
|
logging.info("Service start failed, exit: ", s) |
|
EXIT_EVENT.set() |
|
break |
|
|
|
p = start_service(s) |
|
logging.info("> Find {} stopped, retry {}, {}".format( |
|
s, services_retry[s] + 1, p.pid) |
|
) |
|
processes[s] = p |
|
services_retry[s] += 1 |
|
|
|
def rotate_log_if_need(): |
|
now = datetime.datetime.now() |
|
tm = now.strftime('%H:%M') |
|
if tm != '23:59': |
|
return |
|
suffix = now.strftime('%Y-%m-%d') |
|
services = list(processes.keys()) |
|
services.append('jms') |
|
|
|
for s in services: |
|
log_path = get_log_file_path(s) |
|
log_dir = os.path.dirname(log_path) |
|
filename = os.path.basename(log_path) |
|
pre_log_dir = os.path.join(log_dir, suffix) |
|
if not os.path.exists(pre_log_dir): |
|
os.mkdir(pre_log_dir) |
|
|
|
pre_log_path = os.path.join(pre_log_dir, filename) |
|
if os.path.isfile(log_path) and not os.path.isfile(pre_log_path): |
|
logging.info("Rotate log file: {} => {}".format(log_path, pre_log_path)) |
|
shutil.copy(log_path, pre_log_path) |
|
with open(log_path, 'w') as f: |
|
pass |
|
some_days_ago = now - datetime.timedelta(days=LOG_KEEP_DAYS) |
|
days_ago_dir = os.path.join(LOG_DIR, some_days_ago.strftime('%Y-%m-%d')) |
|
if os.path.exists(days_ago_dir): |
|
logger.info("Remove old log: {}".format(days_ago_dir)) |
|
shutil.rmtree(days_ago_dir, ignore_errors=True) |
|
|
|
while not EXIT_EVENT.is_set(): |
|
try: |
|
with LOCK: |
|
check_services() |
|
retry_start_stopped_services() |
|
rotate_log_if_need() |
|
time.sleep(30) |
|
except KeyboardInterrupt: |
|
time.sleep(1) |
|
break |
|
clean_up() |
|
|
|
|
|
def start_service(s): |
|
services_kwargs = { |
|
"gunicorn": get_start_gunicorn_kwargs, |
|
"celery_ansible": get_start_celery_ansible_kwargs, |
|
"celery_default": get_start_celery_default_kwargs, |
|
"beat": get_start_beat_kwargs, |
|
"flower": get_start_flower_kwargs, |
|
"daphne": get_start_daphne_kwargs, |
|
} |
|
|
|
kwargs = services_kwargs.get(s)() |
|
pid_file = get_pid_file_path(s) |
|
|
|
if os.path.isfile(pid_file): |
|
os.unlink(pid_file) |
|
cmd = kwargs.pop('cmd') |
|
|
|
log_file_path = get_log_file_path(s) |
|
log_file_f = open(log_file_path, 'a') |
|
files_preserve.append(log_file_f) |
|
kwargs['stderr'] = log_file_f |
|
kwargs['stdout'] = log_file_f |
|
p = subprocess.Popen(cmd, **kwargs) |
|
with open(pid_file, 'w') as f: |
|
f.write(str(p.pid)) |
|
return p |
|
|
|
|
|
def start_services_and_watch(s): |
|
logging.info(time.ctime()) |
|
logging.info('Jumpserver version {}, more see https://www.jumpserver.org'.format( |
|
__version__) |
|
) |
|
|
|
services_set = parse_service(s) |
|
for i in services_set: |
|
if is_running(i): |
|
show_service_status(i) |
|
continue |
|
p = start_service(i) |
|
time.sleep(2) |
|
processes[i] = p |
|
|
|
if not DAEMON: |
|
watch_services() |
|
else: |
|
show_service_status(s) |
|
context = get_daemon_context() |
|
with context: |
|
watch_services() |
|
|
|
|
|
def get_daemon_context(): |
|
daemon_pid_file = get_pid_file_path('jms') |
|
daemon_log_f = open(get_log_file_path('jms'), 'a') |
|
files_preserve.append(daemon_log_f) |
|
context = daemon.DaemonContext( |
|
pidfile=pidfile.TimeoutPIDLockFile(daemon_pid_file), |
|
signal_map={ |
|
signal.SIGTERM: lambda x, y: clean_up(), |
|
signal.SIGHUP: 'terminate', |
|
}, |
|
stdout=daemon_log_f, |
|
stderr=daemon_log_f, |
|
files_preserve=files_preserve, |
|
detach_process=True, |
|
) |
|
return context |
|
|
|
|
|
def stop_service(srv, sig=15): |
|
services_set = parse_service(srv) |
|
for s in services_set: |
|
if not is_running(s): |
|
show_service_status(s) |
|
continue |
|
print("Stop service: {}".format(s), end='') |
|
pid = get_pid(s) |
|
os.kill(pid, sig) |
|
for i in range(STOP_TIMEOUT): |
|
if i == STOP_TIMEOUT - 1: |
|
print("\033[31m Error\033[0m") |
|
if not is_running(s): |
|
print("\033[32m Ok\033[0m") |
|
break |
|
else: |
|
time.sleep(1) |
|
continue |
|
|
|
with LOCK: |
|
processes.pop(s, None) |
|
|
|
if srv == "all": |
|
stop_daemon_service() |
|
|
|
|
|
def stop_daemon_service(): |
|
pid = get_pid('jms') |
|
if pid and check_pid(pid): |
|
os.kill(pid, 15) |
|
|
|
|
|
def stop_multi_services(services): |
|
for s in services: |
|
stop_service(s, sig=9) |
|
|
|
|
|
def stop_service_force(s): |
|
stop_service(s, sig=9) |
|
|
|
|
|
def clean_up(): |
|
if not EXIT_EVENT.is_set(): |
|
EXIT_EVENT.set() |
|
processes_dump = {k: v for k, v in processes.items()} |
|
for s1, p1 in processes_dump.items(): |
|
stop_service(s1) |
|
p1.wait() |
|
|
|
|
|
def show_service_status(s): |
|
services_set = parse_service(s) |
|
for ns in services_set: |
|
if is_running(ns): |
|
pid = get_pid(ns) |
|
print("{} is running: {}".format(ns, pid)) |
|
else: |
|
print("{} is stopped".format(ns)) |
|
|
|
|
|
if __name__ == '__main__': |
|
parser = argparse.ArgumentParser( |
|
description=""" |
|
Jumpserver service control tools; |
|
|
|
Example: \r\n |
|
|
|
%(prog)s start all -d; |
|
""" |
|
) |
|
parser.add_argument( |
|
'action', type=str, |
|
choices=("start", "stop", "restart", "status"), |
|
help="Action to run" |
|
) |
|
parser.add_argument( |
|
"service", type=str, default="all", nargs="?", |
|
choices=("all", "web", "task", "gunicorn", "celery", "beat", "celery,beat", "flower", "ws"), |
|
help="The service to start", |
|
) |
|
parser.add_argument('-d', '--daemon', nargs="?", const=1) |
|
parser.add_argument('-w', '--worker', type=int, nargs="?", const=4) |
|
parser.add_argument('-f', '--force', nargs="?", const=1) |
|
args = parser.parse_args() |
|
if args.daemon: |
|
DAEMON = True |
|
|
|
if args.worker: |
|
WORKERS = args.worker |
|
|
|
action = args.action |
|
srv = args.service |
|
|
|
if action == "start": |
|
start_services_and_watch(srv) |
|
os._exit(0) |
|
elif action == "stop": |
|
if args.force: |
|
stop_service_force(srv) |
|
else: |
|
stop_service(srv) |
|
elif action == "restart": |
|
DAEMON = True |
|
stop_service(srv) |
|
time.sleep(5) |
|
start_services_and_watch(srv) |
|
else: |
|
show_service_status(srv)
|
|
|