mirror of
https://gitlab.com/crafty-controller/crafty-4.git
synced 2024-08-30 18:23:09 +00:00
871 lines
36 KiB
Python
871 lines
36 KiB
Python
import os
|
|
import time
|
|
import logging
|
|
import threading
|
|
import asyncio
|
|
import datetime
|
|
import json
|
|
from zoneinfo import ZoneInfoNotFoundError
|
|
from tzlocal import get_localzone
|
|
from apscheduler.events import EVENT_JOB_EXECUTED
|
|
from apscheduler.jobstores.base import JobLookupError
|
|
from apscheduler.schedulers.background import BackgroundScheduler
|
|
from apscheduler.triggers.cron import CronTrigger
|
|
|
|
from app.classes.models.management import HelpersManagement
|
|
from app.classes.models.users import HelperUsers
|
|
from app.classes.controllers.users_controller import UsersController
|
|
from app.classes.shared.console import Console
|
|
from app.classes.shared.file_helpers import FileHelpers
|
|
from app.classes.shared.helpers import Helpers
|
|
from app.classes.shared.main_controller import Controller
|
|
from app.classes.web.tornado_handler import Webserver
|
|
from app.classes.shared.websocket_manager import WebSocketManager
|
|
|
|
logger = logging.getLogger("apscheduler")
|
|
command_log = logging.getLogger("cmd_queue")
|
|
scheduler_intervals = {
|
|
"seconds",
|
|
"minutes",
|
|
"hours",
|
|
"days",
|
|
"weeks",
|
|
"monday",
|
|
"tuesday",
|
|
"wednesday",
|
|
"thursday",
|
|
"friday",
|
|
"saturday",
|
|
"sunday",
|
|
}
|
|
|
|
|
|
class TasksManager:
|
|
controller: Controller
|
|
|
|
def __init__(self, helper, controller, file_helper):
|
|
self.helper: Helpers = helper
|
|
self.controller: Controller = controller
|
|
self.tornado: Webserver = Webserver(helper, controller, self, file_helper)
|
|
try:
|
|
self.tz = get_localzone()
|
|
except ZoneInfoNotFoundError as e:
|
|
logger.error(
|
|
"Could not capture time zone from system. Falling back to Europe/London"
|
|
f" error: {e}"
|
|
)
|
|
self.tz = "Europe/London"
|
|
self.scheduler = BackgroundScheduler(timezone=str(self.tz))
|
|
|
|
self.users_controller: UsersController = self.controller.users
|
|
|
|
self.webserver_thread = threading.Thread(
|
|
target=self.tornado.run_tornado, daemon=True, name="tornado_thread"
|
|
)
|
|
|
|
self.main_thread_exiting = False
|
|
|
|
self.schedule_thread = threading.Thread(
|
|
target=self.scheduler_thread, daemon=True, name="scheduler"
|
|
)
|
|
|
|
self.log_watcher_thread = threading.Thread(
|
|
target=self.log_watcher, daemon=True, name="log_watcher"
|
|
)
|
|
|
|
self.command_thread = threading.Thread(
|
|
target=self.command_watcher, daemon=True, name="command_watcher"
|
|
)
|
|
|
|
self.realtime_thread = threading.Thread(
|
|
target=self.realtime, daemon=True, name="realtime"
|
|
)
|
|
|
|
self.reload_schedule_from_db()
|
|
|
|
def get_main_thread_run_status(self):
|
|
return self.main_thread_exiting
|
|
|
|
def reload_schedule_from_db(self):
|
|
jobs = HelpersManagement.get_schedules_enabled()
|
|
logger.info("Reload from DB called. Current enabled schedules: ")
|
|
for item in jobs:
|
|
logger.info(f"JOB: {item}")
|
|
|
|
def command_watcher(self):
|
|
while True:
|
|
# select any commands waiting to be processed
|
|
command_log.debug(
|
|
"Queue currently has "
|
|
f"{self.controller.management.command_queue.qsize()} queued commands."
|
|
)
|
|
if not self.controller.management.command_queue.empty():
|
|
command_log.info(
|
|
"Current queued commands: "
|
|
f"{list(self.controller.management.command_queue.queue)}"
|
|
)
|
|
cmd = self.controller.management.command_queue.get()
|
|
try:
|
|
svr = self.controller.servers.get_server_instance_by_id(
|
|
cmd["server_id"]
|
|
)
|
|
except:
|
|
logger.error(
|
|
f"Server value {cmd['server_id']} requested does not exist! "
|
|
"Purging item from waiting commands."
|
|
)
|
|
continue
|
|
|
|
user_id = cmd["user_id"]
|
|
command = cmd["command"]
|
|
|
|
if command == "start_server":
|
|
svr.run_threaded_server(user_id)
|
|
|
|
elif command == "stop_server":
|
|
svr.stop_threaded_server()
|
|
|
|
elif command == "restart_server":
|
|
svr.restart_threaded_server(user_id)
|
|
|
|
elif command == "kill_server":
|
|
try:
|
|
svr.kill()
|
|
time.sleep(5)
|
|
svr.cleanup_server_object()
|
|
svr.record_server_stats()
|
|
except Exception as e:
|
|
logger.error(
|
|
f"Could not find PID for requested termsig. Full error: {e}"
|
|
)
|
|
|
|
elif command == "backup_server":
|
|
svr.server_backup_threader(cmd["action_id"])
|
|
|
|
elif command == "update_executable":
|
|
svr.jar_update()
|
|
else:
|
|
svr.send_command(command)
|
|
|
|
time.sleep(1)
|
|
|
|
def _main_graceful_exit(self):
|
|
try:
|
|
os.remove(self.helper.session_file)
|
|
self.controller.servers.stop_all_servers()
|
|
except:
|
|
logger.info("Caught error during shutdown", exc_info=True)
|
|
try:
|
|
temp_dir = os.path.join(self.controller.project_root, "temp")
|
|
FileHelpers.del_dirs(temp_dir)
|
|
except:
|
|
logger.info(
|
|
"Caught error during shutdown - "
|
|
"unable to delete files from Crafty Temp Dir",
|
|
exc_info=True,
|
|
)
|
|
|
|
logger.info("***** Crafty Shutting Down *****\n\n")
|
|
Console.info("***** Crafty Shutting Down *****\n\n")
|
|
self.main_thread_exiting = True
|
|
|
|
def start_webserver(self):
|
|
self.webserver_thread.start()
|
|
|
|
def reload_webserver(self):
|
|
self.tornado.stop_web_server()
|
|
Console.info("Waiting 3 seconds")
|
|
time.sleep(3)
|
|
self.webserver_thread = threading.Thread(
|
|
target=self.tornado.run_tornado, daemon=True, name="tornado_thread"
|
|
)
|
|
self.start_webserver()
|
|
|
|
def stop_webserver(self):
|
|
self.tornado.stop_web_server()
|
|
|
|
def start_scheduler(self):
|
|
logger.info("Launching Scheduler Thread...")
|
|
Console.info("Launching Scheduler Thread...")
|
|
self.schedule_thread.start()
|
|
logger.info("Launching command thread...")
|
|
Console.info("Launching command thread...")
|
|
self.command_thread.start()
|
|
logger.info("Launching log watcher...")
|
|
Console.info("Launching log watcher...")
|
|
self.log_watcher_thread.start()
|
|
logger.info("Launching realtime thread...")
|
|
Console.info("Launching realtime thread...")
|
|
self.realtime_thread.start()
|
|
|
|
def scheduler_thread(self):
|
|
schedules = HelpersManagement.get_schedules_enabled()
|
|
self.scheduler.add_listener(self.schedule_watcher, mask=EVENT_JOB_EXECUTED)
|
|
self.scheduler.start()
|
|
self.check_for_updates()
|
|
self.scheduler.add_job(
|
|
self.check_for_updates,
|
|
"interval",
|
|
hours=12,
|
|
id="update_watcher",
|
|
start_date=datetime.datetime.now(),
|
|
)
|
|
self.scheduler.add_job(
|
|
self.controller.write_auth_tracker,
|
|
"interval",
|
|
minutes=5,
|
|
id="auth_tracker_write",
|
|
start_date=datetime.datetime.now(),
|
|
)
|
|
# self.scheduler.add_job(
|
|
# self.scheduler.print_jobs, "interval", seconds=10, id="-1"
|
|
# )
|
|
|
|
# load schedules from DB
|
|
for schedule in schedules:
|
|
if schedule.interval != "reaction":
|
|
new_job = "error"
|
|
if schedule.cron_string != "":
|
|
try:
|
|
new_job = self.scheduler.add_job(
|
|
self.controller.management.queue_command,
|
|
CronTrigger.from_crontab(
|
|
schedule.cron_string, timezone=str(self.tz)
|
|
),
|
|
id=str(schedule.schedule_id),
|
|
args=[
|
|
{
|
|
"server_id": schedule.server_id.server_id,
|
|
"user_id": self.users_controller.get_id_by_name(
|
|
"system"
|
|
),
|
|
"command": schedule.command,
|
|
"action_id": schedule.action_id,
|
|
}
|
|
],
|
|
)
|
|
except Exception as e:
|
|
new_job = "error"
|
|
Console.error(f"Failed to schedule task with error: {e}.")
|
|
Console.warning("Removing failed task from DB.")
|
|
logger.error(f"Failed to schedule task with error: {e}.")
|
|
logger.warning("Removing failed task from DB.")
|
|
# remove items from DB if task fails to add to apscheduler
|
|
self.controller.management_helper.delete_scheduled_task(
|
|
schedule.schedule_id
|
|
)
|
|
else:
|
|
if schedule.interval_type == "hours":
|
|
new_job = self.scheduler.add_job(
|
|
self.controller.management.queue_command,
|
|
"cron",
|
|
minute=0,
|
|
hour="*/" + str(schedule.interval),
|
|
id=str(schedule.schedule_id),
|
|
args=[
|
|
{
|
|
"server_id": schedule.server_id.server_id,
|
|
"user_id": self.users_controller.get_id_by_name(
|
|
"system"
|
|
),
|
|
"command": schedule.command,
|
|
"action_id": schedule.action_id,
|
|
}
|
|
],
|
|
)
|
|
elif schedule.interval_type == "minutes":
|
|
new_job = self.scheduler.add_job(
|
|
self.controller.management.queue_command,
|
|
"cron",
|
|
minute="*/" + str(schedule.interval),
|
|
id=str(schedule.schedule_id),
|
|
args=[
|
|
{
|
|
"server_id": schedule.server_id.server_id,
|
|
"user_id": self.users_controller.get_id_by_name(
|
|
"system"
|
|
),
|
|
"command": schedule.command,
|
|
"action_id": schedule.action_id,
|
|
}
|
|
],
|
|
)
|
|
elif schedule.interval_type == "days":
|
|
curr_time = schedule.start_time.split(":")
|
|
new_job = self.scheduler.add_job(
|
|
self.controller.management.queue_command,
|
|
"cron",
|
|
day="*/" + str(schedule.interval),
|
|
hour=curr_time[0],
|
|
minute=curr_time[1],
|
|
id=str(schedule.schedule_id),
|
|
args=[
|
|
{
|
|
"server_id": schedule.server_id.server_id,
|
|
"user_id": self.users_controller.get_id_by_name(
|
|
"system"
|
|
),
|
|
"command": schedule.command,
|
|
"action_id": schedule.action_id,
|
|
}
|
|
],
|
|
)
|
|
if new_job != "error":
|
|
task = self.controller.management.get_scheduled_task_model(
|
|
int(new_job.id)
|
|
)
|
|
self.controller.management.update_scheduled_task(
|
|
task.schedule_id,
|
|
{
|
|
"next_run": str(
|
|
new_job.next_run_time.strftime("%m/%d/%Y, %H:%M:%S")
|
|
)
|
|
},
|
|
)
|
|
jobs = self.scheduler.get_jobs()
|
|
logger.info("Loaded schedules. Current enabled schedules: ")
|
|
for item in jobs:
|
|
logger.info(f"JOB: {item}")
|
|
|
|
def schedule_job(self, job_data):
|
|
sch_id = HelpersManagement.create_scheduled_task(
|
|
job_data["server_id"],
|
|
job_data["action"],
|
|
job_data["interval"],
|
|
job_data["interval_type"],
|
|
job_data["start_time"],
|
|
job_data["command"],
|
|
job_data["name"],
|
|
job_data["enabled"],
|
|
job_data["one_time"],
|
|
job_data["cron_string"],
|
|
job_data["parent"],
|
|
job_data["delay"],
|
|
job_data["action_id"],
|
|
)
|
|
|
|
# Checks to make sure some doofus didn't actually make the newly
|
|
# created task a child of itself.
|
|
if (
|
|
str(job_data["parent"]) == str(sch_id)
|
|
or job_data["interval_type"] != "reaction"
|
|
):
|
|
HelpersManagement.update_scheduled_task(sch_id, {"parent": None})
|
|
|
|
# Check to see if it's enabled and is not a chain reaction.
|
|
if job_data["enabled"] and job_data["interval_type"] != "reaction":
|
|
# Lets make sure this can not be mistaken for a reaction
|
|
job_data["parent"] = None
|
|
new_job = "error"
|
|
if job_data["cron_string"] != "":
|
|
try:
|
|
new_job = self.scheduler.add_job(
|
|
self.controller.management.queue_command,
|
|
CronTrigger.from_crontab(
|
|
job_data["cron_string"], timezone=str(self.tz)
|
|
),
|
|
id=str(sch_id),
|
|
args=[
|
|
{
|
|
"server_id": job_data["server_id"],
|
|
"user_id": self.users_controller.get_id_by_name(
|
|
"system"
|
|
),
|
|
"command": job_data["command"],
|
|
"action_id": job_data["action_id"],
|
|
}
|
|
],
|
|
)
|
|
except Exception as e:
|
|
new_job = "error"
|
|
Console.error(f"Failed to schedule task with error: {e}.")
|
|
Console.warning("Removing failed task from DB.")
|
|
logger.error(f"Failed to schedule task with error: {e}.")
|
|
logger.warning("Removing failed task from DB.")
|
|
# remove items from DB if task fails to add to apscheduler
|
|
self.controller.management_helper.delete_scheduled_task(sch_id)
|
|
else:
|
|
if job_data["interval_type"] == "hours":
|
|
new_job = self.scheduler.add_job(
|
|
self.controller.management.queue_command,
|
|
"cron",
|
|
minute=0,
|
|
hour="*/" + str(job_data["interval"]),
|
|
id=str(sch_id),
|
|
args=[
|
|
{
|
|
"server_id": job_data["server_id"],
|
|
"user_id": self.users_controller.get_id_by_name(
|
|
"system"
|
|
),
|
|
"command": job_data["command"],
|
|
"action_id": job_data["action_id"],
|
|
}
|
|
],
|
|
)
|
|
elif job_data["interval_type"] == "minutes":
|
|
new_job = self.scheduler.add_job(
|
|
self.controller.management.queue_command,
|
|
"cron",
|
|
minute="*/" + str(job_data["interval"]),
|
|
id=str(sch_id),
|
|
args=[
|
|
{
|
|
"server_id": job_data["server_id"],
|
|
"user_id": self.users_controller.get_id_by_name(
|
|
"system"
|
|
),
|
|
"command": job_data["command"],
|
|
"action_id": job_data["action_id"],
|
|
}
|
|
],
|
|
)
|
|
elif job_data["interval_type"] == "days":
|
|
curr_time = job_data["start_time"].split(":")
|
|
new_job = self.scheduler.add_job(
|
|
self.controller.management.queue_command,
|
|
"cron",
|
|
day="*/" + str(job_data["interval"]),
|
|
hour=curr_time[0],
|
|
minute=curr_time[1],
|
|
id=str(sch_id),
|
|
args=[
|
|
{
|
|
"server_id": job_data["server_id"],
|
|
"user_id": self.users_controller.get_id_by_name(
|
|
"system"
|
|
),
|
|
"command": job_data["command"],
|
|
"action_id": job_data["action_id"],
|
|
}
|
|
],
|
|
)
|
|
logger.info("Added job. Current enabled schedules: ")
|
|
jobs = self.scheduler.get_jobs()
|
|
if new_job != "error":
|
|
task = self.controller.management.get_scheduled_task_model(
|
|
int(new_job.id)
|
|
)
|
|
self.controller.management.update_scheduled_task(
|
|
task.schedule_id,
|
|
{"next_run": new_job.next_run_time.strftime("%m/%d/%Y, %H:%M:%S")},
|
|
)
|
|
for item in jobs:
|
|
logger.info(f"JOB: {item}")
|
|
return task.schedule_id
|
|
|
|
def remove_all_server_tasks(self, server_id):
|
|
schedules = HelpersManagement.get_schedules_by_server(server_id)
|
|
for schedule in schedules:
|
|
if schedule.interval != "reaction":
|
|
self.remove_job(schedule.schedule_id)
|
|
|
|
def remove_job(self, sch_id):
|
|
job = HelpersManagement.get_scheduled_task_model(sch_id)
|
|
for schedule in HelpersManagement.get_child_schedules(sch_id):
|
|
self.controller.management_helper.update_scheduled_task(
|
|
schedule.schedule_id, {"parent": None}
|
|
)
|
|
self.controller.management_helper.delete_scheduled_task(sch_id)
|
|
if job.enabled and job.interval_type != "reaction":
|
|
self.scheduler.remove_job(str(sch_id))
|
|
logger.info(f"Job with ID {sch_id} was deleted.")
|
|
else:
|
|
logger.info(
|
|
f"Job with ID {sch_id} was deleted from DB, but was not enabled."
|
|
f"Not going to try removing something "
|
|
f"that doesn't exist from active schedules."
|
|
)
|
|
|
|
def update_job(self, sch_id, job_data):
|
|
# Checks to make sure some doofus didn't actually make the newly
|
|
# created task a child of itself.
|
|
interval_type = job_data.get("interval_type")
|
|
if str(job_data.get("parent")) == str(sch_id) or interval_type != "reaction":
|
|
job_data["parent"] = None
|
|
HelpersManagement.update_scheduled_task(sch_id, job_data)
|
|
|
|
if not (
|
|
"interval" in job_data
|
|
and "enabled" in job_data
|
|
and "cron_string" in job_data
|
|
and "interval_type" in job_data
|
|
):
|
|
if not "enabled" in job_data:
|
|
return
|
|
|
|
if job_data["enabled"] is True:
|
|
job_data = HelpersManagement.get_scheduled_task(sch_id)
|
|
job_data["server_id"] = job_data["server_id"]["server_id"]
|
|
else:
|
|
job = HelpersManagement.get_scheduled_task(sch_id)
|
|
if job["interval_type"] != "reaction":
|
|
self.scheduler.remove_job(str(sch_id))
|
|
return
|
|
|
|
try:
|
|
if job_data["interval"] != "reaction":
|
|
self.scheduler.remove_job(str(sch_id))
|
|
except JobLookupError:
|
|
logger.info(
|
|
"No job found in update job. "
|
|
"Assuming it was previously disabled. Starting new job."
|
|
)
|
|
|
|
if job_data["enabled"] and job_data["interval"] != "reaction":
|
|
new_job = "error"
|
|
if job_data["cron_string"] != "":
|
|
try:
|
|
new_job = self.scheduler.add_job(
|
|
self.controller.management.queue_command,
|
|
CronTrigger.from_crontab(
|
|
job_data["cron_string"], timezone=str(self.tz)
|
|
),
|
|
id=str(sch_id),
|
|
args=[
|
|
{
|
|
"server_id": job_data["server_id"],
|
|
"user_id": self.users_controller.get_id_by_name(
|
|
"system"
|
|
),
|
|
"command": job_data["command"],
|
|
"action_id": job_data["action_id"],
|
|
}
|
|
],
|
|
)
|
|
except Exception as e:
|
|
new_job = "error"
|
|
Console.error(f"Failed to schedule task with error: {e}.")
|
|
Console.info("Removing failed task from DB.")
|
|
self.controller.management_helper.delete_scheduled_task(sch_id)
|
|
else:
|
|
if job_data["interval_type"] == "hours":
|
|
new_job = self.scheduler.add_job(
|
|
self.controller.management.queue_command,
|
|
"cron",
|
|
minute=0,
|
|
hour="*/" + str(job_data["interval"]),
|
|
id=str(sch_id),
|
|
args=[
|
|
{
|
|
"server_id": job_data["server_id"],
|
|
"user_id": self.users_controller.get_id_by_name(
|
|
"system"
|
|
),
|
|
"command": job_data["command"],
|
|
"action_id": job_data["action_id"],
|
|
}
|
|
],
|
|
)
|
|
elif job_data["interval_type"] == "minutes":
|
|
new_job = self.scheduler.add_job(
|
|
self.controller.management.queue_command,
|
|
"cron",
|
|
minute="*/" + str(job_data["interval"]),
|
|
id=str(sch_id),
|
|
args=[
|
|
{
|
|
"server_id": job_data["server_id"],
|
|
"user_id": self.users_controller.get_id_by_name(
|
|
"system"
|
|
),
|
|
"command": job_data["command"],
|
|
"action_id": job_data["action_id"],
|
|
}
|
|
],
|
|
)
|
|
elif job_data["interval_type"] == "days":
|
|
curr_time = job_data["start_time"].split(":")
|
|
new_job = self.scheduler.add_job(
|
|
self.controller.management.queue_command,
|
|
"cron",
|
|
day="*/" + str(job_data["interval"]),
|
|
hour=curr_time[0],
|
|
minute=curr_time[1],
|
|
id=str(sch_id),
|
|
args=[
|
|
{
|
|
"server_id": job_data["server_id"],
|
|
"user_id": self.users_controller.get_id_by_name(
|
|
"system"
|
|
),
|
|
"command": job_data["command"],
|
|
"action_id": job_data["action_id"],
|
|
}
|
|
],
|
|
)
|
|
if new_job != "error":
|
|
task = self.controller.management.get_scheduled_task_model(
|
|
int(new_job.id)
|
|
)
|
|
self.controller.management.update_scheduled_task(
|
|
task.schedule_id,
|
|
{"next_run": new_job.next_run_time.strftime("%m/%d/%Y, %H:%M:%S")},
|
|
)
|
|
else:
|
|
try:
|
|
self.scheduler.get_job(str(sch_id))
|
|
self.scheduler.remove_job(str(sch_id))
|
|
except:
|
|
logger.info(
|
|
f"APScheduler found no scheduled job on schedule update for "
|
|
f"schedule with id: {sch_id} Assuming it was already disabled."
|
|
)
|
|
|
|
def schedule_watcher(self, event):
|
|
if not event.exception:
|
|
if str(event.job_id).isnumeric():
|
|
task = self.controller.management.get_scheduled_task_model(
|
|
int(event.job_id)
|
|
)
|
|
self.controller.management.add_to_audit_log_raw(
|
|
"system",
|
|
HelperUsers.get_user_id_by_name("system"),
|
|
task.server_id,
|
|
f"Task with id {task.schedule_id} completed successfully",
|
|
"127.0.0.1",
|
|
)
|
|
# check if the task is a single run.
|
|
if task.one_time:
|
|
self.remove_job(task.schedule_id)
|
|
logger.info("one time task detected. Deleting...")
|
|
elif task.interval_type != "reaction":
|
|
self.controller.management.update_scheduled_task(
|
|
task.schedule_id,
|
|
{
|
|
"next_run": self.scheduler.get_job(
|
|
event.job_id
|
|
).next_run_time.strftime("%m/%d/%Y, %H:%M:%S")
|
|
},
|
|
)
|
|
# check for any child tasks for this. It's kind of backward,
|
|
# but this makes DB management a lot easier. One to one
|
|
# instead of one to many.
|
|
for schedule in HelpersManagement.get_child_schedules_by_server(
|
|
task.schedule_id, task.server_id
|
|
):
|
|
# event job ID's are strings so we need to look at
|
|
# this as the same data type.
|
|
if (
|
|
str(schedule.parent) == str(event.job_id)
|
|
and schedule.interval_type == "reaction"
|
|
):
|
|
if schedule.enabled:
|
|
delaytime = datetime.datetime.now() + datetime.timedelta(
|
|
seconds=schedule.delay
|
|
)
|
|
self.scheduler.add_job(
|
|
self.controller.management.queue_command,
|
|
"date",
|
|
run_date=delaytime,
|
|
id=str(schedule.schedule_id),
|
|
args=[
|
|
{
|
|
"server_id": schedule.server_id.server_id,
|
|
"user_id": self.users_controller.get_id_by_name(
|
|
"system"
|
|
),
|
|
"command": schedule.command,
|
|
"action_id": schedule.action_id,
|
|
}
|
|
],
|
|
)
|
|
else:
|
|
logger.info(
|
|
"Event job ID is not numerical. Assuming it's stats "
|
|
"- not stored in DB. Moving on."
|
|
)
|
|
else:
|
|
logger.error(f"Task failed with error: {event.exception}")
|
|
|
|
def start_stats_recording(self):
|
|
stats_update_frequency = self.helper.get_setting(
|
|
"stats_update_frequency_seconds"
|
|
)
|
|
logger.info(
|
|
f"Stats collection frequency set to {stats_update_frequency} seconds"
|
|
)
|
|
Console.info(
|
|
f"Stats collection frequency set to {stats_update_frequency} seconds"
|
|
)
|
|
|
|
# one for now,
|
|
self.controller.servers.stats.record_stats()
|
|
# one for later
|
|
self.scheduler.add_job(
|
|
self.controller.servers.stats.record_stats,
|
|
"interval",
|
|
seconds=stats_update_frequency,
|
|
id="stats",
|
|
)
|
|
|
|
def big_bucket_cache_refresher(self):
|
|
logger.info("Refreshing big bucket cache on start")
|
|
self.controller.big_bucket.refresh_cache()
|
|
|
|
logger.info("Scheduling big bucket cache refresh service every 12 hours")
|
|
self.scheduler.add_job(
|
|
self.controller.big_bucket.refresh_cache,
|
|
"interval",
|
|
hours=12,
|
|
id="big_bucket",
|
|
)
|
|
|
|
def realtime(self):
|
|
loop = asyncio.new_event_loop()
|
|
asyncio.set_event_loop(loop)
|
|
|
|
host_stats = HelpersManagement.get_latest_hosts_stats()
|
|
|
|
while True:
|
|
if host_stats.get(
|
|
"cpu_usage"
|
|
) != HelpersManagement.get_latest_hosts_stats().get(
|
|
"cpu_usage"
|
|
) or host_stats.get(
|
|
"mem_percent"
|
|
) != HelpersManagement.get_latest_hosts_stats().get(
|
|
"mem_percent"
|
|
):
|
|
# Stats are different
|
|
|
|
host_stats = HelpersManagement.get_latest_hosts_stats()
|
|
|
|
self.controller.management.cpu_usage.set(host_stats.get("cpu_usage"))
|
|
self.controller.management.mem_usage_percent.set(
|
|
host_stats.get("mem_percent")
|
|
)
|
|
|
|
if len(WebSocketManager().clients) > 0:
|
|
# There are clients
|
|
try:
|
|
WebSocketManager().broadcast_page(
|
|
"/panel/dashboard",
|
|
"update_host_stats",
|
|
{
|
|
"cpu_usage": host_stats.get("cpu_usage"),
|
|
"cpu_cores": host_stats.get("cpu_cores"),
|
|
"cpu_cur_freq": host_stats.get("cpu_cur_freq"),
|
|
"cpu_max_freq": host_stats.get("cpu_max_freq"),
|
|
"mem_percent": host_stats.get("mem_percent"),
|
|
"mem_usage": host_stats.get("mem_usage"),
|
|
"disk_usage": json.loads(
|
|
host_stats.get("disk_json").replace("'", '"')
|
|
),
|
|
"mounts": self.helper.get_setting("monitored_mounts"),
|
|
},
|
|
)
|
|
except:
|
|
WebSocketManager().broadcast_page(
|
|
"/panel/dashboard",
|
|
"update_host_stats",
|
|
{
|
|
"cpu_usage": host_stats.get("cpu_usage"),
|
|
"cpu_cores": host_stats.get("cpu_cores"),
|
|
"cpu_cur_freq": host_stats.get("cpu_cur_freq"),
|
|
"cpu_max_freq": host_stats.get("cpu_max_freq"),
|
|
"mem_percent": host_stats.get("mem_percent"),
|
|
"mem_usage": host_stats.get("mem_usage"),
|
|
"disk_usage": {},
|
|
},
|
|
)
|
|
time.sleep(1)
|
|
|
|
def check_for_updates(self):
|
|
logger.info("Checking for Crafty updates...")
|
|
self.helper.update_available = self.helper.check_remote_version()
|
|
remote = self.helper.update_available
|
|
if self.helper.update_available:
|
|
logger.info(f"Found new version {self.helper.update_available}")
|
|
else:
|
|
logger.info(
|
|
"No updates found! You are on the most up to date Crafty version."
|
|
)
|
|
if self.helper.update_available:
|
|
self.helper.update_available = {
|
|
"id": str(remote),
|
|
"title": f"{remote} Update Available",
|
|
"date": "",
|
|
"desc": "Release notes are available by clicking this notification.",
|
|
"link": "https://gitlab.com/crafty-controller/crafty-4/-/releases",
|
|
}
|
|
logger.info("Refreshing Gravatar PFPs...")
|
|
for user in HelperUsers.get_all_users():
|
|
if user.email:
|
|
HelperUsers.update_user(
|
|
user.id, {"pfp": self.helper.get_gravatar_image(user.email)}
|
|
)
|
|
# Search for old files in imports
|
|
self.helper.ensure_dir_exists(
|
|
os.path.join(self.controller.project_root, "import", "upload")
|
|
)
|
|
self.helper.ensure_dir_exists(
|
|
os.path.join(self.controller.project_root, "temp")
|
|
)
|
|
for file in os.listdir(os.path.join(self.controller.project_root, "temp")):
|
|
if self.helper.is_file_older_than_x_days(
|
|
os.path.join(self.controller.project_root, "temp", file)
|
|
):
|
|
try:
|
|
os.remove(os.path.join(file))
|
|
except FileNotFoundError:
|
|
logger.debug("Could not clear out file from temp directory")
|
|
|
|
for file in os.listdir(
|
|
os.path.join(self.controller.project_root, "import", "upload")
|
|
):
|
|
if self.helper.is_file_older_than_x_days(
|
|
os.path.join(self.controller.project_root, "import", "upload", file)
|
|
):
|
|
try:
|
|
os.remove(os.path.join(file))
|
|
except FileNotFoundError:
|
|
logger.debug("Could not clear out file from import directory")
|
|
|
|
def log_watcher(self):
|
|
self.check_for_old_logs()
|
|
self.scheduler.add_job(
|
|
self.check_for_old_logs,
|
|
"interval",
|
|
hours=6,
|
|
id="log-mgmt",
|
|
)
|
|
|
|
def check_for_old_logs(self):
|
|
# check for server logs first
|
|
self.controller.servers.check_for_old_logs()
|
|
try:
|
|
# check for crafty logs now
|
|
logs_path = os.path.join(self.controller.project_root, "logs")
|
|
logs_delete_after = int(
|
|
self.helper.get_setting("crafty_logs_delete_after_days")
|
|
)
|
|
latest_log_files = [
|
|
"session.log",
|
|
"schedule.log",
|
|
"tornado-access.log",
|
|
"session.log",
|
|
"commander.log",
|
|
]
|
|
# we won't delete if delete logs after is set to 0
|
|
if logs_delete_after != 0:
|
|
log_files = list(
|
|
filter(
|
|
lambda val: val not in latest_log_files,
|
|
os.listdir(logs_path),
|
|
)
|
|
)
|
|
for log_file in log_files:
|
|
log_file_path = os.path.join(logs_path, log_file)
|
|
if Helpers.check_file_exists(
|
|
log_file_path
|
|
) and Helpers.is_file_older_than_x_days(
|
|
log_file_path, logs_delete_after
|
|
):
|
|
os.remove(log_file_path)
|
|
except:
|
|
logger.debug(
|
|
"Unable to find project root."
|
|
" If this issue persists please contact support."
|
|
)
|