veloren/server-cli/src/main.rs
2023-10-18 10:50:58 +02:00

343 lines
12 KiB
Rust

#![deny(unsafe_code)]
#![deny(clippy::clone_on_ref_ptr)]
#[cfg(all(
target_os = "windows",
not(feature = "hot-agent"),
not(feature = "hot-site"),
))]
#[global_allocator]
static GLOBAL: mimalloc::MiMalloc = mimalloc::MiMalloc;
/// `server-cli` interface commands not to be confused with the commands sent
/// from the client to the server
mod cli;
mod settings;
mod shutdown_coordinator;
mod tui_runner;
mod tuilog;
mod web;
use crate::{
cli::{Admin, ArgvApp, ArgvCommand, Message, SharedCommand, Shutdown},
shutdown_coordinator::ShutdownCoordinator,
tui_runner::Tui,
tuilog::TuiLog,
};
use common::{clock::Clock, consts::MIN_RECOMMENDED_TOKIO_THREADS};
use common_base::span;
use core::sync::atomic::{AtomicUsize, Ordering};
use server::{persistence::DatabaseSettings, settings::Protocol, Event, Input, Server};
use std::{
io,
sync::{atomic::AtomicBool, mpsc, Arc},
time::{Duration, Instant},
};
use tokio::sync::Notify;
use tracing::{info, trace};
lazy_static::lazy_static! {
pub static ref LOG: TuiLog<'static> = TuiLog::default();
}
const TPS: u64 = 30;
fn main() -> io::Result<()> {
#[cfg(feature = "tracy")]
common_base::tracy_client::Client::start();
use clap::Parser;
let app = ArgvApp::parse();
let basic = !app.tui || app.command.is_some();
let noninteractive = app.non_interactive;
let no_auth = app.no_auth;
let sql_log_mode = app.sql_log_mode;
// noninteractive implies basic
let basic = basic || noninteractive;
let sigusr1_signal = Arc::new(AtomicBool::new(false));
#[cfg(any(target_os = "linux", target_os = "macos"))]
let _ = signal_hook::flag::register(signal_hook::consts::SIGUSR1, Arc::clone(&sigusr1_signal));
let (_guards, _guards2) = if basic {
(Vec::new(), common_frontend::init_stdout(None))
} else {
(common_frontend::init(None, &|| LOG.clone()), Vec::new())
};
// Load settings
let settings = settings::Settings::load();
// Determine folder to save server data in
let server_data_dir = {
let mut path = common_base::userdata_dir_workspace!();
info!("Using userdata folder at {}", path.display());
path.push(server::DEFAULT_DATA_DIR_NAME);
path
};
// We don't need that many threads in the async pool, at least 2 but generally
// 25% of all available will do
// TODO: evaluate std::thread::available_concurrency as a num_cpus replacement
let runtime = Arc::new(
tokio::runtime::Builder::new_multi_thread()
.enable_all()
.worker_threads((num_cpus::get() / 4).max(MIN_RECOMMENDED_TOKIO_THREADS))
.thread_name_fn(|| {
static ATOMIC_ID: AtomicUsize = AtomicUsize::new(0);
let id = ATOMIC_ID.fetch_add(1, Ordering::SeqCst);
format!("tokio-server-{}", id)
})
.build()
.unwrap(),
);
#[cfg(feature = "hot-agent")]
{
agent::init();
}
#[cfg(feature = "hot-site")]
{
world::init();
}
// Load server settings
let mut server_settings = server::Settings::load(&server_data_dir);
let mut editable_settings = server::EditableSettings::load(&server_data_dir);
// Apply no_auth modifier to the settings
if no_auth {
server_settings.auth_server_address = None;
}
// Relative to data_dir
const PERSISTENCE_DB_DIR: &str = "saves";
let database_settings = DatabaseSettings {
db_dir: server_data_dir.join(PERSISTENCE_DB_DIR),
sql_log_mode,
};
let mut bench = None;
if let Some(command) = app.command {
match command {
ArgvCommand::Shared(SharedCommand::Admin { command }) => {
let login_provider = server::login_provider::LoginProvider::new(
server_settings.auth_server_address,
runtime,
);
return match command {
Admin::Add { username, role } => {
// FIXME: Currently the UUID can get returned even if the file didn't
// change, so this can't be relied on as an error
// code; moreover, we do nothing with the UUID
// returned in the success case. Fix the underlying function to return
// enough information that we can reliably return an error code.
let _ = server::add_admin(
&username,
role,
&login_provider,
&mut editable_settings,
&server_data_dir,
);
Ok(())
},
Admin::Remove { username } => {
// FIXME: Currently the UUID can get returned even if the file didn't
// change, so this can't be relied on as an error
// code; moreover, we do nothing with the UUID
// returned in the success case. Fix the underlying function to return
// enough information that we can reliably return an error code.
let _ = server::remove_admin(
&username,
&login_provider,
&mut editable_settings,
&server_data_dir,
);
Ok(())
},
};
},
ArgvCommand::Bench(params) => {
bench = Some(params);
// If we are trying to benchmark, don't limit the server view distance.
server_settings.max_view_distance = None;
// TODO: add setting to adjust wildlife spawn density, note I
// tried but Index setup makes it a bit
// annoying, might require a more involved refactor to get
// working nicely
},
};
}
// Panic hook to ensure that console mode is set back correctly if in non-basic
// mode
if !basic {
let hook = std::panic::take_hook();
std::panic::set_hook(Box::new(move |info| {
Tui::shutdown(basic);
hook(info);
}));
}
let tui = (!noninteractive).then(|| Tui::run(basic));
info!("Starting server...");
let protocols_and_addresses = server_settings.gameserver_protocols.clone();
let web_port = &settings.web_address.port();
// Create server
let mut server = Server::new(
server_settings,
editable_settings,
database_settings,
&server_data_dir,
&|_| {},
Arc::clone(&runtime),
)
.expect("Failed to create server instance!");
let registry = Arc::clone(server.metrics_registry());
let chat = server.chat_cache().clone();
let metrics_shutdown = Arc::new(Notify::new());
let metrics_shutdown_clone = Arc::clone(&metrics_shutdown);
let web_chat_secret = settings.web_chat_secret.clone();
runtime.spawn(async move {
web::run(
registry,
chat,
web_chat_secret,
settings.web_address,
metrics_shutdown_clone.notified(),
)
.await
});
// Collect addresses that the server is listening to log.
let gameserver_addresses = protocols_and_addresses
.into_iter()
.map(|protocol| match protocol {
Protocol::Tcp { address } => ("TCP", address),
Protocol::Quic {
address,
cert_file_path: _,
key_file_path: _,
} => ("QUIC", address),
});
info!(
?web_port,
?gameserver_addresses,
"Server is ready to accept connections."
);
let mut shutdown_coordinator = ShutdownCoordinator::new(Arc::clone(&sigusr1_signal));
// Set up an fps clock
let mut clock = Clock::new(Duration::from_secs_f64(1.0 / TPS as f64));
if let Some(bench) = bench {
#[cfg(feature = "worldgen")]
server.create_centered_persister(bench.view_distance);
}
let mut bench_exit_time = None;
let mut tick_no = 0u64;
loop {
span!(guard, "work");
if let Some(bench) = bench {
if let Some(t) = bench_exit_time {
if Instant::now() > t {
break;
}
} else if tick_no != 0 && !server.chunks_pending() {
println!("Chunk loading complete");
bench_exit_time = Some(Instant::now() + Duration::from_secs(bench.duration.into()));
}
};
tick_no += 1;
// Terminate the server if instructed to do so by the shutdown coordinator
if shutdown_coordinator.check(&mut server, &settings) {
break;
}
let events = server
.tick(Input::default(), clock.dt())
.expect("Failed to tick server");
for event in events {
match event {
Event::ClientConnected { entity: _ } => info!("Client connected!"),
Event::ClientDisconnected { entity: _ } => info!("Client disconnected!"),
Event::Chat { entity: _, msg } => info!("[Client] {}", msg),
}
}
// Clean up the server after a tick.
server.cleanup();
if tick_no.rem_euclid(1000) == 0 {
trace!(?tick_no, "keepalive")
}
if let Some(tui) = tui.as_ref() {
match tui.msg_r.try_recv() {
Ok(msg) => match msg {
Message::Shutdown {
command: Shutdown::Cancel,
} => shutdown_coordinator.abort_shutdown(&mut server),
Message::Shutdown {
command: Shutdown::Graceful { seconds, reason },
} => {
shutdown_coordinator.initiate_shutdown(
&mut server,
Duration::from_secs(seconds),
reason,
);
},
Message::Shutdown {
command: Shutdown::Immediate,
} => {
info!("Closing the server");
break;
},
Message::Shared(SharedCommand::Admin {
command: Admin::Add { username, role },
}) => {
server.add_admin(&username, role);
},
Message::Shared(SharedCommand::Admin {
command: Admin::Remove { username },
}) => {
server.remove_admin(&username);
},
Message::LoadArea { view_distance } => {
#[cfg(feature = "worldgen")]
server.create_centered_persister(view_distance);
},
Message::SqlLogMode { mode } => {
server.set_sql_log_mode(mode);
},
Message::DisconnectAllClients => {
server.disconnect_all_clients();
},
},
Err(mpsc::TryRecvError::Empty) | Err(mpsc::TryRecvError::Disconnected) => {},
}
}
drop(guard);
// Wait for the next tick.
clock.tick();
#[cfg(feature = "tracy")]
common_base::tracy_client::frame_mark();
}
metrics_shutdown.notify_one();
Ok(())
}