* Replaced diesel with rusqlite and refinery

* Added "migration of migrations" to transfer the data from the __diesel_schema_migrations table to the refinery_schema_history table
* Removed all down migrations as refinery does not support down migrations
* Changed all diesel up migrations to refinery naming format
* Added --sql-log-mode parameter to veloren-server-cli to allow SQL tracing and profiling
* Added /disconnect_all_players admin command
* Added disconnectall CLI command
* Fixes for several potential persistence-related race conditions
This commit is contained in:
Ben Wallis 2021-04-13 22:05:47 +00:00
parent 9ccaec1aca
commit 1de94a9979
94 changed files with 1455 additions and 958 deletions

View File

@ -29,6 +29,9 @@ and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0
- Villagers and guards now spawn with potions, and know how to use them. - Villagers and guards now spawn with potions, and know how to use them.
- Combat music in dungeons when within range of enemies. - Combat music in dungeons when within range of enemies.
- New Command: "kit", place a set of items into your inventory - New Command: "kit", place a set of items into your inventory
- Added --sql-log-mode profile/trace parameter to veloren-server-cli
- Added /disconnect_all_players admin command
- Added disconnectall CLI command
### Changed ### Changed
@ -50,6 +53,7 @@ and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0
- Removed infinite armour values from most admin items - Removed infinite armour values from most admin items
- Item tooltips during trades will now inform the user of what ctrl-click and shift-click do - Item tooltips during trades will now inform the user of what ctrl-click and shift-click do
- International keyboards can now display more key names on Linux and Windows instead of `Unknown`. - International keyboards can now display more key names on Linux and Windows instead of `Unknown`.
- There is now a brief period after a character leaves the world where they cannot rejoin until their data is saved
### Removed ### Removed

140
Cargo.lock generated
View File

@ -1325,38 +1325,6 @@ version = "1.2.0"
source = "registry+https://github.com/rust-lang/crates.io-index" source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "c0b7756d6eb729250618a3693b34b3311b282e12aeeee7970ae2a70997c03eb6" checksum = "c0b7756d6eb729250618a3693b34b3311b282e12aeeee7970ae2a70997c03eb6"
[[package]]
name = "diesel"
version = "1.4.6"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "047bfc4d5c3bd2ef6ca6f981941046113524b9a9f9a7cbdfdd7ff40f58e6f542"
dependencies = [
"byteorder",
"diesel_derives",
"libsqlite3-sys",
]
[[package]]
name = "diesel_derives"
version = "1.4.1"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "45f5098f628d02a7a0f68ddba586fb61e80edec3bdc1be3b921f4ceec60858d3"
dependencies = [
"proc-macro2 1.0.26",
"quote 1.0.9",
"syn 1.0.69",
]
[[package]]
name = "diesel_migrations"
version = "1.4.0"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "bf3cde8413353dc7f5d72fa8ce0b99a560a359d2c5ef1e5817ca731cd9008f4c"
dependencies = [
"migrations_internals",
"migrations_macros",
]
[[package]] [[package]]
name = "directories-next" name = "directories-next"
version = "2.0.0" version = "2.0.0"
@ -1559,6 +1527,12 @@ version = "0.2.0"
source = "registry+https://github.com/rust-lang/crates.io-index" source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "4443176a9f2c162692bd3d352d745ef9413eec5782a80d8fd6f8a1ac692a07f7" checksum = "4443176a9f2c162692bd3d352d745ef9413eec5782a80d8fd6f8a1ac692a07f7"
[[package]]
name = "fallible-streaming-iterator"
version = "0.1.9"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "7360491ce676a36bf9bb3c56c1aa791658183a54d2744120f27285738d90465a"
[[package]] [[package]]
name = "fehler" name = "fehler"
version = "1.0.0" version = "1.0.0"
@ -2197,6 +2171,15 @@ dependencies = [
"serde", "serde",
] ]
[[package]]
name = "hashlink"
version = "0.6.0"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "d99cf782f0dc4372d26846bec3de7804ceb5df083c2d4462c0b8d2330e894fa8"
dependencies = [
"hashbrown",
]
[[package]] [[package]]
name = "heapless" name = "heapless"
version = "0.5.6" version = "0.5.6"
@ -2688,9 +2671,9 @@ checksum = "c7d73b3f436185384286bd8098d17ec07c9a7d2388a6599f824d8502b529702a"
[[package]] [[package]]
name = "libsqlite3-sys" name = "libsqlite3-sys"
version = "0.18.0" version = "0.20.1"
source = "registry+https://github.com/rust-lang/crates.io-index" source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "1e704a02bcaecd4a08b93a23f6be59d0bd79cd161e0963e9499165a0a35df7bd" checksum = "64d31059f22935e6c31830db5249ba2b7ecd54fd73a9909286f0a67aa55c2fbd"
dependencies = [ dependencies = [
"cc", "cc",
"pkg-config", "pkg-config",
@ -2879,27 +2862,6 @@ dependencies = [
"autocfg", "autocfg",
] ]
[[package]]
name = "migrations_internals"
version = "1.4.1"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "2b4fc84e4af020b837029e017966f86a1c2d5e83e64b589963d5047525995860"
dependencies = [
"diesel",
]
[[package]]
name = "migrations_macros"
version = "1.4.2"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "9753f12909fd8d923f75ae5c3258cae1ed3c8ec052e1b38c93c21a6d157f789c"
dependencies = [
"migrations_internals",
"proc-macro2 1.0.26",
"quote 1.0.9",
"syn 1.0.69",
]
[[package]] [[package]]
name = "minifb" name = "minifb"
version = "0.19.1" version = "0.19.1"
@ -4108,6 +4070,47 @@ dependencies = [
"redox_syscall 0.2.5", "redox_syscall 0.2.5",
] ]
[[package]]
name = "refinery"
version = "0.5.0"
source = "git+https://gitlab.com/veloren/refinery.git?rev=8ecf4b4772d791e6c8c0a3f9b66a7530fad1af3e#8ecf4b4772d791e6c8c0a3f9b66a7530fad1af3e"
dependencies = [
"refinery-core",
"refinery-macros",
]
[[package]]
name = "refinery-core"
version = "0.5.0"
source = "git+https://gitlab.com/veloren/refinery.git?rev=8ecf4b4772d791e6c8c0a3f9b66a7530fad1af3e#8ecf4b4772d791e6c8c0a3f9b66a7530fad1af3e"
dependencies = [
"async-trait",
"cfg-if 1.0.0",
"chrono",
"lazy_static",
"log",
"regex",
"rusqlite",
"serde",
"siphasher",
"thiserror",
"toml",
"url",
"walkdir 2.3.2",
]
[[package]]
name = "refinery-macros"
version = "0.5.0"
source = "git+https://gitlab.com/veloren/refinery.git?rev=8ecf4b4772d791e6c8c0a3f9b66a7530fad1af3e#8ecf4b4772d791e6c8c0a3f9b66a7530fad1af3e"
dependencies = [
"proc-macro2 1.0.26",
"quote 1.0.9",
"refinery-core",
"regex",
"syn 1.0.69",
]
[[package]] [[package]]
name = "regalloc" name = "regalloc"
version = "0.0.31" version = "0.0.31"
@ -4210,6 +4213,22 @@ version = "0.0.6"
source = "registry+https://github.com/rust-lang/crates.io-index" source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "84348444bd7ad45729d0c49a4240d7cdc11c9d512c06c5ad1835c1ad4acda6db" checksum = "84348444bd7ad45729d0c49a4240d7cdc11c9d512c06c5ad1835c1ad4acda6db"
[[package]]
name = "rusqlite"
version = "0.24.2"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "d5f38ee71cbab2c827ec0ac24e76f82eca723cee92c509a65f67dee393c25112"
dependencies = [
"bitflags",
"fallible-iterator",
"fallible-streaming-iterator",
"hashlink",
"lazy_static",
"libsqlite3-sys",
"memchr",
"smallvec",
]
[[package]] [[package]]
name = "rust-argon2" name = "rust-argon2"
version = "0.8.3" version = "0.8.3"
@ -4621,6 +4640,12 @@ dependencies = [
"libc", "libc",
] ]
[[package]]
name = "siphasher"
version = "0.3.3"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "fa8f3741c7372e75519bd9346068370c9cdaabcc1f9599cbcf2a2719352286b7"
[[package]] [[package]]
name = "slab" name = "slab"
version = "0.4.2" version = "0.4.2"
@ -5619,14 +5644,11 @@ dependencies = [
"authc", "authc",
"chrono", "chrono",
"crossbeam-channel", "crossbeam-channel",
"diesel",
"diesel_migrations",
"dotenv", "dotenv",
"futures-util", "futures-util",
"hashbrown", "hashbrown",
"itertools 0.10.0", "itertools 0.10.0",
"lazy_static", "lazy_static",
"libsqlite3-sys",
"num_cpus", "num_cpus",
"portpicker", "portpicker",
"prometheus", "prometheus",
@ -5634,7 +5656,9 @@ dependencies = [
"rand 0.8.3", "rand 0.8.3",
"rand_distr", "rand_distr",
"rayon", "rayon",
"refinery",
"ron", "ron",
"rusqlite",
"scan_fmt", "scan_fmt",
"serde", "serde",
"serde_json", "serde_json",

View File

@ -94,4 +94,4 @@ debug = 1
[patch.crates-io] [patch.crates-io]
# macos CI fix isn't merged yet # macos CI fix isn't merged yet
winit = { git = "https://gitlab.com/veloren/winit.git", branch = "macos-test-spiffed" } winit = { git = "https://gitlab.com/veloren/winit.git", branch = "macos-test-spiffed" }
vek = { git = "https://gitlab.com/veloren/vek.git", branch = "fix_intrinsics2" } vek = { git = "https://gitlab.com/veloren/vek.git", branch = "fix_intrinsics2" }

View File

@ -4,6 +4,9 @@ const VELOREN_USERDATA_ENV: &str = "VELOREN_USERDATA";
// TODO: consider expanding this to a general install strategy variable that is // TODO: consider expanding this to a general install strategy variable that is
// also used for finding assets // also used for finding assets
// TODO: Ensure there are no NUL (\0) characters in userdata_dir (possible on
// MacOS but not Windows or Linux) as SQLite requires the database path does not
// include this character.
/// # `VELOREN_USERDATA_STRATEGY` environment variable /// # `VELOREN_USERDATA_STRATEGY` environment variable
/// Read during compilation /// Read during compilation
/// Useful to set when compiling for distribution /// Useful to set when compiling for distribution

View File

@ -47,6 +47,7 @@ pub enum ChatCommand {
BuildAreaRemove, BuildAreaRemove,
Campfire, Campfire,
DebugColumn, DebugColumn,
DisconnectAllPlayers,
DropAll, DropAll,
Dummy, Dummy,
Explosion, Explosion,
@ -108,6 +109,7 @@ pub static CHAT_COMMANDS: &[ChatCommand] = &[
ChatCommand::BuildAreaRemove, ChatCommand::BuildAreaRemove,
ChatCommand::Campfire, ChatCommand::Campfire,
ChatCommand::DebugColumn, ChatCommand::DebugColumn,
ChatCommand::DisconnectAllPlayers,
ChatCommand::DropAll, ChatCommand::DropAll,
ChatCommand::Dummy, ChatCommand::Dummy,
ChatCommand::Explosion, ChatCommand::Explosion,
@ -293,6 +295,11 @@ impl ChatCommand {
"Prints some debug information about a column", "Prints some debug information about a column",
NoAdmin, NoAdmin,
), ),
ChatCommand::DisconnectAllPlayers => cmd(
vec![Any("confirm", Required)],
"Disconnects all players from the server",
Admin,
),
ChatCommand::DropAll => cmd(vec![], "Drops all your items on the ground", Admin), ChatCommand::DropAll => cmd(vec![], "Drops all your items on the ground", Admin),
ChatCommand::Dummy => cmd(vec![], "Spawns a training dummy", Admin), ChatCommand::Dummy => cmd(vec![], "Spawns a training dummy", Admin),
ChatCommand::Explosion => cmd( ChatCommand::Explosion => cmd(
@ -538,6 +545,7 @@ impl ChatCommand {
ChatCommand::BuildAreaRemove => "build_area_remove", ChatCommand::BuildAreaRemove => "build_area_remove",
ChatCommand::Campfire => "campfire", ChatCommand::Campfire => "campfire",
ChatCommand::DebugColumn => "debug_column", ChatCommand::DebugColumn => "debug_column",
ChatCommand::DisconnectAllPlayers => "disconnect_all_players",
ChatCommand::DropAll => "dropall", ChatCommand::DropAll => "dropall",
ChatCommand::Dummy => "dummy", ChatCommand::Dummy => "dummy",
ChatCommand::Explosion => "explosion", ChatCommand::Explosion => "explosion",

View File

@ -135,6 +135,7 @@ pub enum ServerEvent {
}, },
CreateWaypoint(Vec3<f32>), CreateWaypoint(Vec3<f32>),
ClientDisconnect(EcsEntity), ClientDisconnect(EcsEntity),
ClientDisconnectWithoutPersistence(EcsEntity),
ChunkRequest(EcsEntity, Vec2<i32>), ChunkRequest(EcsEntity, Vec2<i32>),
ChatCmd(EcsEntity, String), ChatCmd(EcsEntity, String),
/// Send a chat message to the player from an npc or other player /// Send a chat message to the player from an npc or other player

View File

@ -1,4 +1,5 @@
use core::time::Duration; use core::time::Duration;
use server::persistence::SqlLogMode;
use std::sync::mpsc::Sender; use std::sync::mpsc::Sender;
use tracing::{error, info, warn}; use tracing::{error, info, warn};
@ -10,6 +11,8 @@ pub enum Message {
AddAdmin(String), AddAdmin(String),
RemoveAdmin(String), RemoveAdmin(String),
LoadArea(u32), LoadArea(u32),
SetSqlLogMode(SqlLogMode),
DisconnectAllClients,
} }
struct Command<'a> { struct Command<'a> {
@ -22,13 +25,13 @@ struct Command<'a> {
} }
// TODO: maybe we could be using clap here? // TODO: maybe we could be using clap here?
const COMMANDS: [Command; 6] = [ const COMMANDS: [Command; 8] = [
Command { Command {
name: "quit", name: "quit",
description: "Closes the server", description: "Closes the server",
split_spaces: true, split_spaces: true,
args: 0, args: 0,
cmd: |_, sender| sender.send(Message::Quit).unwrap(), cmd: |_, sender| send(sender, Message::Quit),
}, },
Command { Command {
name: "shutdown", name: "shutdown",
@ -38,16 +41,21 @@ const COMMANDS: [Command; 6] = [
args: 1, args: 1,
cmd: |args, sender| { cmd: |args, sender| {
if let Ok(grace_period) = args.first().unwrap().parse::<u64>() { if let Ok(grace_period) = args.first().unwrap().parse::<u64>() {
sender send(sender, Message::Shutdown {
.send(Message::Shutdown { grace_period: Duration::from_secs(grace_period),
grace_period: Duration::from_secs(grace_period), })
})
.unwrap()
} else { } else {
error!("Grace period must an integer") error!("Grace period must an integer")
} }
}, },
}, },
Command {
name: "disconnectall",
description: "Disconnects all connected clients",
split_spaces: true,
args: 0,
cmd: |_, sender| send(sender, Message::DisconnectAllClients),
},
Command { Command {
name: "loadarea", name: "loadarea",
description: "Loads up the chunks in a random area and adds a entity that mimics a player \ description: "Loads up the chunks in a random area and adds a entity that mimics a player \
@ -56,7 +64,7 @@ const COMMANDS: [Command; 6] = [
args: 1, args: 1,
cmd: |args, sender| { cmd: |args, sender| {
if let Ok(view_distance) = args.first().unwrap().parse::<u32>() { if let Ok(view_distance) = args.first().unwrap().parse::<u32>() {
sender.send(Message::LoadArea(view_distance)).unwrap(); send(sender, Message::LoadArea(view_distance));
} else { } else {
error!("View distance must be an integer"); error!("View distance must be an integer");
} }
@ -67,7 +75,7 @@ const COMMANDS: [Command; 6] = [
description: "Aborts a shutdown if one is in progress", description: "Aborts a shutdown if one is in progress",
split_spaces: false, split_spaces: false,
args: 0, args: 0,
cmd: |_, sender| sender.send(Message::AbortShutdown).unwrap(), cmd: |_, sender| send(sender, Message::AbortShutdown),
}, },
Command { Command {
name: "admin", name: "admin",
@ -76,15 +84,38 @@ const COMMANDS: [Command; 6] = [
args: 2, args: 2,
cmd: |args, sender| match args.get(..2) { cmd: |args, sender| match args.get(..2) {
Some([op, username]) if op == "add" => { Some([op, username]) if op == "add" => {
sender.send(Message::AddAdmin(username.clone())).unwrap() send(sender, Message::AddAdmin(username.clone()));
}, },
Some([op, username]) if op == "remove" => { Some([op, username]) if op == "remove" => {
sender.send(Message::RemoveAdmin(username.clone())).unwrap() send(sender, Message::RemoveAdmin(username.clone()));
}, },
Some(_) => error!("First arg must be add or remove"), Some(_) => error!("First arg must be add or remove"),
_ => error!("Not enough args, should be unreachable"), _ => error!("Not enough args, should be unreachable"),
}, },
}, },
Command {
name: "sqllog",
description: "Sets the SQL logging mode, valid values are off, trace and profile",
split_spaces: true,
args: 1,
cmd: |args, sender| match args.get(0) {
Some(arg) => {
let sql_log_mode = match arg.to_lowercase().as_str() {
"off" => Some(SqlLogMode::Disabled),
"profile" => Some(SqlLogMode::Profile),
"trace" => Some(SqlLogMode::Trace),
_ => None,
};
if let Some(sql_log_mode) = sql_log_mode {
send(sender, Message::SetSqlLogMode(sql_log_mode));
} else {
error!("Invalid SQL log mode");
}
},
_ => error!("Not enough args"),
},
},
Command { Command {
name: "help", name: "help",
description: "List all command available", description: "List all command available",
@ -100,6 +131,12 @@ const COMMANDS: [Command; 6] = [
}, },
]; ];
fn send(sender: &mut Sender<Message>, message: Message) {
sender
.send(message)
.unwrap_or_else(|err| error!("Failed to send CLI message, err: {:?}", err));
}
pub fn parse_command(input: &str, msg_s: &mut Sender<Message>) { pub fn parse_command(input: &str, msg_s: &mut Sender<Message>) {
let mut args = input.split_whitespace(); let mut args = input.split_whitespace();

View File

@ -17,7 +17,10 @@ use clap::{App, Arg, SubCommand};
use common::clock::Clock; use common::clock::Clock;
use common_base::span; use common_base::span;
use core::sync::atomic::{AtomicUsize, Ordering}; use core::sync::atomic::{AtomicUsize, Ordering};
use server::{Event, Input, Server}; use server::{
persistence::{DatabaseSettings, SqlLogMode},
Event, Input, Server,
};
use std::{ use std::{
io, io,
sync::{atomic::AtomicBool, mpsc, Arc}, sync::{atomic::AtomicBool, mpsc, Arc},
@ -48,6 +51,11 @@ fn main() -> io::Result<()> {
Arg::with_name("no-auth") Arg::with_name("no-auth")
.long("no-auth") .long("no-auth")
.help("Runs without auth enabled"), .help("Runs without auth enabled"),
Arg::with_name("sql-log-mode")
.long("sql-log-mode")
.help("Enables SQL logging, valid values are \"trace\" and \"profile\"")
.possible_values(&["trace", "profile"])
.takes_value(true)
]) ])
.subcommand( .subcommand(
SubCommand::with_name("admin") SubCommand::with_name("admin")
@ -78,6 +86,12 @@ fn main() -> io::Result<()> {
let noninteractive = matches.is_present("non-interactive"); let noninteractive = matches.is_present("non-interactive");
let no_auth = matches.is_present("no-auth"); let no_auth = matches.is_present("no-auth");
let sql_log_mode = match matches.value_of("sql-log-mode") {
Some("trace") => SqlLogMode::Trace,
Some("profile") => SqlLogMode::Profile,
_ => SqlLogMode::Disabled,
};
// noninteractive implies basic // noninteractive implies basic
let basic = basic || noninteractive; let basic = basic || noninteractive;
@ -118,6 +132,15 @@ fn main() -> io::Result<()> {
// Load server settings // Load server settings
let mut server_settings = server::Settings::load(&server_data_dir); let mut server_settings = server::Settings::load(&server_data_dir);
let mut editable_settings = server::EditableSettings::load(&server_data_dir); let mut editable_settings = server::EditableSettings::load(&server_data_dir);
// Relative to data_dir
const PERSISTENCE_DB_DIR: &str = "saves";
let database_settings = DatabaseSettings {
db_dir: server_data_dir.join(PERSISTENCE_DB_DIR),
sql_log_mode,
};
#[allow(clippy::single_match)] // Note: remove this when there are more subcommands #[allow(clippy::single_match)] // Note: remove this when there are more subcommands
match matches.subcommand() { match matches.subcommand() {
("admin", Some(sub_m)) => { ("admin", Some(sub_m)) => {
@ -157,6 +180,7 @@ fn main() -> io::Result<()> {
let mut server = Server::new( let mut server = Server::new(
server_settings, server_settings,
editable_settings, editable_settings,
database_settings,
&server_data_dir, &server_data_dir,
runtime, runtime,
) )
@ -226,6 +250,12 @@ fn main() -> io::Result<()> {
Message::LoadArea(view_distance) => { Message::LoadArea(view_distance) => {
server.create_centered_persister(view_distance); server.create_centered_persister(view_distance);
}, },
Message::SetSqlLogMode(sql_log_mode) => {
server.set_sql_log_mode(sql_log_mode);
},
Message::DisconnectAllClients => {
server.disconnect_all_clients();
},
}, },
Err(mpsc::TryRecvError::Empty) | Err(mpsc::TryRecvError::Disconnected) => {}, Err(mpsc::TryRecvError::Empty) | Err(mpsc::TryRecvError::Disconnected) => {},
} }

View File

@ -44,12 +44,12 @@ crossbeam-channel = "0.5"
prometheus = { version = "0.12", default-features = false} prometheus = { version = "0.12", default-features = false}
portpicker = { git = "https://github.com/xMAC94x/portpicker-rs", rev = "df6b37872f3586ac3b21d08b56c8ec7cd92fb172" } portpicker = { git = "https://github.com/xMAC94x/portpicker-rs", rev = "df6b37872f3586ac3b21d08b56c8ec7cd92fb172" }
authc = { git = "https://gitlab.com/veloren/auth.git", rev = "fb3dcbc4962b367253f8f2f92760ef44d2679c9a" } authc = { git = "https://gitlab.com/veloren/auth.git", rev = "fb3dcbc4962b367253f8f2f92760ef44d2679c9a" }
libsqlite3-sys = { version = "0.18", features = ["bundled"] }
diesel = { version = "1.4.3", features = ["sqlite"] }
diesel_migrations = "1.4.0"
dotenv = "0.15.0" dotenv = "0.15.0"
slab = "0.4" slab = "0.4"
rand_distr = "0.4.0" rand_distr = "0.4.0"
rusqlite = { version = "0.24.2", features = ["array", "vtab", "bundled", "trace"] }
refinery = { git = "https://gitlab.com/veloren/refinery.git", rev = "8ecf4b4772d791e6c8c0a3f9b66a7530fad1af3e", features = ["rusqlite"] }
# Plugins # Plugins
plugin-api = { package = "veloren-plugin-api", path = "../plugin/api"} plugin-api = { package = "veloren-plugin-api", path = "../plugin/api"}

View File

@ -45,7 +45,7 @@ use world::util::Sampler;
use crate::{client::Client, login_provider::LoginProvider}; use crate::{client::Client, login_provider::LoginProvider};
use scan_fmt::{scan_fmt, scan_fmt_some}; use scan_fmt::{scan_fmt, scan_fmt_some};
use tracing::error; use tracing::{error, info, warn};
pub trait ChatCommandExt { pub trait ChatCommandExt {
fn execute(&self, server: &mut Server, entity: EcsEntity, args: String); fn execute(&self, server: &mut Server, entity: EcsEntity, args: String);
@ -102,6 +102,7 @@ fn get_handler(cmd: &ChatCommand) -> CommandHandler {
ChatCommand::BuildAreaRemove => handle_build_area_remove, ChatCommand::BuildAreaRemove => handle_build_area_remove,
ChatCommand::Campfire => handle_spawn_campfire, ChatCommand::Campfire => handle_spawn_campfire,
ChatCommand::DebugColumn => handle_debug_column, ChatCommand::DebugColumn => handle_debug_column,
ChatCommand::DisconnectAllPlayers => handle_disconnect_all_players,
ChatCommand::DropAll => handle_drop_all, ChatCommand::DropAll => handle_drop_all,
ChatCommand::Dummy => handle_spawn_training_dummy, ChatCommand::Dummy => handle_spawn_training_dummy,
ChatCommand::Explosion => handle_explosion, ChatCommand::Explosion => handle_explosion,
@ -2118,6 +2119,46 @@ spawn_rate {:?} "#,
} }
} }
fn handle_disconnect_all_players(
server: &mut Server,
client: EcsEntity,
_target: EcsEntity,
args: String,
_action: &ChatCommand,
) -> CmdResult<()> {
if args != *"confirm" {
return Err(
"Please run the command again with the second argument of \"confirm\" to confirm that \
you really want to disconnect all players from the server"
.to_string(),
);
}
let ecs = server.state.ecs();
let players = &ecs.read_storage::<comp::Player>();
// TODO: This logging and verification of admin commands would be better moved
// to a more generic method used for auditing -all- admin commands.
let player_name;
if let Some(player) = players.get(client) {
player_name = &*player.alias;
} else {
warn!(
"Failed to get player name for admin who used /disconnect_all_players - ignoring \
command."
);
return Err("You do not exist, so you cannot use this command".to_string());
}
info!(
"Disconnecting all clients due to admin command from {}",
player_name
);
server.disconnect_all_clients_requested = true;
Ok(())
}
fn handle_skill_point( fn handle_skill_point(
server: &mut Server, server: &mut Server,
_client: EcsEntity, _client: EcsEntity,

View File

@ -1,5 +1,5 @@
use crate::persistence::error::PersistenceError;
use network::{NetworkError, ParticipantError, StreamError}; use network::{NetworkError, ParticipantError, StreamError};
use std::fmt::{self, Display}; use std::fmt::{self, Display};
#[derive(Debug)] #[derive(Debug)]
@ -7,7 +7,8 @@ pub enum Error {
NetworkErr(NetworkError), NetworkErr(NetworkError),
ParticipantErr(ParticipantError), ParticipantErr(ParticipantError),
StreamErr(StreamError), StreamErr(StreamError),
DatabaseErr(diesel::result::Error), DatabaseErr(rusqlite::Error),
PersistenceErr(PersistenceError),
Other(String), Other(String),
} }
@ -23,8 +24,13 @@ impl From<StreamError> for Error {
fn from(err: StreamError) -> Self { Error::StreamErr(err) } fn from(err: StreamError) -> Self { Error::StreamErr(err) }
} }
impl From<diesel::result::Error> for Error { // TODO: Don't expose rusqlite::Error from persistence module
fn from(err: diesel::result::Error) -> Self { Error::DatabaseErr(err) } impl From<rusqlite::Error> for Error {
fn from(err: rusqlite::Error) -> Self { Error::DatabaseErr(err) }
}
impl From<PersistenceError> for Error {
fn from(err: PersistenceError) -> Self { Error::PersistenceErr(err) }
} }
impl Display for Error { impl Display for Error {
@ -34,6 +40,7 @@ impl Display for Error {
Self::ParticipantErr(err) => write!(f, "Participant Error: {}", err), Self::ParticipantErr(err) => write!(f, "Participant Error: {}", err),
Self::StreamErr(err) => write!(f, "Stream Error: {}", err), Self::StreamErr(err) => write!(f, "Stream Error: {}", err),
Self::DatabaseErr(err) => write!(f, "Database Error: {}", err), Self::DatabaseErr(err) => write!(f, "Database Error: {}", err),
Self::PersistenceErr(err) => write!(f, "Persistence Error: {}", err),
Self::Other(err) => write!(f, "Error: {}", err), Self::Other(err) => write!(f, "Error: {}", err),
} }
} }

View File

@ -174,7 +174,10 @@ impl Server {
} => handle_create_ship(self, pos, ship, mountable, agent, rtsim_entity), } => handle_create_ship(self, pos, ship, mountable, agent, rtsim_entity),
ServerEvent::CreateWaypoint(pos) => handle_create_waypoint(self, pos), ServerEvent::CreateWaypoint(pos) => handle_create_waypoint(self, pos),
ServerEvent::ClientDisconnect(entity) => { ServerEvent::ClientDisconnect(entity) => {
frontend_events.push(handle_client_disconnect(self, entity)) frontend_events.push(handle_client_disconnect(self, entity, false))
},
ServerEvent::ClientDisconnectWithoutPersistence(entity) => {
frontend_events.push(handle_client_disconnect(self, entity, true))
}, },
ServerEvent::ChunkRequest(entity, key) => { ServerEvent::ChunkRequest(entity, key) => {

View File

@ -1,5 +1,8 @@
use super::Event; use super::Event;
use crate::{client::Client, persistence, presence::Presence, state_ext::StateExt, Server}; use crate::{
client::Client, persistence::character_updater::CharacterUpdater, presence::Presence,
state_ext::StateExt, Server,
};
use common::{ use common::{
comp, comp,
comp::group, comp::group,
@ -93,7 +96,11 @@ pub fn handle_exit_ingame(server: &mut Server, entity: EcsEntity) {
} }
} }
pub fn handle_client_disconnect(server: &mut Server, entity: EcsEntity) -> Event { pub fn handle_client_disconnect(
server: &mut Server,
mut entity: EcsEntity,
skip_persistence: bool,
) -> Event {
span!(_guard, "handle_client_disconnect"); span!(_guard, "handle_client_disconnect");
if let Some(client) = server if let Some(client) = server
.state() .state()
@ -150,30 +157,45 @@ pub fn handle_client_disconnect(server: &mut Server, entity: EcsEntity) -> Event
} }
// Sync the player's character data to the database // Sync the player's character data to the database
let entity = persist_entity(state, entity); if !skip_persistence {
entity = persist_entity(state, entity);
}
// Delete client entity // Delete client entity
if let Err(e) = state.delete_entity_recorded(entity) { if let Err(e) = server.state.delete_entity_recorded(entity) {
error!(?e, ?entity, "Failed to delete disconnected client"); error!(?e, ?entity, "Failed to delete disconnected client");
} }
Event::ClientDisconnected { entity } Event::ClientDisconnected { entity }
} }
// When a player logs out, their data is queued for persistence in the next tick
// of the persistence batch update. The player will be
// temporarily unable to log in during this period to avoid
// the race condition of their login fetching their old data
// and overwriting the data saved here.
fn persist_entity(state: &mut State, entity: EcsEntity) -> EcsEntity { fn persist_entity(state: &mut State, entity: EcsEntity) -> EcsEntity {
if let (Some(presences), Some(stats), Some(inventory), updater) = ( if let (Some(presence), Some(stats), Some(inventory), mut character_updater) = (
state.read_storage::<Presence>().get(entity), state.read_storage::<Presence>().get(entity),
state.read_storage::<comp::Stats>().get(entity), state.read_storage::<comp::Stats>().get(entity),
state.read_storage::<comp::Inventory>().get(entity), state.read_storage::<comp::Inventory>().get(entity),
state state.ecs().fetch_mut::<CharacterUpdater>(),
.ecs()
.read_resource::<persistence::character_updater::CharacterUpdater>(),
) { ) {
if let PresenceKind::Character(character_id) = presences.kind { match presence.kind {
let waypoint_read = state.read_storage::<comp::Waypoint>(); PresenceKind::Character(char_id) => {
let waypoint = waypoint_read.get(entity); let waypoint = state
updater.update(character_id, stats, inventory, waypoint); .ecs()
} .read_storage::<common::comp::Waypoint>()
.get(entity)
.cloned();
character_updater.add_pending_logout_update(
char_id,
(stats.clone(), inventory.clone(), waypoint),
);
},
PresenceKind::Spectator => { /* Do nothing, spectators do not need persisting */ },
};
} }
entity entity

View File

@ -2,6 +2,7 @@
#![allow(clippy::option_map_unit_fn)] #![allow(clippy::option_map_unit_fn)]
#![deny(clippy::clone_on_ref_ptr)] #![deny(clippy::clone_on_ref_ptr)]
#![feature( #![feature(
box_patterns,
label_break_value, label_break_value,
bool_to_option, bool_to_option,
drain_filter, drain_filter,
@ -101,15 +102,17 @@ use tokio::{runtime::Runtime, sync::Notify};
use tracing::{debug, error, info, trace}; use tracing::{debug, error, info, trace};
use vek::*; use vek::*;
use crate::{
persistence::{DatabaseSettings, SqlLogMode},
sys::terrain,
};
use std::sync::RwLock;
#[cfg(feature = "worldgen")] #[cfg(feature = "worldgen")]
use world::{ use world::{
sim::{FileOpts, WorldOpts, DEFAULT_WORLD_MAP}, sim::{FileOpts, WorldOpts, DEFAULT_WORLD_MAP},
IndexOwned, World, IndexOwned, World,
}; };
#[macro_use] extern crate diesel;
#[macro_use] extern crate diesel_migrations;
#[derive(Copy, Clone)] #[derive(Copy, Clone)]
struct SpawnPoint(Vec3<f32>); struct SpawnPoint(Vec3<f32>);
@ -124,6 +127,12 @@ pub struct HwStats {
rayon_threads: u32, rayon_threads: u32,
} }
#[derive(Clone, Copy, PartialEq)]
enum DisconnectType {
WithPersistence,
WithoutPersistence,
}
// Start of Tick, used for metrics // Start of Tick, used for metrics
#[derive(Copy, Clone)] #[derive(Copy, Clone)]
pub struct TickStart(Instant); pub struct TickStart(Instant);
@ -139,6 +148,8 @@ pub struct Server {
runtime: Arc<Runtime>, runtime: Arc<Runtime>,
metrics_shutdown: Arc<Notify>, metrics_shutdown: Arc<Notify>,
database_settings: Arc<RwLock<DatabaseSettings>>,
disconnect_all_clients_requested: bool,
} }
impl Server { impl Server {
@ -148,6 +159,7 @@ impl Server {
pub fn new( pub fn new(
settings: Settings, settings: Settings,
editable_settings: EditableSettings, editable_settings: EditableSettings,
database_settings: DatabaseSettings,
data_dir: &std::path::Path, data_dir: &std::path::Path,
runtime: Arc<Runtime>, runtime: Arc<Runtime>,
) -> Result<Self, Error> { ) -> Result<Self, Error> {
@ -156,15 +168,11 @@ impl Server {
info!("Authentication is disabled"); info!("Authentication is disabled");
} }
// Relative to data_dir
const PERSISTENCE_DB_DIR: &str = "saves";
let persistence_db_dir = data_dir.join(PERSISTENCE_DB_DIR);
// Run pending DB migrations (if any) // Run pending DB migrations (if any)
debug!("Running DB migrations..."); debug!("Running DB migrations...");
if let Some(e) = persistence::run_migrations(&persistence_db_dir).err() { persistence::run_migrations(&database_settings);
panic!("Migration error: {:?}", e);
} let database_settings = Arc::new(RwLock::new(database_settings));
let registry = Arc::new(Registry::new()); let registry = Arc::new(Registry::new());
let chunk_gen_metrics = metrics::ChunkGenMetrics::new(&registry).unwrap(); let chunk_gen_metrics = metrics::ChunkGenMetrics::new(&registry).unwrap();
@ -203,9 +211,10 @@ impl Server {
state state
.ecs_mut() .ecs_mut()
.insert(ChunkGenerator::new(chunk_gen_metrics)); .insert(ChunkGenerator::new(chunk_gen_metrics));
state
.ecs_mut() state.ecs_mut().insert(CharacterUpdater::new(
.insert(CharacterUpdater::new(&persistence_db_dir)?); Arc::<RwLock<DatabaseSettings>>::clone(&database_settings),
)?);
let ability_map = comp::item::tool::AbilityMap::<CharacterAbility>::load_expect_cloned( let ability_map = comp::item::tool::AbilityMap::<CharacterAbility>::load_expect_cloned(
"common.abilities.weapon_ability_manifest", "common.abilities.weapon_ability_manifest",
@ -215,9 +224,9 @@ impl Server {
let msm = comp::inventory::item::MaterialStatManifest::default(); let msm = comp::inventory::item::MaterialStatManifest::default();
state.ecs_mut().insert(msm); state.ecs_mut().insert(msm);
state state.ecs_mut().insert(CharacterLoader::new(
.ecs_mut() Arc::<RwLock<DatabaseSettings>>::clone(&database_settings),
.insert(CharacterLoader::new(&persistence_db_dir)?); )?);
// System schedulers to control execution of systems // System schedulers to control execution of systems
state state
@ -386,6 +395,8 @@ impl Server {
runtime, runtime,
metrics_shutdown, metrics_shutdown,
database_settings,
disconnect_all_clients_requested: false,
}; };
debug!(?settings, "created veloren server with"); debug!(?settings, "created veloren server with");
@ -506,6 +517,10 @@ impl Server {
let before_handle_events = Instant::now(); let before_handle_events = Instant::now();
// Process any pending request to disconnect all clients, the disconnections
// will be processed once handle_events() is called below
let disconnect_type = self.disconnect_all_clients_if_requested();
// Handle game events // Handle game events
frontend_events.append(&mut self.handle_events()); frontend_events.append(&mut self.handle_events());
@ -530,6 +545,14 @@ impl Server {
let before_entity_cleanup = Instant::now(); let before_entity_cleanup = Instant::now();
// In the event of a request to disconnect all players without persistence, we
// must run the terrain system a second time after the messages to
// perform client disconnections have been processed. This ensures that any
// items on the ground are deleted.
if let Some(DisconnectType::WithoutPersistence) = disconnect_type {
run_now::<terrain::Sys>(self.state.ecs_mut());
}
// Remove NPCs that are outside the view distances of all players // Remove NPCs that are outside the view distances of all players
// This is done by removing NPCs in unloaded chunks // This is done by removing NPCs in unloaded chunks
let to_delete = { let to_delete = {
@ -573,6 +596,17 @@ impl Server {
} }
} }
if let Some(DisconnectType::WithoutPersistence) = disconnect_type {
info!(
"Disconnection of all players without persistence complete, signalling to \
persistence thread that character updates may continue to be processed"
);
self.state
.ecs()
.fetch_mut::<CharacterUpdater>()
.disconnected_success();
}
// 7 Persistence updates // 7 Persistence updates
let before_persistence_updates = Instant::now(); let before_persistence_updates = Instant::now();
@ -773,6 +807,58 @@ impl Server {
Ok(Some(entity)) Ok(Some(entity))
} }
/// Disconnects all clients if requested by either an admin command or
/// due to a persistence transaction failure and returns the processed
/// DisconnectionType
fn disconnect_all_clients_if_requested(&mut self) -> Option<DisconnectType> {
let mut character_updater = self.state.ecs().fetch_mut::<CharacterUpdater>();
let disconnect_type = self.get_disconnect_all_clients_requested(&mut character_updater);
if let Some(disconnect_type) = disconnect_type {
let with_persistence = disconnect_type == DisconnectType::WithPersistence;
let clients = self.state.ecs().read_storage::<Client>();
let entities = self.state.ecs().entities();
info!(
"Disconnecting all clients ({} persistence) as requested",
if with_persistence { "with" } else { "without" }
);
for (_, entity) in (&clients, &entities).join() {
info!("Emitting client disconnect event for entity: {:?}", entity);
let event = if with_persistence {
ServerEvent::ClientDisconnect(entity)
} else {
ServerEvent::ClientDisconnectWithoutPersistence(entity)
};
self.state
.ecs()
.read_resource::<EventBus<ServerEvent>>()
.emitter()
.emit(event);
}
self.disconnect_all_clients_requested = false;
}
disconnect_type
}
fn get_disconnect_all_clients_requested(
&self,
character_updater: &mut CharacterUpdater,
) -> Option<DisconnectType> {
let without_persistence_requested = character_updater.disconnect_all_clients_requested();
let with_persistence_requested = self.disconnect_all_clients_requested;
if without_persistence_requested {
return Some(DisconnectType::WithoutPersistence);
};
if with_persistence_requested {
return Some(DisconnectType::WithPersistence);
};
None
}
/// Handle new client connections. /// Handle new client connections.
fn handle_new_connections(&mut self, frontend_events: &mut Vec<Event>) { fn handle_new_connections(&mut self, frontend_events: &mut Vec<Event>) {
while let Ok(sender) = self.connection_handler.info_requester_receiver.try_recv() { while let Ok(sender) = self.connection_handler.info_requester_receiver.try_recv() {
@ -1009,6 +1095,26 @@ impl Server {
.create_persister(pos, view_distance, &self.world, &self.index) .create_persister(pos, view_distance, &self.world, &self.index)
.build(); .build();
} }
/// Sets the SQL log mode at runtime
pub fn set_sql_log_mode(&mut self, sql_log_mode: SqlLogMode) {
// Unwrap is safe here because we only perform a variable assignment with the
// RwLock taken meaning that no panic can occur that would cause the
// RwLock to become poisoned. This justification also means that calling
// unwrap() on the associated read() calls for this RwLock is also safe
// as long as no code that can panic is introduced here.
let mut database_settings = self.database_settings.write().unwrap();
database_settings.sql_log_mode = sql_log_mode;
// Drop the RwLockWriteGuard to avoid performing unnecessary actions (logging)
// with the lock taken.
drop(database_settings);
info!("SQL log mode changed to {:?}", sql_log_mode);
}
pub fn disconnect_all_clients(&mut self) {
info!("Disconnecting all clients due to local console command");
self.disconnect_all_clients_requested = true;
}
} }
impl Drop for Server { impl Drop for Server {

View File

@ -1 +0,0 @@
DROP TABLE IF EXISTS "character";

View File

@ -1 +0,0 @@
DROP TABLE IF EXISTS "body";

View File

@ -1 +0,0 @@
DROP TABLE IF EXISTS "stats";

View File

@ -1,46 +0,0 @@
-- SQLITE v < 3.25 does not support renaming columns.
ALTER TABLE
body RENAME TO body_tmp;
CREATE TABLE IF NOT EXISTS body (
character_id INT NOT NULL PRIMARY KEY,
race SMALLINT NOT NULL,
body_type SMALLINT NOT NULL,
hair_style SMALLINT NOT NULL,
beard SMALLINT NOT NULL,
eyebrows SMALLINT NOT NULL,
accessory SMALLINT NOT NULL,
hair_color SMALLINT NOT NULL,
skin SMALLINT NOT NULL,
eye_color SMALLINT NOT NULL,
FOREIGN KEY(character_id) REFERENCES "character"(id) ON DELETE CASCADE
);
INSERT INTO
body(
character_id,
race,
body_type,
hair_style,
beard,
eyebrows,
accessory,
hair_color,
skin,
eye_color
)
SELECT
character_id,
species,
body_type,
hair_style,
beard,
eyes,
accessory,
hair_color,
skin,
eye_color
FROM
body_tmp;
DROP TABLE body_tmp;

View File

@ -1 +0,0 @@
DROP TABLE IF EXISTS "inventory";

View File

@ -1 +0,0 @@
DROP TABLE IF EXISTS "loadout";

View File

@ -1,41 +0,0 @@
-- This migration downgrades the capacity of existing player inventories from 36 to 18. ITEMS WILL BE REMOVED.
UPDATE
inventory
SET
items = json_object(
'amount',
(
SELECT
json_extract(items, '$.amount')
from
inventory
),
'slots',
json_remove(
(
SELECT
json_extract(items, '$.slots')
from
inventory
),
'$[35]',
'$[34]',
'$[33]',
'$[32]',
'$[31]',
'$[30]',
'$[29]',
'$[28]',
'$[27]',
'$[26]',
'$[25]',
'$[25]',
'$[24]',
'$[23]',
'$[22]',
'$[21]',
'$[20]',
'$[19]',
'$[18]'
)
);

View File

@ -1,22 +0,0 @@
PRAGMA foreign_keys=off;
-- SQLite does not support removing columns from tables so we must rename the current table,
-- recreate the previous version of the table, then copy over the data from the renamed table
ALTER TABLE stats RENAME TO _stats_old;
CREATE TABLE "stats" (
character_id INT NOT NULL PRIMARY KEY,
level INT NOT NULL DEFAULT 1,
exp INT NOT NULL DEFAULT 0,
endurance INT NOT NULL DEFAULT 0,
fitness INT NOT NULL DEFAULT 0,
willpower INT NOT NULL DEFAULT 0,
FOREIGN KEY(character_id) REFERENCES "character"(id) ON DELETE CASCADE
);
INSERT INTO "stats" (character_id, level, exp, endurance, fitness, willpower)
SELECT character_id, level, exp, endurance, fitness, willpower FROM _stats_old;
DROP TABLE _stats_old;
PRAGMA foreign_keys=on;

View File

@ -1 +0,0 @@
-- Nothing to undo since up.sql only creates missing inventory/loadout records

View File

@ -1 +0,0 @@
-- Nothing to undo since up.sql only fixes corrupt JSON in loadouts

View File

@ -1 +0,0 @@
-- No down action for this migration

View File

@ -1 +0,0 @@
-- This file should undo anything in `up.sql`

View File

@ -1 +0,0 @@
-- This file should undo anything in `up.sql`

View File

@ -1 +0,0 @@
-- This file should undo anything in `up.sql`

View File

@ -1 +0,0 @@
-- This file should undo anything in `up.sql`

View File

@ -1,11 +0,0 @@
DROP TABLE stats;
DROP TABLE character;
DROP TABLE body;
DROP TABLE item;
DROP TABLE entity;
ALTER TABLE _body_bak RENAME TO body;
ALTER TABLE _stats_bak RENAME TO stats;
ALTER TABLE _character_bak RENAME TO character;
ALTER TABLE _loadout_bak RENAME TO loadout;
ALTER TABLE _inventory_bak RENAME TO inventory;

View File

@ -1,6 +0,0 @@
-- This file should undo anything in `up.sql`
UPDATE item
SET item_definition_id = 'common.items.weapons.staff.sceptre_velorite_0' WHERE item_definition_id = 'common.items.weapons.sceptre.sceptre_velorite_0';
UPDATE item
SET item_definition_id = 'common.items.weapons.staff.staff_nature' WHERE item_definition_id = 'common.items.weapons.sceptre.staff_nature';

View File

@ -1,6 +0,0 @@
-- This file should undo anything in `up.sql`
UPDATE item
SET item_definition_id = 'common.items.npc_weapons.npcweapon.beast_claws' WHERE item_definition_id = 'common.items.npc_weapons.unique.beast_claws';
UPDATE item
SET item_definition_id = 'common.items.npc_weapons.npcweapon.stone_golems_fist' WHERE item_definition_id = 'common.items.npc_weapons.unique.stone_golems_fist';

View File

@ -1,38 +0,0 @@
-- Put waypoint data back into item table
UPDATE item
SET position = ( SELECT s.waypoint
FROM stats s
WHERE s.stats_id = item.item_id
AND item.item_definition_id = 'veloren.core.pseudo_containers.character'
AND s.waypoint IS NOT NULL)
WHERE EXISTS ( SELECT s.waypoint
FROM stats s
WHERE s.stats_id = item.item_id
AND item.item_definition_id = 'veloren.core.pseudo_containers.character'
AND s.waypoint IS NOT NULL);
-- SQLite does not support dropping columns on tables so the entire table must be
-- dropped and recreated without the 'waypoint' column
CREATE TABLE stats_new
(
stats_id INT NOT NULL
PRIMARY KEY
REFERENCES entity,
level INT NOT NULL,
exp INT NOT NULL,
endurance INT NOT NULL,
fitness INT NOT NULL,
willpower INT NOT NULL
);
INSERT INTO stats_new (stats_id, level, exp, endurance, fitness, willpower)
SELECT stats_id,
level,
exp,
endurance,
fitness,
willpower
FROM stats;
DROP TABLE stats;
ALTER TABLE stats_new RENAME TO stats;

View File

@ -1 +0,0 @@
-- This file should undo anything in `up.sql`

View File

@ -1 +0,0 @@
-- What's a down migration?

View File

@ -1,2 +0,0 @@
UPDATE item
SET item_definition_id = 'common.items.weapons.crafting.shiny_gem' WHERE item_definition_id = 'common.items.crafting_ing.diamond';

View File

@ -1,10 +0,0 @@
UPDATE item
SET item_definition_id = 'common.items.armor.starter.glider' WHERE item_definition_id = 'common.items.glider.glider_cloverleaf';
UPDATE item
SET item_definition_id = 'common.items.armor.starter.lantern' WHERE item_definition_id = 'common.items.lantern.black_0';
UPDATE item
SET item_definition_id = 'common.items.armor.starter.rugged_chest' WHERE item_definition_id = 'common.items.armor.chest.rugged';
UPDATE item
SET item_definition_id = 'common.items.armor.starter.rugged_pants' WHERE item_definition_id = 'common.items.armor.pants.rugged';
UPDATE item
SET item_definition_id = 'common.items.armor.starter.sandals_0' WHERE item_definition_id = 'common.items.armor.foot.sandals_0';

View File

@ -1 +0,0 @@
-- This file should undo anything in `up.sql`

View File

@ -1 +0,0 @@
-- This file should undo anything in `up.sql`

View File

@ -1 +0,0 @@
-- This file should undo anything in `up.sql`

View File

@ -1 +0,0 @@
-- This file should undo anything in `up.sql`

View File

@ -1 +0,0 @@
-- This file should undo anything in `up.sql`

View File

@ -1 +0,0 @@
-- This file should undo anything in `up.sql`

View File

@ -1 +0,0 @@
-- This file should undo anything in `up.sql`

View File

@ -1 +0,0 @@
-- This file should undo anything in `up.sql`

View File

@ -1 +0,0 @@
-- This file should undo anything in `up.sql`

View File

@ -1 +0,0 @@
-- This file should undo anything in `up.sql`

View File

@ -1 +0,0 @@
DATABASE_URL=../../../saves/db.sqlite

View File

@ -4,9 +4,9 @@
//! database updates and loading are communicated via requests to the //! database updates and loading are communicated via requests to the
//! [`CharacterLoader`] and [`CharacterUpdater`] while results/responses are //! [`CharacterLoader`] and [`CharacterUpdater`] while results/responses are
//! polled and handled each server tick. //! polled and handled each server tick.
extern crate diesel; extern crate rusqlite;
use super::{error::Error, models::*, schema, VelorenTransaction}; use super::{error::PersistenceError, models::*};
use crate::{ use crate::{
comp, comp,
comp::{item::MaterialStatManifest, Inventory}, comp::{item::MaterialStatManifest, Inventory},
@ -20,14 +20,14 @@ use crate::{
convert_waypoint_to_database_json, convert_waypoint_to_database_json,
}, },
character_loader::{CharacterCreationResult, CharacterDataResult, CharacterListResult}, character_loader::{CharacterCreationResult, CharacterDataResult, CharacterListResult},
error::Error::DatabaseError, error::PersistenceError::DatabaseError,
PersistedComponents, PersistedComponents,
}, },
}; };
use common::character::{CharacterId, CharacterItem, MAX_CHARACTERS_PER_PLAYER}; use common::character::{CharacterId, CharacterItem, MAX_CHARACTERS_PER_PLAYER};
use core::ops::Range; use core::ops::Range;
use diesel::{prelude::*, sql_query, sql_types::BigInt}; use rusqlite::{types::Value, ToSql, Transaction, NO_PARAMS};
use std::{collections::VecDeque, sync::Arc}; use std::{collections::VecDeque, rc::Rc};
use tracing::{error, trace, warn}; use tracing::{error, trace, warn};
/// Private module for very tightly coupled database conversion methods. In /// Private module for very tightly coupled database conversion methods. In
@ -53,15 +53,38 @@ struct CharacterContainers {
/// BFS the inventory/loadout to ensure that each is topologically sorted in the /// BFS the inventory/loadout to ensure that each is topologically sorted in the
/// sense required by convert_inventory_from_database_items to support recursive /// sense required by convert_inventory_from_database_items to support recursive
/// items /// items
pub fn load_items_bfs(connection: VelorenTransaction, root: i64) -> Result<Vec<Item>, Error> { pub fn load_items_bfs(
use schema::item::dsl::*; connection: &mut Transaction,
root: i64,
) -> Result<Vec<Item>, PersistenceError> {
let mut items = Vec::new(); let mut items = Vec::new();
let mut queue = VecDeque::new(); let mut queue = VecDeque::new();
queue.push_front(root); queue.push_front(root);
#[rustfmt::skip]
let mut stmt = connection.prepare_cached("
SELECT item_id,
parent_container_item_id,
item_definition_id,
stack_size,
position
FROM item
WHERE parent_container_item_id = ?1")?;
while let Some(id) = queue.pop_front() { while let Some(id) = queue.pop_front() {
let frontier = item let frontier = stmt
.filter(parent_container_item_id.eq(id)) .query_map(&[id], |row| {
.load::<Item>(&*connection)?; Ok(Item {
item_id: row.get(0)?,
parent_container_item_id: row.get(1)?,
item_definition_id: row.get(2)?,
stack_size: row.get(3)?,
position: row.get(4)?,
})
})?
.filter_map(Result::ok)
.collect::<Vec<Item>>();
for i in frontier.iter() { for i in frontier.iter() {
queue.push_back(i.item_id); queue.push_back(i.item_id);
} }
@ -77,34 +100,53 @@ pub fn load_items_bfs(connection: VelorenTransaction, root: i64) -> Result<Vec<I
pub fn load_character_data( pub fn load_character_data(
requesting_player_uuid: String, requesting_player_uuid: String,
char_id: CharacterId, char_id: CharacterId,
connection: VelorenTransaction, connection: &mut Transaction,
msm: &MaterialStatManifest, msm: &MaterialStatManifest,
) -> CharacterDataResult { ) -> CharacterDataResult {
use schema::{body::dsl::*, character::dsl::*, skill_group::dsl::*};
let character_containers = get_pseudo_containers(connection, char_id)?; let character_containers = get_pseudo_containers(connection, char_id)?;
let inventory_items = load_items_bfs(connection, character_containers.inventory_container_id)?; let inventory_items = load_items_bfs(connection, character_containers.inventory_container_id)?;
let loadout_items = load_items_bfs(connection, character_containers.loadout_container_id)?; let loadout_items = load_items_bfs(connection, character_containers.loadout_container_id)?;
let character_data = character #[rustfmt::skip]
.filter( let mut stmt = connection.prepare_cached("
schema::character::dsl::character_id SELECT c.character_id,
.eq(char_id) c.alias,
.and(player_uuid.eq(requesting_player_uuid)), c.waypoint,
) b.variant,
.first::<Character>(&*connection)?; b.body_data
FROM character c
JOIN body b ON (c.character_id = b.body_id)
WHERE c.player_uuid = ?1
AND c.character_id = ?2",
)?;
let char_body = body let (body_data, character_data) = stmt.query_row(
.filter(schema::body::dsl::body_id.eq(char_id)) &[requesting_player_uuid.clone(), char_id.to_string()],
.first::<Body>(&*connection)?; |row| {
let character_data = Character {
character_id: row.get(0)?,
player_uuid: requesting_player_uuid,
alias: row.get(1)?,
waypoint: row.get(2)?,
};
let body_data = Body {
body_id: row.get(0)?,
variant: row.get(3)?,
body_data: row.get(4)?,
};
Ok((body_data, character_data))
},
)?;
let char_waypoint = character_data.waypoint.as_ref().and_then(|x| { let char_waypoint = character_data.waypoint.as_ref().and_then(|x| {
match convert_waypoint_from_database_json(&x) { match convert_waypoint_from_database_json(&x) {
Ok(w) => Some(w), Ok(w) => Some(w),
Err(e) => { Err(e) => {
warn!( warn!(
"Error reading waypoint from database for character ID {}, error: {}", "Error reading waypoint from database for character ID
{}, error: {}",
char_id, e char_id, e
); );
None None
@ -112,16 +154,50 @@ pub fn load_character_data(
} }
}); });
let skill_data = schema::skill::dsl::skill #[rustfmt::skip]
.filter(schema::skill::dsl::entity_id.eq(char_id)) let mut stmt = connection.prepare_cached("
.load::<Skill>(&*connection)?; SELECT skill,
level
FROM skill
WHERE entity_id = ?1",
)?;
let skill_group_data = skill_group let skill_data = stmt
.filter(schema::skill_group::dsl::entity_id.eq(char_id)) .query_map(&[char_id], |row| {
.load::<SkillGroup>(&*connection)?; Ok(Skill {
entity_id: char_id,
skill: row.get(0)?,
level: row.get(1)?,
})
})?
.filter_map(Result::ok)
.collect::<Vec<Skill>>();
#[rustfmt::skip]
let mut stmt = connection.prepare_cached("
SELECT skill_group_kind,
exp,
available_sp,
earned_sp
FROM skill_group
WHERE entity_id = ?1",
)?;
let skill_group_data = stmt
.query_map(&[char_id], |row| {
Ok(SkillGroup {
entity_id: char_id,
skill_group_kind: row.get(0)?,
exp: row.get(1)?,
available_sp: row.get(2)?,
earned_sp: row.get(3)?,
})
})?
.filter_map(Result::ok)
.collect::<Vec<SkillGroup>>();
Ok(( Ok((
convert_body_from_database(&char_body)?, convert_body_from_database(&body_data)?,
convert_stats_from_database(character_data.alias, &skill_data, &skill_group_data), convert_stats_from_database(character_data.alias, &skill_data, &skill_group_data),
convert_inventory_from_database_items( convert_inventory_from_database_items(
character_containers.inventory_container_id, character_containers.inventory_container_id,
@ -143,24 +219,56 @@ pub fn load_character_data(
/// returned. /// returned.
pub fn load_character_list( pub fn load_character_list(
player_uuid_: &str, player_uuid_: &str,
connection: VelorenTransaction, connection: &mut Transaction,
msm: &MaterialStatManifest, msm: &MaterialStatManifest,
) -> CharacterListResult { ) -> CharacterListResult {
use schema::{body::dsl::*, character::dsl::*}; let characters;
{
#[rustfmt::skip]
let mut stmt = connection
.prepare_cached("
SELECT character_id,
alias
FROM character
WHERE player_uuid = ?1
ORDER BY character_id")?;
let result = character characters = stmt
.filter(player_uuid.eq(player_uuid_)) .query_map(&[player_uuid_], |row| {
.order(schema::character::dsl::character_id.desc()) Ok(Character {
.load::<Character>(&*connection)?; character_id: row.get(0)?,
alias: row.get(1)?,
result player_uuid: player_uuid_.to_owned(),
waypoint: None, // Not used for character select
})
})?
.map(|x| x.unwrap())
.collect::<Vec<Character>>();
}
characters
.iter() .iter()
.map(|character_data| { .map(|character_data| {
let char = convert_character_from_database(character_data); let char = convert_character_from_database(&character_data);
let db_body = body let db_body;
.filter(schema::body::dsl::body_id.eq(character_data.character_id))
.first::<Body>(&*connection)?; {
#[rustfmt::skip]
let mut stmt = connection
.prepare_cached("\
SELECT body_id,\
variant,\
body_data \
FROM body \
WHERE body_id = ?1")?;
db_body = stmt.query_row(&[char.id], |row| {
Ok(Body {
body_id: row.get(0)?,
variant: row.get(1)?,
body_data: row.get(2)?,
})
})?;
}
let char_body = convert_body_from_database(&db_body)?; let char_body = convert_body_from_database(&db_body)?;
@ -188,15 +296,11 @@ pub fn create_character(
uuid: &str, uuid: &str,
character_alias: &str, character_alias: &str,
persisted_components: PersistedComponents, persisted_components: PersistedComponents,
connection: VelorenTransaction, connection: &mut Transaction,
msm: &MaterialStatManifest, msm: &MaterialStatManifest,
) -> CharacterCreationResult { ) -> CharacterCreationResult {
use schema::item::dsl::*;
check_character_limit(uuid, connection)?; check_character_limit(uuid, connection)?;
use schema::{body, character, skill_group};
let (body, stats, inventory, waypoint) = persisted_components; let (body, stats, inventory, waypoint) = persisted_components;
// Fetch new entity IDs for character, inventory and loadout // Fetch new entity IDs for character, inventory and loadout
@ -230,66 +334,82 @@ pub fn create_character(
position: LOADOUT_PSEUDO_CONTAINER_POSITION.to_owned(), position: LOADOUT_PSEUDO_CONTAINER_POSITION.to_owned(),
}, },
]; ];
let pseudo_container_count = diesel::insert_into(item)
.values(pseudo_containers)
.execute(&*connection)?;
if pseudo_container_count != 3 { #[rustfmt::skip]
return Err(Error::OtherError(format!( let mut stmt = connection.prepare_cached("
"Error inserting initial pseudo containers for character id {} (expected 3, actual {})", INSERT INTO item (item_id,
character_id, pseudo_container_count parent_container_item_id,
))); item_definition_id,
stack_size,
position)
VALUES (?1, ?2, ?3, ?4, ?5)",
)?;
for pseudo_container in pseudo_containers {
stmt.execute(&[
&pseudo_container.item_id as &dyn ToSql,
&pseudo_container.parent_container_item_id,
&pseudo_container.item_definition_id,
&pseudo_container.stack_size,
&pseudo_container.position,
])?;
} }
drop(stmt);
#[rustfmt::skip]
let mut stmt = connection.prepare_cached("
INSERT INTO body (body_id,
variant,
body_data)
VALUES (?1, ?2, ?3)")?;
stmt.execute(&[
&character_id as &dyn ToSql,
&"humanoid".to_string(),
&convert_body_to_database_json(&body)?,
])?;
drop(stmt);
#[rustfmt::skip]
let mut stmt = connection.prepare_cached("
INSERT INTO character (character_id,
player_uuid,
alias,
waypoint)
VALUES (?1, ?2, ?3, ?4)")?;
stmt.execute(&[
&character_id as &dyn ToSql,
&uuid,
&character_alias,
&convert_waypoint_to_database_json(waypoint),
])?;
drop(stmt);
let skill_set = stats.skill_set; let skill_set = stats.skill_set;
// Insert body record
let new_body = Body {
body_id: character_id,
body_data: convert_body_to_database_json(&body)?,
variant: "humanoid".to_string(),
};
let body_count = diesel::insert_into(body::table)
.values(&new_body)
.execute(&*connection)?;
if body_count != 1 {
return Err(Error::OtherError(format!(
"Error inserting into body table for char_id {}",
character_id
)));
}
// Insert character record
let new_character = NewCharacter {
character_id,
player_uuid: uuid,
alias: &character_alias,
waypoint: convert_waypoint_to_database_json(waypoint),
};
let character_count = diesel::insert_into(character::table)
.values(&new_character)
.execute(&*connection)?;
if character_count != 1 {
return Err(Error::OtherError(format!(
"Error inserting into character table for char_id {}",
character_id
)));
}
let db_skill_groups = convert_skill_groups_to_database(character_id, skill_set.skill_groups); let db_skill_groups = convert_skill_groups_to_database(character_id, skill_set.skill_groups);
let skill_groups_count = diesel::insert_into(skill_group::table)
.values(&db_skill_groups)
.execute(&*connection)?;
if skill_groups_count != 1 { #[rustfmt::skip]
return Err(Error::OtherError(format!( let mut stmt = connection.prepare_cached("
"Error inserting into skill_group table for char_id {}", INSERT INTO skill_group (entity_id,
character_id skill_group_kind,
))); exp,
available_sp,
earned_sp)
VALUES (?1, ?2, ?3, ?4, ?5)")?;
for skill_group in db_skill_groups {
stmt.execute(&[
&character_id as &dyn ToSql,
&skill_group.skill_group_kind,
&skill_group.exp,
&skill_group.available_sp,
&skill_group.earned_sp,
])?;
} }
drop(stmt);
// Insert default inventory and loadout item records // Insert default inventory and loadout item records
let mut inserts = Vec::new(); let mut inserts = Vec::new();
@ -305,21 +425,26 @@ pub fn create_character(
next_id next_id
})?; })?;
let expected_inserted_count = inserts.len(); #[rustfmt::skip]
let inserted_items = inserts let mut stmt = connection.prepare_cached("
.into_iter() INSERT INTO item (item_id,
.map(|item_pair| item_pair.model) parent_container_item_id,
.collect::<Vec<_>>(); item_definition_id,
let inserted_count = diesel::insert_into(item) stack_size,
.values(&inserted_items) position)
.execute(&*connection)?; VALUES (?1, ?2, ?3, ?4, ?5)",
)?;
if expected_inserted_count != inserted_count { for item in inserts {
return Err(Error::OtherError(format!( stmt.execute(&[
"Expected insertions={}, actual={}, for char_id {}--unsafe to continue transaction.", &item.model.item_id as &dyn ToSql,
expected_inserted_count, inserted_count, character_id &item.model.parent_container_item_id,
))); &item.model.item_definition_id,
&item.model.stack_size,
&item.model.position,
])?;
} }
drop(stmt);
load_character_list(uuid, connection, msm).map(|list| (character_id, list)) load_character_list(uuid, connection, msm).map(|list| (character_id, list))
} }
@ -328,82 +453,100 @@ pub fn create_character(
pub fn delete_character( pub fn delete_character(
requesting_player_uuid: &str, requesting_player_uuid: &str,
char_id: CharacterId, char_id: CharacterId,
connection: VelorenTransaction, connection: &mut Transaction,
msm: &MaterialStatManifest, msm: &MaterialStatManifest,
) -> CharacterListResult { ) -> CharacterListResult {
use schema::{body::dsl::*, character::dsl::*, skill::dsl::*, skill_group::dsl::*}; #[rustfmt::skip]
let mut stmt = connection.prepare_cached("
SELECT COUNT(1)
FROM character
WHERE character_id = ?1
AND player_uuid = ?2")?;
// Load the character to delete - ensures that the requesting player let result = stmt.query_row(&[&char_id as &dyn ToSql, &requesting_player_uuid], |row| {
// owns the character let y: i64 = row.get(0)?;
let _character_data = character Ok(y)
.filter( })?;
schema::character::dsl::character_id drop(stmt);
.eq(char_id)
.and(player_uuid.eq(requesting_player_uuid)),
)
.first::<Character>(&*connection)?;
if result != 1 {
return Err(PersistenceError::OtherError(
"Requested character to delete does not belong to the requesting player".to_string(),
));
}
// Delete skills // Delete skills
diesel::delete(skill_group.filter(schema::skill_group::dsl::entity_id.eq(char_id))) let mut stmt = connection.prepare_cached(
.execute(&*connection)?; "
DELETE
FROM skill
WHERE entity_id = ?1",
)?;
diesel::delete(skill.filter(schema::skill::dsl::entity_id.eq(char_id))) stmt.execute(&[&char_id])?;
.execute(&*connection)?; drop(stmt);
// Delete skill groups
let mut stmt = connection.prepare_cached(
"
DELETE
FROM skill_group
WHERE entity_id = ?1",
)?;
stmt.execute(&[&char_id])?;
drop(stmt);
// Delete character // Delete character
let character_count = diesel::delete( let mut stmt = connection.prepare_cached(
character "
.filter(schema::character::dsl::character_id.eq(char_id)) DELETE
.filter(player_uuid.eq(requesting_player_uuid)), FROM character
) WHERE character_id = ?1",
.execute(&*connection)?; )?;
if character_count != 1 { stmt.execute(&[&char_id])?;
return Err(Error::OtherError(format!( drop(stmt);
"Error deleting from character table for char_id {}",
char_id
)));
}
// Delete body // Delete body
let body_count = diesel::delete(body.filter(schema::body::dsl::body_id.eq(char_id))) let mut stmt = connection.prepare_cached(
.execute(&*connection)?; "
DELETE
FROM body
WHERE body_id = ?1",
)?;
if body_count != 1 { stmt.execute(&[&char_id])?;
return Err(Error::OtherError(format!( drop(stmt);
"Error deleting from body table for char_id {}",
char_id
)));
}
// Delete all items, recursively walking all containers starting from the // Delete all items, recursively walking all containers starting from the
// "character" pseudo-container that is the root for all items owned by // "character" pseudo-container that is the root for all items owned by
// a character. // a character.
let item_count = diesel::sql_query(format!( let mut stmt = connection.prepare_cached(
" "
WITH RECURSIVE WITH RECURSIVE
parents AS ( parents AS (
SELECT item_id SELECT item_id
FROM item
WHERE item.item_id = ?1 -- Item with character id is the character pseudo-container
UNION ALL
SELECT item.item_id
FROM item,
parents
WHERE item.parent_container_item_id = parents.item_id
)
DELETE
FROM item FROM item
WHERE item.item_id = {} -- Item with character id is the character pseudo-container WHERE EXISTS (SELECT 1 FROM parents WHERE parents.item_id = item.item_id)",
UNION ALL )?;
SELECT item.item_id
FROM item,
parents
WHERE item.parent_container_item_id = parents.item_id
)
DELETE
FROM item
WHERE EXISTS (SELECT 1 FROM parents WHERE parents.item_id = item.item_id)",
char_id
))
.execute(&*connection)?;
if item_count < 3 { let deleted_item_count = stmt.execute(&[&char_id])?;
return Err(Error::OtherError(format!( drop(stmt);
if deleted_item_count < 3 {
return Err(PersistenceError::OtherError(format!(
"Error deleting from item table for char_id {} (expected at least 3 deletions, found \ "Error deleting from item table for char_id {} (expected at least 3 deletions, found \
{})", {})",
char_id, item_count char_id, deleted_item_count
))); )));
} }
@ -412,24 +555,24 @@ pub fn delete_character(
/// Before creating a character, we ensure that the limit on the number of /// Before creating a character, we ensure that the limit on the number of
/// characters has not been exceeded /// characters has not been exceeded
pub fn check_character_limit(uuid: &str, connection: VelorenTransaction) -> Result<(), Error> { pub fn check_character_limit(
use diesel::dsl::count_star; uuid: &str,
use schema::character::dsl::*; connection: &mut Transaction,
) -> Result<(), PersistenceError> {
#[rustfmt::skip]
let mut stmt = connection.prepare_cached("
SELECT COUNT(1)
FROM character
WHERE player_uuid = ?1")?;
let character_count = character #[allow(clippy::needless_question_mark)]
.select(count_star()) let character_count: i64 = stmt.query_row(&[&uuid], |row| Ok(row.get(0)?))?;
.filter(player_uuid.eq(uuid)) drop(stmt);
.load::<i64>(&*connection)?;
match character_count.first() { if character_count < MAX_CHARACTERS_PER_PLAYER as i64 {
Some(count) => { Ok(())
if count < &(MAX_CHARACTERS_PER_PLAYER as i64) { } else {
Ok(()) Err(PersistenceError::CharacterLimitReached)
} else {
Err(Error::CharacterLimitReached)
}
},
_ => Ok(()),
} }
} }
@ -440,48 +583,32 @@ pub fn check_character_limit(uuid: &str, connection: VelorenTransaction) -> Resu
/// ///
/// These are then inserted into the entities table. /// These are then inserted into the entities table.
fn get_new_entity_ids( fn get_new_entity_ids(
conn: VelorenTransaction, conn: &mut Transaction,
mut max: impl FnMut(i64) -> i64, mut max: impl FnMut(i64) -> i64,
) -> Result<Range<EntityId>, Error> { ) -> Result<Range<EntityId>, PersistenceError> {
use super::schema::entity::dsl::*;
#[derive(QueryableByName)]
struct NextEntityId {
#[sql_type = "BigInt"]
entity_id: i64,
}
// The sqlite_sequence table is used here to avoid reusing entity IDs for // The sqlite_sequence table is used here to avoid reusing entity IDs for
// deleted entities. This table always contains the highest used ID for each // deleted entities. This table always contains the highest used ID for
// AUTOINCREMENT column in a SQLite database. // each AUTOINCREMENT column in a SQLite database.
let next_entity_id = sql_query( #[rustfmt::skip]
let mut stmt = conn.prepare_cached(
" "
SELECT seq + 1 AS entity_id SELECT seq + 1 AS entity_id
FROM sqlite_sequence FROM sqlite_sequence
WHERE name = 'entity'", WHERE name = 'entity'",
) )?;
.load::<NextEntityId>(&*conn)?
.pop()
.ok_or_else(|| Error::OtherError("No rows returned for sqlite_sequence query ".to_string()))?
.entity_id;
#[allow(clippy::needless_question_mark)]
let next_entity_id = stmt.query_row(NO_PARAMS, |row| Ok(row.get(0)?))?;
let max_entity_id = max(next_entity_id); let max_entity_id = max(next_entity_id);
// Create a new range of IDs and insert them into the entity table // Create a new range of IDs and insert them into the entity table
let new_ids: Range<EntityId> = next_entity_id..max_entity_id; let new_ids: Range<EntityId> = next_entity_id..max_entity_id;
let new_entities: Vec<Entity> = new_ids.clone().map(|x| Entity { entity_id: x }).collect(); let mut stmt = conn.prepare_cached("INSERT INTO entity (entity_id) VALUES (?1)")?;
let actual_count = diesel::insert_into(entity) // TODO: bulk insert? rarray doesn't seem to work in VALUES clause
.values(&new_entities) for i in new_ids.clone() {
.execute(&*conn)?; stmt.execute(&[i])?;
if actual_count != new_entities.len() {
return Err(Error::OtherError(format!(
"Error updating entity table: expected to add the range {:?}) to entities, but actual \
insertions={}",
new_ids, actual_count
)));
} }
trace!( trace!(
@ -498,9 +625,9 @@ fn get_new_entity_ids(
/// Fetches the pseudo_container IDs for a character /// Fetches the pseudo_container IDs for a character
fn get_pseudo_containers( fn get_pseudo_containers(
connection: VelorenTransaction, connection: &mut Transaction,
character_id: CharacterId, character_id: CharacterId,
) -> Result<CharacterContainers, Error> { ) -> Result<CharacterContainers, PersistenceError> {
let character_containers = CharacterContainers { let character_containers = CharacterContainers {
loadout_container_id: get_pseudo_container_id( loadout_container_id: get_pseudo_container_id(
connection, connection,
@ -518,20 +645,28 @@ fn get_pseudo_containers(
} }
fn get_pseudo_container_id( fn get_pseudo_container_id(
connection: VelorenTransaction, connection: &mut Transaction,
character_id: CharacterId, character_id: CharacterId,
pseudo_container_position: &str, pseudo_container_position: &str,
) -> Result<EntityId, Error> { ) -> Result<EntityId, PersistenceError> {
use super::schema::item::dsl::*; #[rustfmt::skip]
match item let mut stmt = connection.prepare_cached("\
.select(item_id) SELECT item_id
.filter( FROM item
parent_container_item_id WHERE parent_container_item_id = ?1
.eq(character_id) AND position = ?2",
.and(position.eq(pseudo_container_position)), )?;
)
.first::<EntityId>(&*connection) #[allow(clippy::needless_question_mark)]
{ let res = stmt.query_row(
&[
character_id.to_string(),
pseudo_container_position.to_string(),
],
|row| Ok(row.get(0)?),
);
match res {
Ok(id) => Ok(id), Ok(id) => Ok(id),
Err(e) => { Err(e) => {
error!( error!(
@ -550,16 +685,14 @@ pub fn update(
char_stats: comp::Stats, char_stats: comp::Stats,
inventory: comp::Inventory, inventory: comp::Inventory,
char_waypoint: Option<comp::Waypoint>, char_waypoint: Option<comp::Waypoint>,
connection: VelorenTransaction, connection: &mut Transaction,
) -> Result<Vec<Arc<common::comp::item::ItemId>>, Error> { ) -> Result<(), PersistenceError> {
use super::schema::{character::dsl::*, item::dsl::*, skill_group::dsl::*};
let pseudo_containers = get_pseudo_containers(connection, char_id)?; let pseudo_containers = get_pseudo_containers(connection, char_id)?;
let mut upserts = Vec::new(); let mut upserts = Vec::new();
// First, get all the entity IDs for any new items, and identify which slots to // First, get all the entity IDs for any new items, and identify which
// upsert and which ones to delete. // slots to upsert and which ones to delete.
get_new_entity_ids(connection, |mut next_id| { get_new_entity_ids(connection, |mut next_id| {
let upserts_ = convert_items_to_database_items( let upserts_ = convert_items_to_database_items(
pseudo_containers.loadout_container_id, pseudo_containers.loadout_container_id,
@ -573,33 +706,36 @@ pub fn update(
// Next, delete any slots we aren't upserting. // Next, delete any slots we aren't upserting.
trace!("Deleting items for character_id {}", char_id); trace!("Deleting items for character_id {}", char_id);
let mut existing_item_ids: Vec<i64> = vec![ let mut existing_item_ids: Vec<_> = vec![
pseudo_containers.inventory_container_id, Value::from(pseudo_containers.inventory_container_id),
pseudo_containers.loadout_container_id, Value::from(pseudo_containers.loadout_container_id),
]; ];
for it in load_items_bfs(connection, pseudo_containers.inventory_container_id)? { for it in load_items_bfs(connection, pseudo_containers.inventory_container_id)? {
existing_item_ids.push(it.item_id); existing_item_ids.push(Value::from(it.item_id));
} }
for it in load_items_bfs(connection, pseudo_containers.loadout_container_id)? { for it in load_items_bfs(connection, pseudo_containers.loadout_container_id)? {
existing_item_ids.push(it.item_id); existing_item_ids.push(Value::from(it.item_id));
} }
let existing_items = parent_container_item_id.eq_any(existing_item_ids);
let non_upserted_items = item_id.ne_all(
upserts
.iter()
.map(|item_pair| item_pair.model.item_id)
.collect::<Vec<_>>(),
);
let delete_count = diesel::delete(item.filter(existing_items.and(non_upserted_items))) let non_upserted_items = upserts
.execute(&*connection)?; .iter()
.map(|item_pair| Value::from(item_pair.model.item_id))
.collect::<Vec<Value>>();
#[rustfmt::skip]
let mut stmt = connection.prepare_cached("
DELETE
FROM item
WHERE parent_container_item_id
IN rarray(?1)
AND item_id NOT IN rarray(?2)")?;
let delete_count = stmt.execute(&[Rc::new(existing_item_ids), Rc::new(non_upserted_items)])?;
trace!("Deleted {} items", delete_count); trace!("Deleted {} items", delete_count);
// Upsert items // Upsert items
let expected_upsert_count = upserts.len(); let expected_upsert_count = upserts.len();
let mut upserted_comps = Vec::new();
if expected_upsert_count > 0 { if expected_upsert_count > 0 {
let (upserted_items, upserted_comps_): (Vec<_>, Vec<_>) = upserts let (upserted_items, _): (Vec<_>, Vec<_>) = upserts
.into_iter() .into_iter()
.map(|model_pair| { .map(|model_pair| {
debug_assert_eq!( debug_assert_eq!(
@ -609,7 +745,6 @@ pub fn update(
(model_pair.model, model_pair.comp) (model_pair.model, model_pair.comp)
}) })
.unzip(); .unzip();
upserted_comps = upserted_comps_;
trace!( trace!(
"Upserting items {:?} for character_id {}", "Upserting items {:?} for character_id {}",
upserted_items, upserted_items,
@ -617,22 +752,32 @@ pub fn update(
); );
// When moving inventory items around, foreign key constraints on // When moving inventory items around, foreign key constraints on
// `parent_container_item_id` can be temporarily violated by one upsert, but // `parent_container_item_id` can be temporarily violated by one
// restored by another upsert. Deferred constraints allow SQLite to check this // upsert, but restored by another upsert. Deferred constraints
// when committing the transaction. The `defer_foreign_keys` pragma treats the // allow SQLite to check this when committing the transaction.
// foreign key constraints as deferred for the next transaction (it turns itself // The `defer_foreign_keys` pragma treats the foreign key
// constraints as deferred for the next transaction (it turns itself
// off at the commit boundary). https://sqlite.org/foreignkeys.html#fk_deferred // off at the commit boundary). https://sqlite.org/foreignkeys.html#fk_deferred
connection.execute("PRAGMA defer_foreign_keys = ON;")?; connection.pragma_update(None, "defer_foreign_keys", &"ON".to_string())?;
let upsert_count = diesel::replace_into(item)
.values(&upserted_items) #[rustfmt::skip]
.execute(&*connection)?; let mut stmt = connection.prepare_cached("
trace!("upsert_count: {}", upsert_count); REPLACE
if upsert_count != expected_upsert_count { INTO item (item_id,
return Err(Error::OtherError(format!( parent_container_item_id,
"Expected upsertions={}, actual={}, for char_id {}--unsafe to continue \ item_definition_id,
transaction.", stack_size,
expected_upsert_count, upsert_count, char_id position)
))); VALUES (?1, ?2, ?3, ?4, ?5)")?;
for item in upserted_items.iter() {
stmt.execute(&[
&item.item_id as &dyn ToSql,
&item.parent_container_item_id,
&item.item_definition_id,
&item.stack_size,
&item.position,
])?;
} }
} }
@ -640,43 +785,74 @@ pub fn update(
let db_skill_groups = convert_skill_groups_to_database(char_id, char_skill_set.skill_groups); let db_skill_groups = convert_skill_groups_to_database(char_id, char_skill_set.skill_groups);
diesel::replace_into(skill_group) #[rustfmt::skip]
.values(&db_skill_groups) let mut stmt = connection.prepare_cached("
.execute(&*connection)?; REPLACE
INTO skill_group (entity_id,
skill_group_kind,
exp,
available_sp,
earned_sp)
VALUES (?1, ?2, ?3, ?4, ?5)")?;
for skill_group in db_skill_groups {
stmt.execute(&[
&skill_group.entity_id as &dyn ToSql,
&skill_group.skill_group_kind,
&skill_group.exp,
&skill_group.available_sp,
&skill_group.earned_sp,
])?;
}
let db_skills = convert_skills_to_database(char_id, char_skill_set.skills); let db_skills = convert_skills_to_database(char_id, char_skill_set.skills);
let delete_count = diesel::delete( let known_skills = Rc::new(
schema::skill::dsl::skill.filter( db_skills
schema::skill::dsl::entity_id.eq(char_id).and( .iter()
schema::skill::dsl::skill_type.ne_all( .map(|x| Value::from(x.skill.clone()))
db_skills .collect::<Vec<Value>>(),
.iter() );
.map(|x| x.skill_type.clone())
.collect::<Vec<_>>(), #[rustfmt::skip]
), let mut stmt = connection.prepare_cached("
), DELETE
), FROM skill
) WHERE entity_id = ?1
.execute(&*connection)?; AND skill NOT IN rarray(?2)")?;
let delete_count = stmt.execute(&[&char_id as &dyn ToSql, &known_skills])?;
trace!("Deleted {} skills", delete_count); trace!("Deleted {} skills", delete_count);
diesel::replace_into(schema::skill::dsl::skill) #[rustfmt::skip]
.values(&db_skills) let mut stmt = connection.prepare_cached("
.execute(&*connection)?; REPLACE
INTO skill (entity_id,
skill,
level)
VALUES (?1, ?2, ?3)")?;
for skill in db_skills {
stmt.execute(&[&skill.entity_id as &dyn ToSql, &skill.skill, &skill.level])?;
}
let db_waypoint = convert_waypoint_to_database_json(char_waypoint); let db_waypoint = convert_waypoint_to_database_json(char_waypoint);
let waypoint_count =
diesel::update(character.filter(schema::character::dsl::character_id.eq(char_id))) #[rustfmt::skip]
.set(waypoint.eq(db_waypoint)) let mut stmt = connection.prepare_cached("
.execute(&*connection)?; UPDATE character
SET waypoint = ?1
WHERE character_id = ?2
")?;
let waypoint_count = stmt.execute(&[&db_waypoint as &dyn ToSql, &char_id])?;
if waypoint_count != 1 { if waypoint_count != 1 {
return Err(Error::OtherError(format!( return Err(PersistenceError::OtherError(format!(
"Error updating character table for char_id {}", "Error updating character table for char_id {}",
char_id char_id
))); )));
} }
Ok(upserted_comps) Ok(())
} }

View File

@ -4,7 +4,7 @@ use crate::persistence::{
}; };
use crate::persistence::{ use crate::persistence::{
error::Error, error::PersistenceError,
json_models::{self, CharacterPosition, HumanoidBody}, json_models::{self, CharacterPosition, HumanoidBody},
}; };
use common::{ use common::{
@ -23,6 +23,7 @@ use common::{
use core::{convert::TryFrom, num::NonZeroU64}; use core::{convert::TryFrom, num::NonZeroU64};
use hashbrown::HashMap; use hashbrown::HashMap;
use std::{collections::VecDeque, sync::Arc}; use std::{collections::VecDeque, sync::Arc};
use tracing::trace;
#[derive(Debug)] #[derive(Debug)]
pub struct ItemModelPair { pub struct ItemModelPair {
@ -174,17 +175,17 @@ pub fn convert_items_to_database_items(
} }
} }
upserts.sort_by_key(|pair| (depth[&pair.model.item_id], pair.model.item_id)); upserts.sort_by_key(|pair| (depth[&pair.model.item_id], pair.model.item_id));
tracing::debug!("upserts: {:#?}", upserts); trace!("upserts: {:#?}", upserts);
upserts upserts
} }
pub fn convert_body_to_database_json(body: &CompBody) -> Result<String, Error> { pub fn convert_body_to_database_json(body: &CompBody) -> Result<String, PersistenceError> {
let json_model = match body { let json_model = match body {
common::comp::Body::Humanoid(humanoid_body) => HumanoidBody::from(humanoid_body), common::comp::Body::Humanoid(humanoid_body) => HumanoidBody::from(humanoid_body),
_ => unimplemented!("Only humanoid bodies are currently supported for persistence"), _ => unimplemented!("Only humanoid bodies are currently supported for persistence"),
}; };
serde_json::to_string(&json_model).map_err(Error::SerializationError) serde_json::to_string(&json_model).map_err(PersistenceError::SerializationError)
} }
pub fn convert_waypoint_to_database_json(waypoint: Option<Waypoint>) -> Option<String> { pub fn convert_waypoint_to_database_json(waypoint: Option<Waypoint>) -> Option<String> {
@ -196,7 +197,10 @@ pub fn convert_waypoint_to_database_json(waypoint: Option<Waypoint>) -> Option<S
Some( Some(
serde_json::to_string(&charpos) serde_json::to_string(&charpos)
.map_err(|err| { .map_err(|err| {
Error::ConversionError(format!("Error encoding waypoint: {:?}", err)) PersistenceError::ConversionError(format!(
"Error encoding waypoint: {:?}",
err
))
}) })
.ok()?, .ok()?,
) )
@ -205,10 +209,10 @@ pub fn convert_waypoint_to_database_json(waypoint: Option<Waypoint>) -> Option<S
} }
} }
pub fn convert_waypoint_from_database_json(position: &str) -> Result<Waypoint, Error> { pub fn convert_waypoint_from_database_json(position: &str) -> Result<Waypoint, PersistenceError> {
let character_position = let character_position =
serde_json::de::from_str::<CharacterPosition>(position).map_err(|err| { serde_json::de::from_str::<CharacterPosition>(position).map_err(|err| {
Error::ConversionError(format!( PersistenceError::ConversionError(format!(
"Error de-serializing waypoint: {} err: {}", "Error de-serializing waypoint: {} err: {}",
position, err position, err
)) ))
@ -227,7 +231,7 @@ pub fn convert_inventory_from_database_items(
loadout_container_id: i64, loadout_container_id: i64,
loadout_items: &[Item], loadout_items: &[Item],
msm: &MaterialStatManifest, msm: &MaterialStatManifest,
) -> Result<Inventory, Error> { ) -> Result<Inventory, PersistenceError> {
// Loadout items must be loaded before inventory items since loadout items // Loadout items must be loaded before inventory items since loadout items
// provide inventory slots. Since items stored inside loadout items actually // provide inventory slots. Since items stored inside loadout items actually
// have their parent_container_item_id as the loadout pseudo-container we rely // have their parent_container_item_id as the loadout pseudo-container we rely
@ -248,7 +252,7 @@ pub fn convert_inventory_from_database_items(
// Item ID // Item ID
comp.store(Some(NonZeroU64::try_from(db_item.item_id as u64).map_err( comp.store(Some(NonZeroU64::try_from(db_item.item_id as u64).map_err(
|_| Error::ConversionError("Item with zero item_id".to_owned()), |_| PersistenceError::ConversionError("Item with zero item_id".to_owned()),
)?)); )?));
// Stack Size // Stack Size
@ -257,13 +261,15 @@ pub fn convert_inventory_from_database_items(
// (to be dropped next to the player) as this could be the result of // (to be dropped next to the player) as this could be the result of
// a change in the max amount for that item. // a change in the max amount for that item.
item.set_amount(u32::try_from(db_item.stack_size).map_err(|_| { item.set_amount(u32::try_from(db_item.stack_size).map_err(|_| {
Error::ConversionError(format!( PersistenceError::ConversionError(format!(
"Invalid item stack size for stackable={}: {}", "Invalid item stack size for stackable={}: {}",
item.is_stackable(), item.is_stackable(),
&db_item.stack_size &db_item.stack_size
)) ))
})?) })?)
.map_err(|_| Error::ConversionError("Error setting amount for item".to_owned()))?; .map_err(|_| {
PersistenceError::ConversionError("Error setting amount for item".to_owned())
})?;
} }
// Insert item into inventory // Insert item into inventory
@ -271,7 +277,7 @@ pub fn convert_inventory_from_database_items(
// Slot position // Slot position
let slot = |s: &str| { let slot = |s: &str| {
serde_json::from_str::<InvSlotId>(s).map_err(|_| { serde_json::from_str::<InvSlotId>(s).map_err(|_| {
Error::ConversionError(format!( PersistenceError::ConversionError(format!(
"Failed to parse item position: {:?}", "Failed to parse item position: {:?}",
&db_item.position &db_item.position
)) ))
@ -288,7 +294,7 @@ pub fn convert_inventory_from_database_items(
// (to be dropped next to the player) as this could be the // (to be dropped next to the player) as this could be the
// result of a change in the slot capacity for an equipped bag // result of a change in the slot capacity for an equipped bag
// (or a change in the inventory size). // (or a change in the inventory size).
Error::ConversionError(format!( PersistenceError::ConversionError(format!(
"Error inserting item into inventory, position: {:?}", "Error inserting item into inventory, position: {:?}",
slot slot
)) ))
@ -298,7 +304,7 @@ pub fn convert_inventory_from_database_items(
// If inventory.insert returns an item, it means it was swapped for an item that // If inventory.insert returns an item, it means it was swapped for an item that
// already occupied the slot. Multiple items being stored in the database for // already occupied the slot. Multiple items being stored in the database for
// the same slot is an error. // the same slot is an error.
return Err(Error::ConversionError( return Err(PersistenceError::ConversionError(
"Inserted an item into the same slot twice".to_string(), "Inserted an item into the same slot twice".to_string(),
)); ));
} }
@ -306,14 +312,14 @@ pub fn convert_inventory_from_database_items(
if let Some(Some(parent)) = inventory.slot_mut(slot(&inventory_items[j].position)?) { if let Some(Some(parent)) = inventory.slot_mut(slot(&inventory_items[j].position)?) {
parent.add_component(item, msm); parent.add_component(item, msm);
} else { } else {
return Err(Error::ConversionError(format!( return Err(PersistenceError::ConversionError(format!(
"Parent slot {} for component {} was empty even though it occurred earlier in \ "Parent slot {} for component {} was empty even though it occurred earlier in \
the loop?", the loop?",
db_item.parent_container_item_id, db_item.item_id db_item.parent_container_item_id, db_item.item_id
))); )));
} }
} else { } else {
return Err(Error::ConversionError(format!( return Err(PersistenceError::ConversionError(format!(
"Couldn't find parent item {} before item {} in inventory", "Couldn't find parent item {} before item {} in inventory",
db_item.parent_container_item_id, db_item.item_id db_item.parent_container_item_id, db_item.item_id
))); )));
@ -327,7 +333,7 @@ pub fn convert_loadout_from_database_items(
loadout_container_id: i64, loadout_container_id: i64,
database_items: &[Item], database_items: &[Item],
msm: &MaterialStatManifest, msm: &MaterialStatManifest,
) -> Result<Loadout, Error> { ) -> Result<Loadout, PersistenceError> {
let loadout_builder = LoadoutBuilder::new(); let loadout_builder = LoadoutBuilder::new();
let mut loadout = loadout_builder.build(); let mut loadout = loadout_builder.build();
let mut item_indices = HashMap::new(); let mut item_indices = HashMap::new();
@ -340,16 +346,18 @@ pub fn convert_loadout_from_database_items(
// NOTE: item id is currently *unique*, so we can store the ID safely. // NOTE: item id is currently *unique*, so we can store the ID safely.
let comp = item.get_item_id_for_database(); let comp = item.get_item_id_for_database();
comp.store(Some(NonZeroU64::try_from(db_item.item_id as u64).map_err( comp.store(Some(NonZeroU64::try_from(db_item.item_id as u64).map_err(
|_| Error::ConversionError("Item with zero item_id".to_owned()), |_| PersistenceError::ConversionError("Item with zero item_id".to_owned()),
)?)); )?));
let convert_error = |err| match err { let convert_error = |err| match err {
LoadoutError::InvalidPersistenceKey => { LoadoutError::InvalidPersistenceKey => PersistenceError::ConversionError(format!(
Error::ConversionError(format!("Invalid persistence key: {}", &db_item.position)) "Invalid persistence key: {}",
}, &db_item.position
LoadoutError::NoParentAtSlot => { )),
Error::ConversionError(format!("No parent item at slot: {}", &db_item.position)) LoadoutError::NoParentAtSlot => PersistenceError::ConversionError(format!(
}, "No parent item at slot: {}",
&db_item.position
)),
}; };
if db_item.parent_container_item_id == loadout_container_id { if db_item.parent_container_item_id == loadout_container_id {
@ -363,7 +371,7 @@ pub fn convert_loadout_from_database_items(
}) })
.map_err(convert_error)?; .map_err(convert_error)?;
} else { } else {
return Err(Error::ConversionError(format!( return Err(PersistenceError::ConversionError(format!(
"Couldn't find parent item {} before item {} in loadout", "Couldn't find parent item {} before item {} in loadout",
db_item.parent_container_item_id, db_item.item_id db_item.parent_container_item_id, db_item.item_id
))); )));
@ -373,7 +381,7 @@ pub fn convert_loadout_from_database_items(
Ok(loadout) Ok(loadout)
} }
pub fn convert_body_from_database(body: &Body) -> Result<CompBody, Error> { pub fn convert_body_from_database(body: &Body) -> Result<CompBody, PersistenceError> {
Ok(match body.variant.as_str() { Ok(match body.variant.as_str() {
"humanoid" => { "humanoid" => {
let json_model = serde_json::de::from_str::<HumanoidBody>(&body.body_data)?; let json_model = serde_json::de::from_str::<HumanoidBody>(&body.body_data)?;
@ -381,13 +389,16 @@ pub fn convert_body_from_database(body: &Body) -> Result<CompBody, Error> {
species: common::comp::humanoid::ALL_SPECIES species: common::comp::humanoid::ALL_SPECIES
.get(json_model.species as usize) .get(json_model.species as usize)
.ok_or_else(|| { .ok_or_else(|| {
Error::ConversionError(format!("Missing species: {}", json_model.species)) PersistenceError::ConversionError(format!(
"Missing species: {}",
json_model.species
))
})? })?
.to_owned(), .to_owned(),
body_type: common::comp::humanoid::ALL_BODY_TYPES body_type: common::comp::humanoid::ALL_BODY_TYPES
.get(json_model.body_type as usize) .get(json_model.body_type as usize)
.ok_or_else(|| { .ok_or_else(|| {
Error::ConversionError(format!( PersistenceError::ConversionError(format!(
"Missing body_type: {}", "Missing body_type: {}",
json_model.body_type json_model.body_type
)) ))
@ -403,7 +414,7 @@ pub fn convert_body_from_database(body: &Body) -> Result<CompBody, Error> {
}) })
}, },
_ => { _ => {
return Err(Error::ConversionError( return Err(PersistenceError::ConversionError(
"Only humanoid bodies are supported for characters".to_string(), "Only humanoid bodies are supported for characters".to_string(),
)); ));
}, },
@ -439,9 +450,9 @@ pub fn convert_stats_from_database(
new_stats new_stats
} }
fn get_item_from_asset(item_definition_id: &str) -> Result<common::comp::Item, Error> { fn get_item_from_asset(item_definition_id: &str) -> Result<common::comp::Item, PersistenceError> {
common::comp::Item::new_from_asset(item_definition_id).map_err(|err| { common::comp::Item::new_from_asset(item_definition_id).map_err(|err| {
Error::AssetError(format!( PersistenceError::AssetError(format!(
"Error loading item asset: {} - {}", "Error loading item asset: {} - {}",
item_definition_id, item_definition_id,
err.to_string() err.to_string()
@ -467,7 +478,7 @@ fn convert_skill_groups_from_database(skill_groups: &[SkillGroup]) -> Vec<skills
fn convert_skills_from_database(skills: &[Skill]) -> HashMap<skills::Skill, Option<u16>> { fn convert_skills_from_database(skills: &[Skill]) -> HashMap<skills::Skill, Option<u16>> {
let mut new_skills = HashMap::new(); let mut new_skills = HashMap::new();
for skill in skills.iter() { for skill in skills.iter() {
let new_skill = json_models::db_string_to_skill(&skill.skill_type); let new_skill = json_models::db_string_to_skill(&skill.skill);
new_skills.insert(new_skill, skill.level.map(|l| l as u16)); new_skills.insert(new_skill, skill.level.map(|l| l as u16));
} }
new_skills new_skills
@ -497,7 +508,7 @@ pub fn convert_skills_to_database(
.iter() .iter()
.map(|(s, l)| Skill { .map(|(s, l)| Skill {
entity_id, entity_id,
skill_type: json_models::skill_to_db_string(*s), skill: json_models::skill_to_db_string(*s),
level: l.map(|l| l as i32), level: l.map(|l| l as i32),
}) })
.collect() .collect()

View File

@ -1,7 +1,7 @@
use crate::persistence::{ use crate::persistence::{
character::{create_character, delete_character, load_character_data, load_character_list}, character::{create_character, delete_character, load_character_data, load_character_list},
error::Error, error::PersistenceError,
establish_connection, PersistedComponents, establish_connection, DatabaseSettings, PersistedComponents,
}; };
use common::{ use common::{
character::{CharacterId, CharacterItem}, character::{CharacterId, CharacterItem},
@ -9,12 +9,14 @@ use common::{
}; };
use crossbeam_channel::{self, TryIter}; use crossbeam_channel::{self, TryIter};
use lazy_static::lazy_static; use lazy_static::lazy_static;
use std::path::Path; use rusqlite::Transaction;
use tracing::error; use std::sync::{Arc, RwLock};
use tracing::{error, trace};
pub(crate) type CharacterListResult = Result<Vec<CharacterItem>, Error>; pub(crate) type CharacterListResult = Result<Vec<CharacterItem>, PersistenceError>;
pub(crate) type CharacterCreationResult = Result<(CharacterId, Vec<CharacterItem>), Error>; pub(crate) type CharacterCreationResult =
pub(crate) type CharacterDataResult = Result<PersistedComponents, Error>; Result<(CharacterId, Vec<CharacterItem>), PersistenceError>;
pub(crate) type CharacterDataResult = Result<PersistedComponents, PersistenceError>;
type CharacterLoaderRequest = (specs::Entity, CharacterLoaderRequestKind); type CharacterLoaderRequest = (specs::Entity, CharacterLoaderRequestKind);
/// Available database operations when modifying a player's character list /// Available database operations when modifying a player's character list
@ -53,6 +55,17 @@ pub struct CharacterLoaderResponse {
pub result: CharacterLoaderResponseKind, pub result: CharacterLoaderResponseKind,
} }
impl CharacterLoaderResponse {
pub fn is_err(&self) -> bool {
matches!(
&self.result,
CharacterLoaderResponseKind::CharacterData(box Err(_))
| CharacterLoaderResponseKind::CharacterList(Err(_))
| CharacterLoaderResponseKind::CharacterCreation(Err(_))
)
}
}
/// A bi-directional messaging resource for making requests to modify or load /// A bi-directional messaging resource for making requests to modify or load
/// character data in a background thread. /// character data in a background thread.
/// ///
@ -78,84 +91,47 @@ lazy_static! {
} }
impl CharacterLoader { impl CharacterLoader {
pub fn new(db_dir: &Path) -> diesel::QueryResult<Self> { pub fn new(settings: Arc<RwLock<DatabaseSettings>>) -> Result<Self, PersistenceError> {
let (update_tx, internal_rx) = crossbeam_channel::unbounded::<CharacterLoaderRequest>(); let (update_tx, internal_rx) = crossbeam_channel::unbounded::<CharacterLoaderRequest>();
let (internal_tx, update_rx) = crossbeam_channel::unbounded::<CharacterLoaderResponse>(); let (internal_tx, update_rx) = crossbeam_channel::unbounded::<CharacterLoaderResponse>();
let mut conn = establish_connection(db_dir)?;
let builder = std::thread::Builder::new().name("persistence_loader".into()); let builder = std::thread::Builder::new().name("persistence_loader".into());
builder builder
.spawn(move || { .spawn(move || {
for request in internal_rx { // Unwrap here is safe as there is no code that can panic when the write lock is
let (entity, kind) = request; // taken that could cause the RwLock to become poisoned.
let mut conn = establish_connection(&*settings.read().unwrap());
if let Err(e) = internal_tx.send(CharacterLoaderResponse { for request in internal_rx {
entity, conn.update_log_mode(&settings);
result: match kind {
CharacterLoaderRequestKind::CreateCharacter { match conn.connection.transaction() {
player_uuid, Ok(mut transaction) => {
character_alias, let response =
persisted_components, CharacterLoader::process_request(request, &mut transaction);
} => CharacterLoaderResponseKind::CharacterCreation(conn.transaction( if !response.is_err() {
|txn| { match transaction.commit() {
create_character( Ok(()) => {
&player_uuid, trace!("Commit for character loader completed");
&character_alias,
persisted_components,
txn,
&MATERIAL_STATS_MANIFEST,
)
},
)),
CharacterLoaderRequestKind::DeleteCharacter {
player_uuid,
character_id,
} => CharacterLoaderResponseKind::CharacterList(conn.transaction(
|txn| {
delete_character(
&player_uuid,
character_id,
txn,
&MATERIAL_STATS_MANIFEST,
)
},
)),
CharacterLoaderRequestKind::LoadCharacterList { player_uuid } => {
CharacterLoaderResponseKind::CharacterList(conn.transaction(
|txn| {
load_character_list(
&player_uuid,
txn,
&MATERIAL_STATS_MANIFEST,
)
}, },
)) Err(e) => error!(
}, "Failed to commit transaction for character loader, \
CharacterLoaderRequestKind::LoadCharacterData { error: {:?}",
player_uuid, e
character_id, ),
} => { };
let result = conn.transaction(|txn| { };
load_character_data(
player_uuid, if let Err(e) = internal_tx.send(response) {
character_id, error!(?e, "Could not send character loader response");
txn, }
&MATERIAL_STATS_MANIFEST, },
) Err(e) => {
}); error!(
if result.is_err() { "Failed to start transaction for character loader, error: {:?}",
error!( e
?result, )
"Error loading character data for character_id: {}",
character_id
);
}
CharacterLoaderResponseKind::CharacterData(Box::new(result))
},
}, },
}) {
error!(?e, "Could not send send persistence request");
} }
} }
}) })
@ -167,6 +143,66 @@ impl CharacterLoader {
}) })
} }
// TODO: Refactor the way that we send errors to the client to not require a
// specific Result type per CharacterLoaderResponseKind, and remove
// CharacterLoaderResponse::is_err()
fn process_request(
request: CharacterLoaderRequest,
mut transaction: &mut Transaction,
) -> CharacterLoaderResponse {
let (entity, kind) = request;
CharacterLoaderResponse {
entity,
result: match kind {
CharacterLoaderRequestKind::CreateCharacter {
player_uuid,
character_alias,
persisted_components,
} => CharacterLoaderResponseKind::CharacterCreation(create_character(
&player_uuid,
&character_alias,
persisted_components,
&mut transaction,
&MATERIAL_STATS_MANIFEST,
)),
CharacterLoaderRequestKind::DeleteCharacter {
player_uuid,
character_id,
} => CharacterLoaderResponseKind::CharacterList(delete_character(
&player_uuid,
character_id,
&mut transaction,
&MATERIAL_STATS_MANIFEST,
)),
CharacterLoaderRequestKind::LoadCharacterList { player_uuid } => {
CharacterLoaderResponseKind::CharacterList(load_character_list(
&player_uuid,
&mut transaction,
&MATERIAL_STATS_MANIFEST,
))
},
CharacterLoaderRequestKind::LoadCharacterData {
player_uuid,
character_id,
} => {
let result = load_character_data(
player_uuid,
character_id,
&mut transaction,
&MATERIAL_STATS_MANIFEST,
);
if result.is_err() {
error!(
?result,
"Error loading character data for character_id: {}", character_id
);
}
CharacterLoaderResponseKind::CharacterData(Box::new(result))
},
},
}
}
/// Create a new character belonging to the player identified by /// Create a new character belonging to the player identified by
/// `player_uuid` /// `player_uuid`
pub fn create_character( pub fn create_character(

View File

@ -1,36 +1,83 @@
use crate::comp; use crate::comp;
use common::{character::CharacterId, comp::item::ItemId}; use common::character::CharacterId;
use crate::persistence::{establish_connection, VelorenConnection}; use crate::persistence::{
use std::{path::Path, sync::Arc}; error::PersistenceError, establish_connection, DatabaseSettings, VelorenConnection,
use tracing::{error, trace}; };
use rusqlite::DropBehavior;
use std::{
collections::HashMap,
sync::{
atomic::{AtomicBool, Ordering},
Arc, RwLock,
},
};
use tracing::{debug, error, info, trace, warn};
pub type CharacterUpdateData = (comp::Stats, comp::Inventory, Option<comp::Waypoint>); pub type CharacterUpdateData = (comp::Stats, comp::Inventory, Option<comp::Waypoint>);
pub enum CharacterUpdaterEvent {
BatchUpdate(Vec<(CharacterId, CharacterUpdateData)>),
DisconnectedSuccess,
}
/// A unidirectional messaging resource for saving characters in a /// A unidirectional messaging resource for saving characters in a
/// background thread. /// background thread.
/// ///
/// This is used to make updates to a character and their persisted components, /// This is used to make updates to a character and their persisted components,
/// such as inventory, loadout, etc... /// such as inventory, loadout, etc...
pub struct CharacterUpdater { pub struct CharacterUpdater {
update_tx: Option<crossbeam_channel::Sender<Vec<(CharacterId, CharacterUpdateData)>>>, update_tx: Option<crossbeam_channel::Sender<CharacterUpdaterEvent>>,
handle: Option<std::thread::JoinHandle<()>>, handle: Option<std::thread::JoinHandle<()>>,
pending_logout_updates: HashMap<CharacterId, CharacterUpdateData>,
/// Will disconnect all characters (without persistence) on the next tick if
/// set to true
disconnect_all_clients_requested: Arc<AtomicBool>,
} }
impl CharacterUpdater { impl CharacterUpdater {
pub fn new(db_dir: &Path) -> diesel::QueryResult<Self> { pub fn new(settings: Arc<RwLock<DatabaseSettings>>) -> rusqlite::Result<Self> {
let (update_tx, update_rx) = let (update_tx, update_rx) = crossbeam_channel::unbounded::<CharacterUpdaterEvent>();
crossbeam_channel::unbounded::<Vec<(CharacterId, CharacterUpdateData)>>(); let disconnect_all_clients_requested = Arc::new(AtomicBool::new(false));
let disconnect_all_clients_requested_clone = Arc::clone(&disconnect_all_clients_requested);
let mut conn = establish_connection(db_dir)?;
let builder = std::thread::Builder::new().name("persistence_updater".into()); let builder = std::thread::Builder::new().name("persistence_updater".into());
let handle = builder let handle = builder
.spawn(move || { .spawn(move || {
// Unwrap here is safe as there is no code that can panic when the write lock is
// taken that could cause the RwLock to become poisoned.
let mut conn = establish_connection(&*settings.read().unwrap());
while let Ok(updates) = update_rx.recv() { while let Ok(updates) = update_rx.recv() {
trace!("Persistence batch update starting"); match updates {
execute_batch_update(updates, &mut conn); CharacterUpdaterEvent::BatchUpdate(updates) => {
trace!("Persistence batch update finished"); if disconnect_all_clients_requested_clone.load(Ordering::Relaxed) {
debug!(
"Skipping persistence due to pending disconnection of all \
clients"
);
continue;
}
conn.update_log_mode(&settings);
if let Err(e) = execute_batch_update(updates, &mut conn) {
error!(
"Error during character batch update, disconnecting all \
clients to avoid loss of data integrity. Error: {:?}",
e
);
disconnect_all_clients_requested_clone
.store(true, Ordering::Relaxed);
};
},
CharacterUpdaterEvent::DisconnectedSuccess => {
info!(
"CharacterUpdater received DisconnectedSuccess event, resuming \
batch updates"
);
// Reset the disconnection request as we have had confirmation that all
// clients have been disconnected
disconnect_all_clients_requested_clone.store(false, Ordering::Relaxed);
},
}
} }
}) })
.unwrap(); .unwrap();
@ -38,12 +85,49 @@ impl CharacterUpdater {
Ok(Self { Ok(Self {
update_tx: Some(update_tx), update_tx: Some(update_tx),
handle: Some(handle), handle: Some(handle),
pending_logout_updates: HashMap::new(),
disconnect_all_clients_requested,
}) })
} }
/// Adds a character to the list of characters that have recently logged out
/// and will be persisted in the next batch update.
pub fn add_pending_logout_update(
&mut self,
character_id: CharacterId,
update_data: CharacterUpdateData,
) {
if !self
.disconnect_all_clients_requested
.load(Ordering::Relaxed)
{
self.pending_logout_updates
.insert(character_id, update_data);
} else {
warn!(
"Ignoring request to add pending logout update for character ID {} as there is a \
disconnection of all clients in progress",
character_id
);
}
}
/// Returns the character IDs of characters that have recently logged out
/// and are awaiting persistence in the next batch update.
pub fn characters_pending_logout(&self) -> impl Iterator<Item = CharacterId> + '_ {
self.pending_logout_updates.keys().copied()
}
/// Returns a value indicating whether there is a pending request to
/// disconnect all clients due to a batch update transaction failure
pub fn disconnect_all_clients_requested(&self) -> bool {
self.disconnect_all_clients_requested
.load(Ordering::Relaxed)
}
/// Updates a collection of characters based on their id and components /// Updates a collection of characters based on their id and components
pub fn batch_update<'a>( pub fn batch_update<'a>(
&self, &mut self,
updates: impl Iterator< updates: impl Iterator<
Item = ( Item = (
CharacterId, CharacterId,
@ -60,16 +144,22 @@ impl CharacterUpdater {
(stats.clone(), inventory.clone(), waypoint.cloned()), (stats.clone(), inventory.clone(), waypoint.cloned()),
) )
}) })
.chain(self.pending_logout_updates.drain())
.collect::<Vec<_>>(); .collect::<Vec<_>>();
if let Err(e) = self.update_tx.as_ref().unwrap().send(updates) { if let Err(e) = self
.update_tx
.as_ref()
.unwrap()
.send(CharacterUpdaterEvent::BatchUpdate(updates))
{
error!(?e, "Could not send stats updates"); error!(?e, "Could not send stats updates");
} }
} }
/// Updates a single character based on their id and components /// Updates a single character based on their id and components
pub fn update( pub fn update(
&self, &mut self,
character_id: CharacterId, character_id: CharacterId,
stats: &comp::Stats, stats: &comp::Stats,
inventory: &comp::Inventory, inventory: &comp::Inventory,
@ -77,32 +167,37 @@ impl CharacterUpdater {
) { ) {
self.batch_update(std::iter::once((character_id, stats, inventory, waypoint))); self.batch_update(std::iter::once((character_id, stats, inventory, waypoint)));
} }
/// Indicates to the batch update thread that a requested disconnection of
/// all clients has been processed
pub fn disconnected_success(&mut self) {
self.update_tx
.as_ref()
.unwrap()
.send(CharacterUpdaterEvent::DisconnectedSuccess)
.expect(
"Failed to send DisconnectedSuccess event - not sending this event will prevent \
future persistence batches from running",
);
}
} }
fn execute_batch_update( fn execute_batch_update(
updates: Vec<(CharacterId, CharacterUpdateData)>, updates: Vec<(CharacterId, CharacterUpdateData)>,
connection: &mut VelorenConnection, connection: &mut VelorenConnection,
) { ) -> Result<(), PersistenceError> {
let mut inserted_items = Vec::<Arc<ItemId>>::new(); let mut transaction = connection.connection.transaction()?;
transaction.set_drop_behavior(DropBehavior::Rollback);
trace!("Transaction started for character batch update");
updates
.into_iter()
.try_for_each(|(character_id, (stats, inventory, waypoint))| {
super::character::update(character_id, stats, inventory, waypoint, &mut transaction)
})?;
transaction.commit()?;
if let Err(e) = connection.transaction::<_, super::error::Error, _>(|txn| { trace!("Commit for character batch update completed");
for (character_id, (stats, inventory, waypoint)) in updates { Ok(())
inserted_items.append(&mut super::character::update(
character_id,
stats,
inventory,
waypoint,
txn,
)?);
}
Ok(())
}) {
error!(?e, "Error during character batch update transaction");
}
// NOTE: On success, updating thee atomics is already taken care of
// internally.
} }
impl Drop for CharacterUpdater { impl Drop for CharacterUpdater {

View File

@ -1,5 +0,0 @@
# For documentation on how to configure this file,
# see diesel.rs/guides/configuring-diesel-cli
[print_schema]
file = "schema.rs"

View File

@ -0,0 +1,109 @@
use crate::persistence::{error::PersistenceError, VelorenConnection};
use rusqlite::NO_PARAMS;
use tracing::{debug, info};
/// Performs a one-time migration from diesel to refinery migrations. Copies
/// diesel's __diesel_schema_migrations table records to refinery_schema_history
/// and drops __diesel_schema_migrations.
// At some point in the future, when it is deemed no longer necessary to
// support migrations from pre-rusqlite databases this method should be deleted.
pub(crate) fn migrate_from_diesel(
connection: &mut VelorenConnection,
) -> Result<(), PersistenceError> {
let transaction = connection
.connection
.transaction()
.expect("failed to start transaction");
#[rustfmt::skip]
let mut stmt = transaction.prepare("
SELECT COUNT(1)
FROM sqlite_master
WHERE type='table'
AND name='__diesel_schema_migrations';
",
)?;
let diesel_migrations_table_exists = stmt.query_row(NO_PARAMS, |row| {
let row_count: i32 = row.get(0)?;
Ok(row_count > 0)
})?;
drop(stmt);
if !diesel_migrations_table_exists {
debug!(
"__diesel_schema_migrations table does not exist, skipping diesel to refinery \
migration"
);
return Ok(());
}
#[rustfmt::skip]
transaction.execute_batch("
-- Create temporary table to store Diesel > Refinery mapping data in
CREATE TEMP TABLE IF NOT EXISTS _migration_map (
diesel_version VARCHAR(50) NOT NULL,
refinery_version INT4 NOT NULL,
refinery_name VARCHAR(255) NOT NULL,
refinery_checksum VARCHAR(255) NOT NULL
);
-- Insert mapping records to _migration_map
INSERT INTO _migration_map VALUES ('20200411202519',1,'character','18301154882232874638');
INSERT INTO _migration_map VALUES ('20200419025352',2,'body','6687048722955029379');
INSERT INTO _migration_map VALUES ('20200420072214',3,'stats','2322064461300660230');
INSERT INTO _migration_map VALUES ('20200524235534',4,'race_species','16440275012526526388');
INSERT INTO _migration_map VALUES ('20200527145044',5,'inventory','13535458920968937266');
INSERT INTO _migration_map VALUES ('20200528210610',6,'loadout','18209914188629128082');
INSERT INTO _migration_map VALUES ('20200602210738',7,'inv_increase','3368217138206467823');
INSERT INTO _migration_map VALUES ('20200703194516',8,'skills','9202176632428664476');
INSERT INTO _migration_map VALUES ('20200707201052',9,'add_missing_inv_loadout','9127886123837666846');
INSERT INTO _migration_map VALUES ('20200710162552',10,'dash_melee_energy_cost_fix','14010543160640061685');
INSERT INTO _migration_map VALUES ('20200716044718',11,'migrate_armour_stats','1617484395098403184');
INSERT INTO _migration_map VALUES ('20200719223917',12,'update_item_stats','12571040280459413049');
INSERT INTO _migration_map VALUES ('20200724191205',13,'fix_projectile_stats','5178981757717265745');
INSERT INTO _migration_map VALUES ('20200729204534',14,'power_stat_for_weapons','17299186713398844906');
INSERT INTO _migration_map VALUES ('20200806212413',15,'fix_various_problems','17258097957115914749');
INSERT INTO _migration_map VALUES ('20200816130513',16,'item_persistence','18222209741267759587');
INSERT INTO _migration_map VALUES ('20200925200504',17,'move_sceptres','8956411670404874637');
INSERT INTO _migration_map VALUES ('20201107182406',18,'rename_npcweapons','10703468376963165521');
INSERT INTO _migration_map VALUES ('20201116173524',19,'move_waypoint_to_stats','10083555685813984571');
INSERT INTO _migration_map VALUES ('20201128205542',20,'item_storage','11912657465469442777');
INSERT INTO _migration_map VALUES ('20201213172324',21,'shinygem_to_diamond','7188502861698656149');
INSERT INTO _migration_map VALUES ('20210124141845',22,'skills','1249519966980586191');
INSERT INTO _migration_map VALUES ('20210125202618',23,'purge_duplicate_items','10597564860189510441');
INSERT INTO _migration_map VALUES ('20210212054315',24,'remove_duplicate_possess_stick','10774303849135897742');
INSERT INTO _migration_map VALUES ('20210220191847',25,'starter_gear','7937903884108396352');
INSERT INTO _migration_map VALUES ('20210224230149',26,'weapon_replacements','16314806319051099277');
INSERT INTO _migration_map VALUES ('20210301053817',27,'armor_reorganization','17623676960765703100');
INSERT INTO _migration_map VALUES ('20210302023541',28,'fix_sturdy_red_backpack','10808562637001569925');
INSERT INTO _migration_map VALUES ('20210302041950',29,'fix_other_backpacks','3143452502889073613');
INSERT INTO _migration_map VALUES ('20210302182544',30,'fix_leather_set','5238543158379875836');
INSERT INTO _migration_map VALUES ('20210303195917',31,'fix_debug_armor','13228825131487923091');
INSERT INTO _migration_map VALUES ('20210306213310',32,'reset_sceptre_skills','626800208872263587');
INSERT INTO _migration_map VALUES ('20210329012510',33,'fix_amethyst_staff','11008696478673746982');
-- Create refinery_schema_history table
CREATE TABLE refinery_schema_history (
version INT4 PRIMARY KEY,
name VARCHAR(255),
applied_on VARCHAR(255),
checksum VARCHAR(255)
);
-- Migrate diesel migration records to refinery migrations table
INSERT INTO refinery_schema_history
SELECT m.refinery_version,
m.refinery_name,
'2021-03-27T00:00:00.000000000+00:00',
m.refinery_checksum
FROM _migration_map m
JOIN __diesel_schema_migrations d ON (d.version = m.diesel_version);
DROP TABLE __diesel_schema_migrations;"
)?;
transaction.commit()?;
info!("Successfully performed one-time diesel to refinery migration");
Ok(())
}

View File

@ -1,21 +1,19 @@
//! Consolidates Diesel and validation errors under a common error type //! Consolidates rusqlite and validation errors under a common error type
extern crate diesel; extern crate rusqlite;
use std::fmt; use std::fmt;
#[derive(Debug)] #[derive(Debug)]
pub enum Error { pub enum PersistenceError {
// An invalid asset was returned from the database // An invalid asset was returned from the database
AssetError(String), AssetError(String),
// The player has already reached the max character limit // The player has already reached the max character limit
CharacterLimitReached, CharacterLimitReached,
// An error occurred while establish a db connection // An error occurred while establish a db connection
DatabaseConnectionError(diesel::ConnectionError), DatabaseConnectionError(rusqlite::Error),
// An error occurred while running migrations
DatabaseMigrationError(diesel_migrations::RunMigrationsError),
// An error occurred when performing a database action // An error occurred when performing a database action
DatabaseError(diesel::result::Error), DatabaseError(rusqlite::Error),
// Unable to load body or stats for a character // Unable to load body or stats for a character
CharacterDataError, CharacterDataError,
SerializationError(serde_json::Error), SerializationError(serde_json::Error),
@ -23,14 +21,13 @@ pub enum Error {
OtherError(String), OtherError(String),
} }
impl fmt::Display for Error { impl fmt::Display for PersistenceError {
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
write!(f, "{}", match self { write!(f, "{}", match self {
Self::AssetError(error) => error.to_string(), Self::AssetError(error) => error.to_string(),
Self::CharacterLimitReached => String::from("Character limit exceeded"), Self::CharacterLimitReached => String::from("Character limit exceeded"),
Self::DatabaseError(error) => error.to_string(), Self::DatabaseError(error) => error.to_string(),
Self::DatabaseConnectionError(error) => error.to_string(), Self::DatabaseConnectionError(error) => error.to_string(),
Self::DatabaseMigrationError(error) => error.to_string(),
Self::CharacterDataError => String::from("Error while loading character data"), Self::CharacterDataError => String::from("Error while loading character data"),
Self::SerializationError(error) => error.to_string(), Self::SerializationError(error) => error.to_string(),
Self::ConversionError(error) => error.to_string(), Self::ConversionError(error) => error.to_string(),
@ -39,20 +36,12 @@ impl fmt::Display for Error {
} }
} }
impl From<diesel::result::Error> for Error { impl From<rusqlite::Error> for PersistenceError {
fn from(error: diesel::result::Error) -> Error { Error::DatabaseError(error) } fn from(error: rusqlite::Error) -> PersistenceError { PersistenceError::DatabaseError(error) }
} }
impl From<diesel::ConnectionError> for Error { impl From<serde_json::Error> for PersistenceError {
fn from(error: diesel::ConnectionError) -> Error { Error::DatabaseConnectionError(error) } fn from(error: serde_json::Error) -> PersistenceError {
} PersistenceError::SerializationError(error)
impl From<serde_json::Error> for Error {
fn from(error: serde_json::Error) -> Error { Error::SerializationError(error) }
}
impl From<diesel_migrations::RunMigrationsError> for Error {
fn from(error: diesel_migrations::RunMigrationsError) -> Error {
Error::DatabaseMigrationError(error)
} }
} }

View File

@ -1,23 +1,21 @@
//! DB operations and schema migrations //! DB operations and schema migrations
//!
//! This code uses several [`Diesel ORM`](http://diesel.rs/) tools for DB operations:
//! - [`diesel-migrations`](https://docs.rs/diesel_migrations/1.4.0/diesel_migrations/)
//! for managing table migrations
//! - [`diesel-cli`](https://github.com/diesel-rs/diesel/tree/master/diesel_cli/)
//! for generating and testing migrations
pub(in crate::persistence) mod character; pub(in crate::persistence) mod character;
pub mod character_loader; pub mod character_loader;
pub mod character_updater; pub mod character_updater;
mod error; mod diesel_to_rusqlite;
pub mod error;
mod json_models; mod json_models;
mod models; mod models;
mod schema;
use common::comp; use common::comp;
use diesel::{connection::SimpleConnection, prelude::*}; use refinery::Report;
use diesel_migrations::embed_migrations; use rusqlite::{Connection, OpenFlags};
use std::{fs, path::Path}; use std::{
path::PathBuf,
sync::{Arc, RwLock},
time::Duration,
};
use tracing::info; use tracing::info;
/// A tuple of the components that are persisted to the DB for each character /// A tuple of the components that are persisted to the DB for each character
@ -28,92 +26,144 @@ pub type PersistedComponents = (
Option<comp::Waypoint>, Option<comp::Waypoint>,
); );
// See: https://docs.rs/diesel_migrations/1.4.0/diesel_migrations/macro.embed_migrations.html // See: https://docs.rs/refinery/0.5.0/refinery/macro.embed_migrations.html
// This macro is called at build-time, and produces the necessary migration info // This macro is called at build-time, and produces the necessary migration info
// for the `embedded_migrations` call below. // for the `run_migrations` call below.
// mod embedded {
// NOTE: Adding a useless comment to trigger the migrations being run. Alter use refinery::embed_migrations;
// when needed. embed_migrations!("./src/migrations");
embed_migrations!();
struct TracingOut;
impl std::io::Write for TracingOut {
fn write(&mut self, buf: &[u8]) -> std::io::Result<usize> {
info!("{}", String::from_utf8_lossy(buf));
Ok(buf.len())
}
fn flush(&mut self) -> std::io::Result<()> { Ok(()) }
}
/// Runs any pending database migrations. This is executed during server startup
pub fn run_migrations(db_dir: &Path) -> Result<(), diesel_migrations::RunMigrationsError> {
let _ = fs::create_dir(format!("{}/", db_dir.display()));
embedded_migrations::run_with_output(
&establish_connection(db_dir)
.expect(
"If we cannot execute migrations, we should not be allowed to launch the server, \
so we don't populate it with bad data.",
)
.0,
&mut std::io::LineWriter::new(TracingOut),
)
} }
/// A database connection blessed by Veloren. /// A database connection blessed by Veloren.
pub struct VelorenConnection(SqliteConnection); pub(crate) struct VelorenConnection {
connection: Connection,
/// A transaction blessed by Veloren. sql_log_mode: SqlLogMode,
#[derive(Clone, Copy)] }
pub struct VelorenTransaction<'a>(&'a SqliteConnection);
impl VelorenConnection { impl VelorenConnection {
/// Open a transaction in order to be able to run a set of queries against fn new(connection: Connection) -> Self {
/// the database. We require the use of a transaction, rather than Self {
/// allowing direct session access, so that (1) we can control things connection,
/// like the retry process (at a future date), and (2) to avoid sql_log_mode: SqlLogMode::Disabled,
/// accidentally forgetting to open or reuse a transaction. }
/// }
/// We could make things even more foolproof, but we restrict ourselves to
/// this for now. /// Updates the SQLite log mode if DatabaseSetting.sql_log_mode has changed
pub fn transaction<T, E, F>(&mut self, f: F) -> Result<T, E> pub fn update_log_mode(&mut self, database_settings: &Arc<RwLock<DatabaseSettings>>) {
where let settings = database_settings
F: for<'a> FnOnce(VelorenTransaction<'a>) -> Result<T, E>, .read()
E: From<diesel::result::Error>, .expect("DatabaseSettings RwLock was poisoned");
{ if self.sql_log_mode == (*settings).sql_log_mode {
self.0.transaction(|| f(VelorenTransaction(&self.0))) return;
}
set_log_mode(&mut self.connection, (*settings).sql_log_mode);
self.sql_log_mode = (*settings).sql_log_mode;
info!(
"SQL log mode for connection changed to {:?}",
settings.sql_log_mode
);
} }
} }
impl<'a> core::ops::Deref for VelorenTransaction<'a> { fn set_log_mode(connection: &mut Connection, sql_log_mode: SqlLogMode) {
type Target = SqliteConnection; // Rusqlite's trace and profile logging are mutually exclusive and cannot be
// used together
fn deref(&self) -> &Self::Target { &self.0 } match sql_log_mode {
SqlLogMode::Trace => {
connection.trace(Some(rusqlite_trace_callback));
},
SqlLogMode::Profile => {
connection.profile(Some(rusqlite_profile_callback));
},
SqlLogMode::Disabled => {
connection.trace(None);
connection.profile(None);
},
};
} }
pub fn establish_connection(db_dir: &Path) -> QueryResult<VelorenConnection> { #[derive(Clone)]
let database_url = format!("{}/db.sqlite", db_dir.display()); pub struct DatabaseSettings {
pub db_dir: PathBuf,
pub sql_log_mode: SqlLogMode,
}
let connection = SqliteConnection::establish(&database_url) #[derive(Clone, Copy, Debug, PartialEq)]
.unwrap_or_else(|_| panic!("Error connecting to {}", database_url)); pub enum SqlLogMode {
/// Logging is disabled
Disabled,
/// Records timings for each SQL statement
Profile,
/// Prints all executed SQL statements
Trace,
}
/// Runs any pending database migrations. This is executed during server startup
pub fn run_migrations(settings: &DatabaseSettings) {
let mut conn = establish_connection(settings);
diesel_to_rusqlite::migrate_from_diesel(&mut conn)
.expect("One-time migration from Diesel to Refinery failed");
// If migrations fail to run, the server cannot start since the database will
// not be in the required state.
let report: Report = embedded::migrations::runner()
.set_abort_divergent(false)
.run(&mut conn.connection)
.expect("Database migrations failed, server startup aborted");
let applied_migrations = report.applied_migrations().len();
info!("Applied {} database migrations", applied_migrations);
}
// These callbacks use info logging because they are never enabled by default,
// only when explicitly turned on via CLI arguments or interactive CLI commands.
// Setting them to anything other than info would remove the ability to get SQL
// logging from a running server that wasn't started at higher than info.
fn rusqlite_trace_callback(log_message: &str) {
info!("{}", log_message);
}
fn rusqlite_profile_callback(log_message: &str, dur: Duration) {
info!("{} Duration: {:?}", log_message, dur);
}
pub(crate) fn establish_connection(settings: &DatabaseSettings) -> VelorenConnection {
let connection = Connection::open_with_flags(
&settings.db_dir.join("db.sqlite"),
OpenFlags::SQLITE_OPEN_PRIVATE_CACHE | OpenFlags::default(),
)
.unwrap_or_else(|err| {
panic!(
"Error connecting to {}, Error: {:?}",
settings.db_dir.join("db.sqlite").display(),
err
)
});
let mut veloren_connection = VelorenConnection::new(connection);
let connection = &mut veloren_connection.connection;
set_log_mode(connection, settings.sql_log_mode);
veloren_connection.sql_log_mode = settings.sql_log_mode;
rusqlite::vtab::array::load_module(&connection).expect("Failed to load sqlite array module");
connection.set_prepared_statement_cache_capacity(100);
// Use Write-Ahead-Logging for improved concurrency: https://sqlite.org/wal.html // Use Write-Ahead-Logging for improved concurrency: https://sqlite.org/wal.html
// Set a busy timeout (in ms): https://sqlite.org/c3ref/busy_timeout.html // Set a busy timeout (in ms): https://sqlite.org/c3ref/busy_timeout.html
connection connection
.batch_execute( .pragma_update(None, "foreign_keys", &"ON")
" .expect("Failed to set foreign_keys PRAGMA");
PRAGMA foreign_keys = ON; connection
PRAGMA journal_mode = WAL; .pragma_update(None, "journal_mode", &"WAL")
PRAGMA busy_timeout = 250; .expect("Failed to set journal_mode PRAGMA");
", connection
) .pragma_update(None, "busy_timeout", &"250")
.expect( .expect("Failed to set busy_timeout PRAGMA");
"Failed adding PRAGMA statements while establishing sqlite connection, including \
enabling foreign key constraints. We will not allow connecting to the server under \
these conditions.",
);
Ok(VelorenConnection(connection)) veloren_connection
} }

View File

@ -1,25 +1,3 @@
extern crate serde_json;
use super::schema::{body, character, entity, item, skill, skill_group};
#[derive(Debug, Insertable, PartialEq)]
#[table_name = "entity"]
pub struct Entity {
pub entity_id: i64,
}
#[derive(Insertable)]
#[table_name = "character"]
pub struct NewCharacter<'a> {
pub character_id: i64,
pub player_uuid: &'a str,
pub alias: &'a str,
pub waypoint: Option<String>,
}
#[derive(Identifiable, Queryable, Debug)]
#[table_name = "character"]
#[primary_key(character_id)]
pub struct Character { pub struct Character {
pub character_id: i64, pub character_id: i64,
pub player_uuid: String, pub player_uuid: String,
@ -27,9 +5,7 @@ pub struct Character {
pub waypoint: Option<String>, pub waypoint: Option<String>,
} }
#[derive(Debug, Insertable, Queryable, AsChangeset)] #[derive(Debug)]
#[table_name = "item"]
#[primary_key(item_id)]
pub struct Item { pub struct Item {
pub item_id: i64, pub item_id: i64,
pub parent_container_item_id: i64, pub parent_container_item_id: i64,
@ -38,27 +14,18 @@ pub struct Item {
pub position: String, pub position: String,
} }
#[derive(Associations, Identifiable, Insertable, Queryable, Debug)]
#[primary_key(body_id)]
#[table_name = "body"]
pub struct Body { pub struct Body {
pub body_id: i64, pub body_id: i64,
pub variant: String, pub variant: String,
pub body_data: String, pub body_data: String,
} }
#[derive(Associations, Identifiable, Insertable, Queryable, Debug)]
#[primary_key(entity_id, skill_type)]
#[table_name = "skill"]
pub struct Skill { pub struct Skill {
pub entity_id: i64, pub entity_id: i64,
pub skill_type: String, pub skill: String,
pub level: Option<i32>, pub level: Option<i32>,
} }
#[derive(Associations, Identifiable, Insertable, Queryable, Debug)]
#[primary_key(entity_id, skill_group_kind)]
#[table_name = "skill_group"]
pub struct SkillGroup { pub struct SkillGroup {
pub entity_id: i64, pub entity_id: i64,
pub skill_group_kind: String, pub skill_group_kind: String,

View File

@ -1,55 +0,0 @@
table! {
body (body_id) {
body_id -> BigInt,
variant -> Text,
body_data -> Text,
}
}
table! {
character (character_id) {
character_id -> BigInt,
player_uuid -> Text,
alias -> Text,
waypoint -> Nullable<Text>,
}
}
table! {
entity (entity_id) {
entity_id -> BigInt,
}
}
table! {
item (item_id) {
item_id -> BigInt,
parent_container_item_id -> BigInt,
item_definition_id -> Text,
stack_size -> Integer,
position -> Text,
}
}
table! {
skill (entity_id, skill_type) {
entity_id -> BigInt,
#[sql_name = "skill"]
skill_type -> Text,
level -> Nullable<Integer>,
}
}
table! {
skill_group (entity_id, skill_group_kind) {
entity_id -> BigInt,
skill_group_kind -> Text,
exp -> Integer,
available_sp -> Integer,
earned_sp -> Integer,
}
}
joinable!(character -> body (character_id));
allow_tables_to_appear_in_same_query!(body, character, entity, item);

View File

@ -1,6 +1,10 @@
use crate::{ use crate::{
alias_validator::AliasValidator, character_creator, client::Client, alias_validator::AliasValidator,
persistence::character_loader::CharacterLoader, presence::Presence, EditableSettings, character_creator,
client::Client,
persistence::{character_loader::CharacterLoader, character_updater::CharacterUpdater},
presence::Presence,
EditableSettings,
}; };
use common::{ use common::{
comp::{ChatType, Player, UnresolvedChatMsg}, comp::{ChatType, Player, UnresolvedChatMsg},
@ -20,6 +24,7 @@ impl Sys {
entity: specs::Entity, entity: specs::Entity,
client: &Client, client: &Client,
character_loader: &ReadExpect<'_, CharacterLoader>, character_loader: &ReadExpect<'_, CharacterLoader>,
character_updater: &ReadExpect<'_, CharacterUpdater>,
uids: &ReadStorage<'_, Uid>, uids: &ReadStorage<'_, Uid>,
players: &ReadStorage<'_, Player>, players: &ReadStorage<'_, Player>,
presences: &ReadStorage<'_, Presence>, presences: &ReadStorage<'_, Presence>,
@ -40,6 +45,28 @@ impl Sys {
if let Some(player) = players.get(entity) { if let Some(player) = players.get(entity) {
if presences.contains(entity) { if presences.contains(entity) {
debug!("player already ingame, aborting"); debug!("player already ingame, aborting");
} else if character_updater
.characters_pending_logout()
.any(|x| x == character_id)
{
debug!("player recently logged out pending persistence, aborting");
client.send(ServerGeneral::CharacterDataLoadError(
"You have recently logged out, please wait a few seconds and try again"
.to_string(),
))?;
} else if character_updater.disconnect_all_clients_requested() {
// If we're in the middle of disconnecting all clients due to a persistence
// transaction failure, prevent new logins
// temporarily.
debug!(
"Rejecting player login while pending disconnection of all players is \
in progress"
);
client.send(ServerGeneral::CharacterDataLoadError(
"The server is currently recovering from an error, please wait a few \
seconds and try again"
.to_string(),
))?;
} else { } else {
// Send a request to load the character's component data from the // Send a request to load the character's component data from the
// DB. Once loaded, persisted components such as stats and inventory // DB. Once loaded, persisted components such as stats and inventory
@ -127,6 +154,7 @@ impl<'a> System<'a> for Sys {
Entities<'a>, Entities<'a>,
Read<'a, EventBus<ServerEvent>>, Read<'a, EventBus<ServerEvent>>,
ReadExpect<'a, CharacterLoader>, ReadExpect<'a, CharacterLoader>,
ReadExpect<'a, CharacterUpdater>,
ReadStorage<'a, Uid>, ReadStorage<'a, Uid>,
ReadStorage<'a, Client>, ReadStorage<'a, Client>,
ReadStorage<'a, Player>, ReadStorage<'a, Player>,
@ -145,6 +173,7 @@ impl<'a> System<'a> for Sys {
entities, entities,
server_event_bus, server_event_bus,
character_loader, character_loader,
character_updater,
uids, uids,
clients, clients,
players, players,
@ -162,6 +191,7 @@ impl<'a> System<'a> for Sys {
entity, entity,
client, client,
&character_loader, &character_loader,
&character_updater,
&uids, &uids,
&players, &players,
&presences, &presences,

View File

@ -2,7 +2,7 @@ use crate::{persistence::character_updater, presence::Presence, sys::SysSchedule
use common::comp::{Inventory, Stats, Waypoint}; use common::comp::{Inventory, Stats, Waypoint};
use common_ecs::{Job, Origin, Phase, System}; use common_ecs::{Job, Origin, Phase, System};
use common_net::msg::PresenceKind; use common_net::msg::PresenceKind;
use specs::{Join, ReadExpect, ReadStorage, Write}; use specs::{Join, ReadStorage, Write, WriteExpect};
#[derive(Default)] #[derive(Default)]
pub struct Sys; pub struct Sys;
@ -14,7 +14,7 @@ impl<'a> System<'a> for Sys {
ReadStorage<'a, Stats>, ReadStorage<'a, Stats>,
ReadStorage<'a, Inventory>, ReadStorage<'a, Inventory>,
ReadStorage<'a, Waypoint>, ReadStorage<'a, Waypoint>,
ReadExpect<'a, character_updater::CharacterUpdater>, WriteExpect<'a, character_updater::CharacterUpdater>,
Write<'a, SysScheduler<Self>>, Write<'a, SysScheduler<Self>>,
); );
@ -29,7 +29,7 @@ impl<'a> System<'a> for Sys {
player_stats, player_stats,
player_inventories, player_inventories,
player_waypoint, player_waypoint,
updater, mut updater,
mut scheduler, mut scheduler,
): Self::SystemData, ): Self::SystemData,
) { ) {

View File

@ -1,6 +1,9 @@
use common::clock::Clock; use common::clock::Clock;
use crossbeam::channel::{bounded, unbounded, Receiver, Sender, TryRecvError}; use crossbeam::channel::{bounded, unbounded, Receiver, Sender, TryRecvError};
use server::{Error as ServerError, Event, Input, Server}; use server::{
persistence::{DatabaseSettings, SqlLogMode},
Error as ServerError, Event, Input, Server,
};
use std::{ use std::{
sync::{ sync::{
atomic::{AtomicBool, AtomicUsize, Ordering}, atomic::{AtomicBool, AtomicUsize, Ordering},
@ -95,6 +98,17 @@ impl Singleplayer {
let settings2 = settings.clone(); let settings2 = settings.clone();
// Relative to data_dir
const PERSISTENCE_DB_DIR: &str = "saves";
let database_settings = DatabaseSettings {
db_dir: server_data_dir.join(PERSISTENCE_DB_DIR),
sql_log_mode: SqlLogMode::Disabled, /* Voxygen doesn't take in command-line arguments
* so SQL logging can't be enabled for
* singleplayer without changing this line
* manually */
};
let paused = Arc::new(AtomicBool::new(false)); let paused = Arc::new(AtomicBool::new(false));
let paused1 = Arc::clone(&paused); let paused1 = Arc::clone(&paused);
@ -109,6 +123,7 @@ impl Singleplayer {
match Server::new( match Server::new(
settings2, settings2,
editable_settings, editable_settings,
database_settings,
&server_data_dir, &server_data_dir,
Arc::clone(&runtime), Arc::clone(&runtime),
) { ) {