Merge branch 'xMAC94x/update_toolchain' into 'master'

update toolchain to 2023-09-28

See merge request veloren/veloren!4121
This commit is contained in:
Marcel 2023-10-06 19:24:13 +00:00
commit eaeeb0165f
103 changed files with 335 additions and 370 deletions

View File

@ -13,7 +13,7 @@ variables:
# https://docs.gitlab.com/ee/ci/yaml/#shallow-cloning
GIT_DEPTH: 3
GIT_CLEAN_FLAGS: -f
CACHE_IMAGE_TAG: d74ceb0a
CACHE_IMAGE_TAG: d551c14a
TAG_REGEX: '/^v[0-9]+\.[0-9]+\.[0-9]+$/'
default:

46
Cargo.lock generated
View File

@ -2289,7 +2289,7 @@ dependencies = [
[[package]]
name = "gfx-auxil"
version = "0.9.0"
source = "git+https://github.com/gfx-rs/gfx?rev=27a1dae3796d33d23812f2bb8c7e3b5aea18b521#27a1dae3796d33d23812f2bb8c7e3b5aea18b521"
source = "git+https://github.com/Imberflur/gfx.git?tag=veloren-fixes-v1#a8ba0a4859abb5f980b02480cb219030fb64530c"
dependencies = [
"fxhash",
"gfx-hal",
@ -2299,9 +2299,9 @@ dependencies = [
[[package]]
name = "gfx-backend-dx11"
version = "0.8.0"
source = "git+https://github.com/gfx-rs/gfx?rev=27a1dae3796d33d23812f2bb8c7e3b5aea18b521#27a1dae3796d33d23812f2bb8c7e3b5aea18b521"
source = "git+https://github.com/Imberflur/gfx.git?tag=veloren-fixes-v1#a8ba0a4859abb5f980b02480cb219030fb64530c"
dependencies = [
"arrayvec 0.5.2",
"arrayvec 0.7.4",
"bitflags 1.3.2",
"gfx-auxil",
"gfx-hal",
@ -2320,9 +2320,9 @@ dependencies = [
[[package]]
name = "gfx-backend-dx12"
version = "0.8.0"
source = "git+https://github.com/gfx-rs/gfx?rev=27a1dae3796d33d23812f2bb8c7e3b5aea18b521#27a1dae3796d33d23812f2bb8c7e3b5aea18b521"
source = "git+https://github.com/Imberflur/gfx.git?tag=veloren-fixes-v1#a8ba0a4859abb5f980b02480cb219030fb64530c"
dependencies = [
"arrayvec 0.5.2",
"arrayvec 0.7.4",
"bit-set",
"bitflags 1.3.2",
"d3d12",
@ -2341,7 +2341,7 @@ dependencies = [
[[package]]
name = "gfx-backend-empty"
version = "0.8.0"
source = "git+https://github.com/gfx-rs/gfx?rev=27a1dae3796d33d23812f2bb8c7e3b5aea18b521#27a1dae3796d33d23812f2bb8c7e3b5aea18b521"
source = "git+https://github.com/Imberflur/gfx.git?tag=veloren-fixes-v1#a8ba0a4859abb5f980b02480cb219030fb64530c"
dependencies = [
"gfx-hal",
"log",
@ -2351,9 +2351,9 @@ dependencies = [
[[package]]
name = "gfx-backend-gl"
version = "0.8.1"
source = "git+https://github.com/gfx-rs/gfx?rev=27a1dae3796d33d23812f2bb8c7e3b5aea18b521#27a1dae3796d33d23812f2bb8c7e3b5aea18b521"
source = "git+https://github.com/Imberflur/gfx.git?tag=veloren-fixes-v1#a8ba0a4859abb5f980b02480cb219030fb64530c"
dependencies = [
"arrayvec 0.5.2",
"arrayvec 0.7.4",
"bitflags 1.3.2",
"fxhash",
"gfx-auxil",
@ -2374,9 +2374,9 @@ dependencies = [
[[package]]
name = "gfx-backend-metal"
version = "0.8.1"
source = "git+https://github.com/gfx-rs/gfx?rev=27a1dae3796d33d23812f2bb8c7e3b5aea18b521#27a1dae3796d33d23812f2bb8c7e3b5aea18b521"
source = "git+https://github.com/Imberflur/gfx.git?tag=veloren-fixes-v1#a8ba0a4859abb5f980b02480cb219030fb64530c"
dependencies = [
"arrayvec 0.5.2",
"arrayvec 0.7.4",
"bitflags 1.3.2",
"block",
"cocoa-foundation",
@ -2400,9 +2400,9 @@ dependencies = [
[[package]]
name = "gfx-backend-vulkan"
version = "0.8.0"
source = "git+https://github.com/gfx-rs/gfx?rev=27a1dae3796d33d23812f2bb8c7e3b5aea18b521#27a1dae3796d33d23812f2bb8c7e3b5aea18b521"
source = "git+https://github.com/Imberflur/gfx.git?tag=veloren-fixes-v1#a8ba0a4859abb5f980b02480cb219030fb64530c"
dependencies = [
"arrayvec 0.5.2",
"arrayvec 0.7.4",
"ash",
"byteorder",
"core-graphics-types",
@ -2422,7 +2422,7 @@ dependencies = [
[[package]]
name = "gfx-hal"
version = "0.8.0"
source = "git+https://github.com/gfx-rs/gfx?rev=27a1dae3796d33d23812f2bb8c7e3b5aea18b521#27a1dae3796d33d23812f2bb8c7e3b5aea18b521"
source = "git+https://github.com/Imberflur/gfx.git?tag=veloren-fixes-v1#a8ba0a4859abb5f980b02480cb219030fb64530c"
dependencies = [
"bitflags 1.3.2",
"naga",
@ -3250,12 +3250,6 @@ dependencies = [
"windows-sys 0.48.0",
]
[[package]]
name = "libm"
version = "0.1.4"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "7fc7aa29613bd6a620df431842069224d8bc9011086b1db4c0e0cd47fa03ec9a"
[[package]]
name = "libm"
version = "0.2.7"
@ -4005,7 +3999,7 @@ source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "f30b0abd723be7e2ffca1272140fac1a2f084c77ec3e123c192b66af1ee9e6c2"
dependencies = [
"autocfg",
"libm 0.2.7",
"libm",
]
[[package]]
@ -4254,13 +4248,13 @@ dependencies = [
]
[[package]]
name = "packed_simd_2"
version = "0.3.8"
name = "packed_simd"
version = "0.3.9"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "a1914cd452d8fccd6f9db48147b29fd4ae05bea9dc5d9ad578509f72415de282"
checksum = "1f9f08af0c877571712e2e3e686ad79efad9657dbf0f7c3c8ba943ff6c38932d"
dependencies = [
"cfg-if 1.0.0",
"libm 0.1.4",
"num-traits",
]
[[package]]
@ -4835,7 +4829,7 @@ dependencies = [
[[package]]
name = "range-alloc"
version = "0.1.2"
source = "git+https://github.com/gfx-rs/gfx?rev=27a1dae3796d33d23812f2bb8c7e3b5aea18b521#27a1dae3796d33d23812f2bb8c7e3b5aea18b521"
source = "git+https://github.com/Imberflur/gfx.git?tag=veloren-fixes-v1#a8ba0a4859abb5f980b02480cb219030fb64530c"
[[package]]
name = "raw-window-handle"
@ -7315,7 +7309,7 @@ dependencies = [
"num 0.4.1",
"num-traits",
"ordered-float 3.9.1",
"packed_simd_2",
"packed_simd",
"rand 0.8.5",
"rand_chacha 0.3.1",
"rayon",

View File

@ -1,6 +1,7 @@
cargo-features = ["named-profiles","profile-overrides"]
[workspace]
resolver = "2"
members = [
"common",
"common/assets",
@ -165,6 +166,15 @@ wgpu = { git = "https://github.com/pythonesque/wgpu.git", rev = "179ea209374a928
# ntapi 3.7 fails to compile under windows due to the bug https://github.com/MSxDOS/ntapi/pull/12
ntapi = { git = "https://github.com/MSxDOS/ntapi.git", rev = "9f56b149c9e25796739157c0fce3e0007a7de6eb" }
[patch."https://github.com/gfx-rs/gfx"]
gfx-hal = { git = "https://github.com/Imberflur/gfx.git", tag = "veloren-fixes-v1" }
gfx-backend-empty = { git = "https://github.com/Imberflur/gfx.git", tag = "veloren-fixes-v1" }
gfx-backend-vulkan = { git = "https://github.com/Imberflur/gfx.git", tag = "veloren-fixes-v1" }
gfx-backend-gl = { git = "https://github.com/Imberflur/gfx.git", tag = "veloren-fixes-v1" }
gfx-backend-dx12 = { git = "https://github.com/Imberflur/gfx.git", tag = "veloren-fixes-v1" }
gfx-backend-dx11 = { git = "https://github.com/Imberflur/gfx.git", tag = "veloren-fixes-v1" }
gfx-backend-metal = { git = "https://github.com/Imberflur/gfx.git", tag = "veloren-fixes-v1" }
# # use the latest fixes in naga (remove when updates trickle down to wgpu-rs)
# naga = { git = "https://github.com/gfx-rs/naga.git", rev = "3a0f0144112ff621dd7f731bf455adf6cab19164" }
# # use the latest fixes in gfx (remove when updates trickle down to wgpu-rs)

View File

@ -47,6 +47,7 @@ impl Tui {
(Self { _handle: handle }, commands_r)
}
#[allow(clippy::needless_pass_by_ref_mut)]
pub fn process_command(cmd: &str, command_s: &mut async_channel::Sender<Cmd>) -> bool {
let matches = Command::new("veloren-botclient")
.version(common::util::DISPLAY_VERSION_LONG.as_str())

View File

@ -1809,18 +1809,16 @@ impl Client {
// significant changes to this code. Here is the approximate order of
// things. Please update it as this code changes.
//
// 1) Collect input from the frontend, apply input effects to the state
// of the game
// 1) Collect input from the frontend, apply input effects to the state of the
// game
// 2) Handle messages from the server
// 3) Go through any events (timer-driven or otherwise) that need handling
// and apply them to the state of the game
// 4) Perform a single LocalState tick (i.e: update the world and entities
// in the world)
// 5) Go through the terrain update queue and apply all changes
// to the terrain
// 3) Go through any events (timer-driven or otherwise) that need handling and
// apply them to the state of the game
// 4) Perform a single LocalState tick (i.e: update the world and entities in
// the world)
// 5) Go through the terrain update queue and apply all changes to the terrain
// 6) Sync information to the server
// 7) Finish the tick, passing actions of the main thread back
// to the frontend
// 7) Finish the tick, passing actions of the main thread back to the frontend
// 1) Handle input from frontend.
// Pass character actions from frontend input to the player's entity.

View File

@ -749,12 +749,12 @@ pub mod asset_tweak {
run_with_file(tweak_path, |file| {
file.write_all(
br#"
br"
((
such: 5,
field: 35.752346,
))
"#,
",
)
.expect("failed to write to the file");

View File

@ -100,11 +100,11 @@ fn input_validated_string(prompt: &str, check: &dyn Fn(&str) -> bool) -> String
}
fn main() {
let prompt = r#"
let prompt = r"
Stub implementation.
If you want to migrate new assets, edit `v1` and `v2` modules.
If you want to migrate old assets, check commit history.
"#;
";
println!("{prompt}");
let old_dir = input_validated_string(

View File

@ -76,11 +76,11 @@ pub enum AreaKind {
}
lazy_static! {
static ref ALIGNMENTS: Vec<String> = vec!["wild", "enemy", "npc", "pet"]
static ref ALIGNMENTS: Vec<String> = ["wild", "enemy", "npc", "pet"]
.iter()
.map(|s| s.to_string())
.collect();
static ref SKILL_TREES: Vec<String> = vec!["general", "sword", "axe", "hammer", "bow", "staff", "sceptre", "mining"]
static ref SKILL_TREES: Vec<String> = ["general", "sword", "axe", "hammer", "bow", "staff", "sceptre", "mining"]
.iter()
.map(|s| s.to_string())
.collect();
@ -128,14 +128,14 @@ lazy_static! {
.iter()
.map(|o| o.to_string().to_string())
.collect();
static ref TIMES: Vec<String> = vec![
static ref TIMES: Vec<String> = [
"midnight", "night", "dawn", "morning", "day", "noon", "dusk"
]
.iter()
.map(|s| s.to_string())
.collect();
static ref WEATHERS: Vec<String> = vec![
static ref WEATHERS: Vec<String> = [
"clear", "cloudy", "rain", "wind", "storm"
]
.iter()

View File

@ -118,17 +118,16 @@ pub fn members<'a>(
) -> impl Iterator<Item = (specs::Entity, Role)> + 'a {
(entities, groups, alignments, uids)
.join()
.filter_map(move |(e, g, a, u)| {
(*g == group).then(|| {
(
e,
if matches!(a, Alignment::Owned(owner) if owner != u) {
Role::Pet
} else {
Role::Member
},
)
})
.filter(move |&(_e, g, _a, _u)| (*g == group))
.map(|(e, _g, a, u)| {
(
e,
if matches!(a, Alignment::Owned(owner) if owner != u) {
Role::Pet
} else {
Role::Member
},
)
})
}

View File

@ -121,16 +121,8 @@ impl Hands {
Hands::InHands((mainhand, offhand)) => {
let mut from_spec = |i: &ItemSpec| i.try_to_item(rng);
let mainhand = mainhand
.as_ref()
.map(|i| from_spec(i))
.transpose()?
.flatten();
let offhand = offhand
.as_ref()
.map(|i| from_spec(i))
.transpose()?
.flatten();
let mainhand = mainhand.as_ref().map(&mut from_spec).transpose()?.flatten();
let offhand = offhand.as_ref().map(&mut from_spec).transpose()?.flatten();
Ok((mainhand, offhand))
},
Hands::Choice(pairs) => {

View File

@ -649,7 +649,7 @@ impl Inventory {
self.get(inv_slot)
.and_then(|item| self.loadout.get_slot_to_equip_into(&item.kind()))
.map(|equip_slot| self.swap_inventory_loadout(inv_slot, equip_slot, time))
.unwrap_or_else(Vec::new)
.unwrap_or_default()
}
/// Determines how many free inventory slots will be left after equipping an

View File

@ -1185,7 +1185,7 @@ mod tests {
init();
info!("init");
let mut stock: hashbrown::HashMap<Good, f32> = vec![
let mut stock: hashbrown::HashMap<Good, f32> = [
(Good::Ingredients, 50.0),
(Good::Tools, 10.0),
(Good::Armor, 10.0),

View File

@ -517,7 +517,7 @@ impl SkillSet {
// Perform all mutation inside this branch, to avoid triggering a copy
// on write or flagged storage in cases where this matters.
let this_ = to_mut(this_);
let mut this = this_.borrow_mut();
let this = this_.borrow_mut();
// NOTE: Verified to exist previously when we accessed
// this.skill_groups (assuming a non-pathological implementation of
// ToOwned).

View File

@ -1,12 +1,12 @@
use serde::{Deserialize, Serialize};
use std::{
cmp::{Eq, Ord, Ordering, PartialEq, PartialOrd},
fmt, hash,
cmp::{Eq, Ord, PartialEq, PartialOrd},
fmt,
marker::PhantomData,
};
/// Type safe index into Depot
#[derive(Deserialize, Serialize)]
#[derive(Deserialize, Serialize, Copy, Clone, PartialEq, Eq, PartialOrd, Ord, Hash)]
pub struct Id<T> {
idx: u32,
gen: u32,
@ -17,26 +17,6 @@ impl<T> Id<T> {
pub fn id(&self) -> u64 { self.idx as u64 | ((self.gen as u64) << 32) }
}
impl<T> Copy for Id<T> {}
impl<T> Clone for Id<T> {
fn clone(&self) -> Self {
Self {
idx: self.idx,
gen: self.gen,
phantom: PhantomData,
}
}
}
impl<T> Eq for Id<T> {}
impl<T> PartialEq for Id<T> {
fn eq(&self, other: &Self) -> bool { self.idx == other.idx && self.gen == other.gen }
}
impl<T> Ord for Id<T> {
fn cmp(&self, other: &Self) -> Ordering { (self.idx, self.gen).cmp(&(other.idx, other.gen)) }
}
impl<T> PartialOrd for Id<T> {
fn partial_cmp(&self, other: &Self) -> Option<Ordering> { Some(self.cmp(other)) }
}
impl<T> fmt::Debug for Id<T> {
fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
write!(
@ -48,12 +28,6 @@ impl<T> fmt::Debug for Id<T> {
)
}
}
impl<T> hash::Hash for Id<T> {
fn hash<H: hash::Hasher>(&self, h: &mut H) {
self.idx.hash(h);
self.gen.hash(h);
}
}
struct Entry<T> {
gen: u32,

View File

@ -12,9 +12,9 @@
type_alias_impl_trait,
extend_one,
arbitrary_self_types,
int_roundings
int_roundings,
hash_extract_if
)]
#![feature(hash_drain_filter)]
pub use common_assets as assets;
pub use uuid;

View File

@ -133,8 +133,12 @@ pub fn distribute_many<T: Copy + Eq + Hash, I>(
let Some(mut give) = participants
.iter()
.map(|participant| (total_item_amount as f32 * participant.weight / total_weight).ceil() as u32 - participant.recieved_count)
.min() else {
.map(|participant| {
(total_item_amount as f32 * participant.weight / total_weight).ceil() as u32
- participant.recieved_count
})
.min()
else {
tracing::error!("Tried to distribute items to no participants.");
return;
};
@ -152,8 +156,7 @@ pub fn distribute_many<T: Copy + Eq + Hash, I>(
let participant_count = participants.len();
let Some(winner) = participants
.get_mut(index) else {
let Some(winner) = participants.get_mut(index) else {
tracing::error!("Tried to distribute items to no participants.");
return;
};

View File

@ -380,8 +380,9 @@ impl Link for VolumeMounting {
Volume::Terrain => &*terrain_riders,
Volume::Entity(uid) => {
let Some(riders) = entity(uid)
.filter(|entity| is_alive(*entity))
.and_then(|entity| volume_riders.get(entity)) else {
.filter(|entity| is_alive(*entity))
.and_then(|entity| volume_riders.get(entity))
else {
return false;
};
riders

View File

@ -16,7 +16,7 @@ impl<T> Id<T> {
impl<T> Copy for Id<T> {}
impl<T> Clone for Id<T> {
fn clone(&self) -> Self { Self(self.0, PhantomData) }
fn clone(&self) -> Self { *self }
}
impl<T> Eq for Id<T> {}
impl<T> PartialEq for Id<T> {

View File

@ -176,7 +176,7 @@ impl MapSizeLg {
map_size_lg.y + TERRAIN_CHUNK_BLOCKS_LG < 32;
// Assertion on dimensions: product of dimensions must fit in a usize.
let chunks_product_in_range =
matches!(1usize.checked_shl(map_size_lg.x + map_size_lg.y), Some(_));
1usize.checked_shl(map_size_lg.x + map_size_lg.y).is_some();
if blocks_in_range && chunks_product_in_range {
// Cleared all invariants.
Ok(MapSizeLg(map_size_lg))

View File

@ -86,10 +86,10 @@ impl<V, S: VolSize, M> Chunk<V, S, M> {
//
// Rationales:
//
// 1. We have code in the implementation that assumes it. In particular,
// code using `.count_ones()`.
// 2. The maximum group size is `256x256x256`, because there's code that
// stores group relative indices as `u8`.
// 1. We have code in the implementation that assumes it. In particular, code
// using `.count_ones()`.
// 2. The maximum group size is `256x256x256`, because there's code that stores
// group relative indices as `u8`.
// 3. There's code that stores group indices as `u8`.
debug_assert!(S::SIZE.x.is_power_of_two());
debug_assert!(S::SIZE.y.is_power_of_two());

View File

@ -84,7 +84,9 @@ pub(crate) fn wasi_fd_write(
let Ok(cio) = iov_addr
.add_offset(i)
.and_then(|p| p.read(&memory.view(&store)))
else { return Errno::Memviolation as i32; };
else {
return Errno::Memviolation as i32;
};
if let Err(e) = print_impl(env.data(), &store, cio.buf, cio.buf_len) {
return e as i32;
}

View File

@ -178,7 +178,7 @@ impl<T: Event> PreparedEventQuery<T> {
Ok(Self {
bytes: bincode::serialize(&event).map_err(PluginError::Encoding)?,
function_name: event.get_event_name(),
_phantom: PhantomData::default(),
_phantom: PhantomData,
})
}

View File

@ -1,4 +1,4 @@
#![feature(drain_filter, let_chains)]
#![feature(extract_if, let_chains)]
#![allow(clippy::option_map_unit_fn)]
mod aura;

View File

@ -53,27 +53,35 @@ impl<'a> System<'a> for Sys {
// For each mount...
for (entity, is_mount, body) in (&entities, &is_mounts, bodies.maybe()).join() {
// ...find the rider...
let Some((inputs_and_actions, rider)) = id_maps
.uid_entity(is_mount.rider)
.and_then(|rider| {
controllers
.get_mut(rider)
.map(|c| (
// Only take inputs and actions from the rider if the mount is not intelligent (TODO: expand the definition of 'intelligent').
let Some((inputs_and_actions, rider)) =
id_maps.uid_entity(is_mount.rider).and_then(|rider| {
controllers.get_mut(rider).map(|c| {
(
// Only take inputs and actions from the rider if the mount is not
// intelligent (TODO: expand the definition of 'intelligent').
if !matches!(body, Some(Body::Humanoid(_))) {
let actions = c.actions.drain_filter(|action| match action {
ControlAction::StartInput { input: i, .. }
| ControlAction::CancelInput(i) => matches!(i, InputKind::Jump | InputKind::Fly | InputKind::Roll),
_ => false
}).collect();
let actions = c
.actions
.extract_if(|action| match action {
ControlAction::StartInput { input: i, .. }
| ControlAction::CancelInput(i) => matches!(
i,
InputKind::Jump | InputKind::Fly | InputKind::Roll
),
_ => false,
})
.collect();
Some((c.inputs.clone(), actions))
} else {
None
},
rider,
))
)
})
})
else { continue };
else {
continue;
};
// ...apply the mount's position/ori/velocity to the rider...
let pos = positions.get(entity).copied();
@ -151,7 +159,7 @@ impl<'a> System<'a> for Sys {
let inputs = controllers.get_mut(entity).map(|c| {
let actions: Vec<_> = c
.actions
.drain_filter(|action| match action {
.extract_if(|action| match action {
ControlAction::StartInput { input: i, .. }
| ControlAction::CancelInput(i) => {
matches!(i, InputKind::Jump | InputKind::Fly | InputKind::Roll)

View File

@ -222,7 +222,7 @@ impl<'a> PhysicsData<'a> {
}
// Update PreviousPhysCache
for (_, vel, position, ori, mut phys_cache, collider, scale, cs) in (
for (_, vel, position, ori, phys_cache, collider, scale, cs) in (
&self.read.entities,
&self.write.velocities,
&self.write.positions,
@ -797,7 +797,7 @@ impl<'a> PhysicsData<'a> {
ori,
body,
character_state,
mut physics_state,
physics_state,
pos_vel_ori_defer,
previous_cache,
_,

View File

@ -74,7 +74,7 @@ impl<'a> System<'a> for Sys {
let mut rng = rand::thread_rng();
// Attacks
'projectile_loop: for (entity, pos, physics, vel, mut projectile) in (
'projectile_loop: for (entity, pos, physics, vel, projectile) in (
&read_data.entities,
&read_data.positions,
&read_data.physics_states,

View File

@ -141,6 +141,7 @@ pub fn create_player(state: &mut State) -> Entity {
.build()
}
#[allow(clippy::needless_pass_by_ref_mut)]
pub fn generate_chunk(state: &mut State, chunk_pos: Vec2<i32>) {
let (x, y) = chunk_pos.map(|e| e.to_le_bytes()).into_tuple();
let mut rng = SmallRng::from_seed([

View File

@ -1,4 +1,3 @@
#![feature(drain_filter)]
//! Network Protocol
//!
//! a I/O-Free protocol for the veloren network crate.

View File

@ -388,7 +388,8 @@ where
// try to order pending
let mut pending_violated = false;
let mut reliable = vec![];
self.pending_reliable_buffers.drain_filter(|(_, buffer)| {
self.pending_reliable_buffers.retain(|(_, buffer)| {
// try to get Sid without touching buffer
let mut testbuffer = buffer.clone();
match ITFrame::read_frame(&mut testbuffer) {
@ -398,13 +399,13 @@ where
length: _,
})) => {
reliable.push((sid, buffer.clone()));
true
false
},
Ok(Some(_)) | Err(_) => {
pending_violated = true;
true
false
},
Ok(None) => false,
Ok(None) => true,
}
});

View File

@ -500,7 +500,7 @@ pub enum ProtocolsError {
}
///////////////////////////////////////
//// TCP
// TCP
#[derive(Debug)]
pub struct TcpDrain {
half: OwnedWriteHalf,
@ -546,7 +546,7 @@ impl UnreliableSink for TcpSink {
}
///////////////////////////////////////
//// MPSC
// MPSC
#[derive(Debug)]
pub struct MpscDrain {
sender: mpsc::Sender<MpscMsg>,
@ -584,7 +584,7 @@ impl UnreliableSink for MpscSink {
}
///////////////////////////////////////
//// QUIC
// QUIC
#[cfg(feature = "quic")]
type QuicStream = (
BytesMut,

View File

@ -456,7 +456,6 @@ impl BParticipant {
let retrigger = |cid: Cid, mut p: RecvProtocols, map: &mut HashMap<_, _>| {
let hacky_recv_s = hacky_recv_s.clone();
let handle = tokio::spawn(async move {
let cid = cid;
let r = p.recv().await;
let _ = hacky_recv_s.send((cid, r, p)); // ignoring failed
});
@ -862,6 +861,7 @@ mod tests {
)
}
#[allow(clippy::needless_pass_by_ref_mut)]
async fn mock_mpsc(
cid: Cid,
_runtime: &Arc<Runtime>,

View File

@ -27,7 +27,6 @@ impl<T: Eq + Hash> DeferredTracer<T> {
*self.items.entry(t).or_default() += 1;
self.last = Instant::now();
self.last_cnt += 1;
} else {
}
}

View File

@ -94,7 +94,9 @@ impl Data {
.faction
.and_then(|f| this.factions.get(f))
.map(|f| f.good_or_evil)
else { continue };
else {
continue;
};
let rand_wpos = |rng: &mut SmallRng, matches_plot: fn(&PlotKind) -> bool| {
let wpos2d = site2
@ -257,18 +259,23 @@ impl Data {
let Some(species) = [
Some(comp::body::biped_large::Species::Ogre),
Some(comp::body::biped_large::Species::Cyclops),
Some(comp::body::biped_large::Species::Wendigo).filter(|_| biome == BiomeKind::Taiga),
Some(comp::body::biped_large::Species::Wendigo)
.filter(|_| biome == BiomeKind::Taiga),
Some(comp::body::biped_large::Species::Cavetroll),
Some(comp::body::biped_large::Species::Mountaintroll).filter(|_| biome == BiomeKind::Mountain),
Some(comp::body::biped_large::Species::Swamptroll).filter(|_| biome == BiomeKind::Swamp),
Some(comp::body::biped_large::Species::Mountaintroll)
.filter(|_| biome == BiomeKind::Mountain),
Some(comp::body::biped_large::Species::Swamptroll)
.filter(|_| biome == BiomeKind::Swamp),
Some(comp::body::biped_large::Species::Blueoni),
Some(comp::body::biped_large::Species::Redoni),
Some(comp::body::biped_large::Species::Tursus).filter(|_| chunk.temp < CONFIG.snow_temp),
Some(comp::body::biped_large::Species::Tursus)
.filter(|_| chunk.temp < CONFIG.snow_temp),
]
.into_iter()
.flatten()
.choose(&mut rng)
else { continue };
.into_iter()
.flatten()
.choose(&mut rng) else {
continue;
};
this.npcs.create_npc(Npc::new(
rng.gen(),

View File

@ -1 +1 @@
nightly-2023-04-20
nightly-2023-09-28

View File

@ -1,3 +1,7 @@
#![allow(
clippy::needless_pass_by_ref_mut //until we find a better way for specs
)]
use clap::Parser;
use common::comp;
use server::persistence::SqlLogMode;

View File

@ -2924,11 +2924,8 @@ impl<'a> AgentData<'a> {
{
agent.action_state.counters[FCounters::SummonThreshold as usize] -=
SUMMON_THRESHOLD;
if !agent.action_state.conditions[Conditions::AttackToggle as usize] {
agent.action_state.conditions[Conditions::AttackToggle as usize] = true;
} else {
agent.action_state.conditions[Conditions::AttackToggle as usize] = false;
}
agent.action_state.conditions[Conditions::AttackToggle as usize] =
!agent.action_state.conditions[Conditions::AttackToggle as usize];
}
} else {
// If target is in melee range use flamecrush

View File

@ -1,4 +1,7 @@
#![feature(exclusive_range_pattern, let_chains)]
#![allow(
clippy::needless_pass_by_ref_mut //until we find a better way for specs
)]
#[cfg(all(feature = "be-dyn-lib", feature = "use-dyn-lib"))]
compile_error!("Can't use both \"be-dyn-lib\" and \"use-dyn-lib\" features at once");

View File

@ -1,15 +1,10 @@
#![deny(unsafe_code)]
#![allow(clippy::option_map_unit_fn)]
#![deny(clippy::clone_on_ref_ptr)]
#![feature(
box_patterns,
drain_filter,
let_chains,
never_type,
option_zip,
unwrap_infallible
#![allow(
clippy::option_map_unit_fn,
clippy::needless_pass_by_ref_mut //until we find a better way for specs
)]
#![feature(hash_drain_filter)]
#![deny(clippy::clone_on_ref_ptr)]
#![feature(box_patterns, let_chains, never_type, option_zip, unwrap_infallible)]
pub mod automod;
mod character_creator;
@ -710,22 +705,20 @@ impl Server {
// significant changes to this code. Here is the approximate order of
// things. Please update it as this code changes.
//
// 1) Collect input from the frontend, apply input effects to the
// state of the game
// 2) Go through any events (timer-driven or otherwise) that need handling
// and apply them to the state of the game
// 3) Go through all incoming client network communications, apply them to
// the game state
// 4) Perform a single LocalState tick (i.e: update the world and entities
// in the world)
// 5) Go through the terrain update queue and apply all changes to
// the terrain
// 1) Collect input from the frontend, apply input effects to the state of the
// game
// 2) Go through any events (timer-driven or otherwise) that need handling and
// apply them to the state of the game
// 3) Go through all incoming client network communications, apply them to the
// game state
// 4) Perform a single LocalState tick (i.e: update the world and entities in
// the world)
// 5) Go through the terrain update queue and apply all changes to the terrain
// 6) Send relevant state updates to all clients
// 7) Check for persistence updates related to character data, and message the
// relevant entities
// 8) Update Metrics with current data
// 9) Finish the tick, passing control of the main thread back
// to the frontend
// 9) Finish the tick, passing control of the main thread back to the frontend
// 1) Build up a list of events for this frame, to be passed to the frontend.
let mut frontend_events = Vec::new();

View File

@ -1062,7 +1062,7 @@ pub fn update(
// The `defer_foreign_keys` pragma treats the foreign key
// constraints as deferred for the next transaction (it turns itself
// off at the commit boundary). https://sqlite.org/foreignkeys.html#fk_deferred
transaction.pragma_update(None, "defer_foreign_keys", &"ON".to_string())?;
transaction.pragma_update(None, "defer_foreign_keys", "ON")?;
let mut stmt = transaction.prepare_cached(
"

View File

@ -267,8 +267,8 @@ impl CharacterUpdater {
}
pub fn process_batch_completion(&mut self, completed_batch_id: u64) {
self.pending_database_actions.drain_filter(|_, event| {
matches!(event, DatabaseAction::Submitted {
self.pending_database_actions.retain(|_, event| {
!matches!(event, DatabaseAction::Submitted {
batch_id,
} if completed_batch_id == *batch_id)
});

View File

@ -875,8 +875,12 @@ impl StateExt for State {
let mut automod = self.ecs().write_resource::<AutoMod>();
let client = self.ecs().read_storage::<Client>();
let player = self.ecs().read_storage::<Player>();
let Some(client) = client.get(entity) else { return true };
let Some(player) = player.get(entity) else { return true };
let Some(client) = client.get(entity) else {
return true;
};
let Some(player) = player.get(entity) else {
return true;
};
match automod.validate_chat_msg(
player.uuid(),

View File

@ -885,11 +885,7 @@ fn remembers_fight_with(
// read_data: &ReadData,
// agent: &mut Agent,
// target: EcsEntity,
// ) {
// rtsim_entity.is_some().then(|| {
// read_data
// .stats
// .get(target)
// .map(|stats| agent.add_fight_to_memory(&stats.name,
// ) { rtsim_entity.is_some().then(|| { read_data .stats .get(target)
// .map(|stats| agent.add_fight_to_memory(&stats.name,
// read_data.time.0)) });
// }

View File

@ -59,7 +59,7 @@ impl<'a> System<'a> for Sys {
for (pet_entity, owner_pos) in lost_pets.iter() {
let stay = agn.get(*pet_entity).and_then(|x| x.stay_pos).is_some();
if let Some(mut pet_pos) = positions.get_mut(*pet_entity) && !stay{
if let Some(pet_pos) = positions.get_mut(*pet_entity) && !stay{
// Move the pets to their owner's position
// TODO: Create a teleportation event to handle this instead of
// processing the entity position move here

View File

@ -61,8 +61,8 @@ impl<'a> System<'a> for Sys {
// To update subscriptions
// 1. Iterate through clients
// 2. Calculate current chunk position
// 3. If chunk is different (use fuzziness) or the client view distance
// has changed continue, otherwise return
// 3. If chunk is different (use fuzziness) or the client view distance has
// changed continue, otherwise return
// 4. Iterate through subscribed regions
// 5. Check if region is still in range (use fuzziness)
// 6. If not in range
@ -71,7 +71,7 @@ impl<'a> System<'a> for Sys {
// 7. Determine list of regions that are in range and iterate through it
// - check if in hashset (hash calc) if not add it
let mut regions_to_remove = Vec::new();
for (mut subscription, pos, presence, client_entity, client) in (
for (subscription, pos, presence, client_entity, client) in (
&mut subscriptions,
&positions,
&presences,

View File

@ -83,11 +83,14 @@ impl<'a> System<'a> for Sys {
.join()
{
let portal_pos = positions.get(teleporting.portal);
let Some(Object::Portal { target, requires_no_aggro, .. }) = objects
.get(teleporting.portal)
let Some(Object::Portal {
target,
requires_no_aggro,
..
}) = objects.get(teleporting.portal)
else {
cancel_teleporting.push(entity);
continue
continue;
};
if portal_pos.map_or(true, |portal_pos| {

View File

@ -104,7 +104,6 @@ impl Animation for AlphaAnimation {
Quaternion::rotation_y(-0.2 + move1 * -0.3) * Quaternion::rotation_z(0.2);
next.wing_out_r.orientation =
Quaternion::rotation_y(0.2 + move1 * 0.3) * Quaternion::rotation_z(-0.2);
} else {
}
next

View File

@ -137,7 +137,6 @@ impl Animation for ComboAnimation {
* Quaternion::rotation_z(0.2);
next.wing_out_r.orientation = Quaternion::rotation_y(0.2 + move1 * 0.3)
* Quaternion::rotation_z(-0.2);
} else {
}
},
_ => {},

View File

@ -88,7 +88,6 @@ impl Animation for ShockwaveAnimation {
next.foot_l.orientation = Quaternion::rotation_x(0.0);
next.foot_r.position = Vec3::new(s_a.foot.0, s_a.foot.1, s_a.foot.2);
next.foot_r.orientation = Quaternion::rotation_x(0.0);
} else {
}
next

View File

@ -79,7 +79,6 @@ impl Animation for ShootAnimation {
next.foot_l.orientation = Quaternion::rotation_x(movement1abs * 0.3);
next.foot_r.orientation = Quaternion::rotation_x(movement1abs * 0.3);
} else {
}
if velocity.xy().magnitude() < 1.0 {
next.wing_in_l.orientation = Quaternion::rotation_y(-1.0 + movement1abs * 0.8)

View File

@ -103,7 +103,6 @@ impl Animation for SummonAnimation {
next.tail_rear.position = Vec3::new(0.0, s_a.tail_rear.0, s_a.tail_rear.1);
next.tail_rear.orientation =
Quaternion::rotation_x(-movement1abs * 0.1 + movement2abs * 0.1 + twitch2 * 0.02);
} else {
}
next

View File

@ -83,7 +83,6 @@ impl Animation for AlphaAnimation {
Quaternion::rotation_y(-0.2 + move1 * -0.3) * Quaternion::rotation_z(0.2);
next.wing_out_r.orientation =
Quaternion::rotation_y(0.2 + move1 * 0.3) * Quaternion::rotation_z(-0.2);
} else {
}
next

View File

@ -68,7 +68,6 @@ impl Animation for ShockwaveAnimation {
next.leg_l.orientation = Quaternion::rotation_x(0.0);
next.leg_r.position = Vec3::new(s_a.leg.0, s_a.leg.1, s_a.leg.2);
next.leg_r.orientation = Quaternion::rotation_x(0.0);
} else {
}
next

View File

@ -66,7 +66,6 @@ impl Animation for ShootAnimation {
next.leg_l.orientation = Quaternion::rotation_x(movement1abs * -0.5);
next.leg_r.orientation = Quaternion::rotation_x(movement1abs * -0.5);
} else {
}
if velocity.xy().magnitude() < 1.0 {
next.wing_in_l.orientation = Quaternion::rotation_y(-1.0 + movement1abs * 0.8)

View File

@ -84,7 +84,6 @@ impl Animation for SummonAnimation {
next.tail.position = Vec3::new(0.0, s_a.tail.0, s_a.tail.1);
next.tail.orientation =
Quaternion::rotation_x(-movement1abs * 0.1 + movement2abs * 0.1 + twitch2 * 0.02);
} else {
}
next

View File

@ -105,7 +105,6 @@ impl Animation for BeamAnimation {
Quaternion::rotation_x(move1 * 0.1) * Quaternion::rotation_z(move1 * -0.1);
next.shorts.orientation =
Quaternion::rotation_x(move1 * 0.2) * Quaternion::rotation_z(move1 * -0.2);
} else {
};
},
_ => {},

View File

@ -84,7 +84,6 @@ impl Animation for RepeaterAnimation {
* Quaternion::rotation_z(move1 * -0.6 + move3 * 0.8);
next.chest.position = Vec3::new(0.0, s_a.chest.0, s_a.chest.1);
next.chest.orientation = Quaternion::rotation_x(0.0);
} else {
};
next.shorts.position = Vec3::new(0.0, s_a.shorts.0 + move1 * 2.0, s_a.shorts.1);
next.shorts.orientation = Quaternion::rotation_x(move1 * 0.2 + move3 * 0.2);

View File

@ -218,7 +218,6 @@ impl Animation for SneakWieldAnimation {
next.shorts.position = Vec3::new(0.0, 1.0 + s_a.shorts.0, s_a.shorts.1);
next.shorts.orientation =
Quaternion::rotation_x(0.15) * Quaternion::rotation_z(0.25);
} else {
}
next.hand_l.position = Vec3::new(s_a.ahl.0, s_a.ahl.1, s_a.ahl.2);
next.hand_l.orientation =

View File

@ -208,7 +208,6 @@ impl Animation for WieldAnimation {
next.shorts.position = Vec3::new(0.0, 1.0 + s_a.shorts.0, s_a.shorts.1);
next.shorts.orientation =
Quaternion::rotation_x(0.15) * Quaternion::rotation_z(0.25);
} else {
}
next.hand_l.position = Vec3::new(s_a.ahl.0, s_a.ahl.1, s_a.ahl.2);
next.hand_l.orientation =

View File

@ -73,7 +73,6 @@ impl Animation for ShockwaveAnimation {
next.foot_l.position = Vec3::new(-s_a.foot.0, s_a.foot.1, s_a.foot.2 + move2);
next.foot_r.position = Vec3::new(s_a.foot.0, s_a.foot.1, s_a.foot.2 + move2);
} else {
}
next
}

View File

@ -66,7 +66,6 @@ impl Animation for BreatheAnimation {
next.foot_bl.orientation = Quaternion::rotation_y(twitch2 * 0.02);
next.foot_br.orientation = Quaternion::rotation_y(twitch2 * 0.02);
} else {
};
next
}

View File

@ -71,7 +71,6 @@ impl Animation for ShootAnimation {
next.foot_bl.position = Vec3::new(-s_a.feet_b.0, s_a.feet_b.1, s_a.feet_b.2);
next.foot_br.position = Vec3::new(s_a.feet_b.0, s_a.feet_b.1, s_a.feet_b.2);
} else {
};
},
Some("common.abilities.custom.dagon.dagonbombs") => {
@ -117,7 +116,6 @@ impl Animation for ShootAnimation {
next.foot_bl.position = Vec3::new(-s_a.feet_b.0, s_a.feet_b.1, s_a.feet_b.2);
next.foot_br.position = Vec3::new(s_a.feet_b.0, s_a.feet_b.1, s_a.feet_b.2);
} else {
};
},
_ => {
@ -150,7 +148,6 @@ impl Animation for ShootAnimation {
next.foot_bl.position = Vec3::new(-s_a.feet_b.0, s_a.feet_b.1, s_a.feet_b.2);
next.foot_br.position = Vec3::new(s_a.feet_b.0, s_a.feet_b.1, s_a.feet_b.2);
} else {
};
},
}

View File

@ -1,4 +1,7 @@
#![feature(stmt_expr_attributes)]
#![allow(
clippy::needless_pass_by_ref_mut //until we find a better way for specs
)]
#[cfg(all(feature = "be-dyn-lib", feature = "use-dyn-lib"))]
compile_error!("Can't use both \"be-dyn-lib\" and \"use-dyn-lib\" features at once");

View File

@ -407,7 +407,7 @@ impl TabComplete for ArgumentSpec {
.filter(|string| string.starts_with(part))
.map(|c| c.to_string())
.collect(),
ArgumentSpec::Boolean(_, part, _) => vec!["true", "false"]
ArgumentSpec::Boolean(_, part, _) => ["true", "false"]
.iter()
.filter(|string| string.starts_with(part))
.map(|c| c.to_string())

View File

@ -68,7 +68,7 @@ impl<'a> System<'a> for Sys {
.as_mut()
.map(|t| *t += dt.0);
for mut floater in hp_floater_list.floaters.iter_mut() {
for floater in hp_floater_list.floaters.iter_mut() {
// Increment timer
floater.timer += dt.0;
floater.jump_timer += dt.0;

View File

@ -196,7 +196,7 @@ impl<'a> InventoryScroller<'a> {
.set(self.bg_ids.bg_frame, ui);
}
fn title(&mut self, state: &mut ConrodState<'_, InventoryScrollerState>, ui: &mut UiCell<'_>) {
fn title(&mut self, state: &ConrodState<'_, InventoryScrollerState>, ui: &mut UiCell<'_>) {
Text::new(
&self
.localized_strings
@ -371,7 +371,7 @@ impl<'a> InventoryScroller<'a> {
});
}
for (pos, item) in items.into_iter() {
if self.details_mode && !self.is_us && matches!(item, None) {
if self.details_mode && !self.is_us && item.is_none() {
continue;
}
let (x, y) = if self.details_mode {
@ -488,7 +488,7 @@ impl<'a> InventoryScroller<'a> {
fn footer_metrics(
&mut self,
state: &mut ConrodState<'_, InventoryScrollerState>,
state: &ConrodState<'_, InventoryScrollerState>,
ui: &mut UiCell<'_>,
) {
let space_used = self.inventory.populated_slots();

View File

@ -874,13 +874,13 @@ mod tests {
#[test]
fn parse_cmds() {
let expected: Result<(String, Vec<String>), String> = Ok(("help".to_string(), vec![]));
assert_eq!(parse_cmd(r#"help"#), expected);
assert_eq!(parse_cmd(r"help"), expected);
let expected: Result<(String, Vec<String>), String> = Ok(("say".to_string(), vec![
"foo".to_string(),
"bar".to_string(),
]));
assert_eq!(parse_cmd(r#"say foo bar"#), expected);
assert_eq!(parse_cmd(r"say foo bar"), expected);
assert_eq!(parse_cmd(r#"say "foo" "bar""#), expected);
let expected: Result<(String, Vec<String>), String> =

View File

@ -1479,7 +1479,7 @@ impl<'a> Widget for Crafting<'a> {
});
self.inventory
.slots_with_id()
.filter(|(_, item)| item.as_ref().map_or(false, |i| can_repair(i)))
.filter(|(_, item)| item.as_ref().map_or(false, can_repair))
.for_each(|(slot, _)| {
events.push(Event::RepairItem {
slot: Slot::Inventory(slot),
@ -1487,9 +1487,7 @@ impl<'a> Widget for Crafting<'a> {
});
}
let can_perform = repair_slot
.item(self.inventory)
.map_or(false, |item| can_repair(item));
let can_perform = repair_slot.item(self.inventory).map_or(false, can_repair);
(repair_slot.slot, None, can_perform)
},

View File

@ -515,31 +515,27 @@ impl BuffIconKind {
}
impl PartialOrd for BuffIconKind {
fn partial_cmp(&self, other: &Self) -> Option<Ordering> {
fn partial_cmp(&self, other: &Self) -> Option<Ordering> { Some(self.cmp(other)) }
}
impl Ord for BuffIconKind {
fn cmp(&self, other: &Self) -> Ordering {
match (self, other) {
(
BuffIconKind::Buff { kind, .. },
BuffIconKind::Buff {
kind: other_kind, ..
},
) => Some(kind.cmp(other_kind)),
(BuffIconKind::Buff { .. }, BuffIconKind::Stance(_)) => Some(Ordering::Greater),
(BuffIconKind::Stance(_), BuffIconKind::Buff { .. }) => Some(Ordering::Less),
) => kind.cmp(other_kind),
(BuffIconKind::Buff { .. }, BuffIconKind::Stance(_)) => Ordering::Greater,
(BuffIconKind::Stance(_), BuffIconKind::Buff { .. }) => Ordering::Less,
(BuffIconKind::Stance(stance), BuffIconKind::Stance(stance_other)) => {
Some(stance.cmp(stance_other))
stance.cmp(stance_other)
},
}
}
}
impl Ord for BuffIconKind {
fn cmp(&self, other: &Self) -> Ordering {
// We know this is safe since we can look at the partialord implementation and
// see that every variant is wrapped in Some
self.partial_cmp(other).unwrap()
}
}
impl PartialEq for BuffIconKind {
fn eq(&self, other: &Self) -> bool {
match (self, other) {
@ -586,7 +582,7 @@ impl BuffIcon {
buffs
.iter_active()
.filter_map(BuffIcon::from_buffs)
.chain(stance.and_then(BuffIcon::from_stance).into_iter())
.chain(stance.and_then(BuffIcon::from_stance))
.collect::<Vec<_>>()
}

View File

@ -223,7 +223,7 @@ impl<'a> Widget for Quest<'a> {
// [amount, item_desc]
//("common.items.weapons.sword.caladbolg");
let rewards = vec![
let rewards = [
(1, "common.items.weapons.dagger.starter_dagger", "Dagger"),
(4, "common.items.crafting_ing.seashells", "Seashell"),
(

View File

@ -1,17 +1,22 @@
#![deny(unsafe_code)]
#![allow(incomplete_features)]
#![allow(clippy::identity_op, clippy::option_map_unit_fn)]
#![allow(
clippy::identity_op,
clippy::option_map_unit_fn,
clippy::needless_pass_by_ref_mut //until we find a better way for specs
)]
#![deny(clippy::clone_on_ref_ptr)]
#![feature(
array_methods,
array_zip,
drain_filter,
extract_if,
trait_alias,
option_get_or_insert_default,
map_try_insert,
slice_as_chunks,
let_chains,
generic_const_exprs
generic_const_exprs,
maybe_uninit_uninit_array,
maybe_uninit_array_assume_init
)]
#![recursion_limit = "2048"]

View File

@ -314,7 +314,7 @@ impl PlayState for MainMenuState {
password,
server_address,
} => {
let mut net_settings = &mut global_state.settings.networking;
let net_settings = &mut global_state.settings.networking;
let use_quic = net_settings.use_quic;
net_settings.username = username.clone();
net_settings.default_server = server_address.clone();

View File

@ -529,8 +529,8 @@ pub fn generate_mesh<'a>(
(
opaque_deep
.into_iter()
.chain(opaque_shallow.into_iter())
.chain(opaque_surface.into_iter())
.chain(opaque_shallow)
.chain(opaque_surface)
.collect(),
fluid_mesh,
Mesh::new(),

View File

@ -39,7 +39,11 @@ impl Add<Vertex> for Vertex {
fn add(self, other: Self) -> Self::Output {
Self {
pos: self.pos.zip(other.pos).map(|(a, b)| a + b),
pos: [
self.pos[0] + other.pos[0],
self.pos[1] + other.pos[1],
self.pos[2] + other.pos[2],
],
}
}
}

View File

@ -1484,27 +1484,13 @@ impl Renderer {
// _bones: &Consts<figure::BoneData>,
// _lod: &lod_terrain::LodData,
// _locals: &Consts<shadow::Locals>,
// ) {
// // FIXME: Consider reenabling at some point.
// /* let (point_shadow_maps, directed_shadow_maps) =
// if let Some(shadow_map) = &mut self.shadow_map {
// (
// (
// shadow_map.point_res.clone(),
// shadow_map.point_sampler.clone(),
// ),
// (
// shadow_map.directed_res.clone(),
// shadow_map.directed_sampler.clone(),
// ),
// )
// } else {
// (
// (self.noise_tex.srv.clone(), self.noise_tex.sampler.clone()),
// (self.noise_tex.srv.clone(), self.noise_tex.sampler.clone()),
// )
// };
// let model = &model.opaque;
// ) { // FIXME: Consider reenabling at some point. /* let (point_shadow_maps,
// directed_shadow_maps) = if let Some(shadow_map) = &mut self.shadow_map { (
// ( shadow_map.point_res.clone(), shadow_map.point_sampler.clone(), ), (
// shadow_map.directed_res.clone(), shadow_map.directed_sampler.clone(), ), )
// } else { ( (self.noise_tex.srv.clone(), self.noise_tex.sampler.clone()),
// (self.noise_tex.srv.clone(), self.noise_tex.sampler.clone()), ) }; let
// model = &model.opaque;
// self.encoder.draw(
// &gfx::Slice {

View File

@ -445,7 +445,9 @@ impl<'frame> Drawer<'frame> {
/// pending uploads.
fn run_ui_premultiply_passes(&mut self) {
prof_span!("run_ui_premultiply_passes");
let Some(premultiply_alpha) = self.borrow.pipelines.premultiply_alpha() else { return };
let Some(premultiply_alpha) = self.borrow.pipelines.premultiply_alpha() else {
return;
};
let encoder = self.encoder.as_mut().unwrap();
let device = self.borrow.device;

View File

@ -22,6 +22,11 @@ pub struct Locals {
pub postprocess_bind: postprocess::BindGroup,
}
fn arr_zip_map<const N: usize, A, B, C>(a: [A; N], b: [B; N], f: impl Fn(A, B) -> C) -> [C; N] {
let mut b = b.into_iter();
a.map(|a| f(a, b.next().unwrap()))
}
impl Locals {
pub(super) fn new(
device: &wgpu::Device,
@ -58,10 +63,9 @@ impl Locals {
);
let bloom_binds = bloom.map(|bloom| {
bloom
.src_views
.zip(bloom.locals) // zip arrays
.map(|(view, locals)| layouts.bloom.bind(device, view, sampler, locals))
arr_zip_map(bloom.src_views, bloom.locals, |view, locals| {
layouts.bloom.bind(device, view, sampler, locals)
})
});
Self {
@ -107,10 +111,9 @@ impl Locals {
&self.postprocess,
);
self.bloom_binds = bloom.map(|bloom| {
bloom
.src_views
.zip(bloom.locals) // zip arrays
.map(|(view, locals)| layouts.bloom.bind(device, view, sampler, locals))
arr_zip_map(bloom.src_views, bloom.locals, |view, locals| {
layouts.bloom.bind(device, view, sampler, locals)
})
});
}
}

View File

@ -1069,7 +1069,7 @@ fn mesh_hold() -> BoneMeshes {
)
}
/////////
//////
#[derive(Deserialize)]
struct QuadrupedSmallCentralSpec(HashMap<(QSSpecies, QSBodyType), SidedQSCentralVoxSpec>);
@ -1660,7 +1660,7 @@ impl QuadrupedMediumLateralSpec {
}
}
////
//////
#[derive(Deserialize)]
struct BirdMediumCentralSpec(HashMap<(BMSpecies, BMBodyType), SidedBMCentralVoxSpec>);
@ -1914,7 +1914,7 @@ impl BirdMediumLateralSpec {
}
}
////
//////
#[derive(Deserialize)]
struct TheropodCentralSpec(HashMap<(TSpecies, TBodyType), SidedTCentralVoxSpec>);
@ -2244,7 +2244,7 @@ impl TheropodLateralSpec {
}
}
////
//////
#[derive(Deserialize)]
struct ArthropodCentralSpec(HashMap<(ASpecies, ABodyType), SidedACentralVoxSpec>);
@ -2644,7 +2644,7 @@ impl ArthropodLateralSpec {
(lateral, Vec3::from(spec.leg_br.offset))
}
}
////
//////
#[derive(Deserialize)]
struct FishMediumCentralSpec(HashMap<(FMSpecies, FMBodyType), SidedFMCentralVoxSpec>);
@ -2850,7 +2850,7 @@ impl FishMediumLateralSpec {
}
}
////
//////
#[derive(Deserialize)]
struct FishSmallCentralSpec(HashMap<(FSSpecies, FSBodyType), SidedFSCentralVoxSpec>);
@ -2994,7 +2994,7 @@ impl FishSmallLateralSpec {
}
}
////
//////
#[derive(Deserialize)]
struct BipedSmallWeaponSpec(HashMap<ToolKey, ArmorVoxSpec>);
@ -3269,8 +3269,8 @@ impl BipedSmallWeaponSpec {
(tool_kind_segment, offset)
}
}
////
//////
#[derive(Deserialize)]
struct DragonCentralSpec(HashMap<(DSpecies, DBodyType), SidedDCentralVoxSpec>);
@ -3641,7 +3641,7 @@ impl DragonLateralSpec {
}
}
////
//////
#[derive(Deserialize)]
struct BirdLargeCentralSpec(HashMap<(BLASpecies, BLABodyType), SidedBLACentralVoxSpec>);
@ -4044,7 +4044,7 @@ impl BirdLargeLateralSpec {
}
}
////
//////
#[derive(Deserialize)]
struct BipedLargeCentralSpec(HashMap<(BLSpecies, BLBodyType), SidedBLCentralVoxSpec>);
@ -4462,7 +4462,8 @@ impl BipedLargeSecondSpec {
(tool_kind_segment, offset)
}
}
////
//////
#[derive(Deserialize)]
struct GolemCentralSpec(HashMap<(GSpecies, GBodyType), SidedGCentralVoxSpec>);
@ -4772,8 +4773,7 @@ impl GolemLateralSpec {
}
}
/////
//////
#[derive(Deserialize)]
struct QuadrupedLowCentralSpec(HashMap<(QLSpecies, QLBodyType), SidedQLCentralVoxSpec>);
@ -5050,8 +5050,7 @@ impl QuadrupedLowLateralSpec {
}
}
////
//////
#[derive(Deserialize)]
struct ObjectCentralSpec(HashMap<object::Body, SidedObjectCentralVoxSpec>);

View File

@ -648,7 +648,7 @@ impl FigureMgr {
}
let dt = ecs.fetch::<DeltaTime>().0;
let updater = ecs.read_resource::<LazyUpdate>();
for (entity, light_emitter_opt, interpolated, pos, body, mut light_anim) in (
for (entity, light_emitter_opt, interpolated, pos, body, light_anim) in (
&ecs.entities(),
ecs.read_storage::<LightEmitter>().maybe(),
ecs.read_storage::<Interpolated>().maybe(),
@ -1101,9 +1101,9 @@ impl FigureMgr {
let holding_lantern = inventory
.map_or(false, |i| i.equipped(EquipSlot::Lantern).is_some())
&& light_emitter.is_some()
&& !((matches!(second_tool_hand, Some(_))
|| matches!(active_tool_hand, Some(Hands::Two)))
&& character.map_or(false, |c| c.is_wield()))
&& !(second_tool_hand.is_some()
|| matches!(active_tool_hand, Some(Hands::Two))
&& character.map_or(false, |c| c.is_wield()))
&& !character.map_or(false, |c| c.is_using_hands())
&& physics.in_liquid().is_none();

View File

@ -250,7 +250,7 @@ pub fn clip_object_by_plane<T: Float + MulAdd<T, T, Output = T> + core::fmt::Deb
tolerance: T,
) {
let mut intersection_points = Vec::new();
polys.drain_filter(|points| {
polys.retain_mut(|points| {
let len = intersection_points.len();
let outside_first = clip_points_by_plane(points, plane, &mut intersection_points);
// Only remember intersections that are not coplanar with this side; i.e. those
@ -273,7 +273,7 @@ pub fn clip_object_by_plane<T: Float + MulAdd<T, T, Output = T> + core::fmt::Deb
intersection_points.swap(len, len + 1);
}
// Remove polygon if it was clipped away
points.is_empty()
!points.is_empty()
});
// Add a polygon of all intersection points with the plane to close out the
// object.

View File

@ -763,9 +763,9 @@ impl Scene {
renderer.update_consts(&mut self.data.lights, lights);
// Update event lights
self.event_lights.drain_filter(|el| {
self.event_lights.retain_mut(|el| {
el.timeout -= dt;
el.timeout <= 0.0
el.timeout > 0.0
});
// Update shadow constants

View File

@ -137,7 +137,7 @@ impl Scene {
figure_state: None,
backdrop: backdrop.map(|specifier| {
let mut state = FigureState::new(renderer, FixtureSkeleton::default(), ());
let mut state = FigureState::new(renderer, FixtureSkeleton, ());
let mut greedy = FigureModel::make_greedy();
let mut opaque_mesh = Mesh::new();
let (segment, offset) = load_mesh(specifier, Vec3::new(-55.0, -49.5, -2.0));

View File

@ -437,8 +437,8 @@ fn mesh_worker(
(
deep_level
.into_iter()
.chain(shallow_level.into_iter())
.chain(surface_level.into_iter())
.chain(shallow_level)
.chain(surface_level)
.collect(),
alt_indices,
)

View File

@ -59,7 +59,7 @@ impl Interactable {
volume_pos: VolumePos,
interaction: Interaction,
) -> Option<Self> {
let Some(block) = volume_pos.get_block(terrain, id_maps, colliders) else { return None };
let block = volume_pos.get_block(terrain, id_maps, colliders)?;
let block_interaction = match interaction {
Interaction::Collect => {
// Check if this is an unlockable sprite
@ -114,11 +114,11 @@ impl Interactable {
/// interact with if the interact key is pressed
/// Selected in the following order:
/// 1) Targeted items, in order of nearest under cursor:
/// (a) entity (if within range)
/// (b) collectable
/// (c) can be mined, and is a mine sprite (Air) not a weak rock.
/// a) entity (if within range)
/// b) collectable
/// c) can be mined, and is a mine sprite (Air) not a weak rock.
/// 2) outside of targeted cam ray
/// -> closest of nearest interactable entity/block
/// -> closest of nearest interactable entity/block
pub(super) fn select_interactable(
client: &Client,
collect_target: Option<Target<target::Collectable>>,

View File

@ -216,7 +216,7 @@ settings_change_from!(Accessibility);
impl SettingsChange {
pub fn process(self, global_state: &mut GlobalState, session_state: &mut SessionState) {
let mut settings = &mut global_state.settings;
let settings = &mut global_state.settings;
match self {
SettingsChange::Audio(audio_change) => {
@ -366,7 +366,7 @@ impl SettingsChange {
},
SettingsChange::Gamepad(gamepad_change) => match gamepad_change {},
SettingsChange::Gameplay(gameplay_change) => {
let mut window = &mut global_state.window;
let window = &mut global_state.window;
match gameplay_change {
Gameplay::AdjustMousePan(sensitivity) => {
window.pan_sensitivity = sensitivity;

View File

@ -36,7 +36,7 @@ fn load_map(path: &Path) -> Option<SingleplayerWorld> {
let meta_path = path.join("meta.ron");
let Ok(f) = fs::File::open(&meta_path) else {
error!("Failed to open {}", meta_path.to_string_lossy());
error!("Failed to open {}", meta_path.to_string_lossy());
return None;
};

View File

@ -5,7 +5,7 @@ mod widget;
pub use defaults::Defaults;
pub(self) use primitive::Primitive;
use primitive::Primitive;
use super::{
super::graphic::{self, Graphic, TexId},

View File

@ -675,6 +675,7 @@ impl Window {
.game_analog_button_map
.get(&AnalogButton::from((button, code)))
{
#[allow(clippy::never_loop)]
for action in actions {
match *action {}
}
@ -684,6 +685,7 @@ impl Window {
.menu_analog_button_map
.get(&AnalogButton::from((button, code)))
{
#[allow(clippy::never_loop)]
for action in actions {
match *action {}
}

View File

@ -34,7 +34,7 @@ tracing = { workspace = true }
rand = { workspace = true }
rand_chacha = { workspace = true }
arr_macro = "0.2.1"
packed_simd = { package = "packed_simd_2", version = "0.3.8", optional = true }
packed_simd = { version = "0.3.9", optional = true }
rayon = { workspace = true }
serde = { workspace = true }
ron = { workspace = true }

View File

@ -691,7 +691,7 @@ fn main() {
let k = 32;
let sz = world.sim().get_size();
let sites = vec![
let sites = [
("center", sz / 2),
(
"dungeon",

View File

@ -11,7 +11,7 @@ fn main() {
]);
let mut middle = cons.clone();
middle.extend(vec!["tt"]);
let vowel = vec!["o", "e", "a", "i", "u", "au", "ee", "ow", "ay", "ey", "oe"];
let vowel = ["o", "e", "a", "i", "u", "au", "ee", "ow", "ay", "ey", "oe"];
let end = vec![
"et", "ige", "age", "ist", "en", "on", "og", "end", "ind", "ock", "een", "edge", "ist",
"ed", "est", "eed", "ast", "olt", "ey", "ean", "ead", "onk", "ink", "eon", "er", "ow",

View File

@ -196,10 +196,7 @@ fn palette(conn: Connection) -> Result<(), Box<dyn Error>> {
let kind = BlockKind::from_str(&row.get::<_, String>(0)?)?;
let rgb: Rgb<u8> = Rgb::new(row.get(1)?, row.get(2)?, row.get(3)?);
let count: i64 = row.get(4)?;
block_colors
.entry(kind)
.or_insert_with(Vec::new)
.push((rgb, count));
block_colors.entry(kind).or_default().push((rgb, count));
}
for (_, v) in block_colors.iter_mut() {
v.sort_by(|a, b| b.1.cmp(&a.1));
@ -207,7 +204,7 @@ fn palette(conn: Connection) -> Result<(), Box<dyn Error>> {
let mut palettes: HashMap<BlockKind, Vec<Rgb<u8>>> = HashMap::new();
for (kind, colors) in block_colors.iter() {
let palette = palettes.entry(*kind).or_insert_with(Vec::new);
let palette = palettes.entry(*kind).or_default();
if colors.len() <= 256 {
for (color, _) in colors {
palette.push(*color);

View File

@ -741,7 +741,7 @@ impl Civs {
1 << ((to_next_idx as u8 + 4) % 8);
}
for loc in path.iter() {
let mut chunk = ctx.sim.get_mut(loc.0).unwrap();
let chunk = ctx.sim.get_mut(loc.0).unwrap();
let depth = loc.1 * 250.0 - 20.0;
chunk.cave.1.alt =
chunk.alt - depth + ctx.rng.gen_range(-4.0..4.0) * (depth > 10.0) as i32 as f32;
@ -1408,7 +1408,7 @@ impl Civs {
(1 << (to_prev_idx as u8)) | (1 << (to_next_idx as u8));
*/
if randomize_offset {
let mut chunk = ctx.sim.get_mut(locs[1]).unwrap();
let chunk = ctx.sim.get_mut(locs[1]).unwrap();
chunk.path.0.offset =
Vec2::new(ctx.rng.gen_range(-16..17), ctx.rng.gen_range(-16..17));
}
@ -1569,7 +1569,9 @@ fn walk_in_all_dirs(
let adjacents = NEIGHBORS.map(|dir| a + dir);
let Some(a_chunk) = sim.get(a) else { return potential };
let Some(a_chunk) = sim.get(a) else {
return potential;
};
let mut chunks = [None; 8];
for i in 0..8 {
if loc_suitable_for_walking(sim, adjacents[i]) {

View File

@ -145,7 +145,9 @@ impl Tunnel {
}
fn biome_at(&self, wpos: Vec3<i32>, info: &CanvasInfo) -> Biome {
let Some(col) = info.col_or_gen(wpos.xy()) else { return Biome::default() };
let Some(col) = info.col_or_gen(wpos.xy()) else {
return Biome::default();
};
// Below the ground
let below = ((col.alt - wpos.z as f32) / 120.0).clamped(0.0, 1.0);

View File

@ -1036,7 +1036,6 @@ pub fn apply_caverns_to<R: Rng>(canvas: &mut Canvas, dynamic_rng: &mut R) {
}
};
let cavern_top = cavern_top;
let mut last_kind = BlockKind::Rock;
for z in cavern_bottom - 1..cavern_top {
use SpriteKind::*;

View File

@ -2,7 +2,8 @@
#![allow(
clippy::option_map_unit_fn,
clippy::blocks_in_if_conditions,
clippy::identity_op
clippy::identity_op,
clippy::needless_pass_by_ref_mut //until we find a better way for specs
)]
#![allow(clippy::branches_sharing_code)] // TODO: evaluate
#![deny(clippy::clone_on_ref_ptr)]

View File

@ -322,7 +322,7 @@ pub fn get_rivers<F: fmt::Debug + Float + Into<f64>, G: Float + Into<f64>>(
let pass_idx = (-indirection_idx) as usize;
// NOTE: Must exist since this lake had a downhill in the first place.
let neighbor_pass_idx = downhill[pass_idx] as usize/*downhill_idx*/;
let mut lake_neighbor_pass = &mut rivers[neighbor_pass_idx];
let lake_neighbor_pass = &mut rivers[neighbor_pass_idx];
// We definitely shouldn't have encountered this yet!
debug_assert!(lake_neighbor_pass.velocity == Vec3::zero());
// TODO: Rethink making the lake neighbor pass always a river or lake, no matter
@ -388,7 +388,7 @@ pub fn get_rivers<F: fmt::Debug + Float + Into<f64>, G: Float + Into<f64>>(
river_spline_derivative,
)
};
let mut lake = &mut rivers[chunk_idx];
let lake = &mut rivers[chunk_idx];
lake.spline_derivative = river_spline_derivative;
lake.river_kind = Some(RiverKind::Lake {
neighbor_pass_pos: neighbor_pass_pos
@ -495,7 +495,7 @@ pub fn get_rivers<F: fmt::Debug + Float + Into<f64>, G: Float + Into<f64>>(
// CONFIG.river_min_height.
let river = &rivers[chunk_idx];
let is_river = river.is_river() || width >= 0.5 && height >= CONFIG.river_min_height as f64;
let mut downhill_river = &mut rivers[downhill_idx];
let downhill_river = &mut rivers[downhill_idx];
if is_river {
// Provisionally make the downhill chunk a river as well.
@ -532,7 +532,7 @@ pub fn get_rivers<F: fmt::Debug + Float + Into<f64>, G: Float + Into<f64>>(
velocity.normalize();
velocity *= velocity_magnitude;
let mut river = &mut rivers[chunk_idx];
let river = &mut rivers[chunk_idx];
// NOTE: Not trying to do this more cleverly because we want to keep the river's
// neighbors. TODO: Actually put something in the neighbors.
river.velocity = velocity.map(|e| e as f32);
@ -636,31 +636,31 @@ impl m32 {
///
/// This algorithm does this in four steps:
///
/// 1. Sort the nodes in h by height (so the lowest node by altitude is first
/// in the list, and the highest node by altitude is last).
/// 1. Sort the nodes in h by height (so the lowest node by altitude is first in
/// the list, and the highest node by altitude is last).
/// 2. Iterate through the list in *reverse.* For each node, we compute its
/// drainage area as the sum of the drainage areas of its "children" nodes
/// (i.e. the nodes with directed edges to this node). To do this
/// efficiently, we start with the "leaves" (the highest nodes), which
/// have no neighbors higher than them, hence no directed edges to them.
/// We add their area to themselves, and then to all neighbors that they
/// flow into (their "ancestors" in the flow graph); currently, this just
/// means the node immediately downhill of this node. As we go lower, we
/// know that all our "children" already had their areas computed, which
/// means that we can repeat the process in order to derive all the final
/// areas.
/// efficiently, we start with the "leaves" (the highest nodes), which have
/// no neighbors higher than them, hence no directed edges to them. We add
/// their area to themselves, and then to all neighbors that they flow into
/// (their "ancestors" in the flow graph); currently, this just means the
/// node immediately downhill of this node. As we go lower, we know that all
/// our "children" already had their areas computed, which means that we can
/// repeat the process in order to derive all the final areas.
/// 3. Now, iterate through the list in *order.* Whether we used the filling
/// method to compute a "filled" version of each depression, or used the lake
/// connection algorithm described in [1], each node is guaranteed to have
/// zero or one drainage edges out, representing the direction of water flow
/// for that node. For nodes i with zero drainage edges out (boundary nodes
/// and lake bottoms) we set the slope to 0 (so the change in altitude is
/// uplift(i))
/// For nodes with at least one drainage edge out, we take advantage of the
/// fact that we are computing new heights in order and rewrite our equation
/// as (letting j = downhill[i], A[i] be the computed area of point i,
/// p(i) be the x-y position of point i,
/// flux(i) = k * A[i]^m / ((p(i) - p(j)).magnitude()), and δt = 1):
/// uplift(i)).
///
/// For nodes with at least one drainage edge out, we take
/// advantage of the fact that we are computing new heights in order and
/// rewrite our equation as (letting j = downhill[i], A[i] be the computed
/// area of point i, p(i) be the x-y position of point i, flux(i) = k *
/// A[i]^m / ((p(i) - p(j)).magnitude()), and δt = 1):
///
/// h[i](t + dt) = h[i](t) + δt * (uplift[i] + flux(i) * h[j](t + δt)) / (1 +
/// flux(i) * δt).

View File

@ -53,8 +53,8 @@ pub fn map_edge_factor(map_size_lg: MapSizeLg, posi: usize) -> f32 {
/// At some point, we should probably contribute this back to stats-rs.
///
/// 1. [https://www.r-bloggers.com/sums-of-random-variables/][1],
/// 2. Sadooghi-Alvandi, S., A. Nematollahi, & R. Habibi, 2009.
/// On the Distribution of the Sum of Independent Uniform Random Variables.
/// 2. Sadooghi-Alvandi, S., A. Nematollahi, & R. Habibi, 2009. On the
/// Distribution of the Sum of Independent Uniform Random Variables.
/// Statistical Papers, 50, 171-175.
/// 3. [https://en.wikipedia.org/wiki/Cumulative_distribution_function][3]
///

View File

@ -31,7 +31,7 @@ impl<'a, R: Rng> NameGen<'a, R> {
]);
let mut middle = cons.clone();
middle.extend(vec!["tt"]);
let vowel = vec!["o", "e", "a", "i", "u", "au", "ee", "ow", "ay", "ey", "oe"];
let vowel = ["o", "e", "a", "i", "u", "au", "ee", "ow", "ay", "ey", "oe"];
let end = vec![
"et", "ige", "age", "ist", "en", "on", "og", "end", "ind", "ock", "een", "edge", "ist",
"ed", "est", "eed", "ast", "olt", "ey", "ean", "ead", "onk", "ink", "eon", "er", "ow",

View File

@ -641,7 +641,7 @@ impl Archetype for House {
% 6
{
0 => SpriteKind::HangingSign,
1 | 2 | 3 => SpriteKind::HangingBasket,
1..=3 => SpriteKind::HangingBasket,
4 => SpriteKind::WallSconce,
5 => SpriteKind::WallLampSmall,
_ => SpriteKind::DungeonWallDecor,

View File

@ -1414,7 +1414,7 @@ impl Land {
}
}
closed.into_iter().chain(open.into_iter()).collect()
closed.into_iter().chain(open).collect()
}
fn write_path(

Some files were not shown because too many files have changed in this diff Show More