|
| 1 | +use crate::{hint, intrinsics, mem, ptr}; |
| 2 | + |
| 3 | +//#[rustc_const_stable_indirect] |
| 4 | +//#[rustc_allow_const_fn_unstable(const_eval_select)] |
| 5 | +#[rustc_const_unstable(feature = "const_swap_nonoverlapping", issue = "133668")] |
| 6 | +#[inline] |
| 7 | +pub(crate) const unsafe fn swap_nonoverlapping<T>(x: *mut T, y: *mut T, count: usize) { |
| 8 | + intrinsics::const_eval_select!( |
| 9 | + @capture[T] { x: *mut T, y: *mut T, count: usize }: |
| 10 | + if const { |
| 11 | + // At compile-time we want to always copy this in chunks of `T`, to ensure that if there |
| 12 | + // are pointers inside `T` we will copy them in one go rather than trying to copy a part |
| 13 | + // of a pointer (which would not work). |
| 14 | + // SAFETY: Same preconditions as this function |
| 15 | + unsafe { swap_nonoverlapping_const(x, y, count) } |
| 16 | + } else { |
| 17 | + // At runtime we want to make sure not to swap byte-for-byte for types like [u8; 15], |
| 18 | + // and swapping as `MaybeUninit<T>` doesn't actually work as untyped for things like |
| 19 | + // T = (u16, u8), so we type-erase to raw bytes and swap that way. |
| 20 | + // SAFETY: Same preconditions as this function |
| 21 | + unsafe { swap_nonoverlapping_runtime(x, y, count) } |
| 22 | + } |
| 23 | + ) |
| 24 | +} |
| 25 | + |
| 26 | +/// Same behavior and safety conditions as [`swap_nonoverlapping`] |
| 27 | +#[rustc_const_stable_indirect] |
| 28 | +#[inline] |
| 29 | +const unsafe fn swap_nonoverlapping_const<T>(x: *mut T, y: *mut T, count: usize) { |
| 30 | + let x = x.cast::<mem::MaybeUninit<T>>(); |
| 31 | + let y = y.cast::<mem::MaybeUninit<T>>(); |
| 32 | + let mut i = 0; |
| 33 | + while i < count { |
| 34 | + // SAFETY: By precondition, `i` is in-bounds because it's below `n` |
| 35 | + // and because the two input ranges are non-overlapping and read/writeable, |
| 36 | + // these individual items inside them are too. |
| 37 | + unsafe { |
| 38 | + intrinsics::untyped_swap_nonoverlapping::<T>(x.add(i), y.add(i)); |
| 39 | + } |
| 40 | + |
| 41 | + i += 1; |
| 42 | + } |
| 43 | +} |
| 44 | + |
| 45 | +// Scale the monomorphizations with the size of the machine, roughly. |
| 46 | +const MAX_ALIGN: usize = align_of::<usize>().pow(2); |
| 47 | + |
| 48 | +/// Same behavior and safety conditions as [`swap_nonoverlapping`] |
| 49 | +#[inline] |
| 50 | +unsafe fn swap_nonoverlapping_runtime<T>(x: *mut T, y: *mut T, count: usize) { |
| 51 | + let bytes = { |
| 52 | + let slice = ptr::slice_from_raw_parts(x, count); |
| 53 | + unsafe { mem::size_of_val_raw(slice) } |
| 54 | + }; |
| 55 | + |
| 56 | + // Generating *untyped* loops for every type is silly, so we polymorphize away |
| 57 | + // the actual type, but we want to take advantage of alignment if possible, |
| 58 | + // so monomorphize for a restricted set of possible alignments. |
| 59 | + macro_rules! delegate_by_alignment { |
| 60 | + ($($p:pat => $align:expr,)+) => {{ |
| 61 | + #![allow(unreachable_patterns)] |
| 62 | + match const { align_of::<T>() } { |
| 63 | + $( |
| 64 | + $p => { |
| 65 | + swap_nonoverlapping_bytes::<$align>(x.cast(), y.cast(), bytes); |
| 66 | + } |
| 67 | + )+ |
| 68 | + } |
| 69 | + }}; |
| 70 | + } |
| 71 | + |
| 72 | + // SAFETY: |
| 73 | + unsafe { |
| 74 | + delegate_by_alignment! { |
| 75 | + MAX_ALIGN.. => MAX_ALIGN, |
| 76 | + 64.. => 64, |
| 77 | + 32.. => 32, |
| 78 | + 16.. => 16, |
| 79 | + 8.. => 8, |
| 80 | + 4.. => 4, |
| 81 | + 2.. => 2, |
| 82 | + _ => 1, |
| 83 | + } |
| 84 | + } |
| 85 | +} |
| 86 | + |
| 87 | +/// # Safety: |
| 88 | +/// - `x` and `y` must be aligned to `ALIGN` |
| 89 | +/// - `bytes` must be a multiple of `ALIGN` |
| 90 | +/// - They must be readable, writable, and non-overlapping for `bytes` bytes |
| 91 | +#[inline] |
| 92 | +unsafe fn swap_nonoverlapping_bytes<const ALIGN: usize>( |
| 93 | + x: *mut mem::MaybeUninit<u8>, |
| 94 | + y: *mut mem::MaybeUninit<u8>, |
| 95 | + bytes: usize, |
| 96 | +) { |
| 97 | + // SAFETY: Two legal non-overlapping regions can't be bigger than this. |
| 98 | + // (And they couldn't have made allocations any bigger either anyway.) |
| 99 | + // FIXME: Would be nice to have a type for this instead of the assume. |
| 100 | + unsafe { hint::assert_unchecked(bytes < isize::MAX as usize) }; |
| 101 | + |
| 102 | + let mut i = 0; |
| 103 | + macro_rules! swap_next_n { |
| 104 | + ($n:expr) => {{ |
| 105 | + let x: *mut mem::MaybeUninit<[u8; $n]> = x.add(i).cast(); |
| 106 | + let y: *mut mem::MaybeUninit<[u8; $n]> = y.add(i).cast(); |
| 107 | + swap_nonoverlapping_aligned_chunk::<ALIGN, [u8; $n]>( |
| 108 | + x.as_mut_unchecked(), |
| 109 | + y.as_mut_unchecked(), |
| 110 | + ); |
| 111 | + i += $n; |
| 112 | + }}; |
| 113 | + } |
| 114 | + |
| 115 | + while bytes - i >= MAX_ALIGN { |
| 116 | + unsafe { |
| 117 | + swap_next_n!(MAX_ALIGN); |
| 118 | + } |
| 119 | + } |
| 120 | + |
| 121 | + macro_rules! handle_tail { |
| 122 | + ($($n:literal)+) => {$( |
| 123 | + if const { $n % ALIGN == 0 } { |
| 124 | + // Checking this way simplifies the block end to just add+test, |
| 125 | + // rather than needing extra math before the check. |
| 126 | + if (bytes & $n) != 0 { |
| 127 | + unsafe { |
| 128 | + swap_next_n!($n); |
| 129 | + } |
| 130 | + } |
| 131 | + } |
| 132 | + )+}; |
| 133 | + } |
| 134 | + const { assert!(MAX_ALIGN <= 64) }; |
| 135 | + handle_tail!(32 16 8 4 2 1); |
| 136 | + |
| 137 | + debug_assert_eq!(i, bytes); |
| 138 | +} |
| 139 | + |
| 140 | +// Don't let MIR inline this, because we really want it to keep its noalias metadata |
| 141 | +#[rustc_no_mir_inline] |
| 142 | +#[inline] |
| 143 | +unsafe fn swap_nonoverlapping_aligned_chunk<const ALIGN: usize, C: Copy>( |
| 144 | + x: &mut mem::MaybeUninit<C>, |
| 145 | + y: &mut mem::MaybeUninit<C>, |
| 146 | +) { |
| 147 | + assert!(size_of::<C>() % ALIGN == 0); |
| 148 | + |
| 149 | + let x = ptr::from_mut(x); |
| 150 | + let y = ptr::from_mut(y); |
| 151 | + |
| 152 | + unsafe { |
| 153 | + hint::assert_unchecked(x.is_aligned_to(ALIGN)); |
| 154 | + hint::assert_unchecked(y.is_aligned_to(ALIGN)); |
| 155 | + } |
| 156 | + |
| 157 | + unsafe { |
| 158 | + intrinsics::untyped_swap_nonoverlapping::<C>(x, y); |
| 159 | + } |
| 160 | +} |
0 commit comments