|
| 1 | +// compile-flags: -O |
| 2 | +// min-llvm-version: 15.0 (because we're using opaque pointers) |
| 3 | +// ignore-debug (debug assertions in `slice::from_raw_parts` block optimizations) |
| 4 | + |
| 5 | +#![crate_type = "lib"] |
| 6 | + |
| 7 | +// CHECK-LABEL: @align8 |
| 8 | +#[no_mangle] |
| 9 | +pub fn align8(p: *const u8) -> bool { |
| 10 | + // CHECK: ret i1 true |
| 11 | + p.align_offset(8) < 8 |
| 12 | +} |
| 13 | + |
| 14 | +#[repr(align(4))] |
| 15 | +pub struct Align4([u8; 4]); |
| 16 | + |
| 17 | +// CHECK-LABEL: @align_to4 |
| 18 | +#[no_mangle] |
| 19 | +pub fn align_to4(x: &[u8]) -> bool { |
| 20 | + // CHECK: ret i1 true |
| 21 | + let (prefix, _middle, suffix) = unsafe { x.align_to::<Align4>() }; |
| 22 | + prefix.len() < 4 && suffix.len() < 4 |
| 23 | +} |
| 24 | + |
| 25 | +// CHECK-LABEL: @align_offset_byte_ptr(ptr{{.+}}%ptr) |
| 26 | +#[no_mangle] |
| 27 | +pub fn align_offset_byte_ptr(ptr: *const u8) -> usize { |
| 28 | + // CHECK: %[[ADDR:.+]] = ptrtoint ptr %ptr to [[USIZE:i[0-9]+]] |
| 29 | + // CHECK: %[[UP:.+]] = add [[USIZE]] %[[ADDR]], 31 |
| 30 | + // CHECK: %[[ALIGNED:.+]] = and [[USIZE]] %[[UP]], -32 |
| 31 | + // CHECK: %[[OFFSET:.+]] = sub [[USIZE]] %[[ALIGNED]], %[[ADDR]] |
| 32 | + |
| 33 | + // Since we're offsetting a byte pointer, there's no further fixups |
| 34 | + // CHECK-NOT: shr |
| 35 | + // CHECK-NOT: div |
| 36 | + // CHECK-NOT: select |
| 37 | + |
| 38 | + // CHECK: ret [[USIZE]] %[[OFFSET]] |
| 39 | + ptr.align_offset(32) |
| 40 | +} |
| 41 | + |
| 42 | +// CHECK-LABEL: @align_offset_word_slice(ptr{{.+}}align 4{{.+}}%slice.0 |
| 43 | +#[no_mangle] |
| 44 | +pub fn align_offset_word_slice(slice: &[Align4]) -> usize { |
| 45 | + // CHECK: %[[ADDR:.+]] = ptrtoint ptr %slice.0 to [[USIZE]] |
| 46 | + // CHECK: %[[UP:.+]] = add [[USIZE]] %[[ADDR]], 31 |
| 47 | + // CHECK: %[[ALIGNED:.+]] = and [[USIZE]] %[[UP]], -32 |
| 48 | + // CHECK: %[[BOFFSET:.+]] = sub [[USIZE]] %[[ALIGNED]], %[[ADDR]] |
| 49 | + // CHECK: %[[OFFSET:.+]] = lshr exact [[USIZE]] %[[BOFFSET]], 2 |
| 50 | + |
| 51 | + // Slices are known to be aligned, so we don't need the "maybe -1" path |
| 52 | + // CHECK-NOT: select |
| 53 | + |
| 54 | + // CHECK: ret [[USIZE]] %[[OFFSET]] |
| 55 | + slice.as_ptr().align_offset(32) |
| 56 | +} |
| 57 | + |
| 58 | + |
| 59 | +// CHECK-LABEL: @align_offset_word_ptr(ptr{{.+}}%ptr |
| 60 | +#[no_mangle] |
| 61 | +pub fn align_offset_word_ptr(ptr: *const Align4) -> usize { |
| 62 | + // CHECK: %[[ADDR:.+]] = ptrtoint ptr %ptr to [[USIZE]] |
| 63 | + // CHECK: %[[UP:.+]] = add [[USIZE]] %[[ADDR]], 31 |
| 64 | + // CHECK: %[[ALIGNED:.+]] = and [[USIZE]] %[[UP]], -32 |
| 65 | + // CHECK: %[[BOFFSET:.+]] = sub [[USIZE]] %[[ALIGNED]], %[[ADDR]] |
| 66 | + |
| 67 | + // While we can always get a *byte* offset that will work, if the original |
| 68 | + // pointer is unaligned it might be impossible to return an *element* offset |
| 69 | + // that will make it aligned. We want it to be a `select`, not a `br`, so |
| 70 | + // that the assembly will be branchless. |
| 71 | + // CHECK: %[[LOW:.+]] = and [[USIZE]] %[[ADDR]], 3 |
| 72 | + // CHECK: %[[ORIGINAL_ALIGNED:.+]] = icmp eq [[USIZE]] %[[LOW]], 0 |
| 73 | + // CHECK: %[[OFFSET:.+]] = lshr exact [[USIZE]] %[[BOFFSET]], 2 |
| 74 | + // CHECK: %[[R:.+]] = select i1 %[[ORIGINAL_ALIGNED]], [[USIZE]] %[[OFFSET]], [[USIZE]] -1 |
| 75 | + |
| 76 | + // CHECK: ret [[USIZE]] %[[R]] |
| 77 | + ptr.align_offset(32) |
| 78 | +} |
0 commit comments