2016-12-12 02:51:22 +00:00
|
|
|
|
use core::cell::RefCell;
|
2019-12-07 04:18:12 +00:00
|
|
|
|
use core::ptr::*;
|
2014-06-28 20:57:36 +00:00
|
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
|
fn test() {
|
|
|
|
|
unsafe {
|
|
|
|
|
struct Pair {
|
2015-03-26 00:06:52 +00:00
|
|
|
|
fst: isize,
|
2019-12-07 04:18:12 +00:00
|
|
|
|
snd: isize,
|
2014-06-28 20:57:36 +00:00
|
|
|
|
};
|
2019-12-07 04:18:12 +00:00
|
|
|
|
let mut p = Pair { fst: 10, snd: 20 };
|
2014-06-28 20:57:36 +00:00
|
|
|
|
let pptr: *mut Pair = &mut p;
|
2015-07-24 01:04:55 +00:00
|
|
|
|
let iptr: *mut isize = pptr as *mut isize;
|
2014-06-28 20:57:36 +00:00
|
|
|
|
assert_eq!(*iptr, 10);
|
|
|
|
|
*iptr = 30;
|
|
|
|
|
assert_eq!(*iptr, 30);
|
|
|
|
|
assert_eq!(p.fst, 30);
|
|
|
|
|
|
2019-12-07 04:18:12 +00:00
|
|
|
|
*pptr = Pair { fst: 50, snd: 60 };
|
2014-06-28 20:57:36 +00:00
|
|
|
|
assert_eq!(*iptr, 50);
|
|
|
|
|
assert_eq!(p.fst, 50);
|
|
|
|
|
assert_eq!(p.snd, 60);
|
|
|
|
|
|
|
|
|
|
let v0 = vec![32000u16, 32001u16, 32002u16];
|
|
|
|
|
let mut v1 = vec![0u16, 0u16, 0u16];
|
|
|
|
|
|
2015-03-27 18:12:28 +00:00
|
|
|
|
copy(v0.as_ptr().offset(1), v1.as_mut_ptr().offset(1), 1);
|
2019-12-07 04:18:12 +00:00
|
|
|
|
assert!((v1[0] == 0u16 && v1[1] == 32001u16 && v1[2] == 0u16));
|
2015-03-27 18:12:28 +00:00
|
|
|
|
copy(v0.as_ptr().offset(2), v1.as_mut_ptr(), 1);
|
2019-12-07 04:18:12 +00:00
|
|
|
|
assert!((v1[0] == 32002u16 && v1[1] == 32001u16 && v1[2] == 0u16));
|
2015-03-27 18:12:28 +00:00
|
|
|
|
copy(v0.as_ptr(), v1.as_mut_ptr().offset(2), 1);
|
2019-12-07 04:18:12 +00:00
|
|
|
|
assert!((v1[0] == 32002u16 && v1[1] == 32001u16 && v1[2] == 32000u16));
|
2014-06-28 20:57:36 +00:00
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
|
fn test_is_null() {
|
2015-03-26 00:06:52 +00:00
|
|
|
|
let p: *const isize = null();
|
2014-06-28 20:57:36 +00:00
|
|
|
|
assert!(p.is_null());
|
|
|
|
|
|
2019-02-17 11:58:48 +00:00
|
|
|
|
let q = p.wrapping_offset(1);
|
2014-06-28 20:57:36 +00:00
|
|
|
|
assert!(!q.is_null());
|
|
|
|
|
|
2015-03-26 00:06:52 +00:00
|
|
|
|
let mp: *mut isize = null_mut();
|
2014-06-28 20:57:36 +00:00
|
|
|
|
assert!(mp.is_null());
|
|
|
|
|
|
2019-02-17 11:58:48 +00:00
|
|
|
|
let mq = mp.wrapping_offset(1);
|
2014-06-28 20:57:36 +00:00
|
|
|
|
assert!(!mq.is_null());
|
2017-10-10 18:35:41 +00:00
|
|
|
|
|
|
|
|
|
// Pointers to unsized types -- slices
|
|
|
|
|
let s: &mut [u8] = &mut [1, 2, 3];
|
|
|
|
|
let cs: *const [u8] = s;
|
|
|
|
|
assert!(!cs.is_null());
|
|
|
|
|
|
|
|
|
|
let ms: *mut [u8] = s;
|
|
|
|
|
assert!(!ms.is_null());
|
|
|
|
|
|
|
|
|
|
let cz: *const [u8] = &[];
|
|
|
|
|
assert!(!cz.is_null());
|
|
|
|
|
|
|
|
|
|
let mz: *mut [u8] = &mut [];
|
|
|
|
|
assert!(!mz.is_null());
|
|
|
|
|
|
|
|
|
|
let ncs: *const [u8] = null::<[u8; 3]>();
|
|
|
|
|
assert!(ncs.is_null());
|
|
|
|
|
|
|
|
|
|
let nms: *mut [u8] = null_mut::<[u8; 3]>();
|
|
|
|
|
assert!(nms.is_null());
|
|
|
|
|
|
|
|
|
|
// Pointers to unsized types -- trait objects
|
2018-07-13 05:25:22 +00:00
|
|
|
|
let ci: *const dyn ToString = &3;
|
2017-10-10 18:35:41 +00:00
|
|
|
|
assert!(!ci.is_null());
|
|
|
|
|
|
2018-07-13 05:25:22 +00:00
|
|
|
|
let mi: *mut dyn ToString = &mut 3;
|
2017-10-10 18:35:41 +00:00
|
|
|
|
assert!(!mi.is_null());
|
|
|
|
|
|
2018-07-13 05:25:22 +00:00
|
|
|
|
let nci: *const dyn ToString = null::<isize>();
|
2017-10-10 18:35:41 +00:00
|
|
|
|
assert!(nci.is_null());
|
|
|
|
|
|
2018-07-13 05:25:22 +00:00
|
|
|
|
let nmi: *mut dyn ToString = null_mut::<isize>();
|
2017-10-10 18:35:41 +00:00
|
|
|
|
assert!(nmi.is_null());
|
2014-06-28 20:57:36 +00:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
#[test]
|
2014-08-20 20:06:34 +00:00
|
|
|
|
fn test_as_ref() {
|
2014-06-28 20:57:36 +00:00
|
|
|
|
unsafe {
|
2015-03-26 00:06:52 +00:00
|
|
|
|
let p: *const isize = null();
|
2014-08-20 20:06:34 +00:00
|
|
|
|
assert_eq!(p.as_ref(), None);
|
2014-06-28 20:57:36 +00:00
|
|
|
|
|
2015-03-26 00:06:52 +00:00
|
|
|
|
let q: *const isize = &2;
|
2014-08-20 20:06:34 +00:00
|
|
|
|
assert_eq!(q.as_ref().unwrap(), &2);
|
2014-06-28 20:57:36 +00:00
|
|
|
|
|
2015-03-26 00:06:52 +00:00
|
|
|
|
let p: *mut isize = null_mut();
|
2014-08-20 20:06:34 +00:00
|
|
|
|
assert_eq!(p.as_ref(), None);
|
2014-06-28 20:57:36 +00:00
|
|
|
|
|
2015-03-26 00:06:52 +00:00
|
|
|
|
let q: *mut isize = &mut 2;
|
2014-08-20 20:06:34 +00:00
|
|
|
|
assert_eq!(q.as_ref().unwrap(), &2);
|
|
|
|
|
|
|
|
|
|
// Lifetime inference
|
Add trivial cast lints.
This permits all coercions to be performed in casts, but adds lints to warn in those cases.
Part of this patch moves cast checking to a later stage of type checking. We acquire obligations to check casts as part of type checking where we previously checked them. Once we have type checked a function or module, then we check any cast obligations which have been acquired. That means we have more type information available to check casts (this was crucial to making coercions work properly in place of some casts), but it means that casts cannot feed input into type inference.
[breaking change]
* Adds two new lints for trivial casts and trivial numeric casts, these are warn by default, but can cause errors if you build with warnings as errors. Previously, trivial numeric casts and casts to trait objects were allowed.
* The unused casts lint has gone.
* Interactions between casting and type inference have changed in subtle ways. Two ways this might manifest are:
- You may need to 'direct' casts more with extra type information, for example, in some cases where `foo as _ as T` succeeded, you may now need to specify the type for `_`
- Casts do not influence inference of integer types. E.g., the following used to type check:
```
let x = 42;
let y = &x as *const u32;
```
Because the cast would inform inference that `x` must have type `u32`. This no longer applies and the compiler will fallback to `i32` for `x` and thus there will be a type error in the cast. The solution is to add more type information:
```
let x: u32 = 42;
let y = &x as *const u32;
```
2015-03-20 04:15:27 +00:00
|
|
|
|
let u = 2isize;
|
2014-08-20 20:06:34 +00:00
|
|
|
|
{
|
2015-03-26 00:06:52 +00:00
|
|
|
|
let p = &u as *const isize;
|
2014-08-20 20:06:34 +00:00
|
|
|
|
assert_eq!(p.as_ref().unwrap(), &2);
|
|
|
|
|
}
|
2017-09-29 19:36:32 +00:00
|
|
|
|
|
2017-09-30 21:45:18 +00:00
|
|
|
|
// Pointers to unsized types -- slices
|
2017-09-29 19:36:32 +00:00
|
|
|
|
let s: &mut [u8] = &mut [1, 2, 3];
|
|
|
|
|
let cs: *const [u8] = s;
|
|
|
|
|
assert_eq!(cs.as_ref(), Some(&*s));
|
|
|
|
|
|
|
|
|
|
let ms: *mut [u8] = s;
|
|
|
|
|
assert_eq!(ms.as_ref(), Some(&*s));
|
|
|
|
|
|
|
|
|
|
let cz: *const [u8] = &[];
|
|
|
|
|
assert_eq!(cz.as_ref(), Some(&[][..]));
|
|
|
|
|
|
|
|
|
|
let mz: *mut [u8] = &mut [];
|
|
|
|
|
assert_eq!(mz.as_ref(), Some(&[][..]));
|
|
|
|
|
|
2017-09-30 21:45:18 +00:00
|
|
|
|
let ncs: *const [u8] = null::<[u8; 3]>();
|
2017-09-29 19:36:32 +00:00
|
|
|
|
assert_eq!(ncs.as_ref(), None);
|
|
|
|
|
|
2017-09-30 21:45:18 +00:00
|
|
|
|
let nms: *mut [u8] = null_mut::<[u8; 3]>();
|
2017-09-29 19:36:32 +00:00
|
|
|
|
assert_eq!(nms.as_ref(), None);
|
2017-09-30 21:45:18 +00:00
|
|
|
|
|
|
|
|
|
// Pointers to unsized types -- trait objects
|
2018-07-13 05:25:22 +00:00
|
|
|
|
let ci: *const dyn ToString = &3;
|
2017-09-30 21:45:18 +00:00
|
|
|
|
assert!(ci.as_ref().is_some());
|
|
|
|
|
|
2018-07-13 05:25:22 +00:00
|
|
|
|
let mi: *mut dyn ToString = &mut 3;
|
2017-09-30 21:45:18 +00:00
|
|
|
|
assert!(mi.as_ref().is_some());
|
|
|
|
|
|
2018-07-13 05:25:22 +00:00
|
|
|
|
let nci: *const dyn ToString = null::<isize>();
|
2017-09-30 21:45:18 +00:00
|
|
|
|
assert!(nci.as_ref().is_none());
|
|
|
|
|
|
2018-07-13 05:25:22 +00:00
|
|
|
|
let nmi: *mut dyn ToString = null_mut::<isize>();
|
2017-09-30 21:45:18 +00:00
|
|
|
|
assert!(nmi.as_ref().is_none());
|
2014-08-20 20:06:34 +00:00
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
|
fn test_as_mut() {
|
|
|
|
|
unsafe {
|
2015-03-26 00:06:52 +00:00
|
|
|
|
let p: *mut isize = null_mut();
|
2014-08-20 20:06:34 +00:00
|
|
|
|
assert!(p.as_mut() == None);
|
|
|
|
|
|
2015-03-26 00:06:52 +00:00
|
|
|
|
let q: *mut isize = &mut 2;
|
2014-08-20 20:06:34 +00:00
|
|
|
|
assert!(q.as_mut().unwrap() == &mut 2);
|
|
|
|
|
|
|
|
|
|
// Lifetime inference
|
Add trivial cast lints.
This permits all coercions to be performed in casts, but adds lints to warn in those cases.
Part of this patch moves cast checking to a later stage of type checking. We acquire obligations to check casts as part of type checking where we previously checked them. Once we have type checked a function or module, then we check any cast obligations which have been acquired. That means we have more type information available to check casts (this was crucial to making coercions work properly in place of some casts), but it means that casts cannot feed input into type inference.
[breaking change]
* Adds two new lints for trivial casts and trivial numeric casts, these are warn by default, but can cause errors if you build with warnings as errors. Previously, trivial numeric casts and casts to trait objects were allowed.
* The unused casts lint has gone.
* Interactions between casting and type inference have changed in subtle ways. Two ways this might manifest are:
- You may need to 'direct' casts more with extra type information, for example, in some cases where `foo as _ as T` succeeded, you may now need to specify the type for `_`
- Casts do not influence inference of integer types. E.g., the following used to type check:
```
let x = 42;
let y = &x as *const u32;
```
Because the cast would inform inference that `x` must have type `u32`. This no longer applies and the compiler will fallback to `i32` for `x` and thus there will be a type error in the cast. The solution is to add more type information:
```
let x: u32 = 42;
let y = &x as *const u32;
```
2015-03-20 04:15:27 +00:00
|
|
|
|
let mut u = 2isize;
|
2014-08-20 20:06:34 +00:00
|
|
|
|
{
|
2015-03-26 00:06:52 +00:00
|
|
|
|
let p = &mut u as *mut isize;
|
2014-08-20 20:06:34 +00:00
|
|
|
|
assert!(p.as_mut().unwrap() == &mut 2);
|
|
|
|
|
}
|
2017-09-29 19:36:32 +00:00
|
|
|
|
|
2017-09-30 21:45:18 +00:00
|
|
|
|
// Pointers to unsized types -- slices
|
2017-09-29 19:36:32 +00:00
|
|
|
|
let s: &mut [u8] = &mut [1, 2, 3];
|
|
|
|
|
let ms: *mut [u8] = s;
|
2019-08-05 13:30:08 +00:00
|
|
|
|
assert_eq!(ms.as_mut(), Some(&mut [1, 2, 3][..]));
|
2017-09-29 19:36:32 +00:00
|
|
|
|
|
|
|
|
|
let mz: *mut [u8] = &mut [];
|
|
|
|
|
assert_eq!(mz.as_mut(), Some(&mut [][..]));
|
|
|
|
|
|
2017-09-30 21:45:18 +00:00
|
|
|
|
let nms: *mut [u8] = null_mut::<[u8; 3]>();
|
2017-09-29 19:36:32 +00:00
|
|
|
|
assert_eq!(nms.as_mut(), None);
|
2017-09-30 21:45:18 +00:00
|
|
|
|
|
|
|
|
|
// Pointers to unsized types -- trait objects
|
2018-07-13 05:25:22 +00:00
|
|
|
|
let mi: *mut dyn ToString = &mut 3;
|
2017-09-30 21:45:18 +00:00
|
|
|
|
assert!(mi.as_mut().is_some());
|
|
|
|
|
|
2018-07-13 05:25:22 +00:00
|
|
|
|
let nmi: *mut dyn ToString = null_mut::<isize>();
|
2017-09-30 21:45:18 +00:00
|
|
|
|
assert!(nmi.as_mut().is_none());
|
2014-06-28 20:57:36 +00:00
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
|
fn test_ptr_addition() {
|
|
|
|
|
unsafe {
|
2015-07-08 20:52:55 +00:00
|
|
|
|
let xs = vec![5; 16];
|
2014-06-28 20:57:36 +00:00
|
|
|
|
let mut ptr = xs.as_ptr();
|
|
|
|
|
let end = ptr.offset(16);
|
|
|
|
|
|
|
|
|
|
while ptr < end {
|
|
|
|
|
assert_eq!(*ptr, 5);
|
|
|
|
|
ptr = ptr.offset(1);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
let mut xs_mut = xs;
|
|
|
|
|
let mut m_ptr = xs_mut.as_mut_ptr();
|
|
|
|
|
let m_end = m_ptr.offset(16);
|
|
|
|
|
|
|
|
|
|
while m_ptr < m_end {
|
|
|
|
|
*m_ptr += 5;
|
|
|
|
|
m_ptr = m_ptr.offset(1);
|
|
|
|
|
}
|
|
|
|
|
|
2015-07-08 20:52:55 +00:00
|
|
|
|
assert!(xs_mut == vec![10; 16]);
|
2014-06-28 20:57:36 +00:00
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
|
fn test_ptr_subtraction() {
|
|
|
|
|
unsafe {
|
2019-12-07 04:18:12 +00:00
|
|
|
|
let xs = vec![0, 1, 2, 3, 4, 5, 6, 7, 8, 9];
|
2015-03-03 08:42:26 +00:00
|
|
|
|
let mut idx = 9;
|
2014-06-28 20:57:36 +00:00
|
|
|
|
let ptr = xs.as_ptr();
|
|
|
|
|
|
2015-03-03 08:42:26 +00:00
|
|
|
|
while idx >= 0 {
|
2015-03-26 00:06:52 +00:00
|
|
|
|
assert_eq!(*(ptr.offset(idx as isize)), idx as isize);
|
2015-03-03 08:42:26 +00:00
|
|
|
|
idx = idx - 1;
|
2014-06-28 20:57:36 +00:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
let mut xs_mut = xs;
|
|
|
|
|
let m_start = xs_mut.as_mut_ptr();
|
|
|
|
|
let mut m_ptr = m_start.offset(9);
|
|
|
|
|
|
2019-02-17 12:04:48 +00:00
|
|
|
|
loop {
|
2014-06-28 20:57:36 +00:00
|
|
|
|
*m_ptr += *m_ptr;
|
2019-02-17 12:04:48 +00:00
|
|
|
|
if m_ptr == m_start {
|
|
|
|
|
break;
|
|
|
|
|
}
|
2014-06-28 20:57:36 +00:00
|
|
|
|
m_ptr = m_ptr.offset(-1);
|
|
|
|
|
}
|
|
|
|
|
|
2019-12-07 04:18:12 +00:00
|
|
|
|
assert_eq!(xs_mut, [0, 2, 4, 6, 8, 10, 12, 14, 16, 18]);
|
2014-06-28 20:57:36 +00:00
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
|
fn test_set_memory() {
|
2014-12-30 08:19:41 +00:00
|
|
|
|
let mut xs = [0u8; 20];
|
2014-06-28 20:57:36 +00:00
|
|
|
|
let ptr = xs.as_mut_ptr();
|
2019-12-07 04:18:12 +00:00
|
|
|
|
unsafe {
|
|
|
|
|
write_bytes(ptr, 5u8, xs.len());
|
|
|
|
|
}
|
2014-12-30 08:19:41 +00:00
|
|
|
|
assert!(xs == [5u8; 20]);
|
2014-06-28 20:57:36 +00:00
|
|
|
|
}
|
2015-02-10 13:37:44 +00:00
|
|
|
|
|
|
|
|
|
#[test]
|
2017-12-22 18:29:16 +00:00
|
|
|
|
fn test_unsized_nonnull() {
|
2017-04-04 16:31:38 +00:00
|
|
|
|
let xs: &[i32] = &[1, 2, 3];
|
2017-12-22 18:29:16 +00:00
|
|
|
|
let ptr = unsafe { NonNull::new_unchecked(xs as *const [i32] as *mut [i32]) };
|
2017-04-04 16:31:38 +00:00
|
|
|
|
let ys = unsafe { ptr.as_ref() };
|
|
|
|
|
let zs: &[i32] = &[1, 2, 3];
|
2015-02-10 13:37:44 +00:00
|
|
|
|
assert!(ys == zs);
|
|
|
|
|
}
|
2016-07-16 21:15:15 +00:00
|
|
|
|
|
|
|
|
|
#[test]
|
2016-08-29 19:31:28 +00:00
|
|
|
|
#[allow(warnings)]
|
|
|
|
|
// Have a symbol for the test below. It doesn’t need to be an actual variadic function, match the
|
|
|
|
|
// ABI, or even point to an actual executable code, because the function itself is never invoked.
|
|
|
|
|
#[no_mangle]
|
|
|
|
|
pub fn test_variadic_fnptr() {
|
2016-07-16 21:15:15 +00:00
|
|
|
|
use core::hash::{Hash, SipHasher};
|
2019-12-07 04:18:12 +00:00
|
|
|
|
extern "C" {
|
2016-08-29 19:31:28 +00:00
|
|
|
|
fn test_variadic_fnptr(_: u64, ...) -> f64;
|
2016-07-16 21:15:15 +00:00
|
|
|
|
}
|
2019-12-07 04:18:12 +00:00
|
|
|
|
let p: unsafe extern "C" fn(u64, ...) -> f64 = test_variadic_fnptr;
|
2016-07-16 21:15:15 +00:00
|
|
|
|
let q = p.clone();
|
|
|
|
|
assert_eq!(p, q);
|
|
|
|
|
assert!(!(p < q));
|
|
|
|
|
let mut s = SipHasher::new();
|
|
|
|
|
assert_eq!(p.hash(&mut s), q.hash(&mut s));
|
|
|
|
|
}
|
2016-12-12 02:51:22 +00:00
|
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
|
fn write_unaligned_drop() {
|
|
|
|
|
thread_local! {
|
|
|
|
|
static DROPS: RefCell<Vec<u32>> = RefCell::new(Vec::new());
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
struct Dropper(u32);
|
|
|
|
|
|
|
|
|
|
impl Drop for Dropper {
|
|
|
|
|
fn drop(&mut self) {
|
|
|
|
|
DROPS.with(|d| d.borrow_mut().push(self.0));
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
{
|
|
|
|
|
let c = Dropper(0);
|
|
|
|
|
let mut t = Dropper(1);
|
2019-12-07 04:18:12 +00:00
|
|
|
|
unsafe {
|
|
|
|
|
write_unaligned(&mut t, c);
|
|
|
|
|
}
|
2016-12-12 02:51:22 +00:00
|
|
|
|
}
|
|
|
|
|
DROPS.with(|d| assert_eq!(*d.borrow(), [0]));
|
|
|
|
|
}
|
2018-04-29 14:09:56 +00:00
|
|
|
|
|
|
|
|
|
#[test]
|
2019-12-07 11:38:13 +00:00
|
|
|
|
#[cfg_attr(miri, ignore)] // Miri does not compute a maximal `mid` for `align_offset`
|
2018-04-29 14:09:56 +00:00
|
|
|
|
fn align_offset_zst() {
|
|
|
|
|
// For pointers of stride = 0, the pointer is already aligned or it cannot be aligned at
|
|
|
|
|
// all, because no amount of elements will align the pointer.
|
|
|
|
|
let mut p = 1;
|
|
|
|
|
while p < 1024 {
|
|
|
|
|
assert_eq!((p as *const ()).align_offset(p), 0);
|
|
|
|
|
if p != 1 {
|
|
|
|
|
assert_eq!(((p + 1) as *const ()).align_offset(p), !0);
|
|
|
|
|
}
|
|
|
|
|
p = (p + 1).next_power_of_two();
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
#[test]
|
2019-12-07 11:38:13 +00:00
|
|
|
|
#[cfg_attr(miri, ignore)] // Miri does not compute a maximal `mid` for `align_offset`
|
2018-04-29 14:09:56 +00:00
|
|
|
|
fn align_offset_stride1() {
|
|
|
|
|
// For pointers of stride = 1, the pointer can always be aligned. The offset is equal to
|
|
|
|
|
// number of bytes.
|
|
|
|
|
let mut align = 1;
|
|
|
|
|
while align < 1024 {
|
2019-12-07 04:18:12 +00:00
|
|
|
|
for ptr in 1..2 * align {
|
2018-04-29 14:09:56 +00:00
|
|
|
|
let expected = ptr % align;
|
|
|
|
|
let offset = if expected == 0 { 0 } else { align - expected };
|
2019-12-07 04:18:12 +00:00
|
|
|
|
assert_eq!(
|
|
|
|
|
(ptr as *const u8).align_offset(align),
|
|
|
|
|
offset,
|
|
|
|
|
"ptr = {}, align = {}, size = 1",
|
|
|
|
|
ptr,
|
|
|
|
|
align
|
|
|
|
|
);
|
2018-04-29 14:09:56 +00:00
|
|
|
|
}
|
2018-05-17 15:02:47 +00:00
|
|
|
|
align = (align + 1).next_power_of_two();
|
2018-04-29 14:09:56 +00:00
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
#[test]
|
2019-12-07 11:38:13 +00:00
|
|
|
|
#[cfg_attr(miri, ignore)] // Miri is too slow
|
2018-04-29 14:09:56 +00:00
|
|
|
|
fn align_offset_weird_strides() {
|
|
|
|
|
#[repr(packed)]
|
|
|
|
|
struct A3(u16, u8);
|
|
|
|
|
struct A4(u32);
|
|
|
|
|
#[repr(packed)]
|
|
|
|
|
struct A5(u32, u8);
|
|
|
|
|
#[repr(packed)]
|
|
|
|
|
struct A6(u32, u16);
|
|
|
|
|
#[repr(packed)]
|
|
|
|
|
struct A7(u32, u16, u8);
|
|
|
|
|
#[repr(packed)]
|
|
|
|
|
struct A8(u32, u32);
|
|
|
|
|
#[repr(packed)]
|
|
|
|
|
struct A9(u32, u32, u8);
|
|
|
|
|
#[repr(packed)]
|
|
|
|
|
struct A10(u32, u32, u16);
|
|
|
|
|
|
|
|
|
|
unsafe fn test_weird_stride<T>(ptr: *const T, align: usize) -> bool {
|
|
|
|
|
let numptr = ptr as usize;
|
|
|
|
|
let mut expected = usize::max_value();
|
|
|
|
|
// Naive but definitely correct way to find the *first* aligned element of stride::<T>.
|
|
|
|
|
for el in 0..align {
|
|
|
|
|
if (numptr + el * ::std::mem::size_of::<T>()) % align == 0 {
|
|
|
|
|
expected = el;
|
|
|
|
|
break;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
let got = ptr.align_offset(align);
|
|
|
|
|
if got != expected {
|
2019-12-07 04:18:12 +00:00
|
|
|
|
eprintln!(
|
|
|
|
|
"aligning {:p} (with stride of {}) to {}, expected {}, got {}",
|
|
|
|
|
ptr,
|
|
|
|
|
::std::mem::size_of::<T>(),
|
|
|
|
|
align,
|
|
|
|
|
expected,
|
|
|
|
|
got
|
|
|
|
|
);
|
2018-04-29 14:09:56 +00:00
|
|
|
|
return true;
|
|
|
|
|
}
|
|
|
|
|
return false;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
// For pointers of stride != 1, we verify the algorithm against the naivest possible
|
|
|
|
|
// implementation
|
|
|
|
|
let mut align = 1;
|
|
|
|
|
let mut x = false;
|
|
|
|
|
while align < 1024 {
|
2019-12-07 04:18:12 +00:00
|
|
|
|
for ptr in 1usize..4 * align {
|
2018-04-29 14:09:56 +00:00
|
|
|
|
unsafe {
|
|
|
|
|
x |= test_weird_stride::<A3>(ptr as *const A3, align);
|
|
|
|
|
x |= test_weird_stride::<A4>(ptr as *const A4, align);
|
|
|
|
|
x |= test_weird_stride::<A5>(ptr as *const A5, align);
|
|
|
|
|
x |= test_weird_stride::<A6>(ptr as *const A6, align);
|
|
|
|
|
x |= test_weird_stride::<A7>(ptr as *const A7, align);
|
|
|
|
|
x |= test_weird_stride::<A8>(ptr as *const A8, align);
|
|
|
|
|
x |= test_weird_stride::<A9>(ptr as *const A9, align);
|
|
|
|
|
x |= test_weird_stride::<A10>(ptr as *const A10, align);
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
align = (align + 1).next_power_of_two();
|
|
|
|
|
}
|
|
|
|
|
assert!(!x);
|
|
|
|
|
}
|