mirror of
https://github.com/rust-lang/rust.git
synced 2024-11-22 14:55:26 +00:00
581 lines
17 KiB
Rust
581 lines
17 KiB
Rust
#![feature(
|
|
core_intrinsics,
|
|
coroutines,
|
|
stmt_expr_attributes,
|
|
coroutine_trait,
|
|
repr_simd,
|
|
tuple_trait,
|
|
unboxed_closures
|
|
)]
|
|
#![allow(internal_features)]
|
|
|
|
#[cfg(target_arch = "x86_64")]
|
|
use std::arch::x86_64::*;
|
|
use std::hint::black_box;
|
|
use std::io::Write;
|
|
use std::ops::Coroutine;
|
|
|
|
fn main() {
|
|
println!("{:?}", std::env::args().collect::<Vec<_>>());
|
|
|
|
let mutex = std::sync::Mutex::new(());
|
|
let _guard = mutex.lock().unwrap();
|
|
|
|
let _ = ::std::iter::repeat('a' as u8).take(10).collect::<Vec<_>>();
|
|
let stderr = ::std::io::stderr();
|
|
let mut stderr = stderr.lock();
|
|
|
|
std::thread::spawn(move || {
|
|
println!("Hello from another thread!");
|
|
});
|
|
|
|
writeln!(stderr, "some {} text", "<unknown>").unwrap();
|
|
|
|
let _ = std::process::Command::new("true").env("c", "d").spawn();
|
|
|
|
println!("cargo:rustc-link-lib=z");
|
|
|
|
static ONCE: std::sync::Once = std::sync::Once::new();
|
|
ONCE.call_once(|| {});
|
|
|
|
let _eq = LoopState::Continue(()) == LoopState::Break(());
|
|
|
|
// Make sure ByValPair values with differently sized components are correctly passed
|
|
map(None::<(u8, Box<Instruction>)>);
|
|
|
|
println!("{}", 2.3f32.exp());
|
|
println!("{}", 2.3f32.exp2());
|
|
println!("{}", 2.3f32.abs());
|
|
println!("{}", 2.3f32.sqrt());
|
|
println!("{}", 2.3f32.floor());
|
|
println!("{}", 2.3f32.ceil());
|
|
println!("{}", 2.3f32.min(1.0));
|
|
println!("{}", 2.3f32.max(1.0));
|
|
println!("{}", 2.3f32.powi(2));
|
|
println!("{}", 2.3f32.log2());
|
|
assert_eq!(2.3f32.copysign(-1.0), -2.3f32);
|
|
println!("{}", 2.3f32.powf(2.0));
|
|
|
|
assert_eq!(i64::MAX.checked_mul(2), None);
|
|
|
|
assert_eq!(-128i8, (-128i8).saturating_sub(1));
|
|
assert_eq!(127i8, 127i8.saturating_sub(-128));
|
|
assert_eq!(-128i8, (-128i8).saturating_add(-128));
|
|
assert_eq!(127i8, 127i8.saturating_add(1));
|
|
|
|
assert_eq!(0b0000000000000000000000000010000010000000000000000000000000000000_0000000000100000000000000000000000001000000000000100000000000000u128.leading_zeros(), 26);
|
|
assert_eq!(0b0000000000000000000000000010000000000000000000000000000000000000_0000000000000000000000000000000000001000000000000000000010000000u128.trailing_zeros(), 7);
|
|
assert_eq!(
|
|
core::intrinsics::saturating_sub(0, -170141183460469231731687303715884105728i128),
|
|
170141183460469231731687303715884105727i128
|
|
);
|
|
|
|
std::hint::black_box(std::hint::black_box(7571400400375753350092698930310845914i128) * 10);
|
|
assert!(0i128.checked_div(2i128).is_some());
|
|
assert!(0u128.checked_div(2u128).is_some());
|
|
assert_eq!(1u128 + 2, 3);
|
|
|
|
assert_eq!(0b100010000000000000000000000000000u128 >> 10, 0b10001000000000000000000u128);
|
|
assert_eq!(0xFEDCBA987654321123456789ABCDEFu128 >> 64, 0xFEDCBA98765432u128);
|
|
assert_eq!(0xFEDCBA987654321123456789ABCDEFu128 as i128 >> 64, 0xFEDCBA98765432i128);
|
|
|
|
let tmp = 353985398u128;
|
|
assert_eq!(tmp * 932490u128, 330087843781020u128);
|
|
|
|
let tmp = -0x1234_5678_9ABC_DEF0i64;
|
|
assert_eq!(tmp as i128, -0x1234_5678_9ABC_DEF0i128);
|
|
|
|
// Check that all u/i128 <-> float casts work correctly.
|
|
let houndred_u128 = 100u128;
|
|
let houndred_i128 = 100i128;
|
|
let houndred_f32 = 100.0f32;
|
|
let houndred_f64 = 100.0f64;
|
|
assert_eq!(houndred_u128 as f32, 100.0);
|
|
assert_eq!(houndred_u128 as f64, 100.0);
|
|
assert_eq!(houndred_f32 as u128, 100);
|
|
assert_eq!(houndred_f64 as u128, 100);
|
|
assert_eq!(houndred_i128 as f32, 100.0);
|
|
assert_eq!(houndred_i128 as f64, 100.0);
|
|
assert_eq!(houndred_f32 as i128, 100);
|
|
assert_eq!(houndred_f64 as i128, 100);
|
|
assert_eq!(1u128.rotate_left(2), 4);
|
|
|
|
assert_eq!(black_box(f32::NAN) as i128, 0);
|
|
assert_eq!(black_box(f32::NAN) as u128, 0);
|
|
|
|
// Test signed 128bit comparing
|
|
let max = usize::MAX as i128;
|
|
if 100i128 < 0i128 || 100i128 > max {
|
|
panic!();
|
|
}
|
|
|
|
test_checked_mul();
|
|
|
|
let _a = 1u32 << 2u8;
|
|
|
|
let empty: [i32; 0] = [];
|
|
assert!(empty.is_sorted());
|
|
|
|
println!("{:?}", std::intrinsics::caller_location());
|
|
|
|
#[cfg(target_arch = "x86_64")]
|
|
unsafe {
|
|
test_simd();
|
|
}
|
|
|
|
Box::pin(
|
|
#[coroutine]
|
|
move |mut _task_context| {
|
|
yield ();
|
|
},
|
|
)
|
|
.as_mut()
|
|
.resume(0);
|
|
|
|
#[derive(Copy, Clone)]
|
|
enum Nums {
|
|
NegOne = -1,
|
|
}
|
|
|
|
let kind = Nums::NegOne;
|
|
assert_eq!(-1i128, kind as i128);
|
|
|
|
let options = [1u128];
|
|
match options[0] {
|
|
1 => (),
|
|
0 => loop {},
|
|
v => panic(v),
|
|
};
|
|
|
|
if black_box(false) {
|
|
// Based on https://github.com/rust-lang/rust/blob/2f320a224e827b400be25966755a621779f797cc/src/test/ui/debuginfo/debuginfo_with_uninhabitable_field_and_unsized.rs
|
|
let _ = Foo::<dyn Send>::new();
|
|
|
|
#[allow(dead_code)]
|
|
struct Foo<T: ?Sized> {
|
|
base: Never,
|
|
value: T,
|
|
}
|
|
|
|
impl<T: ?Sized> Foo<T> {
|
|
pub fn new() -> Box<Foo<T>> {
|
|
todo!()
|
|
}
|
|
}
|
|
|
|
enum Never {}
|
|
}
|
|
|
|
foo(I64X2([0, 0]));
|
|
|
|
transmute_wide_pointer();
|
|
|
|
rust_call_abi();
|
|
|
|
const fn no_str() -> Option<Box<str>> {
|
|
None
|
|
}
|
|
|
|
static STATIC_WITH_MAYBE_NESTED_BOX: &Option<Box<str>> = &no_str();
|
|
|
|
println!("{:?}", STATIC_WITH_MAYBE_NESTED_BOX);
|
|
}
|
|
|
|
fn panic(_: u128) {
|
|
panic!();
|
|
}
|
|
|
|
use std::mem::transmute;
|
|
|
|
#[cfg(target_pointer_width = "32")]
|
|
type TwoPtrs = i64;
|
|
#[cfg(target_pointer_width = "64")]
|
|
type TwoPtrs = i128;
|
|
|
|
fn transmute_wide_pointer() -> TwoPtrs {
|
|
unsafe { transmute::<_, TwoPtrs>("true !") }
|
|
}
|
|
|
|
extern "rust-call" fn rust_call_abi_callee<T: std::marker::Tuple>(_: T) {}
|
|
|
|
fn rust_call_abi() {
|
|
rust_call_abi_callee(());
|
|
rust_call_abi_callee((1, 2));
|
|
}
|
|
|
|
#[repr(simd)]
|
|
struct I64X2([i64; 2]);
|
|
|
|
#[allow(improper_ctypes_definitions)]
|
|
extern "C" fn foo(_a: I64X2) {}
|
|
|
|
#[cfg(target_arch = "x86_64")]
|
|
#[target_feature(enable = "sse4.2")]
|
|
#[cfg(not(jit))]
|
|
unsafe fn test_crc32() {
|
|
assert!(is_x86_feature_detected!("sse4.2"));
|
|
|
|
let a = 42u32;
|
|
let b = 0xdeadbeefu64;
|
|
|
|
assert_eq!(_mm_crc32_u8(a, b as u8), 4135334616);
|
|
assert_eq!(_mm_crc32_u16(a, b as u16), 1200687288);
|
|
assert_eq!(_mm_crc32_u32(a, b as u32), 2543798776);
|
|
assert_eq!(_mm_crc32_u64(a as u64, b as u64), 241952147);
|
|
}
|
|
|
|
#[cfg(target_arch = "x86_64")]
|
|
#[target_feature(enable = "sse2")]
|
|
unsafe fn test_simd() {
|
|
assert!(is_x86_feature_detected!("sse2"));
|
|
|
|
let x = _mm_setzero_si128();
|
|
let y = _mm_set1_epi16(7);
|
|
let or = _mm_or_si128(x, y);
|
|
let cmp_eq = _mm_cmpeq_epi8(y, y);
|
|
let cmp_lt = _mm_cmplt_epi8(y, y);
|
|
|
|
let (zero0, zero1) = std::mem::transmute::<_, (u64, u64)>(x);
|
|
assert_eq!((zero0, zero1), (0, 0));
|
|
assert_eq!(std::mem::transmute::<_, [u16; 8]>(or), [7, 7, 7, 7, 7, 7, 7, 7]);
|
|
assert_eq!(std::mem::transmute::<_, [u16; 8]>(cmp_eq), [
|
|
0xffff, 0xffff, 0xffff, 0xffff, 0xffff, 0xffff, 0xffff, 0xffff
|
|
]);
|
|
assert_eq!(std::mem::transmute::<_, [u16; 8]>(cmp_lt), [0, 0, 0, 0, 0, 0, 0, 0]);
|
|
|
|
test_mm_slli_si128();
|
|
test_mm_movemask_epi8();
|
|
test_mm256_movemask_epi8();
|
|
test_mm_add_epi8();
|
|
test_mm_add_pd();
|
|
test_mm_cvtepi8_epi16();
|
|
#[cfg(not(jit))]
|
|
test_mm_cvtps_epi32();
|
|
test_mm_cvttps_epi32();
|
|
test_mm_cvtsi128_si64();
|
|
|
|
test_mm_extract_epi8();
|
|
test_mm_insert_epi16();
|
|
test_mm_shuffle_epi8();
|
|
|
|
#[cfg(not(jit))]
|
|
test_mm_cmpestri();
|
|
|
|
test_mm256_shuffle_epi8();
|
|
test_mm256_permute2x128_si256();
|
|
test_mm256_permutevar8x32_epi32();
|
|
|
|
#[rustfmt::skip]
|
|
let mask1 = _mm_movemask_epi8(dbg!(_mm_setr_epi8(255u8 as i8, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0)));
|
|
assert_eq!(mask1, 1);
|
|
|
|
#[cfg(not(jit))]
|
|
test_crc32();
|
|
}
|
|
|
|
#[cfg(target_arch = "x86_64")]
|
|
#[target_feature(enable = "sse2")]
|
|
unsafe fn test_mm_slli_si128() {
|
|
#[rustfmt::skip]
|
|
let a = _mm_setr_epi8(
|
|
1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16,
|
|
);
|
|
let r = _mm_slli_si128(a, 1);
|
|
let e = _mm_setr_epi8(0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
|
|
assert_eq_m128i(r, e);
|
|
|
|
#[rustfmt::skip]
|
|
let a = _mm_setr_epi8(
|
|
1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16,
|
|
);
|
|
let r = _mm_slli_si128(a, 15);
|
|
let e = _mm_setr_epi8(0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1);
|
|
assert_eq_m128i(r, e);
|
|
|
|
#[rustfmt::skip]
|
|
let a = _mm_setr_epi8(
|
|
1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16,
|
|
);
|
|
let r = _mm_slli_si128(a, 16);
|
|
assert_eq_m128i(r, _mm_set1_epi8(0));
|
|
}
|
|
|
|
#[cfg(target_arch = "x86_64")]
|
|
#[target_feature(enable = "sse2")]
|
|
unsafe fn test_mm_movemask_epi8() {
|
|
#[rustfmt::skip]
|
|
let a = _mm_setr_epi8(
|
|
0b1000_0000u8 as i8, 0b0, 0b1000_0000u8 as i8, 0b01,
|
|
0b0101, 0b1111_0000u8 as i8, 0, 0,
|
|
0, 0, 0b1111_0000u8 as i8, 0b0101,
|
|
0b01, 0b1000_0000u8 as i8, 0b0, 0b1000_0000u8 as i8,
|
|
);
|
|
let r = _mm_movemask_epi8(a);
|
|
assert_eq!(r, 0b10100100_00100101);
|
|
}
|
|
|
|
#[cfg(target_arch = "x86_64")]
|
|
#[target_feature(enable = "avx2")]
|
|
unsafe fn test_mm256_movemask_epi8() {
|
|
let a = _mm256_set1_epi8(-1);
|
|
let r = _mm256_movemask_epi8(a);
|
|
let e = -1;
|
|
assert_eq!(r, e);
|
|
}
|
|
|
|
#[cfg(target_arch = "x86_64")]
|
|
#[target_feature(enable = "sse2")]
|
|
unsafe fn test_mm_add_epi8() {
|
|
let a = _mm_setr_epi8(0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
|
|
#[rustfmt::skip]
|
|
let b = _mm_setr_epi8(
|
|
16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31,
|
|
);
|
|
let r = _mm_add_epi8(a, b);
|
|
#[rustfmt::skip]
|
|
let e = _mm_setr_epi8(
|
|
16, 18, 20, 22, 24, 26, 28, 30, 32, 34, 36, 38, 40, 42, 44, 46,
|
|
);
|
|
assert_eq_m128i(r, e);
|
|
}
|
|
|
|
#[cfg(target_arch = "x86_64")]
|
|
#[target_feature(enable = "sse2")]
|
|
unsafe fn test_mm_add_pd() {
|
|
let a = _mm_setr_pd(1.0, 2.0);
|
|
let b = _mm_setr_pd(5.0, 10.0);
|
|
let r = _mm_add_pd(a, b);
|
|
assert_eq_m128d(r, _mm_setr_pd(6.0, 12.0));
|
|
}
|
|
|
|
#[cfg(target_arch = "x86_64")]
|
|
fn assert_eq_m128i(x: std::arch::x86_64::__m128i, y: std::arch::x86_64::__m128i) {
|
|
unsafe {
|
|
assert_eq!(std::mem::transmute::<_, [u8; 16]>(x), std::mem::transmute::<_, [u8; 16]>(y));
|
|
}
|
|
}
|
|
|
|
#[cfg(target_arch = "x86_64")]
|
|
#[target_feature(enable = "sse2")]
|
|
pub unsafe fn assert_eq_m128d(a: __m128d, b: __m128d) {
|
|
if _mm_movemask_pd(_mm_cmpeq_pd(a, b)) != 0b11 {
|
|
panic!("{:?} != {:?}", a, b);
|
|
}
|
|
}
|
|
|
|
#[cfg(target_arch = "x86_64")]
|
|
#[target_feature(enable = "avx")]
|
|
pub unsafe fn assert_eq_m256i(a: __m256i, b: __m256i) {
|
|
assert_eq!(std::mem::transmute::<_, [u64; 4]>(a), std::mem::transmute::<_, [u64; 4]>(b))
|
|
}
|
|
|
|
#[cfg(target_arch = "x86_64")]
|
|
#[target_feature(enable = "sse2")]
|
|
unsafe fn test_mm_cvtsi128_si64() {
|
|
let r = _mm_cvtsi128_si64(std::mem::transmute::<[i64; 2], _>([5, 0]));
|
|
assert_eq!(r, 5);
|
|
}
|
|
|
|
#[cfg(target_arch = "x86_64")]
|
|
#[target_feature(enable = "sse4.1")]
|
|
unsafe fn test_mm_cvtepi8_epi16() {
|
|
let a = _mm_set1_epi8(10);
|
|
let r = _mm_cvtepi8_epi16(a);
|
|
let e = _mm_set1_epi16(10);
|
|
assert_eq_m128i(r, e);
|
|
let a = _mm_set1_epi8(-10);
|
|
let r = _mm_cvtepi8_epi16(a);
|
|
let e = _mm_set1_epi16(-10);
|
|
assert_eq_m128i(r, e);
|
|
}
|
|
|
|
#[cfg(target_arch = "x86_64")]
|
|
#[target_feature(enable = "sse4.1")]
|
|
unsafe fn test_mm_extract_epi8() {
|
|
#[rustfmt::skip]
|
|
let a = _mm_setr_epi8(
|
|
-1, 1, 2, 3, 4, 5, 6, 7,
|
|
8, 9, 10, 11, 12, 13, 14, 15
|
|
);
|
|
let r1 = _mm_extract_epi8(a, 0);
|
|
let r2 = _mm_extract_epi8(a, 3);
|
|
assert_eq!(r1, 0xFF);
|
|
assert_eq!(r2, 3);
|
|
}
|
|
|
|
#[cfg(target_arch = "x86_64")]
|
|
#[target_feature(enable = "sse2")]
|
|
unsafe fn test_mm_insert_epi16() {
|
|
let a = _mm_setr_epi16(0, 1, 2, 3, 4, 5, 6, 7);
|
|
let r = _mm_insert_epi16(a, 9, 0);
|
|
let e = _mm_setr_epi16(9, 1, 2, 3, 4, 5, 6, 7);
|
|
assert_eq_m128i(r, e);
|
|
}
|
|
|
|
#[cfg(target_arch = "x86_64")]
|
|
#[target_feature(enable = "ssse3")]
|
|
unsafe fn test_mm_shuffle_epi8() {
|
|
#[rustfmt::skip]
|
|
let a = _mm_setr_epi8(
|
|
1, 2, 3, 4, 5, 6, 7, 8,
|
|
9, 10, 11, 12, 13, 14, 15, 16,
|
|
);
|
|
#[rustfmt::skip]
|
|
let b = _mm_setr_epi8(
|
|
4, 128_u8 as i8, 4, 3,
|
|
24, 12, 6, 19,
|
|
12, 5, 5, 10,
|
|
4, 1, 8, 0,
|
|
);
|
|
let expected = _mm_setr_epi8(5, 0, 5, 4, 9, 13, 7, 4, 13, 6, 6, 11, 5, 2, 9, 1);
|
|
let r = _mm_shuffle_epi8(a, b);
|
|
assert_eq_m128i(r, expected);
|
|
}
|
|
|
|
// Currently one cannot `load` a &[u8] that is less than 16
|
|
// in length. This makes loading strings less than 16 in length
|
|
// a bit difficult. Rather than `load` and mutate the __m128i,
|
|
// it is easier to memcpy the given string to a local slice with
|
|
// length 16 and `load` the local slice.
|
|
#[cfg(not(jit))]
|
|
#[cfg(target_arch = "x86_64")]
|
|
#[target_feature(enable = "sse4.2")]
|
|
unsafe fn str_to_m128i(s: &[u8]) -> __m128i {
|
|
assert!(s.len() <= 16);
|
|
let slice = &mut [0u8; 16];
|
|
std::ptr::copy_nonoverlapping(s.as_ptr(), slice.as_mut_ptr(), s.len());
|
|
_mm_loadu_si128(slice.as_ptr() as *const _)
|
|
}
|
|
|
|
#[cfg(not(jit))]
|
|
#[cfg(target_arch = "x86_64")]
|
|
#[target_feature(enable = "sse4.2")]
|
|
unsafe fn test_mm_cmpestri() {
|
|
let a = str_to_m128i(b"bar - garbage");
|
|
let b = str_to_m128i(b"foobar");
|
|
let i = _mm_cmpestri::<_SIDD_CMP_EQUAL_ORDERED>(a, 3, b, 6);
|
|
assert_eq!(3, i);
|
|
}
|
|
|
|
#[cfg(target_arch = "x86_64")]
|
|
#[target_feature(enable = "avx2")]
|
|
unsafe fn test_mm256_shuffle_epi8() {
|
|
#[rustfmt::skip]
|
|
let a = _mm256_setr_epi8(
|
|
1, 2, 3, 4, 5, 6, 7, 8,
|
|
9, 10, 11, 12, 13, 14, 15, 16,
|
|
17, 18, 19, 20, 21, 22, 23, 24,
|
|
25, 26, 27, 28, 29, 30, 31, 32,
|
|
);
|
|
#[rustfmt::skip]
|
|
let b = _mm256_setr_epi8(
|
|
4, 128u8 as i8, 4, 3, 24, 12, 6, 19,
|
|
12, 5, 5, 10, 4, 1, 8, 0,
|
|
4, 128u8 as i8, 4, 3, 24, 12, 6, 19,
|
|
12, 5, 5, 10, 4, 1, 8, 0,
|
|
);
|
|
#[rustfmt::skip]
|
|
let expected = _mm256_setr_epi8(
|
|
5, 0, 5, 4, 9, 13, 7, 4,
|
|
13, 6, 6, 11, 5, 2, 9, 1,
|
|
21, 0, 21, 20, 25, 29, 23, 20,
|
|
29, 22, 22, 27, 21, 18, 25, 17,
|
|
);
|
|
let r = _mm256_shuffle_epi8(a, b);
|
|
assert_eq_m256i(r, expected);
|
|
}
|
|
|
|
#[cfg(target_arch = "x86_64")]
|
|
#[target_feature(enable = "avx2")]
|
|
unsafe fn test_mm256_permute2x128_si256() {
|
|
let a = _mm256_setr_epi64x(100, 200, 500, 600);
|
|
let b = _mm256_setr_epi64x(300, 400, 700, 800);
|
|
let r = _mm256_permute2x128_si256::<0b00_01_00_11>(a, b);
|
|
let e = _mm256_setr_epi64x(700, 800, 500, 600);
|
|
assert_eq_m256i(r, e);
|
|
}
|
|
|
|
#[cfg(target_arch = "x86_64")]
|
|
#[target_feature(enable = "avx2")]
|
|
unsafe fn test_mm256_permutevar8x32_epi32() {
|
|
let a = _mm256_setr_epi32(100, 200, 300, 400, 500, 600, 700, 800);
|
|
let idx = _mm256_setr_epi32(7, 6, 5, 4, 3, 2, 1, 0);
|
|
let r = _mm256_setr_epi32(800, 700, 600, 500, 400, 300, 200, 100);
|
|
let e = _mm256_permutevar8x32_epi32(a, idx);
|
|
assert_eq_m256i(r, e);
|
|
}
|
|
|
|
#[cfg(target_arch = "x86_64")]
|
|
#[target_feature(enable = "avx2")]
|
|
#[cfg(not(jit))]
|
|
unsafe fn test_mm_cvtps_epi32() {
|
|
let floats: [f32; 4] = [1.5, -2.5, i32::MAX as f32 + 1.0, f32::NAN];
|
|
|
|
let float_vec = _mm_loadu_ps(floats.as_ptr());
|
|
let int_vec = _mm_cvtps_epi32(float_vec);
|
|
|
|
let mut ints: [i32; 4] = [0; 4];
|
|
_mm_storeu_si128(ints.as_mut_ptr() as *mut __m128i, int_vec);
|
|
|
|
// this is very different from `floats.map(|f| f as i32)`!
|
|
let expected_ints: [i32; 4] = [2, -2, i32::MIN, i32::MIN];
|
|
|
|
assert_eq!(ints, expected_ints);
|
|
}
|
|
|
|
#[cfg(target_arch = "x86_64")]
|
|
#[target_feature(enable = "avx2")]
|
|
unsafe fn test_mm_cvttps_epi32() {
|
|
let floats: [f32; 4] = [1.5, -2.5, i32::MAX as f32 + 1.0, f32::NAN];
|
|
|
|
let float_vec = _mm_loadu_ps(floats.as_ptr());
|
|
let int_vec = _mm_cvttps_epi32(float_vec);
|
|
|
|
let mut ints: [i32; 4] = [0; 4];
|
|
_mm_storeu_si128(ints.as_mut_ptr() as *mut __m128i, int_vec);
|
|
|
|
// this is very different from `floats.map(|f| f as i32)`!
|
|
let expected_ints: [i32; 4] = [1, -2, i32::MIN, i32::MIN];
|
|
|
|
assert_eq!(ints, expected_ints);
|
|
}
|
|
|
|
fn test_checked_mul() {
|
|
let u: Option<u8> = u8::from_str_radix("1000", 10).ok();
|
|
assert_eq!(u, None);
|
|
|
|
assert_eq!(1u8.checked_mul(255u8), Some(255u8));
|
|
assert_eq!(255u8.checked_mul(255u8), None);
|
|
assert_eq!(1i8.checked_mul(127i8), Some(127i8));
|
|
assert_eq!(127i8.checked_mul(127i8), None);
|
|
assert_eq!((-1i8).checked_mul(-127i8), Some(127i8));
|
|
assert_eq!(1i8.checked_mul(-128i8), Some(-128i8));
|
|
assert_eq!((-128i8).checked_mul(-128i8), None);
|
|
|
|
assert_eq!(1u64.checked_mul(u64::MAX), Some(u64::MAX));
|
|
assert_eq!(u64::MAX.checked_mul(u64::MAX), None);
|
|
assert_eq!(1i64.checked_mul(i64::MAX), Some(i64::MAX));
|
|
assert_eq!(i64::MAX.checked_mul(i64::MAX), None);
|
|
assert_eq!((-1i64).checked_mul(i64::MIN + 1), Some(i64::MAX));
|
|
assert_eq!(1i64.checked_mul(i64::MIN), Some(i64::MIN));
|
|
assert_eq!(i64::MIN.checked_mul(i64::MIN), None);
|
|
}
|
|
|
|
#[derive(PartialEq)]
|
|
enum LoopState {
|
|
Continue(()),
|
|
Break(()),
|
|
}
|
|
|
|
pub enum Instruction {
|
|
Increment,
|
|
Loop,
|
|
}
|
|
|
|
fn map(a: Option<(u8, Box<Instruction>)>) -> Option<Box<Instruction>> {
|
|
match a {
|
|
None => None,
|
|
Some((_, instr)) => Some(instr),
|
|
}
|
|
}
|