Split out SIMD into a separate crate
This commit is contained in:
parent
eb9b4a2a06
commit
0b0c913332
|
@ -516,6 +516,7 @@ dependencies = [
|
|||
"euclid 0.19.4 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||
"lyon_geom 0.12.2 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||
"lyon_path 0.12.0 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||
"pathfinder_simd 0.3.0",
|
||||
"serde 1.0.84 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||
"serde_derive 1.0.84 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||
"smallvec 0.6.8 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||
|
@ -551,11 +552,16 @@ dependencies = [
|
|||
"fixedbitset 0.1.9 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||
"hashbrown 0.1.7 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||
"pathfinder_geometry 0.3.0",
|
||||
"pathfinder_simd 0.3.0",
|
||||
"quickcheck 0.7.2 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||
"rayon 1.0.3 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||
"smallvec 0.6.8 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "pathfinder_simd"
|
||||
version = "0.3.0"
|
||||
|
||||
[[package]]
|
||||
name = "pathfinder_svg"
|
||||
version = "0.1.0"
|
||||
|
|
|
@ -5,6 +5,7 @@ members = [
|
|||
"gfx-utils",
|
||||
"gl",
|
||||
"renderer",
|
||||
"simd",
|
||||
"svg",
|
||||
"utils/area-lut",
|
||||
"utils/gamma-lut",
|
||||
|
|
|
@ -13,3 +13,6 @@ lyon_path = "0.12"
|
|||
serde = "1.0"
|
||||
serde_derive = "1.0"
|
||||
smallvec = "0.6"
|
||||
|
||||
[dependencies.pathfinder_simd]
|
||||
path = "../simd"
|
||||
|
|
|
@ -25,7 +25,6 @@ pub mod outline;
|
|||
pub mod point;
|
||||
pub mod segment;
|
||||
pub mod segments;
|
||||
pub mod simd;
|
||||
pub mod stroke;
|
||||
pub mod transform;
|
||||
pub mod transform3d;
|
||||
|
|
|
@ -11,8 +11,8 @@
|
|||
//! Line segment types, optimized with SIMD.
|
||||
|
||||
use crate::point::Point2DF32;
|
||||
use crate::simd::F32x4;
|
||||
use crate::util;
|
||||
use pathfinder_simd::default::F32x4;
|
||||
use std::ops::Sub;
|
||||
|
||||
#[derive(Clone, Copy, Debug, PartialEq, Default)]
|
||||
|
|
|
@ -10,8 +10,8 @@
|
|||
|
||||
//! A SIMD-optimized point type.
|
||||
|
||||
use crate::simd::F32x4;
|
||||
use euclid::Point2D;
|
||||
use pathfinder_simd::default::F32x4;
|
||||
use std::ops::{Add, AddAssign, Mul, Sub};
|
||||
|
||||
// 2D points.
|
||||
|
|
|
@ -12,8 +12,8 @@
|
|||
|
||||
use crate::line_segment::LineSegmentF32;
|
||||
use crate::point::Point2DF32;
|
||||
use crate::simd::F32x4;
|
||||
use lyon_path::PathEvent;
|
||||
use pathfinder_simd::default::F32x4;
|
||||
|
||||
#[derive(Clone, Copy, Debug, PartialEq)]
|
||||
pub struct Segment {
|
||||
|
|
|
@ -1,735 +0,0 @@
|
|||
// pathfinder/geometry/src/simd.rs
|
||||
//
|
||||
// Copyright © 2019 The Pathfinder Project Developers.
|
||||
//
|
||||
// Licensed under the Apache License, Version 2.0 <LICENSE-APACHE or
|
||||
// http://www.apache.org/licenses/LICENSE-2.0> or the MIT license
|
||||
// <LICENSE-MIT or http://opensource.org/licenses/MIT>, at your
|
||||
// option. This file may not be copied, modified, or distributed
|
||||
// except according to those terms.
|
||||
|
||||
#![allow(dead_code)]
|
||||
|
||||
#[cfg(any(feature = "pf-no-simd", all(not(target_arch = "x86"), not(target_arch = "x86_64"))))]
|
||||
pub type F32x4 = scalar::F32x4;
|
||||
#[cfg(any(feature = "pf-no-simd", all(not(target_arch = "x86"), not(target_arch = "x86_64"))))]
|
||||
pub type I32x4 = scalar::I32x4;
|
||||
#[cfg(any(feature = "pf-no-simd", all(not(target_arch = "x86"), not(target_arch = "x86_64"))))]
|
||||
pub type U32x4 = scalar::U32x4;
|
||||
#[cfg(any(feature = "pf-no-simd", all(not(target_arch = "x86"), not(target_arch = "x86_64"))))]
|
||||
pub type U8x16 = scalar::U8x16;
|
||||
|
||||
#[cfg(all(not(feature = "pf-no-simd"), any(target_arch = "x86", target_arch = "x86_64")))]
|
||||
pub type F32x4 = x86::F32x4;
|
||||
#[cfg(all(not(feature = "pf-no-simd"), any(target_arch = "x86", target_arch = "x86_64")))]
|
||||
pub type I32x4 = x86::I32x4;
|
||||
#[cfg(all(not(feature = "pf-no-simd"), any(target_arch = "x86", target_arch = "x86_64")))]
|
||||
pub type U32x4 = x86::U32x4;
|
||||
#[cfg(all(not(feature = "pf-no-simd"), any(target_arch = "x86", target_arch = "x86_64")))]
|
||||
pub type U8x16 = x86::U8x16;
|
||||
|
||||
mod scalar {
|
||||
use std::f32;
|
||||
use std::fmt::{self, Debug, Formatter};
|
||||
use std::mem;
|
||||
use std::ops::{Add, Index, IndexMut, Mul, Sub};
|
||||
|
||||
// 32-bit floats
|
||||
|
||||
#[derive(Clone, Copy, Default, PartialEq)]
|
||||
pub struct F32x4(pub [f32; 4]);
|
||||
|
||||
impl F32x4 {
|
||||
#[inline]
|
||||
pub fn new(a: f32, b: f32, c: f32, d: f32) -> F32x4 {
|
||||
F32x4([a, b, c, d])
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn splat(x: f32) -> F32x4 {
|
||||
F32x4([x; 4])
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn min(self, other: F32x4) -> F32x4 {
|
||||
F32x4([
|
||||
self[0].min(other[0]),
|
||||
self[1].min(other[1]),
|
||||
self[2].min(other[2]),
|
||||
self[3].min(other[3]),
|
||||
])
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn max(self, other: F32x4) -> F32x4 {
|
||||
F32x4([
|
||||
self[0].max(other[0]),
|
||||
self[1].max(other[1]),
|
||||
self[2].max(other[2]),
|
||||
self[3].max(other[3]),
|
||||
])
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn packed_eq(self, other: F32x4) -> U32x4 {
|
||||
U32x4([
|
||||
if self[0] == other[0] { !0 } else { 0 },
|
||||
if self[1] == other[1] { !0 } else { 0 },
|
||||
if self[2] == other[2] { !0 } else { 0 },
|
||||
if self[3] == other[3] { !0 } else { 0 },
|
||||
])
|
||||
}
|
||||
|
||||
// Casts these packed floats to 64-bit floats.
|
||||
//
|
||||
// NB: This is a pure bitcast and does no actual conversion; only use this if you know what
|
||||
// you're doing.
|
||||
#[inline]
|
||||
pub fn as_f64x2(self) -> F64x2 {
|
||||
unsafe {
|
||||
F64x2(*mem::transmute::<&[f32; 4], &[f64; 2]>(&self.0))
|
||||
}
|
||||
}
|
||||
|
||||
// Converts these packed floats to integers.
|
||||
#[inline]
|
||||
pub fn to_i32x4(self) -> I32x4 {
|
||||
unsafe {
|
||||
I32x4(*mem::transmute::<&[f32; 4], &[i32; 4]>(&self.0))
|
||||
}
|
||||
}
|
||||
|
||||
// Shuffles
|
||||
|
||||
#[inline]
|
||||
pub fn xxyy(self) -> F32x4 {
|
||||
F32x4([self[0], self[0], self[1], self[1]])
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn xyxy(self) -> F32x4 {
|
||||
F32x4([self[0], self[1], self[0], self[1]])
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn xyyx(self) -> F32x4 {
|
||||
F32x4([self[0], self[1], self[1], self[0]])
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn xzxz(self) -> F32x4 {
|
||||
F32x4([self[0], self[2], self[0], self[2]])
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn ywyw(self) -> F32x4 {
|
||||
F32x4([self[1], self[3], self[1], self[3]])
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn zzww(self) -> F32x4 {
|
||||
F32x4([self[2], self[2], self[3], self[3]])
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn zwxy(self) -> F32x4 {
|
||||
F32x4([self[2], self[3], self[0], self[1]])
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn zwzw(self) -> F32x4 {
|
||||
F32x4([self[2], self[3], self[2], self[3]])
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn wxyz(self) -> F32x4 {
|
||||
F32x4([self[3], self[0], self[1], self[2]])
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn interleave(self, other: F32x4) -> (F32x4, F32x4) {
|
||||
(F32x4([self[0], other[0], self[1], other[1]]),
|
||||
F32x4([self[2], other[2], self[3], other[3]]))
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn transpose4(a: &mut F32x4, b: &mut F32x4, c: &mut F32x4, d: &mut F32x4) {
|
||||
unimplemented!()
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn cross(&self, other: F32x4) -> F32x4 {
|
||||
unimplemented!()
|
||||
}
|
||||
}
|
||||
|
||||
impl Index<usize> for F32x4 {
|
||||
type Output = f32;
|
||||
#[inline]
|
||||
fn index(&self, index: usize) -> &f32 {
|
||||
&self.0[index]
|
||||
}
|
||||
}
|
||||
|
||||
impl IndexMut<usize> for F32x4 {
|
||||
#[inline]
|
||||
fn index_mut(&mut self, index: usize) -> &mut f32 {
|
||||
&mut self.0[index]
|
||||
}
|
||||
}
|
||||
|
||||
impl Debug for F32x4 {
|
||||
#[inline]
|
||||
fn fmt(&self, f: &mut Formatter) -> Result<(), fmt::Error> {
|
||||
write!(f, "<{}, {}, {}, {}>", self[0], self[1], self[2], self[3])
|
||||
}
|
||||
}
|
||||
|
||||
impl Add<F32x4> for F32x4 {
|
||||
type Output = F32x4;
|
||||
#[inline]
|
||||
fn add(self, other: F32x4) -> F32x4 {
|
||||
F32x4([self[0] + other[0], self[1] + other[1], self[2] + other[2], self[3] + other[3]])
|
||||
}
|
||||
}
|
||||
|
||||
impl Mul<F32x4> for F32x4 {
|
||||
type Output = F32x4;
|
||||
#[inline]
|
||||
fn mul(self, other: F32x4) -> F32x4 {
|
||||
F32x4([self[0] * other[0], self[1] * other[1], self[2] * other[2], self[3] * other[3]])
|
||||
}
|
||||
}
|
||||
|
||||
impl Sub<F32x4> for F32x4 {
|
||||
type Output = F32x4;
|
||||
#[inline]
|
||||
fn sub(self, other: F32x4) -> F32x4 {
|
||||
F32x4([self[0] - other[0], self[1] - other[1], self[2] - other[2], self[3] - other[3]])
|
||||
}
|
||||
}
|
||||
|
||||
// 64-bit floats
|
||||
|
||||
#[derive(Clone, Copy)]
|
||||
pub struct F64x2(pub [f64; 2]);
|
||||
|
||||
impl F64x2 {
|
||||
// Shuffles
|
||||
|
||||
#[inline]
|
||||
pub fn interleave(self, other: F64x2) -> (F64x2, F64x2) {
|
||||
(F64x2([self.0[0], other.0[0]]), F64x2([self.0[1], other.0[1]]))
|
||||
}
|
||||
|
||||
// Creates `<self[0], other[1]>`.
|
||||
#[inline]
|
||||
pub fn combine_low_high(self, other: F64x2) -> F64x2 {
|
||||
F64x2([self.0[0], other.0[1]])
|
||||
}
|
||||
|
||||
// Casts these packed floats to 32-bit floats.
|
||||
//
|
||||
// NB: This is a pure bitcast and does no actual conversion; only use this if you know what
|
||||
// you're doing.
|
||||
#[inline]
|
||||
pub fn as_f32x4(self) -> F32x4 {
|
||||
unsafe {
|
||||
let vector: &[f32; 4] = mem::transmute::<&[f64; 2], &[f32; 4]>(&self.0);
|
||||
F32x4([vector[0], vector[1], vector[2], vector[3]])
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// 32-bit signed integers
|
||||
|
||||
#[derive(Clone, Copy)]
|
||||
pub struct I32x4([i32; 4]);
|
||||
|
||||
impl I32x4 {
|
||||
#[inline]
|
||||
pub fn new(a: i32, b: i32, c: i32, d: i32) -> I32x4 {
|
||||
I32x4([a, b, c, d])
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn splat(x: i32) -> I32x4 {
|
||||
I32x4([x; 4])
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn as_u8x16(self) -> U8x16 {
|
||||
unsafe {
|
||||
U8x16(*mem::transmute::<&[i32; 4], &[u8; 16]>(&self.0))
|
||||
}
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn min(self, other: I32x4) -> I32x4 {
|
||||
I32x4([
|
||||
self[0].min(other[0]),
|
||||
self[1].min(other[1]),
|
||||
self[2].min(other[2]),
|
||||
self[3].min(other[3]),
|
||||
])
|
||||
}
|
||||
}
|
||||
|
||||
impl Index<usize> for I32x4 {
|
||||
type Output = i32;
|
||||
#[inline]
|
||||
fn index(&self, index: usize) -> &i32 {
|
||||
&self.0[index]
|
||||
}
|
||||
}
|
||||
|
||||
impl Sub<I32x4> for I32x4 {
|
||||
type Output = I32x4;
|
||||
#[inline]
|
||||
fn sub(self, other: I32x4) -> I32x4 {
|
||||
I32x4([self[0] - other[0], self[1] - other[1], self[2] - other[2], self[3] - other[3]])
|
||||
}
|
||||
}
|
||||
|
||||
// 32-bit unsigned integers
|
||||
|
||||
#[derive(Clone, Copy)]
|
||||
pub struct U32x4(pub [u32; 4]);
|
||||
|
||||
impl U32x4 {
|
||||
#[inline]
|
||||
fn is_all_ones(&self) -> bool {
|
||||
self[0] == !0 && self[1] == !0 && self[2] == !0 && self[3] == !0
|
||||
}
|
||||
}
|
||||
|
||||
impl Index<usize> for U32x4 {
|
||||
type Output = u32;
|
||||
#[inline]
|
||||
fn index(&self, index: usize) -> &u32 {
|
||||
&self.0[index]
|
||||
}
|
||||
}
|
||||
|
||||
// 8-bit unsigned integers
|
||||
|
||||
#[derive(Clone, Copy)]
|
||||
pub struct U8x16([u8; 16]);
|
||||
|
||||
impl U8x16 {
|
||||
#[inline]
|
||||
pub fn as_i32x4(self) -> I32x4 {
|
||||
unsafe {
|
||||
I32x4(*mem::transmute::<&[u8; 16], &[i32; 4]>(&self.0))
|
||||
}
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn shuffle(self, table: U8x16) -> U8x16 {
|
||||
let mut result = [0; 16];
|
||||
for index in 0..16 {
|
||||
result[index] = table.0[index]
|
||||
}
|
||||
U8x16(result)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
#[cfg(all(not(feature = "pf-no-simd"), any(target_arch = "x86", target_arch = "x86_64")))]
|
||||
mod x86 {
|
||||
use std::arch::x86_64::{self, __m128, __m128i};
|
||||
use std::cmp::PartialEq;
|
||||
use std::fmt::{self, Debug, Formatter};
|
||||
use std::mem;
|
||||
use std::ops::{Add, AddAssign, Index, IndexMut, Mul, Neg, Sub};
|
||||
|
||||
// 32-bit floats
|
||||
|
||||
#[derive(Clone, Copy)]
|
||||
pub struct F32x4(pub __m128);
|
||||
|
||||
impl F32x4 {
|
||||
#[inline]
|
||||
pub fn new(a: f32, b: f32, c: f32, d: f32) -> F32x4 {
|
||||
unsafe {
|
||||
let vector = [a, b, c, d];
|
||||
F32x4(x86_64::_mm_loadu_ps(vector.as_ptr()))
|
||||
}
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn splat(x: f32) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_set1_ps(x)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn min(self, other: F32x4) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_min_ps(self.0, other.0)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn max(self, other: F32x4) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_max_ps(self.0, other.0)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn abs(self) -> F32x4 {
|
||||
unsafe {
|
||||
let tmp = x86_64::_mm_srli_epi32(I32x4::splat(-1).0, 1);
|
||||
F32x4(x86_64::_mm_and_ps(x86_64::_mm_castsi128_ps(tmp), self.0))
|
||||
}
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn packed_eq(self, other: F32x4) -> U32x4 {
|
||||
unsafe {
|
||||
U32x4(x86_64::_mm_castps_si128(x86_64::_mm_cmpeq_ps(
|
||||
self.0, other.0,
|
||||
)))
|
||||
}
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn packed_gt(self, other: F32x4) -> U32x4 {
|
||||
unsafe {
|
||||
U32x4(x86_64::_mm_castps_si128(x86_64::_mm_cmpgt_ps(
|
||||
self.0, other.0,
|
||||
)))
|
||||
}
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn approx_eq(self, other: F32x4, epsilon: f32) -> bool {
|
||||
(self - other).abs().packed_gt(F32x4::splat(epsilon)).is_all_zeroes()
|
||||
}
|
||||
|
||||
// Converts these packed floats to integers.
|
||||
#[inline]
|
||||
pub fn to_i32x4(self) -> I32x4 {
|
||||
unsafe { I32x4(x86_64::_mm_cvtps_epi32(self.0)) }
|
||||
}
|
||||
|
||||
// Shuffles
|
||||
|
||||
#[inline]
|
||||
pub fn xxyy(self) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, self.0, 0b0101_0000)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn xxzz(self) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, self.0, 0b1010_0000)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn xyxy(self) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, self.0, 0b0100_0100)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn xwxw(self) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, self.0, 0b1100_1100)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn xyyx(self) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, self.0, 0b0001_0100)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn xzxz(self) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, self.0, 0b1000_1000)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn xzyw(self) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, self.0, 0b1101_1000)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn yxwz(self) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, self.0, 0b1011_0001)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn yyww(self) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, self.0, 0b1111_0101)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn yzxw(self) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, self.0, 0b1100_1001)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn ywyw(self) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, self.0, 0b1101_1101)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn zxyw(self) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, self.0, 0b1101_0010)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn zyzy(self) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, self.0, 0b0110_0110)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn zzww(self) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, self.0, 0b1111_1010)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn zwxy(self) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, self.0, 0b0100_1110)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn zwzw(self) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, self.0, 0b1110_1110)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn wxyz(self) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, self.0, 0b1001_0011)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn wyzx(self) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, self.0, 0b0010_0111)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn combine_axbxayby(self, other: F32x4) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_unpacklo_ps(self.0, other.0)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn combine_axaybxby(self, other: F32x4) -> F32x4 {
|
||||
unsafe {
|
||||
let this = x86_64::_mm_castps_pd(self.0);
|
||||
let other = x86_64::_mm_castps_pd(other.0);
|
||||
let result = x86_64::_mm_unpacklo_pd(this, other);
|
||||
F32x4(x86_64::_mm_castpd_ps(result))
|
||||
}
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn combine_axaybzbw(self, other: F32x4) -> F32x4 {
|
||||
unsafe {
|
||||
let this = x86_64::_mm_castps_pd(self.0);
|
||||
let other = x86_64::_mm_castps_pd(other.0);
|
||||
let result = x86_64::_mm_shuffle_pd(this, other, 0b10);
|
||||
F32x4(x86_64::_mm_castpd_ps(result))
|
||||
}
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn combine_axazbxbz(self, other: F32x4) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, other.0, 0b1000_1000)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn combine_ayawbybw(self, other: F32x4) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, other.0, 0b1101_1101)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn combine_azawbzbw(self, other: F32x4) -> F32x4 {
|
||||
unsafe {
|
||||
let this = x86_64::_mm_castps_pd(self.0);
|
||||
let other = x86_64::_mm_castps_pd(other.0);
|
||||
let result = x86_64::_mm_unpackhi_pd(this, other);
|
||||
F32x4(x86_64::_mm_castpd_ps(result))
|
||||
}
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn combine_azbzawbw(self, other: F32x4) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_unpackhi_ps(self.0, other.0)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn combine_awazbybx(self, other: F32x4) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, other.0, 0b0001_1011)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn transpose_4x4(a: &mut F32x4, b: &mut F32x4, c: &mut F32x4, d: &mut F32x4) {
|
||||
unsafe {
|
||||
x86_64::_MM_TRANSPOSE4_PS(&mut a.0, &mut b.0, &mut c.0, &mut d.0)
|
||||
}
|
||||
}
|
||||
|
||||
// FIXME(pcwalton): Move to `Point4DF32`!
|
||||
#[inline]
|
||||
pub fn cross(&self, other: F32x4) -> F32x4 {
|
||||
self.yzxw() * other.zxyw() - self.zxyw() * other.yzxw()
|
||||
}
|
||||
}
|
||||
|
||||
impl Default for F32x4 {
|
||||
#[inline]
|
||||
fn default() -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_setzero_ps()) }
|
||||
}
|
||||
}
|
||||
|
||||
impl Index<usize> for F32x4 {
|
||||
type Output = f32;
|
||||
#[inline]
|
||||
fn index(&self, index: usize) -> &f32 {
|
||||
unsafe { &mem::transmute::<&__m128, &[f32; 4]>(&self.0)[index] }
|
||||
}
|
||||
}
|
||||
|
||||
impl IndexMut<usize> for F32x4 {
|
||||
#[inline]
|
||||
fn index_mut(&mut self, index: usize) -> &mut f32 {
|
||||
unsafe { &mut mem::transmute::<&mut __m128, &mut [f32; 4]>(&mut self.0)[index] }
|
||||
}
|
||||
}
|
||||
|
||||
impl Debug for F32x4 {
|
||||
#[inline]
|
||||
fn fmt(&self, f: &mut Formatter) -> Result<(), fmt::Error> {
|
||||
write!(f, "<{}, {}, {}, {}>", self[0], self[1], self[2], self[3])
|
||||
}
|
||||
}
|
||||
|
||||
impl PartialEq for F32x4 {
|
||||
#[inline]
|
||||
fn eq(&self, other: &F32x4) -> bool {
|
||||
self.packed_eq(*other).is_all_ones()
|
||||
}
|
||||
}
|
||||
|
||||
impl Add<F32x4> for F32x4 {
|
||||
type Output = F32x4;
|
||||
#[inline]
|
||||
fn add(self, other: F32x4) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_add_ps(self.0, other.0)) }
|
||||
}
|
||||
}
|
||||
|
||||
impl AddAssign for F32x4 {
|
||||
#[inline]
|
||||
fn add_assign(&mut self, other: F32x4) {
|
||||
unsafe { self.0 = x86_64::_mm_add_ps(self.0, other.0) }
|
||||
}
|
||||
}
|
||||
|
||||
impl Mul<F32x4> for F32x4 {
|
||||
type Output = F32x4;
|
||||
#[inline]
|
||||
fn mul(self, other: F32x4) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_mul_ps(self.0, other.0)) }
|
||||
}
|
||||
}
|
||||
|
||||
impl Sub<F32x4> for F32x4 {
|
||||
type Output = F32x4;
|
||||
#[inline]
|
||||
fn sub(self, other: F32x4) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_sub_ps(self.0, other.0)) }
|
||||
}
|
||||
}
|
||||
|
||||
impl Neg for F32x4 {
|
||||
type Output = F32x4;
|
||||
#[inline]
|
||||
fn neg(self) -> F32x4 {
|
||||
F32x4::default() - self
|
||||
}
|
||||
}
|
||||
|
||||
// 32-bit signed integers
|
||||
|
||||
#[derive(Clone, Copy)]
|
||||
pub struct I32x4(pub __m128i);
|
||||
|
||||
impl I32x4 {
|
||||
#[inline]
|
||||
pub fn new(a: i32, b: i32, c: i32, d: i32) -> I32x4 {
|
||||
unsafe {
|
||||
let vector = [a, b, c, d];
|
||||
I32x4(x86_64::_mm_loadu_si128(vector.as_ptr() as *const __m128i))
|
||||
}
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn splat(x: i32) -> I32x4 {
|
||||
unsafe { I32x4(x86_64::_mm_set1_epi32(x)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn as_u8x16(self) -> U8x16 {
|
||||
U8x16(self.0)
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn min(self, other: I32x4) -> I32x4 {
|
||||
unsafe { I32x4(x86_64::_mm_min_epi32(self.0, other.0)) }
|
||||
}
|
||||
}
|
||||
|
||||
impl Index<usize> for I32x4 {
|
||||
type Output = i32;
|
||||
#[inline]
|
||||
fn index(&self, index: usize) -> &i32 {
|
||||
unsafe { &mem::transmute::<&__m128i, &[i32; 4]>(&self.0)[index] }
|
||||
}
|
||||
}
|
||||
|
||||
impl Sub<I32x4> for I32x4 {
|
||||
type Output = I32x4;
|
||||
#[inline]
|
||||
fn sub(self, other: I32x4) -> I32x4 {
|
||||
unsafe { I32x4(x86_64::_mm_sub_epi32(self.0, other.0)) }
|
||||
}
|
||||
}
|
||||
|
||||
// 32-bit unsigned integers
|
||||
|
||||
#[derive(Clone, Copy)]
|
||||
pub struct U32x4(pub __m128i);
|
||||
|
||||
impl U32x4 {
|
||||
#[inline]
|
||||
fn is_all_ones(&self) -> bool {
|
||||
unsafe { x86_64::_mm_test_all_ones(self.0) != 0 }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
fn is_all_zeroes(&self) -> bool {
|
||||
unsafe { x86_64::_mm_test_all_zeros(self.0, self.0) != 0 }
|
||||
}
|
||||
}
|
||||
|
||||
impl Index<usize> for U32x4 {
|
||||
type Output = u32;
|
||||
#[inline]
|
||||
fn index(&self, index: usize) -> &u32 {
|
||||
unsafe { &mem::transmute::<&__m128i, &[u32; 4]>(&self.0)[index] }
|
||||
}
|
||||
}
|
||||
|
||||
// 8-bit unsigned integers
|
||||
|
||||
#[derive(Clone, Copy)]
|
||||
pub struct U8x16(pub __m128i);
|
||||
|
||||
impl U8x16 {
|
||||
#[inline]
|
||||
pub fn as_i32x4(self) -> I32x4 {
|
||||
I32x4(self.0)
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn shuffle(self, indices: U8x16) -> U8x16 {
|
||||
unsafe { U8x16(x86_64::_mm_shuffle_epi8(self.0, indices.0)) }
|
||||
}
|
||||
}
|
||||
}
|
|
@ -12,10 +12,10 @@
|
|||
|
||||
use crate::point::Point2DF32;
|
||||
use crate::segment::Segment;
|
||||
use crate::simd::F32x4;
|
||||
use crate::transform3d::Transform3DF32;
|
||||
use euclid::{Point2D, Rect, Size2D, Transform2D};
|
||||
use lyon_path::PathEvent;
|
||||
use pathfinder_simd::default::F32x4;
|
||||
use std::ops::Sub;
|
||||
|
||||
/// A 2x2 matrix, optimized with SIMD, in column-major order.
|
||||
|
|
|
@ -12,9 +12,9 @@
|
|||
|
||||
use crate::point::{Point2DF32, Point3DF32, Point4DF32};
|
||||
use crate::segment::Segment;
|
||||
use crate::simd::F32x4;
|
||||
use crate::transform::Matrix2x2F32;
|
||||
use euclid::{Point2D, Rect, Size2D};
|
||||
use pathfinder_simd::default::F32x4;
|
||||
use std::ops::{Add, Neg};
|
||||
|
||||
/// An transform, optimized with SIMD.
|
||||
|
|
|
@ -15,5 +15,8 @@ smallvec = "0.6"
|
|||
[dependencies.pathfinder_geometry]
|
||||
path = "../geometry"
|
||||
|
||||
[dependencies.pathfinder_simd]
|
||||
path = "../simd"
|
||||
|
||||
[dev-dependencies]
|
||||
quickcheck = "0.7"
|
||||
|
|
|
@ -16,8 +16,8 @@ use euclid::Rect;
|
|||
use fixedbitset::FixedBitSet;
|
||||
use pathfinder_geometry::line_segment::{LineSegmentF32, LineSegmentU4, LineSegmentU8};
|
||||
use pathfinder_geometry::point::Point2DF32;
|
||||
use pathfinder_geometry::simd::{F32x4, I32x4};
|
||||
use pathfinder_geometry::util;
|
||||
use pathfinder_simd::default::{F32x4, I32x4};
|
||||
|
||||
#[derive(Debug)]
|
||||
pub struct BuiltObject {
|
||||
|
|
|
@ -0,0 +1,7 @@
|
|||
[package]
|
||||
name = "pathfinder_simd"
|
||||
version = "0.3.0"
|
||||
edition = "2018"
|
||||
authors = ["Patrick Walton <pcwalton@mimiga.net>"]
|
||||
|
||||
[dependencies]
|
|
@ -0,0 +1,19 @@
|
|||
// pathfinder/simd/src/lib.rs
|
||||
//
|
||||
// Copyright © 2019 The Pathfinder Project Developers.
|
||||
//
|
||||
// Licensed under the Apache License, Version 2.0 <LICENSE-APACHE or
|
||||
// http://www.apache.org/licenses/LICENSE-2.0> or the MIT license
|
||||
// <LICENSE-MIT or http://opensource.org/licenses/MIT>, at your
|
||||
// option. This file may not be copied, modified, or distributed
|
||||
// except according to those terms.
|
||||
|
||||
//! A minimal SIMD abstraction, usable outside of Pathfinder.
|
||||
|
||||
#[cfg(any(feature = "pf-no-simd", all(not(target_arch = "x86"), not(target_arch = "x86_64"))))]
|
||||
pub use crate::scalar as default;
|
||||
#[cfg(all(not(feature = "pf-no-simd"), any(target_arch = "x86", target_arch = "x86_64")))]
|
||||
pub use crate::x86 as default;
|
||||
|
||||
pub mod scalar;
|
||||
pub mod x86;
|
|
@ -0,0 +1,314 @@
|
|||
// pathfinder/simd/src/scalar.rs
|
||||
//
|
||||
// Copyright © 2019 The Pathfinder Project Developers.
|
||||
//
|
||||
// Licensed under the Apache License, Version 2.0 <LICENSE-APACHE or
|
||||
// http://www.apache.org/licenses/LICENSE-2.0> or the MIT license
|
||||
// <LICENSE-MIT or http://opensource.org/licenses/MIT>, at your
|
||||
// option. This file may not be copied, modified, or distributed
|
||||
// except according to those terms.
|
||||
|
||||
use std::f32;
|
||||
use std::fmt::{self, Debug, Formatter};
|
||||
use std::mem;
|
||||
use std::ops::{Add, Index, IndexMut, Mul, Sub};
|
||||
|
||||
// 32-bit floats
|
||||
|
||||
#[derive(Clone, Copy, Default, PartialEq)]
|
||||
pub struct F32x4(pub [f32; 4]);
|
||||
|
||||
impl F32x4 {
|
||||
#[inline]
|
||||
pub fn new(a: f32, b: f32, c: f32, d: f32) -> F32x4 {
|
||||
F32x4([a, b, c, d])
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn splat(x: f32) -> F32x4 {
|
||||
F32x4([x; 4])
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn min(self, other: F32x4) -> F32x4 {
|
||||
F32x4([
|
||||
self[0].min(other[0]),
|
||||
self[1].min(other[1]),
|
||||
self[2].min(other[2]),
|
||||
self[3].min(other[3]),
|
||||
])
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn max(self, other: F32x4) -> F32x4 {
|
||||
F32x4([
|
||||
self[0].max(other[0]),
|
||||
self[1].max(other[1]),
|
||||
self[2].max(other[2]),
|
||||
self[3].max(other[3]),
|
||||
])
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn packed_eq(self, other: F32x4) -> U32x4 {
|
||||
U32x4([
|
||||
if self[0] == other[0] { !0 } else { 0 },
|
||||
if self[1] == other[1] { !0 } else { 0 },
|
||||
if self[2] == other[2] { !0 } else { 0 },
|
||||
if self[3] == other[3] { !0 } else { 0 },
|
||||
])
|
||||
}
|
||||
|
||||
// Casts these packed floats to 64-bit floats.
|
||||
//
|
||||
// NB: This is a pure bitcast and does no actual conversion; only use this if you know what
|
||||
// you're doing.
|
||||
#[inline]
|
||||
pub fn as_f64x2(self) -> F64x2 {
|
||||
unsafe {
|
||||
F64x2(*mem::transmute::<&[f32; 4], &[f64; 2]>(&self.0))
|
||||
}
|
||||
}
|
||||
|
||||
// Converts these packed floats to integers.
|
||||
#[inline]
|
||||
pub fn to_i32x4(self) -> I32x4 {
|
||||
unsafe {
|
||||
I32x4(*mem::transmute::<&[f32; 4], &[i32; 4]>(&self.0))
|
||||
}
|
||||
}
|
||||
|
||||
// Shuffles
|
||||
|
||||
#[inline]
|
||||
pub fn xxyy(self) -> F32x4 {
|
||||
F32x4([self[0], self[0], self[1], self[1]])
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn xyxy(self) -> F32x4 {
|
||||
F32x4([self[0], self[1], self[0], self[1]])
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn xyyx(self) -> F32x4 {
|
||||
F32x4([self[0], self[1], self[1], self[0]])
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn xzxz(self) -> F32x4 {
|
||||
F32x4([self[0], self[2], self[0], self[2]])
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn ywyw(self) -> F32x4 {
|
||||
F32x4([self[1], self[3], self[1], self[3]])
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn zzww(self) -> F32x4 {
|
||||
F32x4([self[2], self[2], self[3], self[3]])
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn zwxy(self) -> F32x4 {
|
||||
F32x4([self[2], self[3], self[0], self[1]])
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn zwzw(self) -> F32x4 {
|
||||
F32x4([self[2], self[3], self[2], self[3]])
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn wxyz(self) -> F32x4 {
|
||||
F32x4([self[3], self[0], self[1], self[2]])
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn interleave(self, other: F32x4) -> (F32x4, F32x4) {
|
||||
(F32x4([self[0], other[0], self[1], other[1]]),
|
||||
F32x4([self[2], other[2], self[3], other[3]]))
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn transpose4(a: &mut F32x4, b: &mut F32x4, c: &mut F32x4, d: &mut F32x4) {
|
||||
unimplemented!()
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn cross(&self, other: F32x4) -> F32x4 {
|
||||
unimplemented!()
|
||||
}
|
||||
}
|
||||
|
||||
impl Index<usize> for F32x4 {
|
||||
type Output = f32;
|
||||
#[inline]
|
||||
fn index(&self, index: usize) -> &f32 {
|
||||
&self.0[index]
|
||||
}
|
||||
}
|
||||
|
||||
impl IndexMut<usize> for F32x4 {
|
||||
#[inline]
|
||||
fn index_mut(&mut self, index: usize) -> &mut f32 {
|
||||
&mut self.0[index]
|
||||
}
|
||||
}
|
||||
|
||||
impl Debug for F32x4 {
|
||||
#[inline]
|
||||
fn fmt(&self, f: &mut Formatter) -> Result<(), fmt::Error> {
|
||||
write!(f, "<{}, {}, {}, {}>", self[0], self[1], self[2], self[3])
|
||||
}
|
||||
}
|
||||
|
||||
impl Add<F32x4> for F32x4 {
|
||||
type Output = F32x4;
|
||||
#[inline]
|
||||
fn add(self, other: F32x4) -> F32x4 {
|
||||
F32x4([self[0] + other[0], self[1] + other[1], self[2] + other[2], self[3] + other[3]])
|
||||
}
|
||||
}
|
||||
|
||||
impl Mul<F32x4> for F32x4 {
|
||||
type Output = F32x4;
|
||||
#[inline]
|
||||
fn mul(self, other: F32x4) -> F32x4 {
|
||||
F32x4([self[0] * other[0], self[1] * other[1], self[2] * other[2], self[3] * other[3]])
|
||||
}
|
||||
}
|
||||
|
||||
impl Sub<F32x4> for F32x4 {
|
||||
type Output = F32x4;
|
||||
#[inline]
|
||||
fn sub(self, other: F32x4) -> F32x4 {
|
||||
F32x4([self[0] - other[0], self[1] - other[1], self[2] - other[2], self[3] - other[3]])
|
||||
}
|
||||
}
|
||||
|
||||
// 64-bit floats
|
||||
|
||||
#[derive(Clone, Copy)]
|
||||
pub struct F64x2(pub [f64; 2]);
|
||||
|
||||
impl F64x2 {
|
||||
// Shuffles
|
||||
|
||||
#[inline]
|
||||
pub fn interleave(self, other: F64x2) -> (F64x2, F64x2) {
|
||||
(F64x2([self.0[0], other.0[0]]), F64x2([self.0[1], other.0[1]]))
|
||||
}
|
||||
|
||||
// Creates `<self[0], other[1]>`.
|
||||
#[inline]
|
||||
pub fn combine_low_high(self, other: F64x2) -> F64x2 {
|
||||
F64x2([self.0[0], other.0[1]])
|
||||
}
|
||||
|
||||
// Casts these packed floats to 32-bit floats.
|
||||
//
|
||||
// NB: This is a pure bitcast and does no actual conversion; only use this if you know what
|
||||
// you're doing.
|
||||
#[inline]
|
||||
pub fn as_f32x4(self) -> F32x4 {
|
||||
unsafe {
|
||||
let vector: &[f32; 4] = mem::transmute::<&[f64; 2], &[f32; 4]>(&self.0);
|
||||
F32x4([vector[0], vector[1], vector[2], vector[3]])
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// 32-bit signed integers
|
||||
|
||||
#[derive(Clone, Copy)]
|
||||
pub struct I32x4([i32; 4]);
|
||||
|
||||
impl I32x4 {
|
||||
#[inline]
|
||||
pub fn new(a: i32, b: i32, c: i32, d: i32) -> I32x4 {
|
||||
I32x4([a, b, c, d])
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn splat(x: i32) -> I32x4 {
|
||||
I32x4([x; 4])
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn as_u8x16(self) -> U8x16 {
|
||||
unsafe {
|
||||
U8x16(*mem::transmute::<&[i32; 4], &[u8; 16]>(&self.0))
|
||||
}
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn min(self, other: I32x4) -> I32x4 {
|
||||
I32x4([
|
||||
self[0].min(other[0]),
|
||||
self[1].min(other[1]),
|
||||
self[2].min(other[2]),
|
||||
self[3].min(other[3]),
|
||||
])
|
||||
}
|
||||
}
|
||||
|
||||
impl Index<usize> for I32x4 {
|
||||
type Output = i32;
|
||||
#[inline]
|
||||
fn index(&self, index: usize) -> &i32 {
|
||||
&self.0[index]
|
||||
}
|
||||
}
|
||||
|
||||
impl Sub<I32x4> for I32x4 {
|
||||
type Output = I32x4;
|
||||
#[inline]
|
||||
fn sub(self, other: I32x4) -> I32x4 {
|
||||
I32x4([self[0] - other[0], self[1] - other[1], self[2] - other[2], self[3] - other[3]])
|
||||
}
|
||||
}
|
||||
|
||||
// 32-bit unsigned integers
|
||||
|
||||
#[derive(Clone, Copy)]
|
||||
pub struct U32x4(pub [u32; 4]);
|
||||
|
||||
impl U32x4 {
|
||||
#[inline]
|
||||
fn is_all_ones(&self) -> bool {
|
||||
self[0] == !0 && self[1] == !0 && self[2] == !0 && self[3] == !0
|
||||
}
|
||||
}
|
||||
|
||||
impl Index<usize> for U32x4 {
|
||||
type Output = u32;
|
||||
#[inline]
|
||||
fn index(&self, index: usize) -> &u32 {
|
||||
&self.0[index]
|
||||
}
|
||||
}
|
||||
|
||||
// 8-bit unsigned integers
|
||||
|
||||
#[derive(Clone, Copy)]
|
||||
pub struct U8x16([u8; 16]);
|
||||
|
||||
impl U8x16 {
|
||||
#[inline]
|
||||
pub fn as_i32x4(self) -> I32x4 {
|
||||
unsafe {
|
||||
I32x4(*mem::transmute::<&[u8; 16], &[i32; 4]>(&self.0))
|
||||
}
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn shuffle(self, table: U8x16) -> U8x16 {
|
||||
let mut result = [0; 16];
|
||||
for index in 0..16 {
|
||||
result[index] = table.0[index]
|
||||
}
|
||||
U8x16(result)
|
||||
}
|
||||
}
|
|
@ -0,0 +1,405 @@
|
|||
// pathfinder/simd/src/x86.rs
|
||||
//
|
||||
// Copyright © 2019 The Pathfinder Project Developers.
|
||||
//
|
||||
// Licensed under the Apache License, Version 2.0 <LICENSE-APACHE or
|
||||
// http://www.apache.org/licenses/LICENSE-2.0> or the MIT license
|
||||
// <LICENSE-MIT or http://opensource.org/licenses/MIT>, at your
|
||||
// option. This file may not be copied, modified, or distributed
|
||||
// except according to those terms.
|
||||
|
||||
use std::arch::x86_64::{self, __m128, __m128i};
|
||||
use std::cmp::PartialEq;
|
||||
use std::fmt::{self, Debug, Formatter};
|
||||
use std::mem;
|
||||
use std::ops::{Add, AddAssign, Index, IndexMut, Mul, Neg, Sub};
|
||||
|
||||
// 32-bit floats
|
||||
|
||||
#[derive(Clone, Copy)]
|
||||
pub struct F32x4(pub __m128);
|
||||
|
||||
impl F32x4 {
|
||||
#[inline]
|
||||
pub fn new(a: f32, b: f32, c: f32, d: f32) -> F32x4 {
|
||||
unsafe {
|
||||
let vector = [a, b, c, d];
|
||||
F32x4(x86_64::_mm_loadu_ps(vector.as_ptr()))
|
||||
}
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn splat(x: f32) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_set1_ps(x)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn min(self, other: F32x4) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_min_ps(self.0, other.0)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn max(self, other: F32x4) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_max_ps(self.0, other.0)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn abs(self) -> F32x4 {
|
||||
unsafe {
|
||||
let tmp = x86_64::_mm_srli_epi32(I32x4::splat(-1).0, 1);
|
||||
F32x4(x86_64::_mm_and_ps(x86_64::_mm_castsi128_ps(tmp), self.0))
|
||||
}
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn packed_eq(self, other: F32x4) -> U32x4 {
|
||||
unsafe {
|
||||
U32x4(x86_64::_mm_castps_si128(x86_64::_mm_cmpeq_ps(
|
||||
self.0, other.0,
|
||||
)))
|
||||
}
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn packed_gt(self, other: F32x4) -> U32x4 {
|
||||
unsafe {
|
||||
U32x4(x86_64::_mm_castps_si128(x86_64::_mm_cmpgt_ps(
|
||||
self.0, other.0,
|
||||
)))
|
||||
}
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn approx_eq(self, other: F32x4, epsilon: f32) -> bool {
|
||||
(self - other).abs().packed_gt(F32x4::splat(epsilon)).is_all_zeroes()
|
||||
}
|
||||
|
||||
// Converts these packed floats to integers.
|
||||
#[inline]
|
||||
pub fn to_i32x4(self) -> I32x4 {
|
||||
unsafe { I32x4(x86_64::_mm_cvtps_epi32(self.0)) }
|
||||
}
|
||||
|
||||
// Shuffles
|
||||
|
||||
#[inline]
|
||||
pub fn xxyy(self) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, self.0, 0b0101_0000)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn xxzz(self) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, self.0, 0b1010_0000)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn xyxy(self) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, self.0, 0b0100_0100)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn xwxw(self) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, self.0, 0b1100_1100)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn xyyx(self) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, self.0, 0b0001_0100)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn xzxz(self) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, self.0, 0b1000_1000)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn xzyw(self) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, self.0, 0b1101_1000)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn yxwz(self) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, self.0, 0b1011_0001)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn yyww(self) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, self.0, 0b1111_0101)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn yzxw(self) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, self.0, 0b1100_1001)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn ywyw(self) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, self.0, 0b1101_1101)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn zxyw(self) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, self.0, 0b1101_0010)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn zyzy(self) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, self.0, 0b0110_0110)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn zzww(self) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, self.0, 0b1111_1010)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn zwxy(self) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, self.0, 0b0100_1110)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn zwzw(self) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, self.0, 0b1110_1110)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn wxyz(self) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, self.0, 0b1001_0011)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn wyzx(self) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, self.0, 0b0010_0111)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn combine_axbxayby(self, other: F32x4) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_unpacklo_ps(self.0, other.0)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn combine_axaybxby(self, other: F32x4) -> F32x4 {
|
||||
unsafe {
|
||||
let this = x86_64::_mm_castps_pd(self.0);
|
||||
let other = x86_64::_mm_castps_pd(other.0);
|
||||
let result = x86_64::_mm_unpacklo_pd(this, other);
|
||||
F32x4(x86_64::_mm_castpd_ps(result))
|
||||
}
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn combine_axaybzbw(self, other: F32x4) -> F32x4 {
|
||||
unsafe {
|
||||
let this = x86_64::_mm_castps_pd(self.0);
|
||||
let other = x86_64::_mm_castps_pd(other.0);
|
||||
let result = x86_64::_mm_shuffle_pd(this, other, 0b10);
|
||||
F32x4(x86_64::_mm_castpd_ps(result))
|
||||
}
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn combine_axazbxbz(self, other: F32x4) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, other.0, 0b1000_1000)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn combine_ayawbybw(self, other: F32x4) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, other.0, 0b1101_1101)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn combine_azawbzbw(self, other: F32x4) -> F32x4 {
|
||||
unsafe {
|
||||
let this = x86_64::_mm_castps_pd(self.0);
|
||||
let other = x86_64::_mm_castps_pd(other.0);
|
||||
let result = x86_64::_mm_unpackhi_pd(this, other);
|
||||
F32x4(x86_64::_mm_castpd_ps(result))
|
||||
}
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn combine_azbzawbw(self, other: F32x4) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_unpackhi_ps(self.0, other.0)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn combine_awazbybx(self, other: F32x4) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_shuffle_ps(self.0, other.0, 0b0001_1011)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn transpose_4x4(a: &mut F32x4, b: &mut F32x4, c: &mut F32x4, d: &mut F32x4) {
|
||||
unsafe {
|
||||
x86_64::_MM_TRANSPOSE4_PS(&mut a.0, &mut b.0, &mut c.0, &mut d.0)
|
||||
}
|
||||
}
|
||||
|
||||
// FIXME(pcwalton): Move to `Point4DF32`!
|
||||
#[inline]
|
||||
pub fn cross(&self, other: F32x4) -> F32x4 {
|
||||
self.yzxw() * other.zxyw() - self.zxyw() * other.yzxw()
|
||||
}
|
||||
}
|
||||
|
||||
impl Default for F32x4 {
|
||||
#[inline]
|
||||
fn default() -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_setzero_ps()) }
|
||||
}
|
||||
}
|
||||
|
||||
impl Index<usize> for F32x4 {
|
||||
type Output = f32;
|
||||
#[inline]
|
||||
fn index(&self, index: usize) -> &f32 {
|
||||
unsafe { &mem::transmute::<&__m128, &[f32; 4]>(&self.0)[index] }
|
||||
}
|
||||
}
|
||||
|
||||
impl IndexMut<usize> for F32x4 {
|
||||
#[inline]
|
||||
fn index_mut(&mut self, index: usize) -> &mut f32 {
|
||||
unsafe { &mut mem::transmute::<&mut __m128, &mut [f32; 4]>(&mut self.0)[index] }
|
||||
}
|
||||
}
|
||||
|
||||
impl Debug for F32x4 {
|
||||
#[inline]
|
||||
fn fmt(&self, f: &mut Formatter) -> Result<(), fmt::Error> {
|
||||
write!(f, "<{}, {}, {}, {}>", self[0], self[1], self[2], self[3])
|
||||
}
|
||||
}
|
||||
|
||||
impl PartialEq for F32x4 {
|
||||
#[inline]
|
||||
fn eq(&self, other: &F32x4) -> bool {
|
||||
self.packed_eq(*other).is_all_ones()
|
||||
}
|
||||
}
|
||||
|
||||
impl Add<F32x4> for F32x4 {
|
||||
type Output = F32x4;
|
||||
#[inline]
|
||||
fn add(self, other: F32x4) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_add_ps(self.0, other.0)) }
|
||||
}
|
||||
}
|
||||
|
||||
impl AddAssign for F32x4 {
|
||||
#[inline]
|
||||
fn add_assign(&mut self, other: F32x4) {
|
||||
unsafe { self.0 = x86_64::_mm_add_ps(self.0, other.0) }
|
||||
}
|
||||
}
|
||||
|
||||
impl Mul<F32x4> for F32x4 {
|
||||
type Output = F32x4;
|
||||
#[inline]
|
||||
fn mul(self, other: F32x4) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_mul_ps(self.0, other.0)) }
|
||||
}
|
||||
}
|
||||
|
||||
impl Sub<F32x4> for F32x4 {
|
||||
type Output = F32x4;
|
||||
#[inline]
|
||||
fn sub(self, other: F32x4) -> F32x4 {
|
||||
unsafe { F32x4(x86_64::_mm_sub_ps(self.0, other.0)) }
|
||||
}
|
||||
}
|
||||
|
||||
impl Neg for F32x4 {
|
||||
type Output = F32x4;
|
||||
#[inline]
|
||||
fn neg(self) -> F32x4 {
|
||||
F32x4::default() - self
|
||||
}
|
||||
}
|
||||
|
||||
// 32-bit signed integers
|
||||
|
||||
#[derive(Clone, Copy)]
|
||||
pub struct I32x4(pub __m128i);
|
||||
|
||||
impl I32x4 {
|
||||
#[inline]
|
||||
pub fn new(a: i32, b: i32, c: i32, d: i32) -> I32x4 {
|
||||
unsafe {
|
||||
let vector = [a, b, c, d];
|
||||
I32x4(x86_64::_mm_loadu_si128(vector.as_ptr() as *const __m128i))
|
||||
}
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn splat(x: i32) -> I32x4 {
|
||||
unsafe { I32x4(x86_64::_mm_set1_epi32(x)) }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn as_u8x16(self) -> U8x16 {
|
||||
U8x16(self.0)
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn min(self, other: I32x4) -> I32x4 {
|
||||
unsafe { I32x4(x86_64::_mm_min_epi32(self.0, other.0)) }
|
||||
}
|
||||
}
|
||||
|
||||
impl Index<usize> for I32x4 {
|
||||
type Output = i32;
|
||||
#[inline]
|
||||
fn index(&self, index: usize) -> &i32 {
|
||||
unsafe { &mem::transmute::<&__m128i, &[i32; 4]>(&self.0)[index] }
|
||||
}
|
||||
}
|
||||
|
||||
impl Sub<I32x4> for I32x4 {
|
||||
type Output = I32x4;
|
||||
#[inline]
|
||||
fn sub(self, other: I32x4) -> I32x4 {
|
||||
unsafe { I32x4(x86_64::_mm_sub_epi32(self.0, other.0)) }
|
||||
}
|
||||
}
|
||||
|
||||
// 32-bit unsigned integers
|
||||
|
||||
#[derive(Clone, Copy)]
|
||||
pub struct U32x4(pub __m128i);
|
||||
|
||||
impl U32x4 {
|
||||
#[inline]
|
||||
fn is_all_ones(&self) -> bool {
|
||||
unsafe { x86_64::_mm_test_all_ones(self.0) != 0 }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
fn is_all_zeroes(&self) -> bool {
|
||||
unsafe { x86_64::_mm_test_all_zeros(self.0, self.0) != 0 }
|
||||
}
|
||||
}
|
||||
|
||||
impl Index<usize> for U32x4 {
|
||||
type Output = u32;
|
||||
#[inline]
|
||||
fn index(&self, index: usize) -> &u32 {
|
||||
unsafe { &mem::transmute::<&__m128i, &[u32; 4]>(&self.0)[index] }
|
||||
}
|
||||
}
|
||||
|
||||
// 8-bit unsigned integers
|
||||
|
||||
#[derive(Clone, Copy)]
|
||||
pub struct U8x16(pub __m128i);
|
||||
|
||||
impl U8x16 {
|
||||
#[inline]
|
||||
pub fn as_i32x4(self) -> I32x4 {
|
||||
I32x4(self.0)
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn shuffle(self, indices: U8x16) -> U8x16 {
|
||||
unsafe { U8x16(x86_64::_mm_shuffle_epi8(self.0, indices.0)) }
|
||||
}
|
||||
}
|
Loading…
Reference in New Issue