Source-Changes-HG archive
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index][Old Index]
[src/trunk]: src/sys/crypto/aes/arch/arm Tweak clang neon intrinsics so they ...
details: https://anonhg.NetBSD.org/src/rev/054f8ff9bced
branches: trunk
changeset: 973434:054f8ff9bced
user: riastradh <riastradh%NetBSD.org@localhost>
date: Tue Jun 30 21:24:00 2020 +0000
description:
Tweak clang neon intrinsics so they build.
(this file is still a kludge)
diffstat:
sys/crypto/aes/arch/arm/arm_neon.h | 134 ++++++++++++++++++++++--------------
1 files changed, 80 insertions(+), 54 deletions(-)
diffs (274 lines):
diff -r 4ee0a98fcce7 -r 054f8ff9bced sys/crypto/aes/arch/arm/arm_neon.h
--- a/sys/crypto/aes/arch/arm/arm_neon.h Tue Jun 30 21:22:19 2020 +0000
+++ b/sys/crypto/aes/arch/arm/arm_neon.h Tue Jun 30 21:24:00 2020 +0000
@@ -1,4 +1,4 @@
-/* $NetBSD: arm_neon.h,v 1.1 2020/06/29 23:56:31 riastradh Exp $ */
+/* $NetBSD: arm_neon.h,v 1.2 2020/06/30 21:24:00 riastradh Exp $ */
/*-
* Copyright (c) 2020 The NetBSD Foundation, Inc.
@@ -73,6 +73,8 @@
typedef __attribute__((neon_vector_type(16))) uint8_t uint8x16_t;
typedef __attribute__((neon_vector_type(2))) uint64_t uint64x2_t;
typedef __attribute__((neon_vector_type(4))) uint32_t uint32x4_t;
+
+typedef __attribute__((neon_vector_type(8))) uint8_t uint8x8_t;
typedef struct { uint8x8_t val[2]; } uint8x8x2_t;
#ifdef __LITTLE_ENDIAN__
@@ -118,11 +120,11 @@
};
}
+#if defined(__GNUC__) && !defined(__clang__)
_INTRINSATTR
static __inline uint32x4_t
vextq_u32(uint32x4_t __lo, uint32x4_t __hi, uint8_t __i)
{
-#if defined(__GNUC__) && !defined(__clang__)
#if defined(__AARCH64EB__) || defined(__ARM_BIG_ENDIAN)
return __builtin_shuffle(__hi, __lo,
(uint32x4_t) { 4 - __i, 5 - __i, 6 - __i, 7 - __i });
@@ -130,25 +132,31 @@
return __builtin_shuffle(__lo, __hi,
(uint32x4_t) { __i + 0, __i + 1, __i + 2, __i + 3 });
#endif
+}
#elif defined(__clang__)
#ifdef __LITTLE_ENDIAN__
- return __builtin_neon_vextq_v((int8x16_t)__lo, (int8x16_t)__hi, __i,
- 50);
+#define vextq_u32(__lo, __hi, __i) \
+ (uint32x4_t)__builtin_neon_vextq_v((int8x16_t)(__lo), \
+ (int8x16_t)(__hi), (__i), 50)
#else
- uint32x4_t __lo_r = __builtin_shufflevector(__lo, __lo, 3, 2, 1, 0);
- uint32x4_t __hi_r = __builtin_shufflevector(__hi, __hi, 3, 2, 1, 0);
- uint32x4_t __r = __builtin_neon_vextq_v((int8x16_t)__lo_r,
- (int8x16_t)__hi_r, __i, 50);
- return __builtin_shufflevector(__r, __r, 3, 2, 1, 0);
+#define vextq_u32(__lo, __hi, __i) ( \
+{ \
+ uint32x4_t __tlo = (__lo); \
+ uint32x4_t __thi = (__hi); \
+ uint32x4_t __lo_r = __builtin_shufflevector(__tlo, __tlo, 3,2,1,0); \
+ uint32x4_t __hi_r = __builtin_shufflevector(__thi, __thi, 3,2,1,0); \
+ uint32x4_t __r = __builtin_neon_vextq_v((int8x16_t)__lo_r, \
+ (int8x16_t)__hi_r, __i, 50); \
+ __builtin_shufflevector(__r, __r, 3,2,1,0); \
+})
+#endif /* __LITTLE_ENDIAN__ */
#endif
-#endif
-}
+#if defined(__GNUC__) && !defined(__clang__)
_INTRINSATTR
static __inline uint8x16_t
vextq_u8(uint8x16_t __lo, uint8x16_t __hi, uint8_t __i)
{
-#if defined(__GNUC__) && !defined(__clang__)
#if defined(__AARCH64EB__) || defined(__ARM_BIG_ENDIAN)
return __builtin_shuffle(__hi, __lo,
(uint8x16_t) {
@@ -166,38 +174,45 @@
__i + 12, __i + 13, __i + 14, __i + 15,
});
#endif
+}
#elif defined(__clang__)
#ifdef __LITTLE_ENDIAN__
- return __builtin_neon_vextq_v((int8x16_t)__lo, (int8x16_t)__hi, __i,
- 48);
+#define vextq_u8(__lo, __hi, __i) \
+ (uint8x16_t)__builtin_neon_vextq_v((int8x16_t)(__lo), \
+ (int8x16_t)(__hi), (__i), 48)
#else
- uint32x4_t __lo_r = __builtin_shufflevector(__lo, __lo,
- 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
- uint32x4_t __hi_r = __builtin_shufflevector(__hi, __hi,
- 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
- uint32x4_t __r = __builtin_neon_vextq_v((int8x16_t)__lo_r,
- (int8x16_t)__hi_r, __i, 50);
- return __builtin_shufflevector(__r, __r,
- 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
+#define vextq_u8(__lo, __hi, __i) ( \
+{ \
+ uint8x16_t __tlo = (__lo); \
+ uint8x16_t __thi = (__hi); \
+ uint8x16_t __lo_r = __builtin_shufflevector(__tlo, __tlo, \
+ 15,14,13,12,11,10,9,8,7,6,5,4,3,2,1,0); \
+ uint8x16_t __hi_r = __builtin_shufflevector(__thi, __thi, \
+ 15,14,13,12,11,10,9,8,7,6,5,4,3,2,1,0); \
+ uint8x16_t __r = __builtin_neon_vextq_v((int8x16_t)__lo_r, \
+ (int8x16_t)__hi_r, (__i), 48); \
+ return __builtin_shufflevector(__r, __r, \
+ 15,14,13,12,11,10,9,8,7,6,5,4,3,2,1,0); \
+})
+#endif /* __LITTLE_ENDIAN */
#endif
-#endif
-}
+#if defined(__GNUC__) && !defined(__clang__)
_INTRINSATTR
static __inline uint32_t
vgetq_lane_u32(uint32x4_t __v, uint8_t __i)
{
-#if defined(__GNUC__) && !defined(__clang__)
#ifdef __aarch64__
return __v[__i];
#else
return (uint32_t)__builtin_neon_vget_laneuv4si((int32x4_t)__v, __i);
#endif
+}
#elif defined(__clang__)
- return (uint32_t)__builtin_neon_vgetq_lane_i32((int32x4_t)__v,
- __neon_lane_index(__v, __i));
+#define vgetq_lane_u32(__v, __i) \
+ (uint32_t)__builtin_neon_vgetq_lane_i32((int32x4_t)(__v), \
+ __neon_lane_index(__v, __i))
#endif
-}
_INTRINSATTR
static __inline uint8x16_t
@@ -215,7 +230,7 @@
return (uint8x16_t)__builtin_neon_vld1v16qi(__p);
#endif
#elif defined(__clang__)
- return (uint8x16_t)__builtin_neon_vld1q_v(__p, 48);
+ return (uint8x16_t)__builtin_neon_vld1q_v(__p8, 48);
#endif
}
@@ -260,13 +275,13 @@
(int8x16_t)__idx, 48);
#else
uint32x4_t __lo_r = __builtin_shufflevector(__lo, __lo,
- 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
+ 15,14,13,12,11,10,9,8,7,6,5,4,3,2,1,0);
uint32x4_t __hi_r = __builtin_shufflevector(__hi, __hi,
- 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
+ 15,14,13,12,11,10,9,8,7,6,5,4,3,2,1,0);
uint32x4_t __r = __builtin_neon_vqtbl1q_v((int8x16_t)__tab,
(int8x16_t)__idx, __i, 48);
return __builtin_shufflevector(__r, __r,
- 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
+ 15,14,13,12,11,10,9,8,7,6,5,4,3,2,1,0);
#endif
#endif
}
@@ -313,71 +328,82 @@
return (uint8x16_t)__v;
}
+#if defined(__GNUC__) && !defined(__clang__)
_INTRINSATTR
static __inline uint32x4_t
vsetq_lane_u32(uint32_t __x, uint32x4_t __v, uint8_t __i)
{
-#if defined(__GNUC__) && !defined(__clang__)
__v[__neon_lane_index(__v, __i)] = __x;
return __v;
+}
#elif defined(__clang__)
- return (uint32x4_t)__builtin_neon_vsetq_lane_i32(__x, (int32x4_t)__v,
- __neon_lane_index(__v, __i));
+#define vsetq_lane_u32(__x, __v, __i) \
+ (uint32x4_t)__builtin_neon_vsetq_lane_i32((__x), (int32x4_t)(__v), \
+ __neon_lane_index(__v, __i))
#endif
-}
+#if defined(__GNUC__) && !defined(__clang__)
_INTRINSATTR
static __inline uint64x2_t
vsetq_lane_u64(uint64_t __x, uint64x2_t __v, uint8_t __i)
{
-#if defined(__GNUC__) && !defined(__clang__)
__v[__neon_lane_index(__v, __i)] = __x;
return __v;
+}
#elif defined(__clang__)
- return (uint64x2_t)__builtin_neon_vsetq_lane_i32(__x, (int64x2_t)__v,
+#define vsetq_lane_u64(__x, __v, __i) \
+ (uint64x2_t)__builtin_neon_vsetq_lane_i32((__x), (int64x2_t)(__v), \
__neon_lane_index(__v, __i));
#endif
-}
+#if defined(__GNUC__) && !defined(__clang__)
_INTRINSATTR
static __inline uint8x16_t
vshrq_n_u8(uint8x16_t __v, uint8_t __bits)
{
-#if defined(__GNUC__) && !defined(__clang__)
#ifdef __aarch64__
return (uint8x16_t)__builtin_aarch64_lshrv16qi((int8x16_t)__v, __bits);
#else
return (uint8x16_t)__builtin_neon_vshru_nv16qi((int8x16_t)__v, __bits);
#endif
+}
#elif defined(__clang__)
- return __builtin_neon_vshrq_n_v((int8x16_t)__v, __bits, 48);
+#define vshrq_n_u8(__v, __bits) \
+ (uint8x16_t)__builtin_neon_vshrq_n_v((int8x16_t)(__v), (__bits), 48)
#endif
-}
+#if defined(__GNUC__) && !defined(__clang__)
_INTRINSATTR
static __inline int32x4_t
vsliq_n_s32(int32x4_t __vins, int32x4_t __vsh, uint8_t __bits)
{
-#if defined(__GNUC__) && !defined(__clang__)
#ifdef __aarch64__
return (int32x4_t)__builtin_aarch64_ssli_nv4si(__vins, __vsh, __bits);
#else
return (int32x4_t)__builtin_neon_vsli_nv4si(__vins, __vsh, __bits);
#endif
+}
#elif defined(__clang__)
#ifdef __LITTLE_ENDIAN__
- return __builtin_neon_vsliq_n_v(__vins_r, __vsh_r, __bits, 34);
+#define vsliq_n_s32(__vins, __vsh, __bits) \
+ (int32x4_t)__builtin_neon_vsliq_n_v((int32x4_t)(__vins), \
+ (int32x4_t)(__vsh), (__bits), 34)
#else
- int32x4_t __vins_r = __builtin_shufflevector(__vins_r, __vins_r,
- 3, 2, 1, 0);
- int32x4_t __vsh_r = __builtin_shufflevector(__vsh_r, __vsh_r,
- 3, 2, 1, 0);
- int32x4_t __r = __builtin_neon_vsliq_n_v(__vins_r, __vsh_r, __bits,
- 34);
- return __builtin_shufflevector(__r, __r, 3, 2, 1, 0);
+#define vsliq_n_s32(__vins, __vsh, __bits) ( \
+{ \
+ int32x4_t __tvins = (__vins); \
+ int32x4_t __tvsh = (__vsh); \
+ uint8_t __tbits = (__bits); \
+ int32x4_t __vins_r = __builtin_shufflevector(__tvins, __tvins, \
+ 3,2,1,0); \
+ int32x4_t __vsh_r = __builtin_shufflevector(__tvsh, __tvsh, \
+ 3,2,1,0); \
+ int32x4_t __r = __builtin_neon_vsliq_n_v(__tvins, __tvsh, __tbits, \
+ 34); \
+ __builtin_shufflevector(__r, __r, 3,2,1,0); \
+})
+#endif /* __LITTLE_ENDIAN__ */
#endif
-#endif
-}
_INTRINSATTR
static __inline void
@@ -398,7 +424,7 @@
__v = __builtin_shufflevector(__v, __v,
15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
#endif
- __builtin_neon_vst1q_v(__p, __v);
+ __builtin_neon_vst1q_v(__p8, __v, 48);
#endif
}
Home |
Main Index |
Thread Index |
Old Index