mirror of
https://gitee.com/bianbu-linux/linux-6.6
synced 2025-04-24 14:07:52 -04:00
Modern compilers are perfectly capable of extracting parallelism from the XOR routines, provided that the prototypes reflect the nature of the input accurately, in particular, the fact that the input vectors are expected not to overlap. This is not documented explicitly, but is implied by the interchangeability of the various C routines, some of which use temporary variables while others don't: this means that these routines only behave identically for non-overlapping inputs. So let's decorate these input vectors with the __restrict modifier, which informs the compiler that there is no overlap. While at it, make the input-only vectors pointer-to-const as well. Tested-by: Nathan Chancellor <nathan@kernel.org> Signed-off-by: Ard Biesheuvel <ardb@kernel.org> Reviewed-by: Nick Desaulniers <ndesaulniers@google.com> Link: https://github.com/ClangBuiltLinux/linux/issues/563 Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
225 lines
5.6 KiB
C
225 lines
5.6 KiB
C
/* SPDX-License-Identifier: GPL-2.0-only */
|
|
/*
|
|
* arch/arm/include/asm/xor.h
|
|
*
|
|
* Copyright (C) 2001 Russell King
|
|
*/
|
|
#include <linux/hardirq.h>
|
|
#include <asm-generic/xor.h>
|
|
#include <asm/hwcap.h>
|
|
#include <asm/neon.h>
|
|
|
|
#define __XOR(a1, a2) a1 ^= a2
|
|
|
|
#define GET_BLOCK_2(dst) \
|
|
__asm__("ldmia %0, {%1, %2}" \
|
|
: "=r" (dst), "=r" (a1), "=r" (a2) \
|
|
: "0" (dst))
|
|
|
|
#define GET_BLOCK_4(dst) \
|
|
__asm__("ldmia %0, {%1, %2, %3, %4}" \
|
|
: "=r" (dst), "=r" (a1), "=r" (a2), "=r" (a3), "=r" (a4) \
|
|
: "0" (dst))
|
|
|
|
#define XOR_BLOCK_2(src) \
|
|
__asm__("ldmia %0!, {%1, %2}" \
|
|
: "=r" (src), "=r" (b1), "=r" (b2) \
|
|
: "0" (src)); \
|
|
__XOR(a1, b1); __XOR(a2, b2);
|
|
|
|
#define XOR_BLOCK_4(src) \
|
|
__asm__("ldmia %0!, {%1, %2, %3, %4}" \
|
|
: "=r" (src), "=r" (b1), "=r" (b2), "=r" (b3), "=r" (b4) \
|
|
: "0" (src)); \
|
|
__XOR(a1, b1); __XOR(a2, b2); __XOR(a3, b3); __XOR(a4, b4)
|
|
|
|
#define PUT_BLOCK_2(dst) \
|
|
__asm__ __volatile__("stmia %0!, {%2, %3}" \
|
|
: "=r" (dst) \
|
|
: "0" (dst), "r" (a1), "r" (a2))
|
|
|
|
#define PUT_BLOCK_4(dst) \
|
|
__asm__ __volatile__("stmia %0!, {%2, %3, %4, %5}" \
|
|
: "=r" (dst) \
|
|
: "0" (dst), "r" (a1), "r" (a2), "r" (a3), "r" (a4))
|
|
|
|
static void
|
|
xor_arm4regs_2(unsigned long bytes, unsigned long * __restrict p1,
|
|
const unsigned long * __restrict p2)
|
|
{
|
|
unsigned int lines = bytes / sizeof(unsigned long) / 4;
|
|
register unsigned int a1 __asm__("r4");
|
|
register unsigned int a2 __asm__("r5");
|
|
register unsigned int a3 __asm__("r6");
|
|
register unsigned int a4 __asm__("r7");
|
|
register unsigned int b1 __asm__("r8");
|
|
register unsigned int b2 __asm__("r9");
|
|
register unsigned int b3 __asm__("ip");
|
|
register unsigned int b4 __asm__("lr");
|
|
|
|
do {
|
|
GET_BLOCK_4(p1);
|
|
XOR_BLOCK_4(p2);
|
|
PUT_BLOCK_4(p1);
|
|
} while (--lines);
|
|
}
|
|
|
|
static void
|
|
xor_arm4regs_3(unsigned long bytes, unsigned long * __restrict p1,
|
|
const unsigned long * __restrict p2,
|
|
const unsigned long * __restrict p3)
|
|
{
|
|
unsigned int lines = bytes / sizeof(unsigned long) / 4;
|
|
register unsigned int a1 __asm__("r4");
|
|
register unsigned int a2 __asm__("r5");
|
|
register unsigned int a3 __asm__("r6");
|
|
register unsigned int a4 __asm__("r7");
|
|
register unsigned int b1 __asm__("r8");
|
|
register unsigned int b2 __asm__("r9");
|
|
register unsigned int b3 __asm__("ip");
|
|
register unsigned int b4 __asm__("lr");
|
|
|
|
do {
|
|
GET_BLOCK_4(p1);
|
|
XOR_BLOCK_4(p2);
|
|
XOR_BLOCK_4(p3);
|
|
PUT_BLOCK_4(p1);
|
|
} while (--lines);
|
|
}
|
|
|
|
static void
|
|
xor_arm4regs_4(unsigned long bytes, unsigned long * __restrict p1,
|
|
const unsigned long * __restrict p2,
|
|
const unsigned long * __restrict p3,
|
|
const unsigned long * __restrict p4)
|
|
{
|
|
unsigned int lines = bytes / sizeof(unsigned long) / 2;
|
|
register unsigned int a1 __asm__("r8");
|
|
register unsigned int a2 __asm__("r9");
|
|
register unsigned int b1 __asm__("ip");
|
|
register unsigned int b2 __asm__("lr");
|
|
|
|
do {
|
|
GET_BLOCK_2(p1);
|
|
XOR_BLOCK_2(p2);
|
|
XOR_BLOCK_2(p3);
|
|
XOR_BLOCK_2(p4);
|
|
PUT_BLOCK_2(p1);
|
|
} while (--lines);
|
|
}
|
|
|
|
static void
|
|
xor_arm4regs_5(unsigned long bytes, unsigned long * __restrict p1,
|
|
const unsigned long * __restrict p2,
|
|
const unsigned long * __restrict p3,
|
|
const unsigned long * __restrict p4,
|
|
const unsigned long * __restrict p5)
|
|
{
|
|
unsigned int lines = bytes / sizeof(unsigned long) / 2;
|
|
register unsigned int a1 __asm__("r8");
|
|
register unsigned int a2 __asm__("r9");
|
|
register unsigned int b1 __asm__("ip");
|
|
register unsigned int b2 __asm__("lr");
|
|
|
|
do {
|
|
GET_BLOCK_2(p1);
|
|
XOR_BLOCK_2(p2);
|
|
XOR_BLOCK_2(p3);
|
|
XOR_BLOCK_2(p4);
|
|
XOR_BLOCK_2(p5);
|
|
PUT_BLOCK_2(p1);
|
|
} while (--lines);
|
|
}
|
|
|
|
static struct xor_block_template xor_block_arm4regs = {
|
|
.name = "arm4regs",
|
|
.do_2 = xor_arm4regs_2,
|
|
.do_3 = xor_arm4regs_3,
|
|
.do_4 = xor_arm4regs_4,
|
|
.do_5 = xor_arm4regs_5,
|
|
};
|
|
|
|
#undef XOR_TRY_TEMPLATES
|
|
#define XOR_TRY_TEMPLATES \
|
|
do { \
|
|
xor_speed(&xor_block_arm4regs); \
|
|
xor_speed(&xor_block_8regs); \
|
|
xor_speed(&xor_block_32regs); \
|
|
NEON_TEMPLATES; \
|
|
} while (0)
|
|
|
|
#ifdef CONFIG_KERNEL_MODE_NEON
|
|
|
|
extern struct xor_block_template const xor_block_neon_inner;
|
|
|
|
static void
|
|
xor_neon_2(unsigned long bytes, unsigned long * __restrict p1,
|
|
const unsigned long * __restrict p2)
|
|
{
|
|
if (in_interrupt()) {
|
|
xor_arm4regs_2(bytes, p1, p2);
|
|
} else {
|
|
kernel_neon_begin();
|
|
xor_block_neon_inner.do_2(bytes, p1, p2);
|
|
kernel_neon_end();
|
|
}
|
|
}
|
|
|
|
static void
|
|
xor_neon_3(unsigned long bytes, unsigned long * __restrict p1,
|
|
const unsigned long * __restrict p2,
|
|
const unsigned long * __restrict p3)
|
|
{
|
|
if (in_interrupt()) {
|
|
xor_arm4regs_3(bytes, p1, p2, p3);
|
|
} else {
|
|
kernel_neon_begin();
|
|
xor_block_neon_inner.do_3(bytes, p1, p2, p3);
|
|
kernel_neon_end();
|
|
}
|
|
}
|
|
|
|
static void
|
|
xor_neon_4(unsigned long bytes, unsigned long * __restrict p1,
|
|
const unsigned long * __restrict p2,
|
|
const unsigned long * __restrict p3,
|
|
const unsigned long * __restrict p4)
|
|
{
|
|
if (in_interrupt()) {
|
|
xor_arm4regs_4(bytes, p1, p2, p3, p4);
|
|
} else {
|
|
kernel_neon_begin();
|
|
xor_block_neon_inner.do_4(bytes, p1, p2, p3, p4);
|
|
kernel_neon_end();
|
|
}
|
|
}
|
|
|
|
static void
|
|
xor_neon_5(unsigned long bytes, unsigned long * __restrict p1,
|
|
const unsigned long * __restrict p2,
|
|
const unsigned long * __restrict p3,
|
|
const unsigned long * __restrict p4,
|
|
const unsigned long * __restrict p5)
|
|
{
|
|
if (in_interrupt()) {
|
|
xor_arm4regs_5(bytes, p1, p2, p3, p4, p5);
|
|
} else {
|
|
kernel_neon_begin();
|
|
xor_block_neon_inner.do_5(bytes, p1, p2, p3, p4, p5);
|
|
kernel_neon_end();
|
|
}
|
|
}
|
|
|
|
static struct xor_block_template xor_block_neon = {
|
|
.name = "neon",
|
|
.do_2 = xor_neon_2,
|
|
.do_3 = xor_neon_3,
|
|
.do_4 = xor_neon_4,
|
|
.do_5 = xor_neon_5
|
|
};
|
|
|
|
#define NEON_TEMPLATES \
|
|
do { if (cpu_has_neon()) xor_speed(&xor_block_neon); } while (0)
|
|
#else
|
|
#define NEON_TEMPLATES
|
|
#endif
|