svn commit: r358688 - head/lib/libc/powerpc64/string
Justin Hibbits
jhibbits at FreeBSD.org
Fri Mar 6 01:45:04 UTC 2020
Author: jhibbits
Date: Fri Mar 6 01:45:03 2020
New Revision: 358688
URL: https://svnweb.freebsd.org/changeset/base/358688
Log:
powerpc/powerpc64: Enforce natural alignment in memcpy
Summary:
POWER architecture CPUs (Book-S) require natural alignment for
cache-inhibited storage accesses. Since we can't know the caching model
for a page ahead of time, always enforce natural alignment in memcpy.
This fixes a SIGBUS in X with acceleration enabled on POWER9.
As part of this, revert r358672, it's no longer necessary with this fix.
Regression tested by alfredo.
Reviewed by: alfredo
Differential Revision: https://reviews.freebsd.org/D23969
Modified:
head/lib/libc/powerpc64/string/bcopy_resolver.c
head/lib/libc/powerpc64/string/memcpy.S
head/lib/libc/powerpc64/string/memcpy_vsx.S
Modified: head/lib/libc/powerpc64/string/bcopy_resolver.c
==============================================================================
--- head/lib/libc/powerpc64/string/bcopy_resolver.c Thu Mar 5 22:45:16 2020 (r358687)
+++ head/lib/libc/powerpc64/string/bcopy_resolver.c Fri Mar 6 01:45:03 2020 (r358688)
@@ -66,7 +66,7 @@ DEFINE_UIFUNC(, FN_RET, FN_NAME, FN_PARAMS)
* Since POWER ISA 2.07B this is solved transparently
* by the hardware
*/
- if (cpu_features2 & PPC_FEATURE2_ARCH_2_07)
+ if (cpu_features2 & PPC_FEATURE_HAS_VSX)
return (FN_NAME_VSX);
else
return (FN_NAME_NOVSX);
Modified: head/lib/libc/powerpc64/string/memcpy.S
==============================================================================
--- head/lib/libc/powerpc64/string/memcpy.S Thu Mar 5 22:45:16 2020 (r358687)
+++ head/lib/libc/powerpc64/string/memcpy.S Fri Mar 6 01:45:03 2020 (r358688)
@@ -39,6 +39,11 @@ WEAK_REFERENCE(__memcpy, memcpy);
#define BLOCK_BYTES (1 << BLOCK_BITS)
#define BLOCK_MASK (BLOCK_BYTES - 1)
+/* Minimum 8 byte alignment, to avoid cache-inhibited alignment faults. */
+#ifndef ALIGN_MASK
+#define ALIGN_MASK 0x7
+#endif
+
/*
* r3: dst
* r4: src
@@ -48,6 +53,12 @@ ENTRY(FN_NAME)
cmpdi %r5, 0 /* len == 0? nothing to do */
beqlr-
+ /* If src and dst are relatively misaligned, do byte copies. */
+ andi. %r8, %r3, ALIGN_MASK
+ andi. %r7, %r4, ALIGN_MASK
+ cmpd %r8, %r7
+ mr %r7, %r5
+ bne+ .Lcopy_remaining_fix_index_byte
mr %r8, %r3 /* save dst */
/* align src */
Modified: head/lib/libc/powerpc64/string/memcpy_vsx.S
==============================================================================
--- head/lib/libc/powerpc64/string/memcpy_vsx.S Thu Mar 5 22:45:16 2020 (r358687)
+++ head/lib/libc/powerpc64/string/memcpy_vsx.S Fri Mar 6 01:45:03 2020 (r358688)
@@ -30,6 +30,7 @@
#define FN_NAME __memcpy_vsx
#define BLOCK_BITS 6
+#define ALIGN_MASK 0xf
/*
* r5: bytes to copy (multiple of BLOCK_BYTES)
More information about the svn-src-head
mailing list