dnl AMD64 SSSE3/XOP mpn_hamdist -- hamming distance. dnl Copyright 2010-2017 Free Software Foundation, Inc. dnl This file is part of the GNU MP Library. dnl dnl The GNU MP Library is free software; you can redistribute it and/or modify dnl it under the terms of either: dnl dnl * the GNU Lesser General Public License as published by the Free dnl Software Foundation; either version 3 of the License, or (at your dnl option) any later version. dnl dnl or dnl dnl * the GNU General Public License as published by the Free Software dnl Foundation; either version 2 of the License, or (at your option) any dnl later version. dnl dnl or both in parallel, as here. dnl dnl The GNU MP Library is distributed in the hope that it will be useful, but dnl WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY dnl or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License dnl for more details. dnl dnl You should have received copies of the GNU General Public License and the dnl GNU Lesser General Public License along with the GNU MP Library. If not, dnl see https://www.gnu.org/licenses/. include(`../config.m4') C cycles/limb good for cpu? C AMD K8,K9 n/a C AMD K10 n/a C AMD bd1 1.51-2.0 y C AMD bd2 1.50-1.9 y C AMD bd3 ? C AMD bd4 ? C AMD zen n/a C AMD bobcat n/a C AMD jaguar n/a C Intel P4 n/a C Intel PNR n/a C Intel NHM n/a C Intel SBR n/a C Intel IBR n/a C Intel HWL n/a C Intel BWL n/a C Intel SKL n/a C Intel atom n/a C Intel SLM n/a C VIA nano n/a C TODO C * We need to use .byte for vpshlb, vpperm, vphaddubq, and all popcnt if we C intend to support old systems. C We use vpshlb and vpperm below, which are XOP extensions to AVX. Some C systems, e.g., NetBSD, set OSXSAVE but nevertheless trigger SIGILL for AVX. C We fall back to the core2 code. ifdef(`GMP_AVX_NOT_REALLY_AVAILABLE',` MULFUNC_PROLOGUE(mpn_hamdist) include_mpn(`x86_64/core2/hamdist.asm') ',` define(`up', `%rdi') define(`vp', `%rsi') define(`n', `%rdx') ABI_SUPPORT(DOS64) ABI_SUPPORT(STD64) ASM_START() TEXT ALIGN(32) PROLOGUE(mpn_hamdist) FUNC_ENTRY(3) cmp $5, n jl L(sma) lea L(cnsts)(%rip), %r9 xor R32(%r10), R32(%r10) test $8, R8(vp) jz L(ali) mov (up), %r8 xor (vp), %r8 add $8, up add $8, vp dec n popcnt %r8, %r10 L(ali): ifdef(`PIC', `define(`OFF1',16) define(`OFF2',32) define(`OFF3',48)', `define(`OFF1',32) define(`OFF2',48) define(`OFF3',64)') movdqa OFF1`'(%r9), %xmm7 C nibble counts table movdqa OFF2`'(%r9), %xmm6 C splat shift counts movdqa OFF3`'(%r9), %xmm5 C masks pxor %xmm4, %xmm4 pxor %xmm8, %xmm8 C grand total count mov R32(n), R32(%rax) and $6, R32(%rax) lea -64(up,%rax,8), up lea -64(vp,%rax,8), vp ifdef(`PIC',` movslq (%r9,%rax,2), %r11 add %r9, %r11 jmp *%r11 ',` jmp *(%r9,%rax,4) ') L(0): add $64, up add $64, vp sub $2, n ALIGN(32) L(top): lddqu (up), %xmm0 pxor (vp), %xmm0 .byte 0x8f,0xe9,0x48,0x94,0xc8 C vpshlb %xmm6, %xmm0, %xmm1 pand %xmm5, %xmm0 pand %xmm5, %xmm1 .byte 0x8f,0xe8,0x40,0xa3,0xd7,0x00 C vpperm %xmm0,%xmm7,%xmm7,%xmm2 .byte 0x8f,0xe8,0x40,0xa3,0xdf,0x10 C vpperm %xmm1,%xmm7,%xmm7,%xmm3 paddb %xmm2, %xmm3 paddb %xmm3, %xmm4 L(6): lddqu 16(up), %xmm0 pxor 16(vp), %xmm0 .byte 0x8f,0xe9,0x48,0x94,0xc8 C vpshlb %xmm6, %xmm0, %xmm1 pand %xmm5, %xmm0 pand %xmm5, %xmm1 .byte 0x8f,0xe8,0x40,0xa3,0xd7,0x00 C vpperm %xmm0,%xmm7,%xmm7,%xmm2 .byte 0x8f,0xe8,0x40,0xa3,0xdf,0x10 C vpperm %xmm1,%xmm7,%xmm7,%xmm3 paddb %xmm2, %xmm3 paddb %xmm3, %xmm4 L(4): lddqu 32(up), %xmm0 pxor 32(vp), %xmm0 .byte 0x8f,0xe9,0x48,0x94,0xc8 C vpshlb %xmm6, %xmm0, %xmm1 pand %xmm5, %xmm0 pand %xmm5, %xmm1 .byte 0x8f,0xe8,0x40,0xa3,0xd7,0x00 C vpperm %xmm0,%xmm7,%xmm7,%xmm2 .byte 0x8f,0xe9,0x78,0xd3,0xc4 C vphaddubq %xmm4, %xmm0 .byte 0x8f,0xe8,0x40,0xa3,0xe7,0x10 C vpperm %xmm1,%xmm7,%xmm7,%xmm4 paddb %xmm2, %xmm3 paddb %xmm2, %xmm4 paddq %xmm0, %xmm8 C sum to 2 x 64-bit counts L(2): mov 48(up), %r8 mov 56(up), %r9 add $64, up xor 48(vp), %r8 xor 56(vp), %r9 add $64, vp popcnt %r8, %r8 popcnt %r9, %r9 add %r8, %r10 add %r9, %r10 sub $8, n jg L(top) test $1, R8(n) jz L(x) mov (up), %r8 xor (vp), %r8 popcnt %r8, %r8 add %r8, %r10 L(x): .byte 0x8f,0xe9,0x78,0xd3,0xc4 C vphaddubq %xmm4, %xmm0 paddq %xmm0, %xmm8 pshufd $14, %xmm8, %xmm0 paddq %xmm8, %xmm0 movd %xmm0, %rax add %r10, %rax FUNC_EXIT() ret L(sma): mov (up), %r8 xor (vp), %r8 popcnt %r8, %rax dec n jz L(ed) L(tp): mov 8(up), %r8 add $8, up xor 8(vp), %r8 add $8, vp popcnt %r8, %r8 add %r8, %rax dec n jnz L(tp) L(ed): FUNC_EXIT() ret EPILOGUE() DEF_OBJECT(L(cnsts),16,`JUMPTABSECT') JMPENT( L(0), L(cnsts)) JMPENT( L(2), L(cnsts)) JMPENT( L(4), L(cnsts)) JMPENT( L(6), L(cnsts)) .byte 0x00,0x01,0x01,0x02,0x01,0x02,0x02,0x03 .byte 0x01,0x02,0x02,0x03,0x02,0x03,0x03,0x04 .byte -4,-4,-4,-4,-4,-4,-4,-4 .byte -4,-4,-4,-4,-4,-4,-4,-4 .byte 0x0f,0x0f,0x0f,0x0f,0x0f,0x0f,0x0f,0x0f .byte 0x0f,0x0f,0x0f,0x0f,0x0f,0x0f,0x0f,0x0f END_OBJECT(L(cnsts)) ')