1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188
|
/* This file is part of the gf2x library.
Copyright 2007, 2008, 2009, 2010, 2013, 2015
Richard Brent, Pierrick Gaudry, Emmanuel Thome', Paul Zimmermann
This program is free software; you can redistribute it and/or modify it
under the terms of either:
- If the archive contains a file named toom-gpl.c (not a trivial
placeholder), the GNU General Public License as published by the Free
Software Foundation; either version 3 of the License, or (at your
option) any later version.
- If the archive contains a file named toom-gpl.c which is a trivial
placeholder, the GNU Lesser General Public License as published by
the Free Software Foundation; either version 2.1 of the License, or
(at your option) any later version.
This program is distributed in the hope that it will be useful, but WITHOUT
ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
FITNESS FOR A PARTICULAR PURPOSE. See the license text for more details.
You should have received a copy of the GNU General Public License as
well as the GNU Lesser General Public License along with this program;
see the files COPYING and COPYING.LIB. If not, write to the Free
Software Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA
02110-1301, USA.
*/
/* Implements 96x96 -> 192 bit product using SSE2 instructions. */
#ifndef GF2X_MUL3_H_
#define GF2X_MUL3_H_
#include "gf2x.h"
/* All gf2x source files for lowlevel functions must include gf2x-small.h
* This is mandatory for the tuning mechanism. */
#include "gf2x/gf2x-small.h"
#if GF2X_WORDSIZE != 32
#error "This code is for 32-bit only"
#endif
#ifndef GF2X_HAVE_SSE2_SUPPORT
#error "This code needs sse-2 support"
#endif
#if defined(__GNUC__) && __GNUC__ == 4 && __GNUC_MINOR__ == 3 && (__GNUC_PATCHLEVEL__ == 0 || __GNUC_PATCHLEVEL__ == 1)
#warning "Your GCC version is buggy. Binary fields may fail randomly"
/* Gcc bug reports 37101 and 37340 -- the only convenient fix is to
* upgrade to 4.3.2 */
#endif
/* mul4t is so fast on pentium-4s that it's even faster than mul3k. So we
* provide a stripped-down version (much similar to mul4 of course) */
/* This has been edited without testing */
GF2X_STORAGE_CLASS_mul3
void gf2x_mul3(unsigned long *t, unsigned long const *s1,
unsigned long const *s2)
{
#define SHL(x, r) _mm_slli_epi64((x), (r))
#define SHR(x, r) _mm_srli_epi64((x), (r))
#define SHLD(x, r) _mm_slli_si128((x), (r) >> 3)
#define SHRD(x, r) _mm_srli_si128((x), (r) >> 3)
#define PZERO _mm_setzero_si128()
#define PADD(x, y) _mm_add_epi64((x), (y))
#define PSUB(x, y) _mm_sub_epi64((x), (y))
#define PNEG(x) PSUB(PZERO, (x))
#define PXOR(lop, rop) _mm_xor_si128((lop), (rop))
#define XOREQ(lop, rop) lop = _mm_xor_si128((lop), (rop))
#define PAND(lop, rop) _mm_and_si128((lop), (rop))
__m128i u;
__m128i t0;
__m128i t1;
__m128i t2;
__m128i g[16];
/* sequence update walk */
g[0] = _mm_setzero_si128();
g[1] = _gf2x_mm_setr_epi32(s2[0], s2[1], s2[2], 0);
g[2] = SHL(g[1], 1);
g[3] = PXOR(g[2], g[1]);
g[4] = SHL(g[2], 1);
g[5] = PXOR(g[4], g[1]);
g[6] = SHL(g[3], 1);
g[7] = PXOR(g[6], g[1]);
g[8] = SHL(g[4], 1);
g[9] = PXOR(g[8], g[1]);
g[10] = SHL(g[5], 1);
g[11] = PXOR(g[10], g[1]);
g[12] = SHL(g[6], 1);
g[13] = PXOR(g[12], g[1]);
g[14] = SHL(g[7], 1);
g[15] = PXOR(g[14], g[1]);
/* round 0 */
u = g[s1[0] & 15];
t0 = u;
u = g[s1[0] >> 4 & 15];
XOREQ(t0, SHL(u, 4)); t1 = SHR(u, 60);
u = g[s1[0] >> 8 & 15];
XOREQ(t0, SHL(u, 8)); XOREQ(t1, SHR(u, 56));
u = g[s1[0] >> 12 & 15];
XOREQ(t0, SHL(u, 12)); XOREQ(t1, SHR(u, 52));
u = g[s1[0] >> 16 & 15];
XOREQ(t0, SHL(u, 16)); XOREQ(t1, SHR(u, 48));
u = g[s1[0] >> 20 & 15];
XOREQ(t0, SHL(u, 20)); XOREQ(t1, SHR(u, 44));
u = g[s1[0] >> 24 & 15];
XOREQ(t0, SHL(u, 24)); XOREQ(t1, SHR(u, 40));
u = g[s1[0] >> 28 & 15];
XOREQ(t0, SHL(u, 28)); XOREQ(t1, SHR(u, 36));
u = g[s1[1] & 15];
XOREQ(t0, SHL(u, 32)); XOREQ(t1, SHR(u, 32));
u = g[s1[1] >> 4 & 15];
XOREQ(t0, SHL(u, 36)); XOREQ(t1, SHR(u, 28));
u = g[s1[1] >> 8 & 15];
XOREQ(t0, SHL(u, 40)); XOREQ(t1, SHR(u, 24));
u = g[s1[1] >> 12 & 15];
XOREQ(t0, SHL(u, 44)); XOREQ(t1, SHR(u, 20));
u = g[s1[1] >> 16 & 15];
XOREQ(t0, SHL(u, 48)); XOREQ(t1, SHR(u, 16));
u = g[s1[1] >> 20 & 15];
XOREQ(t0, SHL(u, 52)); XOREQ(t1, SHR(u, 12));
u = g[s1[1] >> 24 & 15];
XOREQ(t0, SHL(u, 56)); XOREQ(t1, SHR(u, 8));
u = g[s1[1] >> 28 & 15];
XOREQ(t0, SHL(u, 60)); XOREQ(t1, SHR(u, 4));
/* round 1 */
u = g[s1[2] & 15];
XOREQ(t1, u);
u = g[s1[2] >> 4 & 15];
XOREQ(t1, SHL(u, 4)); t2 = SHR(u, 60);
u = g[s1[2] >> 8 & 15];
XOREQ(t1, SHL(u, 8)); XOREQ(t2, SHR(u, 56));
u = g[s1[2] >> 12 & 15];
XOREQ(t1, SHL(u, 12)); XOREQ(t2, SHR(u, 52));
u = g[s1[2] >> 16 & 15];
XOREQ(t1, SHL(u, 16)); XOREQ(t2, SHR(u, 48));
u = g[s1[2] >> 20 & 15];
XOREQ(t1, SHL(u, 20)); XOREQ(t2, SHR(u, 44));
u = g[s1[2] >> 24 & 15];
XOREQ(t1, SHL(u, 24)); XOREQ(t2, SHR(u, 40));
u = g[s1[2] >> 28 & 15];
XOREQ(t1, SHL(u, 28)); XOREQ(t2, SHR(u, 36));
/* end */
/* repair steps */
/* repair section 200711-200803 */
__m128i v1 = SHR(_gf2x_mm_setr_epi32(s1[0], s1[1], s1[0], s1[1]), 1);
__m128i v2 = SHR(_gf2x_mm_setr_epi32(s1[2], 0, s1[2], 0), 1);
__m128i m = _gf2x_mm_set1_epi32_c(0x77777777);
__m128i w = PNEG(SHR(g[1],63));
v1 = PAND(v1, m);
XOREQ(t1, PAND(v1, w));
v2 = PAND(v2, m);
XOREQ(t2, PAND(v2, w));
w = PNEG(SHR(g[2],63));
v1 = SHR(v1, 1) & m;
XOREQ(t1, PAND(v1, w));
v2 = SHR(v2, 1) & m;
XOREQ(t2, PAND(v2, w));
w = PNEG(SHR(g[4],63));
v1 = SHR(v1, 1) & m;
XOREQ(t1, PAND(v1, w));
v2 = SHR(v2, 1) & m;
XOREQ(t2, PAND(v2, w));
/* store result */
_mm_storeu_si128((__m128i *)(t), PXOR(t0, _mm_unpacklo_epi64(PZERO, t1)));
XOREQ(t2, _mm_unpackhi_epi64(t1, PZERO));
t[4] = _mm_cvtsi128_si32(t2);
t[5] = _mm_cvtsi128_si32(SHRD(t2, 32));
#undef PAND
#undef XOREQ
#undef PZERO
#undef PADD
#undef PSUB
#undef PNEG
#undef PXOR
#undef SHL
#undef SHR
#undef SHLD
#undef SHRD
}
#endif /* GF2X_MUL3_H_ */
|