Chris@10: /* Chris@10: * Copyright (c) 2003, 2007-11 Matteo Frigo Chris@10: * Copyright (c) 2003, 2007-11 Massachusetts Institute of Technology Chris@10: * Chris@10: * This program is free software; you can redistribute it and/or modify Chris@10: * it under the terms of the GNU General Public License as published by Chris@10: * the Free Software Foundation; either version 2 of the License, or Chris@10: * (at your option) any later version. Chris@10: * Chris@10: * This program is distributed in the hope that it will be useful, Chris@10: * but WITHOUT ANY WARRANTY; without even the implied warranty of Chris@10: * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the Chris@10: * GNU General Public License for more details. Chris@10: * Chris@10: * You should have received a copy of the GNU General Public License Chris@10: * along with this program; if not, write to the Free Software Chris@10: * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA Chris@10: * Chris@10: */ Chris@10: Chris@10: /* This file was automatically generated --- DO NOT EDIT */ Chris@10: /* Generated on Sun Nov 25 07:39:05 EST 2012 */ Chris@10: Chris@10: #include "codelet-dft.h" Chris@10: Chris@10: #ifdef HAVE_FMA Chris@10: Chris@10: /* Generated by: ../../../genfft/gen_twiddle_c.native -fma -reorder-insns -schedule-for-pipeline -simd -compact -variables 4 -pipeline-latency 8 -n 64 -name t1bv_64 -include t1b.h -sign 1 */ Chris@10: Chris@10: /* Chris@10: * This function contains 519 FP additions, 384 FP multiplications, Chris@10: * (or, 261 additions, 126 multiplications, 258 fused multiply/add), Chris@10: * 187 stack variables, 15 constants, and 128 memory accesses Chris@10: */ Chris@10: #include "t1b.h" Chris@10: Chris@10: static void t1bv_64(R *ri, R *ii, const R *W, stride rs, INT mb, INT me, INT ms) Chris@10: { Chris@10: DVK(KP773010453, +0.773010453362736960810906609758469800971041293); Chris@10: DVK(KP995184726, +0.995184726672196886244836953109479921575474869); Chris@10: DVK(KP820678790, +0.820678790828660330972281985331011598767386482); Chris@10: DVK(KP098491403, +0.098491403357164253077197521291327432293052451); Chris@10: DVK(KP956940335, +0.956940335732208864935797886980269969482849206); Chris@10: DVK(KP881921264, +0.881921264348355029712756863660388349508442621); Chris@10: DVK(KP303346683, +0.303346683607342391675883946941299872384187453); Chris@10: DVK(KP534511135, +0.534511135950791641089685961295362908582039528); Chris@10: DVK(KP831469612, +0.831469612302545237078788377617905756738560812); Chris@10: DVK(KP980785280, +0.980785280403230449126182236134239036973933731); Chris@10: DVK(KP668178637, +0.668178637919298919997757686523080761552472251); Chris@10: DVK(KP198912367, +0.198912367379658006911597622644676228597850501); Chris@10: DVK(KP923879532, +0.923879532511286756128183189396788286822416626); Chris@10: DVK(KP414213562, +0.414213562373095048801688724209698078569671875); Chris@10: DVK(KP707106781, +0.707106781186547524400844362104849039284835938); Chris@10: { Chris@10: INT m; Chris@10: R *x; Chris@10: x = ii; Chris@10: for (m = mb, W = W + (mb * ((TWVL / VL) * 126)); m < me; m = m + VL, x = x + (VL * ms), W = W + (TWVL * 126), MAKE_VOLATILE_STRIDE(64, rs)) { Chris@10: V T6L, T6M, T6O, T6P, T75, T6V, T5A, T6A, T72, T6K, T6t, T6D, T6w, T6B, T6h; Chris@10: V T6E; Chris@10: { Chris@10: V Ta, T3U, T3V, T37, T7a, T58, T7B, T6l, T1v, T24, T5Q, T7o, T5F, T7l, T43; Chris@10: V T4F, T2i, T2R, T6b, T7v, T60, T7s, T4a, T4I, T5u, T7h, T5x, T7g, T1i, T3b; Chris@10: V T4m, T4C, T7e, T5l, T7d, T5o, T3a, TV, T4B, T4j, T3X, T3Y, T6o, T7b, T5f; Chris@10: V T7C, Tx, T38, T2p, T61, T2n, T65, T2D, T7p, T5M, T7m, T5T, T4G, T46, T25; Chris@10: V T1S, T2q, T2u, T2w; Chris@10: { Chris@10: V T5q, T10, T5v, T15, T1b, T5s, T1c, T1e; Chris@10: { Chris@10: V T1V, T1p, T5B, T5O, T1u, T1X, T20, T21; Chris@10: { Chris@10: V T1, T2, T7, T5, T32, T34, T2X, T2Z; Chris@10: T1 = LD(&(x[0]), ms, &(x[0])); Chris@10: T2 = LD(&(x[WS(rs, 32)]), ms, &(x[0])); Chris@10: T7 = LD(&(x[WS(rs, 48)]), ms, &(x[0])); Chris@10: T5 = LD(&(x[WS(rs, 16)]), ms, &(x[0])); Chris@10: T32 = LD(&(x[WS(rs, 56)]), ms, &(x[0])); Chris@10: T34 = LD(&(x[WS(rs, 24)]), ms, &(x[0])); Chris@10: T2X = LD(&(x[WS(rs, 8)]), ms, &(x[0])); Chris@10: T2Z = LD(&(x[WS(rs, 40)]), ms, &(x[0])); Chris@10: { Chris@10: V T1m, T54, T6j, T36, T56, T31, T55, T1n, T1q, T1s, T4, T9; Chris@10: { Chris@10: V T3, T8, T6, T33, T35, T2Y, T30, T1l; Chris@10: T1l = LD(&(x[WS(rs, 1)]), ms, &(x[WS(rs, 1)])); Chris@10: T3 = BYTW(&(W[TWVL * 62]), T2); Chris@10: T8 = BYTW(&(W[TWVL * 94]), T7); Chris@10: T6 = BYTW(&(W[TWVL * 30]), T5); Chris@10: T33 = BYTW(&(W[TWVL * 110]), T32); Chris@10: T35 = BYTW(&(W[TWVL * 46]), T34); Chris@10: T2Y = BYTW(&(W[TWVL * 14]), T2X); Chris@10: T30 = BYTW(&(W[TWVL * 78]), T2Z); Chris@10: T1m = BYTW(&(W[0]), T1l); Chris@10: T54 = VSUB(T1, T3); Chris@10: T4 = VADD(T1, T3); Chris@10: T6j = VSUB(T6, T8); Chris@10: T9 = VADD(T6, T8); Chris@10: T36 = VADD(T33, T35); Chris@10: T56 = VSUB(T33, T35); Chris@10: T31 = VADD(T2Y, T30); Chris@10: T55 = VSUB(T2Y, T30); Chris@10: T1n = LD(&(x[WS(rs, 33)]), ms, &(x[WS(rs, 1)])); Chris@10: } Chris@10: T1q = LD(&(x[WS(rs, 17)]), ms, &(x[WS(rs, 1)])); Chris@10: T1s = LD(&(x[WS(rs, 49)]), ms, &(x[WS(rs, 1)])); Chris@10: Ta = VSUB(T4, T9); Chris@10: T3U = VADD(T4, T9); Chris@10: { Chris@10: V T57, T6k, T1o, T1r, T1t, T1W, T1U, T1Z; Chris@10: T1U = LD(&(x[WS(rs, 9)]), ms, &(x[WS(rs, 1)])); Chris@10: T3V = VADD(T31, T36); Chris@10: T37 = VSUB(T31, T36); Chris@10: T57 = VADD(T55, T56); Chris@10: T6k = VSUB(T55, T56); Chris@10: T1o = BYTW(&(W[TWVL * 64]), T1n); Chris@10: T1r = BYTW(&(W[TWVL * 32]), T1q); Chris@10: T1t = BYTW(&(W[TWVL * 96]), T1s); Chris@10: T1V = BYTW(&(W[TWVL * 16]), T1U); Chris@10: T1W = LD(&(x[WS(rs, 41)]), ms, &(x[WS(rs, 1)])); Chris@10: T1Z = LD(&(x[WS(rs, 57)]), ms, &(x[WS(rs, 1)])); Chris@10: T7a = VFNMS(LDK(KP707106781), T57, T54); Chris@10: T58 = VFMA(LDK(KP707106781), T57, T54); Chris@10: T7B = VFNMS(LDK(KP707106781), T6k, T6j); Chris@10: T6l = VFMA(LDK(KP707106781), T6k, T6j); Chris@10: T1p = VADD(T1m, T1o); Chris@10: T5B = VSUB(T1m, T1o); Chris@10: T5O = VSUB(T1r, T1t); Chris@10: T1u = VADD(T1r, T1t); Chris@10: T1X = BYTW(&(W[TWVL * 80]), T1W); Chris@10: T20 = BYTW(&(W[TWVL * 112]), T1Z); Chris@10: T21 = LD(&(x[WS(rs, 25)]), ms, &(x[WS(rs, 1)])); Chris@10: } Chris@10: } Chris@10: } Chris@10: { Chris@10: V T5W, T2N, T69, T2L, T5Y, T2P, T48, T2c, T2h; Chris@10: { Chris@10: V T41, T1Y, T5C, T22, T2d, T29, T2b, T2f, T28, T2a, T2H, T2J; Chris@10: T28 = LD(&(x[WS(rs, 63)]), ms, &(x[WS(rs, 1)])); Chris@10: T2a = LD(&(x[WS(rs, 31)]), ms, &(x[WS(rs, 1)])); Chris@10: T1v = VSUB(T1p, T1u); Chris@10: T41 = VADD(T1p, T1u); Chris@10: T1Y = VADD(T1V, T1X); Chris@10: T5C = VSUB(T1V, T1X); Chris@10: T22 = BYTW(&(W[TWVL * 48]), T21); Chris@10: T2d = LD(&(x[WS(rs, 15)]), ms, &(x[WS(rs, 1)])); Chris@10: T29 = BYTW(&(W[TWVL * 124]), T28); Chris@10: T2b = BYTW(&(W[TWVL * 60]), T2a); Chris@10: T2f = LD(&(x[WS(rs, 47)]), ms, &(x[WS(rs, 1)])); Chris@10: T2H = LD(&(x[WS(rs, 55)]), ms, &(x[WS(rs, 1)])); Chris@10: T2J = LD(&(x[WS(rs, 23)]), ms, &(x[WS(rs, 1)])); Chris@10: { Chris@10: V T23, T5D, T2e, T2g, T2I, T2K, T2M; Chris@10: T2M = LD(&(x[WS(rs, 7)]), ms, &(x[WS(rs, 1)])); Chris@10: T23 = VADD(T20, T22); Chris@10: T5D = VSUB(T20, T22); Chris@10: T2e = BYTW(&(W[TWVL * 28]), T2d); Chris@10: T2c = VADD(T29, T2b); Chris@10: T5W = VSUB(T29, T2b); Chris@10: T2g = BYTW(&(W[TWVL * 92]), T2f); Chris@10: T2I = BYTW(&(W[TWVL * 108]), T2H); Chris@10: T2K = BYTW(&(W[TWVL * 44]), T2J); Chris@10: T2N = BYTW(&(W[TWVL * 12]), T2M); Chris@10: { Chris@10: V T5E, T5P, T42, T2O; Chris@10: T5E = VADD(T5C, T5D); Chris@10: T5P = VSUB(T5C, T5D); Chris@10: T24 = VSUB(T1Y, T23); Chris@10: T42 = VADD(T1Y, T23); Chris@10: T69 = VSUB(T2g, T2e); Chris@10: T2h = VADD(T2e, T2g); Chris@10: T2O = LD(&(x[WS(rs, 39)]), ms, &(x[WS(rs, 1)])); Chris@10: T2L = VADD(T2I, T2K); Chris@10: T5Y = VSUB(T2I, T2K); Chris@10: T5Q = VFMA(LDK(KP707106781), T5P, T5O); Chris@10: T7o = VFNMS(LDK(KP707106781), T5P, T5O); Chris@10: T5F = VFMA(LDK(KP707106781), T5E, T5B); Chris@10: T7l = VFNMS(LDK(KP707106781), T5E, T5B); Chris@10: T43 = VADD(T41, T42); Chris@10: T4F = VSUB(T41, T42); Chris@10: T2P = BYTW(&(W[TWVL * 76]), T2O); Chris@10: } Chris@10: } Chris@10: } Chris@10: T2i = VSUB(T2c, T2h); Chris@10: T48 = VADD(T2c, T2h); Chris@10: { Chris@10: V TW, TY, T11, T2Q, T5X, T13; Chris@10: TW = LD(&(x[WS(rs, 62)]), ms, &(x[0])); Chris@10: TY = LD(&(x[WS(rs, 30)]), ms, &(x[0])); Chris@10: T11 = LD(&(x[WS(rs, 14)]), ms, &(x[0])); Chris@10: T2Q = VADD(T2N, T2P); Chris@10: T5X = VSUB(T2N, T2P); Chris@10: T13 = LD(&(x[WS(rs, 46)]), ms, &(x[0])); Chris@10: { Chris@10: V T12, T5Z, T6a, T49, T14, T18, T1a; Chris@10: { Chris@10: V T17, T19, TX, TZ; Chris@10: T17 = LD(&(x[WS(rs, 54)]), ms, &(x[0])); Chris@10: T19 = LD(&(x[WS(rs, 22)]), ms, &(x[0])); Chris@10: TX = BYTW(&(W[TWVL * 122]), TW); Chris@10: TZ = BYTW(&(W[TWVL * 58]), TY); Chris@10: T12 = BYTW(&(W[TWVL * 26]), T11); Chris@10: T5Z = VADD(T5X, T5Y); Chris@10: T6a = VSUB(T5Y, T5X); Chris@10: T2R = VSUB(T2L, T2Q); Chris@10: T49 = VADD(T2Q, T2L); Chris@10: T14 = BYTW(&(W[TWVL * 90]), T13); Chris@10: T18 = BYTW(&(W[TWVL * 106]), T17); Chris@10: T5q = VSUB(TX, TZ); Chris@10: T10 = VADD(TX, TZ); Chris@10: T1a = BYTW(&(W[TWVL * 42]), T19); Chris@10: } Chris@10: T6b = VFMA(LDK(KP707106781), T6a, T69); Chris@10: T7v = VFNMS(LDK(KP707106781), T6a, T69); Chris@10: T60 = VFMA(LDK(KP707106781), T5Z, T5W); Chris@10: T7s = VFNMS(LDK(KP707106781), T5Z, T5W); Chris@10: T4a = VADD(T48, T49); Chris@10: T4I = VSUB(T48, T49); Chris@10: T5v = VSUB(T14, T12); Chris@10: T15 = VADD(T12, T14); Chris@10: T1b = VADD(T18, T1a); Chris@10: T5s = VSUB(T18, T1a); Chris@10: } Chris@10: T1c = LD(&(x[WS(rs, 6)]), ms, &(x[0])); Chris@10: T1e = LD(&(x[WS(rs, 38)]), ms, &(x[0])); Chris@10: } Chris@10: } Chris@10: } Chris@10: { Chris@10: V Th, T59, Tf, Tv, T5d, Tj, Tm, To; Chris@10: { Chris@10: V T5h, TQ, T5m, T5i, TO, TS, TJ, T4h, TD, TI; Chris@10: { Chris@10: V T4k, T16, TB, T1d, T1f, TE, TG, TA, Tz, TK, TM, TC; Chris@10: Tz = LD(&(x[WS(rs, 2)]), ms, &(x[0])); Chris@10: T4k = VADD(T10, T15); Chris@10: T16 = VSUB(T10, T15); Chris@10: TB = LD(&(x[WS(rs, 34)]), ms, &(x[0])); Chris@10: T1d = BYTW(&(W[TWVL * 10]), T1c); Chris@10: T1f = BYTW(&(W[TWVL * 74]), T1e); Chris@10: TE = LD(&(x[WS(rs, 18)]), ms, &(x[0])); Chris@10: TG = LD(&(x[WS(rs, 50)]), ms, &(x[0])); Chris@10: TA = BYTW(&(W[TWVL * 2]), Tz); Chris@10: TK = LD(&(x[WS(rs, 10)]), ms, &(x[0])); Chris@10: TM = LD(&(x[WS(rs, 42)]), ms, &(x[0])); Chris@10: TC = BYTW(&(W[TWVL * 66]), TB); Chris@10: { Chris@10: V T1g, T5r, TF, TH, TL, TN, TP; Chris@10: TP = LD(&(x[WS(rs, 58)]), ms, &(x[0])); Chris@10: T1g = VADD(T1d, T1f); Chris@10: T5r = VSUB(T1d, T1f); Chris@10: TF = BYTW(&(W[TWVL * 34]), TE); Chris@10: TH = BYTW(&(W[TWVL * 98]), TG); Chris@10: TL = BYTW(&(W[TWVL * 18]), TK); Chris@10: TN = BYTW(&(W[TWVL * 82]), TM); Chris@10: T5h = VSUB(TA, TC); Chris@10: TD = VADD(TA, TC); Chris@10: TQ = BYTW(&(W[TWVL * 114]), TP); Chris@10: { Chris@10: V T5w, T5t, T4l, T1h, TR; Chris@10: T5w = VSUB(T5s, T5r); Chris@10: T5t = VADD(T5r, T5s); Chris@10: T4l = VADD(T1g, T1b); Chris@10: T1h = VSUB(T1b, T1g); Chris@10: T5m = VSUB(TF, TH); Chris@10: TI = VADD(TF, TH); Chris@10: T5i = VSUB(TL, TN); Chris@10: TO = VADD(TL, TN); Chris@10: TR = LD(&(x[WS(rs, 26)]), ms, &(x[0])); Chris@10: T5u = VFMA(LDK(KP707106781), T5t, T5q); Chris@10: T7h = VFNMS(LDK(KP707106781), T5t, T5q); Chris@10: T5x = VFMA(LDK(KP707106781), T5w, T5v); Chris@10: T7g = VFNMS(LDK(KP707106781), T5w, T5v); Chris@10: T1i = VFNMS(LDK(KP414213562), T1h, T16); Chris@10: T3b = VFMA(LDK(KP414213562), T16, T1h); Chris@10: T4m = VADD(T4k, T4l); Chris@10: T4C = VSUB(T4k, T4l); Chris@10: TS = BYTW(&(W[TWVL * 50]), TR); Chris@10: } Chris@10: } Chris@10: } Chris@10: TJ = VSUB(TD, TI); Chris@10: T4h = VADD(TD, TI); Chris@10: { Chris@10: V Tb, Td, Tr, T5j, TT, Tt, Tg; Chris@10: Tb = LD(&(x[WS(rs, 4)]), ms, &(x[0])); Chris@10: Td = LD(&(x[WS(rs, 36)]), ms, &(x[0])); Chris@10: Tr = LD(&(x[WS(rs, 12)]), ms, &(x[0])); Chris@10: T5j = VSUB(TQ, TS); Chris@10: TT = VADD(TQ, TS); Chris@10: Tt = LD(&(x[WS(rs, 44)]), ms, &(x[0])); Chris@10: Tg = LD(&(x[WS(rs, 20)]), ms, &(x[0])); Chris@10: { Chris@10: V Ti, Tc, Te, Ts; Chris@10: Ti = LD(&(x[WS(rs, 52)]), ms, &(x[0])); Chris@10: Tc = BYTW(&(W[TWVL * 6]), Tb); Chris@10: Te = BYTW(&(W[TWVL * 70]), Td); Chris@10: Ts = BYTW(&(W[TWVL * 22]), Tr); Chris@10: { Chris@10: V T5k, T5n, TU, T4i, Tu; Chris@10: T5k = VADD(T5i, T5j); Chris@10: T5n = VSUB(T5i, T5j); Chris@10: TU = VSUB(TO, TT); Chris@10: T4i = VADD(TO, TT); Chris@10: Tu = BYTW(&(W[TWVL * 86]), Tt); Chris@10: Th = BYTW(&(W[TWVL * 38]), Tg); Chris@10: T59 = VSUB(Tc, Te); Chris@10: Tf = VADD(Tc, Te); Chris@10: T7e = VFNMS(LDK(KP707106781), T5k, T5h); Chris@10: T5l = VFMA(LDK(KP707106781), T5k, T5h); Chris@10: T7d = VFNMS(LDK(KP707106781), T5n, T5m); Chris@10: T5o = VFMA(LDK(KP707106781), T5n, T5m); Chris@10: T3a = VFMA(LDK(KP414213562), TJ, TU); Chris@10: TV = VFNMS(LDK(KP414213562), TU, TJ); Chris@10: T4B = VSUB(T4h, T4i); Chris@10: T4j = VADD(T4h, T4i); Chris@10: Tv = VADD(Ts, Tu); Chris@10: T5d = VSUB(Tu, Ts); Chris@10: Tj = BYTW(&(W[TWVL * 102]), Ti); Chris@10: } Chris@10: } Chris@10: Tm = LD(&(x[WS(rs, 60)]), ms, &(x[0])); Chris@10: To = LD(&(x[WS(rs, 28)]), ms, &(x[0])); Chris@10: } Chris@10: } Chris@10: { Chris@10: V T5b, T6m, Tl, T1A, T5G, T1Q, T5K, T1C, T1D, T5e, T6n, Tw, T1H, T1J; Chris@10: { Chris@10: V T1w, T1y, T1M, T1O, Tq, T5c, T1B; Chris@10: T1w = LD(&(x[WS(rs, 5)]), ms, &(x[WS(rs, 1)])); Chris@10: T1y = LD(&(x[WS(rs, 37)]), ms, &(x[WS(rs, 1)])); Chris@10: T1M = LD(&(x[WS(rs, 13)]), ms, &(x[WS(rs, 1)])); Chris@10: T1O = LD(&(x[WS(rs, 45)]), ms, &(x[WS(rs, 1)])); Chris@10: T1B = LD(&(x[WS(rs, 21)]), ms, &(x[WS(rs, 1)])); Chris@10: { Chris@10: V Tk, T5a, Tn, Tp; Chris@10: Tk = VADD(Th, Tj); Chris@10: T5a = VSUB(Th, Tj); Chris@10: Tn = BYTW(&(W[TWVL * 118]), Tm); Chris@10: Tp = BYTW(&(W[TWVL * 54]), To); Chris@10: { Chris@10: V T1x, T1z, T1N, T1P; Chris@10: T1x = BYTW(&(W[TWVL * 8]), T1w); Chris@10: T1z = BYTW(&(W[TWVL * 72]), T1y); Chris@10: T1N = BYTW(&(W[TWVL * 24]), T1M); Chris@10: T1P = BYTW(&(W[TWVL * 88]), T1O); Chris@10: T5b = VFNMS(LDK(KP414213562), T5a, T59); Chris@10: T6m = VFMA(LDK(KP414213562), T59, T5a); Chris@10: T3X = VADD(Tf, Tk); Chris@10: Tl = VSUB(Tf, Tk); Chris@10: Tq = VADD(Tn, Tp); Chris@10: T5c = VSUB(Tn, Tp); Chris@10: T1A = VADD(T1x, T1z); Chris@10: T5G = VSUB(T1x, T1z); Chris@10: T1Q = VADD(T1N, T1P); Chris@10: T5K = VSUB(T1N, T1P); Chris@10: T1C = BYTW(&(W[TWVL * 40]), T1B); Chris@10: } Chris@10: } Chris@10: T1D = LD(&(x[WS(rs, 53)]), ms, &(x[WS(rs, 1)])); Chris@10: T5e = VFNMS(LDK(KP414213562), T5d, T5c); Chris@10: T6n = VFMA(LDK(KP414213562), T5c, T5d); Chris@10: T3Y = VADD(Tq, Tv); Chris@10: Tw = VSUB(Tq, Tv); Chris@10: T1H = LD(&(x[WS(rs, 61)]), ms, &(x[WS(rs, 1)])); Chris@10: T1J = LD(&(x[WS(rs, 29)]), ms, &(x[WS(rs, 1)])); Chris@10: } Chris@10: { Chris@10: V T1I, T1K, T1F, T5H, T2k, T2l, T2z, T2B, T2j, T1E; Chris@10: T2j = LD(&(x[WS(rs, 3)]), ms, &(x[WS(rs, 1)])); Chris@10: T1E = BYTW(&(W[TWVL * 104]), T1D); Chris@10: T6o = VSUB(T6m, T6n); Chris@10: T7b = VADD(T6m, T6n); Chris@10: T5f = VADD(T5b, T5e); Chris@10: T7C = VSUB(T5b, T5e); Chris@10: Tx = VADD(Tl, Tw); Chris@10: T38 = VSUB(Tl, Tw); Chris@10: T1I = BYTW(&(W[TWVL * 120]), T1H); Chris@10: T1K = BYTW(&(W[TWVL * 56]), T1J); Chris@10: T1F = VADD(T1C, T1E); Chris@10: T5H = VSUB(T1C, T1E); Chris@10: T2k = BYTW(&(W[TWVL * 4]), T2j); Chris@10: T2l = LD(&(x[WS(rs, 35)]), ms, &(x[WS(rs, 1)])); Chris@10: T2z = LD(&(x[WS(rs, 11)]), ms, &(x[WS(rs, 1)])); Chris@10: T2B = LD(&(x[WS(rs, 43)]), ms, &(x[WS(rs, 1)])); Chris@10: { Chris@10: V T5I, T5R, T44, T1G, T2m, T2A, T2C, T5S, T5L, T1R, T45, T2o, T5J, T1L; Chris@10: T2o = LD(&(x[WS(rs, 19)]), ms, &(x[WS(rs, 1)])); Chris@10: T5J = VSUB(T1I, T1K); Chris@10: T1L = VADD(T1I, T1K); Chris@10: T5I = VFNMS(LDK(KP414213562), T5H, T5G); Chris@10: T5R = VFMA(LDK(KP414213562), T5G, T5H); Chris@10: T44 = VADD(T1A, T1F); Chris@10: T1G = VSUB(T1A, T1F); Chris@10: T2m = BYTW(&(W[TWVL * 68]), T2l); Chris@10: T2A = BYTW(&(W[TWVL * 20]), T2z); Chris@10: T2C = BYTW(&(W[TWVL * 84]), T2B); Chris@10: T5S = VFNMS(LDK(KP414213562), T5J, T5K); Chris@10: T5L = VFMA(LDK(KP414213562), T5K, T5J); Chris@10: T1R = VSUB(T1L, T1Q); Chris@10: T45 = VADD(T1L, T1Q); Chris@10: T2p = BYTW(&(W[TWVL * 36]), T2o); Chris@10: T61 = VSUB(T2k, T2m); Chris@10: T2n = VADD(T2k, T2m); Chris@10: T65 = VSUB(T2C, T2A); Chris@10: T2D = VADD(T2A, T2C); Chris@10: T7p = VSUB(T5I, T5L); Chris@10: T5M = VADD(T5I, T5L); Chris@10: T7m = VSUB(T5R, T5S); Chris@10: T5T = VADD(T5R, T5S); Chris@10: T4G = VSUB(T44, T45); Chris@10: T46 = VADD(T44, T45); Chris@10: T25 = VSUB(T1G, T1R); Chris@10: T1S = VADD(T1G, T1R); Chris@10: T2q = LD(&(x[WS(rs, 51)]), ms, &(x[WS(rs, 1)])); Chris@10: } Chris@10: T2u = LD(&(x[WS(rs, 59)]), ms, &(x[WS(rs, 1)])); Chris@10: T2w = LD(&(x[WS(rs, 27)]), ms, &(x[WS(rs, 1)])); Chris@10: } Chris@10: } Chris@10: } Chris@10: } Chris@10: { Chris@10: V T67, T7w, T6e, T7t, T3s, T3E, T39, T3D, T1k, T3k, T3t, T3c, T1T, T3v, T3w; Chris@10: V T26, T2G, T3y, T3z, T2T; Chris@10: { Chris@10: V T4A, T4N, T47, T4v, T2r, T2v, T2x, T4s, T40, T3W, T3Z; Chris@10: T4A = VSUB(T3U, T3V); Chris@10: T3W = VADD(T3U, T3V); Chris@10: T3Z = VADD(T3X, T3Y); Chris@10: T4N = VSUB(T3X, T3Y); Chris@10: T47 = VSUB(T43, T46); Chris@10: T4v = VADD(T43, T46); Chris@10: T2r = BYTW(&(W[TWVL * 100]), T2q); Chris@10: T2v = BYTW(&(W[TWVL * 116]), T2u); Chris@10: T2x = BYTW(&(W[TWVL * 52]), T2w); Chris@10: T4s = VADD(T3W, T3Z); Chris@10: T40 = VSUB(T3W, T3Z); Chris@10: { Chris@10: V T4O, T4n, T4Q, T4H, T4E, T4W, T4u, T4y, T4d, T4J, T2F, T2S; Chris@10: { Chris@10: V T6c, T63, T2t, T4b, T6d, T66, T2E, T4c; Chris@10: { Chris@10: V T4D, T62, T2s, T64, T2y, T4t; Chris@10: T4O = VSUB(T4B, T4C); Chris@10: T4D = VADD(T4B, T4C); Chris@10: T62 = VSUB(T2r, T2p); Chris@10: T2s = VADD(T2p, T2r); Chris@10: T64 = VSUB(T2v, T2x); Chris@10: T2y = VADD(T2v, T2x); Chris@10: T4t = VADD(T4j, T4m); Chris@10: T4n = VSUB(T4j, T4m); Chris@10: T4Q = VFMA(LDK(KP414213562), T4F, T4G); Chris@10: T4H = VFNMS(LDK(KP414213562), T4G, T4F); Chris@10: T4E = VFMA(LDK(KP707106781), T4D, T4A); Chris@10: T4W = VFNMS(LDK(KP707106781), T4D, T4A); Chris@10: T6c = VFNMS(LDK(KP414213562), T61, T62); Chris@10: T63 = VFMA(LDK(KP414213562), T62, T61); Chris@10: T2t = VSUB(T2n, T2s); Chris@10: T4b = VADD(T2n, T2s); Chris@10: T6d = VFMA(LDK(KP414213562), T64, T65); Chris@10: T66 = VFNMS(LDK(KP414213562), T65, T64); Chris@10: T2E = VSUB(T2y, T2D); Chris@10: T4c = VADD(T2y, T2D); Chris@10: T4u = VSUB(T4s, T4t); Chris@10: T4y = VADD(T4s, T4t); Chris@10: } Chris@10: T67 = VADD(T63, T66); Chris@10: T7w = VSUB(T66, T63); Chris@10: T6e = VADD(T6c, T6d); Chris@10: T7t = VSUB(T6d, T6c); Chris@10: T4d = VADD(T4b, T4c); Chris@10: T4J = VSUB(T4c, T4b); Chris@10: T2F = VADD(T2t, T2E); Chris@10: T2S = VSUB(T2E, T2t); Chris@10: } Chris@10: { Chris@10: V Ty, T1j, T4R, T4K; Chris@10: Ty = VFMA(LDK(KP707106781), Tx, Ta); Chris@10: T3s = VFNMS(LDK(KP707106781), Tx, Ta); Chris@10: T3E = VSUB(TV, T1i); Chris@10: T1j = VADD(TV, T1i); Chris@10: T39 = VFMA(LDK(KP707106781), T38, T37); Chris@10: T3D = VFNMS(LDK(KP707106781), T38, T37); Chris@10: T4R = VFMA(LDK(KP414213562), T4I, T4J); Chris@10: T4K = VFNMS(LDK(KP414213562), T4J, T4I); Chris@10: { Chris@10: V T4w, T4e, T4P, T4Z; Chris@10: T4w = VADD(T4a, T4d); Chris@10: T4e = VSUB(T4a, T4d); Chris@10: T4P = VFMA(LDK(KP707106781), T4O, T4N); Chris@10: T4Z = VFNMS(LDK(KP707106781), T4O, T4N); Chris@10: T1k = VFMA(LDK(KP923879532), T1j, Ty); Chris@10: T3k = VFNMS(LDK(KP923879532), T1j, Ty); Chris@10: { Chris@10: V T4L, T50, T4S, T4X; Chris@10: T4L = VADD(T4H, T4K); Chris@10: T50 = VSUB(T4H, T4K); Chris@10: T4S = VSUB(T4Q, T4R); Chris@10: T4X = VADD(T4Q, T4R); Chris@10: { Chris@10: V T4f, T4o, T4x, T4z; Chris@10: T4f = VADD(T47, T4e); Chris@10: T4o = VSUB(T47, T4e); Chris@10: T4x = VSUB(T4v, T4w); Chris@10: T4z = VADD(T4v, T4w); Chris@10: { Chris@10: V T53, T51, T4M, T4U; Chris@10: T53 = VFNMS(LDK(KP923879532), T50, T4Z); Chris@10: T51 = VFMA(LDK(KP923879532), T50, T4Z); Chris@10: T4M = VFNMS(LDK(KP923879532), T4L, T4E); Chris@10: T4U = VFMA(LDK(KP923879532), T4L, T4E); Chris@10: { Chris@10: V T52, T4Y, T4T, T4V; Chris@10: T52 = VFMA(LDK(KP923879532), T4X, T4W); Chris@10: T4Y = VFNMS(LDK(KP923879532), T4X, T4W); Chris@10: T4T = VFNMS(LDK(KP923879532), T4S, T4P); Chris@10: T4V = VFMA(LDK(KP923879532), T4S, T4P); Chris@10: { Chris@10: V T4p, T4r, T4g, T4q; Chris@10: T4p = VFNMS(LDK(KP707106781), T4o, T4n); Chris@10: T4r = VFMA(LDK(KP707106781), T4o, T4n); Chris@10: T4g = VFNMS(LDK(KP707106781), T4f, T40); Chris@10: T4q = VFMA(LDK(KP707106781), T4f, T40); Chris@10: ST(&(x[0]), VADD(T4y, T4z), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 32)]), VSUB(T4y, T4z), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 16)]), VFMAI(T4x, T4u), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 48)]), VFNMSI(T4x, T4u), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 44)]), VFNMSI(T51, T4Y), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 20)]), VFMAI(T51, T4Y), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 52)]), VFMAI(T53, T52), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 12)]), VFNMSI(T53, T52), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 4)]), VFMAI(T4V, T4U), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 60)]), VFNMSI(T4V, T4U), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 36)]), VFMAI(T4T, T4M), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 28)]), VFNMSI(T4T, T4M), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 56)]), VFNMSI(T4r, T4q), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 8)]), VFMAI(T4r, T4q), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 40)]), VFMAI(T4p, T4g), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 24)]), VFNMSI(T4p, T4g), ms, &(x[0])); Chris@10: T3t = VADD(T3a, T3b); Chris@10: T3c = VSUB(T3a, T3b); Chris@10: } Chris@10: } Chris@10: } Chris@10: } Chris@10: } Chris@10: } Chris@10: T1T = VFMA(LDK(KP707106781), T1S, T1v); Chris@10: T3v = VFNMS(LDK(KP707106781), T1S, T1v); Chris@10: T3w = VFNMS(LDK(KP707106781), T25, T24); Chris@10: T26 = VFMA(LDK(KP707106781), T25, T24); Chris@10: T2G = VFMA(LDK(KP707106781), T2F, T2i); Chris@10: T3y = VFNMS(LDK(KP707106781), T2F, T2i); Chris@10: T3z = VFNMS(LDK(KP707106781), T2S, T2R); Chris@10: T2T = VFMA(LDK(KP707106781), T2S, T2R); Chris@10: } Chris@10: } Chris@10: } Chris@10: { Chris@10: V T3u, T3M, T3F, T3P, T3x, T3G, T3q, T3m, T3h, T3j, T3r, T3p, T2W, T3i; Chris@10: { Chris@10: V T3d, T3n, T27, T3e, T2U, T3f; Chris@10: T3d = VFMA(LDK(KP923879532), T3c, T39); Chris@10: T3n = VFNMS(LDK(KP923879532), T3c, T39); Chris@10: T27 = VFNMS(LDK(KP198912367), T26, T1T); Chris@10: T3e = VFMA(LDK(KP198912367), T1T, T26); Chris@10: T2U = VFNMS(LDK(KP198912367), T2T, T2G); Chris@10: T3f = VFMA(LDK(KP198912367), T2G, T2T); Chris@10: T3u = VFMA(LDK(KP923879532), T3t, T3s); Chris@10: T3M = VFNMS(LDK(KP923879532), T3t, T3s); Chris@10: { Chris@10: V T3g, T3l, T2V, T3o; Chris@10: T3g = VSUB(T3e, T3f); Chris@10: T3l = VADD(T3e, T3f); Chris@10: T2V = VADD(T27, T2U); Chris@10: T3o = VSUB(T27, T2U); Chris@10: T3F = VFNMS(LDK(KP923879532), T3E, T3D); Chris@10: T3P = VFMA(LDK(KP923879532), T3E, T3D); Chris@10: T3x = VFMA(LDK(KP668178637), T3w, T3v); Chris@10: T3G = VFNMS(LDK(KP668178637), T3v, T3w); Chris@10: T3q = VFMA(LDK(KP980785280), T3l, T3k); Chris@10: T3m = VFNMS(LDK(KP980785280), T3l, T3k); Chris@10: T3h = VFNMS(LDK(KP980785280), T3g, T3d); Chris@10: T3j = VFMA(LDK(KP980785280), T3g, T3d); Chris@10: T3r = VFNMS(LDK(KP980785280), T3o, T3n); Chris@10: T3p = VFMA(LDK(KP980785280), T3o, T3n); Chris@10: T2W = VFNMS(LDK(KP980785280), T2V, T1k); Chris@10: T3i = VFMA(LDK(KP980785280), T2V, T1k); Chris@10: } Chris@10: } Chris@10: { Chris@10: V T7n, T7Z, T8j, T89, T7k, T7O, T8g, T7Y, T7H, T7R, T80, T7q, T7u, T82, T83; Chris@10: V T7x; Chris@10: { Chris@10: V T7c, T7W, T7D, T87, T7f, T7E, T3A, T3H, T7F, T7i; Chris@10: T7c = VFNMS(LDK(KP923879532), T7b, T7a); Chris@10: T7W = VFMA(LDK(KP923879532), T7b, T7a); Chris@10: T7D = VFMA(LDK(KP923879532), T7C, T7B); Chris@10: T87 = VFNMS(LDK(KP923879532), T7C, T7B); Chris@10: T7f = VFNMS(LDK(KP668178637), T7e, T7d); Chris@10: T7E = VFMA(LDK(KP668178637), T7d, T7e); Chris@10: ST(&(x[WS(rs, 46)]), VFNMSI(T3p, T3m), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 18)]), VFMAI(T3p, T3m), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 50)]), VFMAI(T3r, T3q), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 14)]), VFNMSI(T3r, T3q), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 2)]), VFMAI(T3j, T3i), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 62)]), VFNMSI(T3j, T3i), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 34)]), VFMAI(T3h, T2W), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 30)]), VFNMSI(T3h, T2W), ms, &(x[0])); Chris@10: T3A = VFMA(LDK(KP668178637), T3z, T3y); Chris@10: T3H = VFNMS(LDK(KP668178637), T3y, T3z); Chris@10: T7F = VFMA(LDK(KP668178637), T7g, T7h); Chris@10: T7i = VFNMS(LDK(KP668178637), T7h, T7g); Chris@10: T7n = VFNMS(LDK(KP923879532), T7m, T7l); Chris@10: T7Z = VFMA(LDK(KP923879532), T7m, T7l); Chris@10: { Chris@10: V T3I, T3N, T3B, T3Q; Chris@10: T3I = VSUB(T3G, T3H); Chris@10: T3N = VADD(T3G, T3H); Chris@10: T3B = VADD(T3x, T3A); Chris@10: T3Q = VSUB(T3x, T3A); Chris@10: { Chris@10: V T7j, T88, T7G, T7X; Chris@10: T7j = VADD(T7f, T7i); Chris@10: T88 = VSUB(T7f, T7i); Chris@10: T7G = VSUB(T7E, T7F); Chris@10: T7X = VADD(T7E, T7F); Chris@10: { Chris@10: V T3S, T3O, T3J, T3L; Chris@10: T3S = VFNMS(LDK(KP831469612), T3N, T3M); Chris@10: T3O = VFMA(LDK(KP831469612), T3N, T3M); Chris@10: T3J = VFNMS(LDK(KP831469612), T3I, T3F); Chris@10: T3L = VFMA(LDK(KP831469612), T3I, T3F); Chris@10: { Chris@10: V T3T, T3R, T3C, T3K; Chris@10: T3T = VFMA(LDK(KP831469612), T3Q, T3P); Chris@10: T3R = VFNMS(LDK(KP831469612), T3Q, T3P); Chris@10: T3C = VFNMS(LDK(KP831469612), T3B, T3u); Chris@10: T3K = VFMA(LDK(KP831469612), T3B, T3u); Chris@10: T8j = VFNMS(LDK(KP831469612), T88, T87); Chris@10: T89 = VFMA(LDK(KP831469612), T88, T87); Chris@10: T7k = VFNMS(LDK(KP831469612), T7j, T7c); Chris@10: T7O = VFMA(LDK(KP831469612), T7j, T7c); Chris@10: T8g = VFNMS(LDK(KP831469612), T7X, T7W); Chris@10: T7Y = VFMA(LDK(KP831469612), T7X, T7W); Chris@10: T7H = VFMA(LDK(KP831469612), T7G, T7D); Chris@10: T7R = VFNMS(LDK(KP831469612), T7G, T7D); Chris@10: ST(&(x[WS(rs, 42)]), VFMAI(T3R, T3O), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 22)]), VFNMSI(T3R, T3O), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 54)]), VFNMSI(T3T, T3S), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 10)]), VFMAI(T3T, T3S), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 58)]), VFMAI(T3L, T3K), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 6)]), VFNMSI(T3L, T3K), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 26)]), VFMAI(T3J, T3C), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 38)]), VFNMSI(T3J, T3C), ms, &(x[0])); Chris@10: T80 = VFNMS(LDK(KP923879532), T7p, T7o); Chris@10: T7q = VFMA(LDK(KP923879532), T7p, T7o); Chris@10: } Chris@10: } Chris@10: } Chris@10: } Chris@10: T7u = VFNMS(LDK(KP923879532), T7t, T7s); Chris@10: T82 = VFMA(LDK(KP923879532), T7t, T7s); Chris@10: T83 = VFNMS(LDK(KP923879532), T7w, T7v); Chris@10: T7x = VFMA(LDK(KP923879532), T7w, T7v); Chris@10: } Chris@10: { Chris@10: V T5g, T6I, T6p, T6T, T5p, T6q, T6r, T5y; Chris@10: T5g = VFMA(LDK(KP923879532), T5f, T58); Chris@10: T6I = VFNMS(LDK(KP923879532), T5f, T58); Chris@10: { Chris@10: V T7r, T7I, T7y, T7J; Chris@10: T7r = VFNMS(LDK(KP534511135), T7q, T7n); Chris@10: T7I = VFMA(LDK(KP534511135), T7n, T7q); Chris@10: T7y = VFNMS(LDK(KP534511135), T7x, T7u); Chris@10: T7J = VFMA(LDK(KP534511135), T7u, T7x); Chris@10: { Chris@10: V T81, T8a, T84, T8b; Chris@10: T81 = VFMA(LDK(KP303346683), T80, T7Z); Chris@10: T8a = VFNMS(LDK(KP303346683), T7Z, T80); Chris@10: T84 = VFMA(LDK(KP303346683), T83, T82); Chris@10: T8b = VFNMS(LDK(KP303346683), T82, T83); Chris@10: T6p = VFMA(LDK(KP923879532), T6o, T6l); Chris@10: T6T = VFNMS(LDK(KP923879532), T6o, T6l); Chris@10: T5p = VFNMS(LDK(KP198912367), T5o, T5l); Chris@10: T6q = VFMA(LDK(KP198912367), T5l, T5o); Chris@10: { Chris@10: V T7K, T7P, T7z, T7S; Chris@10: T7K = VSUB(T7I, T7J); Chris@10: T7P = VADD(T7I, T7J); Chris@10: T7z = VADD(T7r, T7y); Chris@10: T7S = VSUB(T7r, T7y); Chris@10: { Chris@10: V T8c, T8h, T85, T8k; Chris@10: T8c = VSUB(T8a, T8b); Chris@10: T8h = VADD(T8a, T8b); Chris@10: T85 = VADD(T81, T84); Chris@10: T8k = VSUB(T81, T84); Chris@10: { Chris@10: V T7Q, T7U, T7L, T7N; Chris@10: T7Q = VFNMS(LDK(KP881921264), T7P, T7O); Chris@10: T7U = VFMA(LDK(KP881921264), T7P, T7O); Chris@10: T7L = VFNMS(LDK(KP881921264), T7K, T7H); Chris@10: T7N = VFMA(LDK(KP881921264), T7K, T7H); Chris@10: { Chris@10: V T7T, T7V, T7A, T7M; Chris@10: T7T = VFMA(LDK(KP881921264), T7S, T7R); Chris@10: T7V = VFNMS(LDK(KP881921264), T7S, T7R); Chris@10: T7A = VFNMS(LDK(KP881921264), T7z, T7k); Chris@10: T7M = VFMA(LDK(KP881921264), T7z, T7k); Chris@10: { Chris@10: V T8i, T8m, T8d, T8f; Chris@10: T8i = VFMA(LDK(KP956940335), T8h, T8g); Chris@10: T8m = VFNMS(LDK(KP956940335), T8h, T8g); Chris@10: T8d = VFNMS(LDK(KP956940335), T8c, T89); Chris@10: T8f = VFMA(LDK(KP956940335), T8c, T89); Chris@10: { Chris@10: V T8l, T8n, T86, T8e; Chris@10: T8l = VFNMS(LDK(KP956940335), T8k, T8j); Chris@10: T8n = VFMA(LDK(KP956940335), T8k, T8j); Chris@10: T86 = VFNMS(LDK(KP956940335), T85, T7Y); Chris@10: T8e = VFMA(LDK(KP956940335), T85, T7Y); Chris@10: ST(&(x[WS(rs, 53)]), VFMAI(T7V, T7U), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 11)]), VFNMSI(T7V, T7U), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 43)]), VFNMSI(T7T, T7Q), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 21)]), VFMAI(T7T, T7Q), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 5)]), VFMAI(T7N, T7M), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 59)]), VFNMSI(T7N, T7M), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 37)]), VFMAI(T7L, T7A), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 27)]), VFNMSI(T7L, T7A), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 51)]), VFNMSI(T8n, T8m), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 13)]), VFMAI(T8n, T8m), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 45)]), VFMAI(T8l, T8i), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 19)]), VFNMSI(T8l, T8i), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 61)]), VFMAI(T8f, T8e), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 3)]), VFNMSI(T8f, T8e), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 29)]), VFMAI(T8d, T86), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 35)]), VFNMSI(T8d, T86), ms, &(x[WS(rs, 1)])); Chris@10: T6r = VFMA(LDK(KP198912367), T5u, T5x); Chris@10: T5y = VFNMS(LDK(KP198912367), T5x, T5u); Chris@10: } Chris@10: } Chris@10: } Chris@10: } Chris@10: } Chris@10: } Chris@10: } Chris@10: } Chris@10: { Chris@10: V T5N, T5U, T68, T5z, T6U, T6f; Chris@10: T5N = VFMA(LDK(KP923879532), T5M, T5F); Chris@10: T6L = VFNMS(LDK(KP923879532), T5M, T5F); Chris@10: T6M = VFNMS(LDK(KP923879532), T5T, T5Q); Chris@10: T5U = VFMA(LDK(KP923879532), T5T, T5Q); Chris@10: T68 = VFMA(LDK(KP923879532), T67, T60); Chris@10: T6O = VFNMS(LDK(KP923879532), T67, T60); Chris@10: T5z = VADD(T5p, T5y); Chris@10: T6U = VSUB(T5p, T5y); Chris@10: T6P = VFNMS(LDK(KP923879532), T6e, T6b); Chris@10: T6f = VFMA(LDK(KP923879532), T6e, T6b); Chris@10: { Chris@10: V T5V, T6u, T6g, T6v, T6s, T6J; Chris@10: T6s = VSUB(T6q, T6r); Chris@10: T6J = VADD(T6q, T6r); Chris@10: T5V = VFNMS(LDK(KP098491403), T5U, T5N); Chris@10: T6u = VFMA(LDK(KP098491403), T5N, T5U); Chris@10: T75 = VFMA(LDK(KP980785280), T6U, T6T); Chris@10: T6V = VFNMS(LDK(KP980785280), T6U, T6T); Chris@10: T5A = VFMA(LDK(KP980785280), T5z, T5g); Chris@10: T6A = VFNMS(LDK(KP980785280), T5z, T5g); Chris@10: T6g = VFNMS(LDK(KP098491403), T6f, T68); Chris@10: T6v = VFMA(LDK(KP098491403), T68, T6f); Chris@10: T72 = VFNMS(LDK(KP980785280), T6J, T6I); Chris@10: T6K = VFMA(LDK(KP980785280), T6J, T6I); Chris@10: T6t = VFMA(LDK(KP980785280), T6s, T6p); Chris@10: T6D = VFNMS(LDK(KP980785280), T6s, T6p); Chris@10: T6w = VSUB(T6u, T6v); Chris@10: T6B = VADD(T6u, T6v); Chris@10: T6h = VADD(T5V, T6g); Chris@10: T6E = VSUB(T5V, T6g); Chris@10: } Chris@10: } Chris@10: } Chris@10: } Chris@10: } Chris@10: } Chris@10: } Chris@10: { Chris@10: V T6W, T6N, T6G, T6C, T6z, T6x, T6H, T6F, T6y, T6i, T6X, T6Q; Chris@10: T6W = VFNMS(LDK(KP820678790), T6L, T6M); Chris@10: T6N = VFMA(LDK(KP820678790), T6M, T6L); Chris@10: T6G = VFMA(LDK(KP995184726), T6B, T6A); Chris@10: T6C = VFNMS(LDK(KP995184726), T6B, T6A); Chris@10: T6z = VFMA(LDK(KP995184726), T6w, T6t); Chris@10: T6x = VFNMS(LDK(KP995184726), T6w, T6t); Chris@10: T6H = VFNMS(LDK(KP995184726), T6E, T6D); Chris@10: T6F = VFMA(LDK(KP995184726), T6E, T6D); Chris@10: T6y = VFMA(LDK(KP995184726), T6h, T5A); Chris@10: T6i = VFNMS(LDK(KP995184726), T6h, T5A); Chris@10: T6X = VFNMS(LDK(KP820678790), T6O, T6P); Chris@10: T6Q = VFMA(LDK(KP820678790), T6P, T6O); Chris@10: { Chris@10: V T73, T6Y, T76, T6R; Chris@10: ST(&(x[WS(rs, 49)]), VFMAI(T6H, T6G), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 15)]), VFNMSI(T6H, T6G), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 47)]), VFNMSI(T6F, T6C), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 17)]), VFMAI(T6F, T6C), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 1)]), VFMAI(T6z, T6y), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 63)]), VFNMSI(T6z, T6y), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 33)]), VFMAI(T6x, T6i), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 31)]), VFNMSI(T6x, T6i), ms, &(x[WS(rs, 1)])); Chris@10: T73 = VADD(T6W, T6X); Chris@10: T6Y = VSUB(T6W, T6X); Chris@10: T76 = VSUB(T6N, T6Q); Chris@10: T6R = VADD(T6N, T6Q); Chris@10: { Chris@10: V T78, T74, T71, T6Z, T79, T77, T70, T6S; Chris@10: T78 = VFNMS(LDK(KP773010453), T73, T72); Chris@10: T74 = VFMA(LDK(KP773010453), T73, T72); Chris@10: T71 = VFMA(LDK(KP773010453), T6Y, T6V); Chris@10: T6Z = VFNMS(LDK(KP773010453), T6Y, T6V); Chris@10: T79 = VFMA(LDK(KP773010453), T76, T75); Chris@10: T77 = VFNMS(LDK(KP773010453), T76, T75); Chris@10: T70 = VFMA(LDK(KP773010453), T6R, T6K); Chris@10: T6S = VFNMS(LDK(KP773010453), T6R, T6K); Chris@10: ST(&(x[WS(rs, 55)]), VFNMSI(T79, T78), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 9)]), VFMAI(T79, T78), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 41)]), VFMAI(T77, T74), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 23)]), VFNMSI(T77, T74), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 57)]), VFMAI(T71, T70), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 7)]), VFNMSI(T71, T70), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 25)]), VFMAI(T6Z, T6S), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 39)]), VFNMSI(T6Z, T6S), ms, &(x[WS(rs, 1)])); Chris@10: } Chris@10: } Chris@10: } Chris@10: } Chris@10: } Chris@10: VLEAVE(); Chris@10: } Chris@10: Chris@10: static const tw_instr twinstr[] = { Chris@10: VTW(0, 1), Chris@10: VTW(0, 2), Chris@10: VTW(0, 3), Chris@10: VTW(0, 4), Chris@10: VTW(0, 5), Chris@10: VTW(0, 6), Chris@10: VTW(0, 7), Chris@10: VTW(0, 8), Chris@10: VTW(0, 9), Chris@10: VTW(0, 10), Chris@10: VTW(0, 11), Chris@10: VTW(0, 12), Chris@10: VTW(0, 13), Chris@10: VTW(0, 14), Chris@10: VTW(0, 15), Chris@10: VTW(0, 16), Chris@10: VTW(0, 17), Chris@10: VTW(0, 18), Chris@10: VTW(0, 19), Chris@10: VTW(0, 20), Chris@10: VTW(0, 21), Chris@10: VTW(0, 22), Chris@10: VTW(0, 23), Chris@10: VTW(0, 24), Chris@10: VTW(0, 25), Chris@10: VTW(0, 26), Chris@10: VTW(0, 27), Chris@10: VTW(0, 28), Chris@10: VTW(0, 29), Chris@10: VTW(0, 30), Chris@10: VTW(0, 31), Chris@10: VTW(0, 32), Chris@10: VTW(0, 33), Chris@10: VTW(0, 34), Chris@10: VTW(0, 35), Chris@10: VTW(0, 36), Chris@10: VTW(0, 37), Chris@10: VTW(0, 38), Chris@10: VTW(0, 39), Chris@10: VTW(0, 40), Chris@10: VTW(0, 41), Chris@10: VTW(0, 42), Chris@10: VTW(0, 43), Chris@10: VTW(0, 44), Chris@10: VTW(0, 45), Chris@10: VTW(0, 46), Chris@10: VTW(0, 47), Chris@10: VTW(0, 48), Chris@10: VTW(0, 49), Chris@10: VTW(0, 50), Chris@10: VTW(0, 51), Chris@10: VTW(0, 52), Chris@10: VTW(0, 53), Chris@10: VTW(0, 54), Chris@10: VTW(0, 55), Chris@10: VTW(0, 56), Chris@10: VTW(0, 57), Chris@10: VTW(0, 58), Chris@10: VTW(0, 59), Chris@10: VTW(0, 60), Chris@10: VTW(0, 61), Chris@10: VTW(0, 62), Chris@10: VTW(0, 63), Chris@10: {TW_NEXT, VL, 0} Chris@10: }; Chris@10: Chris@10: static const ct_desc desc = { 64, XSIMD_STRING("t1bv_64"), twinstr, &GENUS, {261, 126, 258, 0}, 0, 0, 0 }; Chris@10: Chris@10: void XSIMD(codelet_t1bv_64) (planner *p) { Chris@10: X(kdft_dit_register) (p, t1bv_64, &desc); Chris@10: } Chris@10: #else /* HAVE_FMA */ Chris@10: Chris@10: /* Generated by: ../../../genfft/gen_twiddle_c.native -simd -compact -variables 4 -pipeline-latency 8 -n 64 -name t1bv_64 -include t1b.h -sign 1 */ Chris@10: Chris@10: /* Chris@10: * This function contains 519 FP additions, 250 FP multiplications, Chris@10: * (or, 467 additions, 198 multiplications, 52 fused multiply/add), Chris@10: * 107 stack variables, 15 constants, and 128 memory accesses Chris@10: */ Chris@10: #include "t1b.h" Chris@10: Chris@10: static void t1bv_64(R *ri, R *ii, const R *W, stride rs, INT mb, INT me, INT ms) Chris@10: { Chris@10: DVK(KP290284677, +0.290284677254462367636192375817395274691476278); Chris@10: DVK(KP956940335, +0.956940335732208864935797886980269969482849206); Chris@10: DVK(KP471396736, +0.471396736825997648556387625905254377657460319); Chris@10: DVK(KP881921264, +0.881921264348355029712756863660388349508442621); Chris@10: DVK(KP634393284, +0.634393284163645498215171613225493370675687095); Chris@10: DVK(KP773010453, +0.773010453362736960810906609758469800971041293); Chris@10: DVK(KP098017140, +0.098017140329560601994195563888641845861136673); Chris@10: DVK(KP995184726, +0.995184726672196886244836953109479921575474869); Chris@10: DVK(KP195090322, +0.195090322016128267848284868477022240927691618); Chris@10: DVK(KP980785280, +0.980785280403230449126182236134239036973933731); Chris@10: DVK(KP555570233, +0.555570233019602224742830813948532874374937191); Chris@10: DVK(KP831469612, +0.831469612302545237078788377617905756738560812); Chris@10: DVK(KP382683432, +0.382683432365089771728459984030398866761344562); Chris@10: DVK(KP923879532, +0.923879532511286756128183189396788286822416626); Chris@10: DVK(KP707106781, +0.707106781186547524400844362104849039284835938); Chris@10: { Chris@10: INT m; Chris@10: R *x; Chris@10: x = ii; Chris@10: for (m = mb, W = W + (mb * ((TWVL / VL) * 126)); m < me; m = m + VL, x = x + (VL * ms), W = W + (TWVL * 126), MAKE_VOLATILE_STRIDE(64, rs)) { Chris@10: V Tg, T4B, T6v, T7G, T3r, T4w, T5q, T7F, T5Y, T62, T28, T4d, T2g, T4a, T7g; Chris@10: V T7Y, T6f, T6j, T2Z, T4k, T37, T4h, T7n, T81, T7w, T7x, T7y, T5M, T6q, T1k; Chris@10: V T4s, T1r, T4t, T7t, T7u, T7v, T5F, T6p, TV, T4p, T12, T4q, T7A, T7B, TD; Chris@10: V T4x, T3k, T4C, T5x, T6s, T1R, T4b, T7j, T7Z, T2j, T4e, T5V, T63, T2I, T4i; Chris@10: V T7q, T82, T3a, T4l, T6c, T6k; Chris@10: { Chris@10: V T1, T3, T3p, T3n, Tb, Td, Te, T6, T8, T9, T2, T3o, T3m; Chris@10: T1 = LD(&(x[0]), ms, &(x[0])); Chris@10: T2 = LD(&(x[WS(rs, 32)]), ms, &(x[0])); Chris@10: T3 = BYTW(&(W[TWVL * 62]), T2); Chris@10: T3o = LD(&(x[WS(rs, 48)]), ms, &(x[0])); Chris@10: T3p = BYTW(&(W[TWVL * 94]), T3o); Chris@10: T3m = LD(&(x[WS(rs, 16)]), ms, &(x[0])); Chris@10: T3n = BYTW(&(W[TWVL * 30]), T3m); Chris@10: { Chris@10: V Ta, Tc, T5, T7; Chris@10: Ta = LD(&(x[WS(rs, 56)]), ms, &(x[0])); Chris@10: Tb = BYTW(&(W[TWVL * 110]), Ta); Chris@10: Tc = LD(&(x[WS(rs, 24)]), ms, &(x[0])); Chris@10: Td = BYTW(&(W[TWVL * 46]), Tc); Chris@10: Te = VSUB(Tb, Td); Chris@10: T5 = LD(&(x[WS(rs, 8)]), ms, &(x[0])); Chris@10: T6 = BYTW(&(W[TWVL * 14]), T5); Chris@10: T7 = LD(&(x[WS(rs, 40)]), ms, &(x[0])); Chris@10: T8 = BYTW(&(W[TWVL * 78]), T7); Chris@10: T9 = VSUB(T6, T8); Chris@10: } Chris@10: { Chris@10: V T4, Tf, T6t, T6u; Chris@10: T4 = VSUB(T1, T3); Chris@10: Tf = VMUL(LDK(KP707106781), VADD(T9, Te)); Chris@10: Tg = VSUB(T4, Tf); Chris@10: T4B = VADD(T4, Tf); Chris@10: T6t = VADD(T6, T8); Chris@10: T6u = VADD(Tb, Td); Chris@10: T6v = VSUB(T6t, T6u); Chris@10: T7G = VADD(T6t, T6u); Chris@10: } Chris@10: { Chris@10: V T3l, T3q, T5o, T5p; Chris@10: T3l = VMUL(LDK(KP707106781), VSUB(T9, Te)); Chris@10: T3q = VSUB(T3n, T3p); Chris@10: T3r = VSUB(T3l, T3q); Chris@10: T4w = VADD(T3q, T3l); Chris@10: T5o = VADD(T1, T3); Chris@10: T5p = VADD(T3n, T3p); Chris@10: T5q = VSUB(T5o, T5p); Chris@10: T7F = VADD(T5o, T5p); Chris@10: } Chris@10: } Chris@10: { Chris@10: V T24, T26, T61, T2b, T2d, T60, T1W, T5W, T21, T5X, T22, T27; Chris@10: { Chris@10: V T23, T25, T2a, T2c; Chris@10: T23 = LD(&(x[WS(rs, 17)]), ms, &(x[WS(rs, 1)])); Chris@10: T24 = BYTW(&(W[TWVL * 32]), T23); Chris@10: T25 = LD(&(x[WS(rs, 49)]), ms, &(x[WS(rs, 1)])); Chris@10: T26 = BYTW(&(W[TWVL * 96]), T25); Chris@10: T61 = VADD(T24, T26); Chris@10: T2a = LD(&(x[WS(rs, 1)]), ms, &(x[WS(rs, 1)])); Chris@10: T2b = BYTW(&(W[0]), T2a); Chris@10: T2c = LD(&(x[WS(rs, 33)]), ms, &(x[WS(rs, 1)])); Chris@10: T2d = BYTW(&(W[TWVL * 64]), T2c); Chris@10: T60 = VADD(T2b, T2d); Chris@10: } Chris@10: { Chris@10: V T1T, T1V, T1S, T1U; Chris@10: T1S = LD(&(x[WS(rs, 9)]), ms, &(x[WS(rs, 1)])); Chris@10: T1T = BYTW(&(W[TWVL * 16]), T1S); Chris@10: T1U = LD(&(x[WS(rs, 41)]), ms, &(x[WS(rs, 1)])); Chris@10: T1V = BYTW(&(W[TWVL * 80]), T1U); Chris@10: T1W = VSUB(T1T, T1V); Chris@10: T5W = VADD(T1T, T1V); Chris@10: } Chris@10: { Chris@10: V T1Y, T20, T1X, T1Z; Chris@10: T1X = LD(&(x[WS(rs, 57)]), ms, &(x[WS(rs, 1)])); Chris@10: T1Y = BYTW(&(W[TWVL * 112]), T1X); Chris@10: T1Z = LD(&(x[WS(rs, 25)]), ms, &(x[WS(rs, 1)])); Chris@10: T20 = BYTW(&(W[TWVL * 48]), T1Z); Chris@10: T21 = VSUB(T1Y, T20); Chris@10: T5X = VADD(T1Y, T20); Chris@10: } Chris@10: T5Y = VSUB(T5W, T5X); Chris@10: T62 = VSUB(T60, T61); Chris@10: T22 = VMUL(LDK(KP707106781), VSUB(T1W, T21)); Chris@10: T27 = VSUB(T24, T26); Chris@10: T28 = VSUB(T22, T27); Chris@10: T4d = VADD(T27, T22); Chris@10: { Chris@10: V T2e, T2f, T7e, T7f; Chris@10: T2e = VSUB(T2b, T2d); Chris@10: T2f = VMUL(LDK(KP707106781), VADD(T1W, T21)); Chris@10: T2g = VSUB(T2e, T2f); Chris@10: T4a = VADD(T2e, T2f); Chris@10: T7e = VADD(T60, T61); Chris@10: T7f = VADD(T5W, T5X); Chris@10: T7g = VSUB(T7e, T7f); Chris@10: T7Y = VADD(T7e, T7f); Chris@10: } Chris@10: } Chris@10: { Chris@10: V T2V, T2X, T6i, T32, T34, T6h, T2N, T6d, T2S, T6e, T2T, T2Y; Chris@10: { Chris@10: V T2U, T2W, T31, T33; Chris@10: T2U = LD(&(x[WS(rs, 15)]), ms, &(x[WS(rs, 1)])); Chris@10: T2V = BYTW(&(W[TWVL * 28]), T2U); Chris@10: T2W = LD(&(x[WS(rs, 47)]), ms, &(x[WS(rs, 1)])); Chris@10: T2X = BYTW(&(W[TWVL * 92]), T2W); Chris@10: T6i = VADD(T2V, T2X); Chris@10: T31 = LD(&(x[WS(rs, 63)]), ms, &(x[WS(rs, 1)])); Chris@10: T32 = BYTW(&(W[TWVL * 124]), T31); Chris@10: T33 = LD(&(x[WS(rs, 31)]), ms, &(x[WS(rs, 1)])); Chris@10: T34 = BYTW(&(W[TWVL * 60]), T33); Chris@10: T6h = VADD(T32, T34); Chris@10: } Chris@10: { Chris@10: V T2K, T2M, T2J, T2L; Chris@10: T2J = LD(&(x[WS(rs, 7)]), ms, &(x[WS(rs, 1)])); Chris@10: T2K = BYTW(&(W[TWVL * 12]), T2J); Chris@10: T2L = LD(&(x[WS(rs, 39)]), ms, &(x[WS(rs, 1)])); Chris@10: T2M = BYTW(&(W[TWVL * 76]), T2L); Chris@10: T2N = VSUB(T2K, T2M); Chris@10: T6d = VADD(T2K, T2M); Chris@10: } Chris@10: { Chris@10: V T2P, T2R, T2O, T2Q; Chris@10: T2O = LD(&(x[WS(rs, 55)]), ms, &(x[WS(rs, 1)])); Chris@10: T2P = BYTW(&(W[TWVL * 108]), T2O); Chris@10: T2Q = LD(&(x[WS(rs, 23)]), ms, &(x[WS(rs, 1)])); Chris@10: T2R = BYTW(&(W[TWVL * 44]), T2Q); Chris@10: T2S = VSUB(T2P, T2R); Chris@10: T6e = VADD(T2P, T2R); Chris@10: } Chris@10: T6f = VSUB(T6d, T6e); Chris@10: T6j = VSUB(T6h, T6i); Chris@10: T2T = VMUL(LDK(KP707106781), VSUB(T2N, T2S)); Chris@10: T2Y = VSUB(T2V, T2X); Chris@10: T2Z = VSUB(T2T, T2Y); Chris@10: T4k = VADD(T2Y, T2T); Chris@10: { Chris@10: V T35, T36, T7l, T7m; Chris@10: T35 = VSUB(T32, T34); Chris@10: T36 = VMUL(LDK(KP707106781), VADD(T2N, T2S)); Chris@10: T37 = VSUB(T35, T36); Chris@10: T4h = VADD(T35, T36); Chris@10: T7l = VADD(T6h, T6i); Chris@10: T7m = VADD(T6d, T6e); Chris@10: T7n = VSUB(T7l, T7m); Chris@10: T81 = VADD(T7l, T7m); Chris@10: } Chris@10: } Chris@10: { Chris@10: V T1g, T1i, T5K, T1m, T1o, T5J, T18, T5G, T1d, T5H, T5I, T5L; Chris@10: { Chris@10: V T1f, T1h, T1l, T1n; Chris@10: T1f = LD(&(x[WS(rs, 14)]), ms, &(x[0])); Chris@10: T1g = BYTW(&(W[TWVL * 26]), T1f); Chris@10: T1h = LD(&(x[WS(rs, 46)]), ms, &(x[0])); Chris@10: T1i = BYTW(&(W[TWVL * 90]), T1h); Chris@10: T5K = VADD(T1g, T1i); Chris@10: T1l = LD(&(x[WS(rs, 62)]), ms, &(x[0])); Chris@10: T1m = BYTW(&(W[TWVL * 122]), T1l); Chris@10: T1n = LD(&(x[WS(rs, 30)]), ms, &(x[0])); Chris@10: T1o = BYTW(&(W[TWVL * 58]), T1n); Chris@10: T5J = VADD(T1m, T1o); Chris@10: } Chris@10: { Chris@10: V T15, T17, T14, T16; Chris@10: T14 = LD(&(x[WS(rs, 6)]), ms, &(x[0])); Chris@10: T15 = BYTW(&(W[TWVL * 10]), T14); Chris@10: T16 = LD(&(x[WS(rs, 38)]), ms, &(x[0])); Chris@10: T17 = BYTW(&(W[TWVL * 74]), T16); Chris@10: T18 = VSUB(T15, T17); Chris@10: T5G = VADD(T15, T17); Chris@10: } Chris@10: { Chris@10: V T1a, T1c, T19, T1b; Chris@10: T19 = LD(&(x[WS(rs, 54)]), ms, &(x[0])); Chris@10: T1a = BYTW(&(W[TWVL * 106]), T19); Chris@10: T1b = LD(&(x[WS(rs, 22)]), ms, &(x[0])); Chris@10: T1c = BYTW(&(W[TWVL * 42]), T1b); Chris@10: T1d = VSUB(T1a, T1c); Chris@10: T5H = VADD(T1a, T1c); Chris@10: } Chris@10: T7w = VADD(T5J, T5K); Chris@10: T7x = VADD(T5G, T5H); Chris@10: T7y = VSUB(T7w, T7x); Chris@10: T5I = VSUB(T5G, T5H); Chris@10: T5L = VSUB(T5J, T5K); Chris@10: T5M = VFNMS(LDK(KP382683432), T5L, VMUL(LDK(KP923879532), T5I)); Chris@10: T6q = VFMA(LDK(KP923879532), T5L, VMUL(LDK(KP382683432), T5I)); Chris@10: { Chris@10: V T1e, T1j, T1p, T1q; Chris@10: T1e = VMUL(LDK(KP707106781), VSUB(T18, T1d)); Chris@10: T1j = VSUB(T1g, T1i); Chris@10: T1k = VSUB(T1e, T1j); Chris@10: T4s = VADD(T1j, T1e); Chris@10: T1p = VSUB(T1m, T1o); Chris@10: T1q = VMUL(LDK(KP707106781), VADD(T18, T1d)); Chris@10: T1r = VSUB(T1p, T1q); Chris@10: T4t = VADD(T1p, T1q); Chris@10: } Chris@10: } Chris@10: { Chris@10: V TR, TT, T5A, TX, TZ, T5z, TJ, T5C, TO, T5D, T5B, T5E; Chris@10: { Chris@10: V TQ, TS, TW, TY; Chris@10: TQ = LD(&(x[WS(rs, 18)]), ms, &(x[0])); Chris@10: TR = BYTW(&(W[TWVL * 34]), TQ); Chris@10: TS = LD(&(x[WS(rs, 50)]), ms, &(x[0])); Chris@10: TT = BYTW(&(W[TWVL * 98]), TS); Chris@10: T5A = VADD(TR, TT); Chris@10: TW = LD(&(x[WS(rs, 2)]), ms, &(x[0])); Chris@10: TX = BYTW(&(W[TWVL * 2]), TW); Chris@10: TY = LD(&(x[WS(rs, 34)]), ms, &(x[0])); Chris@10: TZ = BYTW(&(W[TWVL * 66]), TY); Chris@10: T5z = VADD(TX, TZ); Chris@10: } Chris@10: { Chris@10: V TG, TI, TF, TH; Chris@10: TF = LD(&(x[WS(rs, 10)]), ms, &(x[0])); Chris@10: TG = BYTW(&(W[TWVL * 18]), TF); Chris@10: TH = LD(&(x[WS(rs, 42)]), ms, &(x[0])); Chris@10: TI = BYTW(&(W[TWVL * 82]), TH); Chris@10: TJ = VSUB(TG, TI); Chris@10: T5C = VADD(TG, TI); Chris@10: } Chris@10: { Chris@10: V TL, TN, TK, TM; Chris@10: TK = LD(&(x[WS(rs, 58)]), ms, &(x[0])); Chris@10: TL = BYTW(&(W[TWVL * 114]), TK); Chris@10: TM = LD(&(x[WS(rs, 26)]), ms, &(x[0])); Chris@10: TN = BYTW(&(W[TWVL * 50]), TM); Chris@10: TO = VSUB(TL, TN); Chris@10: T5D = VADD(TL, TN); Chris@10: } Chris@10: T7t = VADD(T5z, T5A); Chris@10: T7u = VADD(T5C, T5D); Chris@10: T7v = VSUB(T7t, T7u); Chris@10: T5B = VSUB(T5z, T5A); Chris@10: T5E = VSUB(T5C, T5D); Chris@10: T5F = VFMA(LDK(KP382683432), T5B, VMUL(LDK(KP923879532), T5E)); Chris@10: T6p = VFNMS(LDK(KP382683432), T5E, VMUL(LDK(KP923879532), T5B)); Chris@10: { Chris@10: V TP, TU, T10, T11; Chris@10: TP = VMUL(LDK(KP707106781), VSUB(TJ, TO)); Chris@10: TU = VSUB(TR, TT); Chris@10: TV = VSUB(TP, TU); Chris@10: T4p = VADD(TU, TP); Chris@10: T10 = VSUB(TX, TZ); Chris@10: T11 = VMUL(LDK(KP707106781), VADD(TJ, TO)); Chris@10: T12 = VSUB(T10, T11); Chris@10: T4q = VADD(T10, T11); Chris@10: } Chris@10: } Chris@10: { Chris@10: V Tl, T5r, TB, T5u, Tq, T5s, Tw, T5v, Tr, TC; Chris@10: { Chris@10: V Ti, Tk, Th, Tj; Chris@10: Th = LD(&(x[WS(rs, 4)]), ms, &(x[0])); Chris@10: Ti = BYTW(&(W[TWVL * 6]), Th); Chris@10: Tj = LD(&(x[WS(rs, 36)]), ms, &(x[0])); Chris@10: Tk = BYTW(&(W[TWVL * 70]), Tj); Chris@10: Tl = VSUB(Ti, Tk); Chris@10: T5r = VADD(Ti, Tk); Chris@10: } Chris@10: { Chris@10: V Ty, TA, Tx, Tz; Chris@10: Tx = LD(&(x[WS(rs, 60)]), ms, &(x[0])); Chris@10: Ty = BYTW(&(W[TWVL * 118]), Tx); Chris@10: Tz = LD(&(x[WS(rs, 28)]), ms, &(x[0])); Chris@10: TA = BYTW(&(W[TWVL * 54]), Tz); Chris@10: TB = VSUB(Ty, TA); Chris@10: T5u = VADD(Ty, TA); Chris@10: } Chris@10: { Chris@10: V Tn, Tp, Tm, To; Chris@10: Tm = LD(&(x[WS(rs, 20)]), ms, &(x[0])); Chris@10: Tn = BYTW(&(W[TWVL * 38]), Tm); Chris@10: To = LD(&(x[WS(rs, 52)]), ms, &(x[0])); Chris@10: Tp = BYTW(&(W[TWVL * 102]), To); Chris@10: Tq = VSUB(Tn, Tp); Chris@10: T5s = VADD(Tn, Tp); Chris@10: } Chris@10: { Chris@10: V Tt, Tv, Ts, Tu; Chris@10: Ts = LD(&(x[WS(rs, 12)]), ms, &(x[0])); Chris@10: Tt = BYTW(&(W[TWVL * 22]), Ts); Chris@10: Tu = LD(&(x[WS(rs, 44)]), ms, &(x[0])); Chris@10: Tv = BYTW(&(W[TWVL * 86]), Tu); Chris@10: Tw = VSUB(Tt, Tv); Chris@10: T5v = VADD(Tt, Tv); Chris@10: } Chris@10: T7A = VADD(T5r, T5s); Chris@10: T7B = VADD(T5u, T5v); Chris@10: Tr = VFMA(LDK(KP382683432), Tl, VMUL(LDK(KP923879532), Tq)); Chris@10: TC = VFNMS(LDK(KP382683432), TB, VMUL(LDK(KP923879532), Tw)); Chris@10: TD = VSUB(Tr, TC); Chris@10: T4x = VADD(Tr, TC); Chris@10: { Chris@10: V T3i, T3j, T5t, T5w; Chris@10: T3i = VFNMS(LDK(KP382683432), Tq, VMUL(LDK(KP923879532), Tl)); Chris@10: T3j = VFMA(LDK(KP923879532), TB, VMUL(LDK(KP382683432), Tw)); Chris@10: T3k = VSUB(T3i, T3j); Chris@10: T4C = VADD(T3i, T3j); Chris@10: T5t = VSUB(T5r, T5s); Chris@10: T5w = VSUB(T5u, T5v); Chris@10: T5x = VMUL(LDK(KP707106781), VADD(T5t, T5w)); Chris@10: T6s = VMUL(LDK(KP707106781), VSUB(T5t, T5w)); Chris@10: } Chris@10: } Chris@10: { Chris@10: V T1z, T5P, T1P, T5T, T1E, T5Q, T1K, T5S; Chris@10: { Chris@10: V T1w, T1y, T1v, T1x; Chris@10: T1v = LD(&(x[WS(rs, 5)]), ms, &(x[WS(rs, 1)])); Chris@10: T1w = BYTW(&(W[TWVL * 8]), T1v); Chris@10: T1x = LD(&(x[WS(rs, 37)]), ms, &(x[WS(rs, 1)])); Chris@10: T1y = BYTW(&(W[TWVL * 72]), T1x); Chris@10: T1z = VSUB(T1w, T1y); Chris@10: T5P = VADD(T1w, T1y); Chris@10: } Chris@10: { Chris@10: V T1M, T1O, T1L, T1N; Chris@10: T1L = LD(&(x[WS(rs, 13)]), ms, &(x[WS(rs, 1)])); Chris@10: T1M = BYTW(&(W[TWVL * 24]), T1L); Chris@10: T1N = LD(&(x[WS(rs, 45)]), ms, &(x[WS(rs, 1)])); Chris@10: T1O = BYTW(&(W[TWVL * 88]), T1N); Chris@10: T1P = VSUB(T1M, T1O); Chris@10: T5T = VADD(T1M, T1O); Chris@10: } Chris@10: { Chris@10: V T1B, T1D, T1A, T1C; Chris@10: T1A = LD(&(x[WS(rs, 21)]), ms, &(x[WS(rs, 1)])); Chris@10: T1B = BYTW(&(W[TWVL * 40]), T1A); Chris@10: T1C = LD(&(x[WS(rs, 53)]), ms, &(x[WS(rs, 1)])); Chris@10: T1D = BYTW(&(W[TWVL * 104]), T1C); Chris@10: T1E = VSUB(T1B, T1D); Chris@10: T5Q = VADD(T1B, T1D); Chris@10: } Chris@10: { Chris@10: V T1H, T1J, T1G, T1I; Chris@10: T1G = LD(&(x[WS(rs, 61)]), ms, &(x[WS(rs, 1)])); Chris@10: T1H = BYTW(&(W[TWVL * 120]), T1G); Chris@10: T1I = LD(&(x[WS(rs, 29)]), ms, &(x[WS(rs, 1)])); Chris@10: T1J = BYTW(&(W[TWVL * 56]), T1I); Chris@10: T1K = VSUB(T1H, T1J); Chris@10: T5S = VADD(T1H, T1J); Chris@10: } Chris@10: { Chris@10: V T1F, T1Q, T7h, T7i; Chris@10: T1F = VFNMS(LDK(KP382683432), T1E, VMUL(LDK(KP923879532), T1z)); Chris@10: T1Q = VFMA(LDK(KP923879532), T1K, VMUL(LDK(KP382683432), T1P)); Chris@10: T1R = VSUB(T1F, T1Q); Chris@10: T4b = VADD(T1F, T1Q); Chris@10: T7h = VADD(T5P, T5Q); Chris@10: T7i = VADD(T5S, T5T); Chris@10: T7j = VSUB(T7h, T7i); Chris@10: T7Z = VADD(T7h, T7i); Chris@10: } Chris@10: { Chris@10: V T2h, T2i, T5R, T5U; Chris@10: T2h = VFMA(LDK(KP382683432), T1z, VMUL(LDK(KP923879532), T1E)); Chris@10: T2i = VFNMS(LDK(KP382683432), T1K, VMUL(LDK(KP923879532), T1P)); Chris@10: T2j = VSUB(T2h, T2i); Chris@10: T4e = VADD(T2h, T2i); Chris@10: T5R = VSUB(T5P, T5Q); Chris@10: T5U = VSUB(T5S, T5T); Chris@10: T5V = VMUL(LDK(KP707106781), VSUB(T5R, T5U)); Chris@10: T63 = VMUL(LDK(KP707106781), VADD(T5R, T5U)); Chris@10: } Chris@10: } Chris@10: { Chris@10: V T2q, T66, T2G, T6a, T2v, T67, T2B, T69; Chris@10: { Chris@10: V T2n, T2p, T2m, T2o; Chris@10: T2m = LD(&(x[WS(rs, 3)]), ms, &(x[WS(rs, 1)])); Chris@10: T2n = BYTW(&(W[TWVL * 4]), T2m); Chris@10: T2o = LD(&(x[WS(rs, 35)]), ms, &(x[WS(rs, 1)])); Chris@10: T2p = BYTW(&(W[TWVL * 68]), T2o); Chris@10: T2q = VSUB(T2n, T2p); Chris@10: T66 = VADD(T2n, T2p); Chris@10: } Chris@10: { Chris@10: V T2D, T2F, T2C, T2E; Chris@10: T2C = LD(&(x[WS(rs, 11)]), ms, &(x[WS(rs, 1)])); Chris@10: T2D = BYTW(&(W[TWVL * 20]), T2C); Chris@10: T2E = LD(&(x[WS(rs, 43)]), ms, &(x[WS(rs, 1)])); Chris@10: T2F = BYTW(&(W[TWVL * 84]), T2E); Chris@10: T2G = VSUB(T2D, T2F); Chris@10: T6a = VADD(T2D, T2F); Chris@10: } Chris@10: { Chris@10: V T2s, T2u, T2r, T2t; Chris@10: T2r = LD(&(x[WS(rs, 19)]), ms, &(x[WS(rs, 1)])); Chris@10: T2s = BYTW(&(W[TWVL * 36]), T2r); Chris@10: T2t = LD(&(x[WS(rs, 51)]), ms, &(x[WS(rs, 1)])); Chris@10: T2u = BYTW(&(W[TWVL * 100]), T2t); Chris@10: T2v = VSUB(T2s, T2u); Chris@10: T67 = VADD(T2s, T2u); Chris@10: } Chris@10: { Chris@10: V T2y, T2A, T2x, T2z; Chris@10: T2x = LD(&(x[WS(rs, 59)]), ms, &(x[WS(rs, 1)])); Chris@10: T2y = BYTW(&(W[TWVL * 116]), T2x); Chris@10: T2z = LD(&(x[WS(rs, 27)]), ms, &(x[WS(rs, 1)])); Chris@10: T2A = BYTW(&(W[TWVL * 52]), T2z); Chris@10: T2B = VSUB(T2y, T2A); Chris@10: T69 = VADD(T2y, T2A); Chris@10: } Chris@10: { Chris@10: V T2w, T2H, T7o, T7p; Chris@10: T2w = VFNMS(LDK(KP382683432), T2v, VMUL(LDK(KP923879532), T2q)); Chris@10: T2H = VFMA(LDK(KP923879532), T2B, VMUL(LDK(KP382683432), T2G)); Chris@10: T2I = VSUB(T2w, T2H); Chris@10: T4i = VADD(T2w, T2H); Chris@10: T7o = VADD(T66, T67); Chris@10: T7p = VADD(T69, T6a); Chris@10: T7q = VSUB(T7o, T7p); Chris@10: T82 = VADD(T7o, T7p); Chris@10: } Chris@10: { Chris@10: V T38, T39, T68, T6b; Chris@10: T38 = VFMA(LDK(KP382683432), T2q, VMUL(LDK(KP923879532), T2v)); Chris@10: T39 = VFNMS(LDK(KP382683432), T2B, VMUL(LDK(KP923879532), T2G)); Chris@10: T3a = VSUB(T38, T39); Chris@10: T4l = VADD(T38, T39); Chris@10: T68 = VSUB(T66, T67); Chris@10: T6b = VSUB(T69, T6a); Chris@10: T6c = VMUL(LDK(KP707106781), VSUB(T68, T6b)); Chris@10: T6k = VMUL(LDK(KP707106781), VADD(T68, T6b)); Chris@10: } Chris@10: } Chris@10: { Chris@10: V T7s, T7R, T7M, T7U, T7D, T7T, T7J, T7Q; Chris@10: { Chris@10: V T7k, T7r, T7K, T7L; Chris@10: T7k = VFNMS(LDK(KP382683432), T7j, VMUL(LDK(KP923879532), T7g)); Chris@10: T7r = VFMA(LDK(KP923879532), T7n, VMUL(LDK(KP382683432), T7q)); Chris@10: T7s = VSUB(T7k, T7r); Chris@10: T7R = VADD(T7k, T7r); Chris@10: T7K = VFMA(LDK(KP382683432), T7g, VMUL(LDK(KP923879532), T7j)); Chris@10: T7L = VFNMS(LDK(KP382683432), T7n, VMUL(LDK(KP923879532), T7q)); Chris@10: T7M = VSUB(T7K, T7L); Chris@10: T7U = VADD(T7K, T7L); Chris@10: } Chris@10: { Chris@10: V T7z, T7C, T7H, T7I; Chris@10: T7z = VMUL(LDK(KP707106781), VSUB(T7v, T7y)); Chris@10: T7C = VSUB(T7A, T7B); Chris@10: T7D = VSUB(T7z, T7C); Chris@10: T7T = VADD(T7C, T7z); Chris@10: T7H = VSUB(T7F, T7G); Chris@10: T7I = VMUL(LDK(KP707106781), VADD(T7v, T7y)); Chris@10: T7J = VSUB(T7H, T7I); Chris@10: T7Q = VADD(T7H, T7I); Chris@10: } Chris@10: { Chris@10: V T7E, T7N, T7W, T7X; Chris@10: T7E = VBYI(VSUB(T7s, T7D)); Chris@10: T7N = VSUB(T7J, T7M); Chris@10: ST(&(x[WS(rs, 20)]), VADD(T7E, T7N), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 44)]), VSUB(T7N, T7E), ms, &(x[0])); Chris@10: T7W = VSUB(T7Q, T7R); Chris@10: T7X = VBYI(VSUB(T7U, T7T)); Chris@10: ST(&(x[WS(rs, 36)]), VSUB(T7W, T7X), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 28)]), VADD(T7W, T7X), ms, &(x[0])); Chris@10: } Chris@10: { Chris@10: V T7O, T7P, T7S, T7V; Chris@10: T7O = VBYI(VADD(T7D, T7s)); Chris@10: T7P = VADD(T7J, T7M); Chris@10: ST(&(x[WS(rs, 12)]), VADD(T7O, T7P), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 52)]), VSUB(T7P, T7O), ms, &(x[0])); Chris@10: T7S = VADD(T7Q, T7R); Chris@10: T7V = VBYI(VADD(T7T, T7U)); Chris@10: ST(&(x[WS(rs, 60)]), VSUB(T7S, T7V), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 4)]), VADD(T7S, T7V), ms, &(x[0])); Chris@10: } Chris@10: } Chris@10: { Chris@10: V T84, T8c, T8l, T8n, T87, T8h, T8b, T8g, T8i, T8m; Chris@10: { Chris@10: V T80, T83, T8j, T8k; Chris@10: T80 = VSUB(T7Y, T7Z); Chris@10: T83 = VSUB(T81, T82); Chris@10: T84 = VMUL(LDK(KP707106781), VSUB(T80, T83)); Chris@10: T8c = VMUL(LDK(KP707106781), VADD(T80, T83)); Chris@10: T8j = VADD(T7Y, T7Z); Chris@10: T8k = VADD(T81, T82); Chris@10: T8l = VBYI(VSUB(T8j, T8k)); Chris@10: T8n = VADD(T8j, T8k); Chris@10: } Chris@10: { Chris@10: V T85, T86, T89, T8a; Chris@10: T85 = VADD(T7t, T7u); Chris@10: T86 = VADD(T7w, T7x); Chris@10: T87 = VSUB(T85, T86); Chris@10: T8h = VADD(T85, T86); Chris@10: T89 = VADD(T7F, T7G); Chris@10: T8a = VADD(T7A, T7B); Chris@10: T8b = VSUB(T89, T8a); Chris@10: T8g = VADD(T89, T8a); Chris@10: } Chris@10: T8i = VSUB(T8g, T8h); Chris@10: ST(&(x[WS(rs, 48)]), VSUB(T8i, T8l), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 16)]), VADD(T8i, T8l), ms, &(x[0])); Chris@10: T8m = VADD(T8g, T8h); Chris@10: ST(&(x[WS(rs, 32)]), VSUB(T8m, T8n), ms, &(x[0])); Chris@10: ST(&(x[0]), VADD(T8m, T8n), ms, &(x[0])); Chris@10: { Chris@10: V T88, T8d, T8e, T8f; Chris@10: T88 = VBYI(VSUB(T84, T87)); Chris@10: T8d = VSUB(T8b, T8c); Chris@10: ST(&(x[WS(rs, 24)]), VADD(T88, T8d), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 40)]), VSUB(T8d, T88), ms, &(x[0])); Chris@10: T8e = VBYI(VADD(T87, T84)); Chris@10: T8f = VADD(T8b, T8c); Chris@10: ST(&(x[WS(rs, 8)]), VADD(T8e, T8f), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 56)]), VSUB(T8f, T8e), ms, &(x[0])); Chris@10: } Chris@10: } Chris@10: { Chris@10: V T5O, T6H, T6x, T6F, T6n, T6I, T6A, T6E; Chris@10: { Chris@10: V T5y, T5N, T6r, T6w; Chris@10: T5y = VSUB(T5q, T5x); Chris@10: T5N = VSUB(T5F, T5M); Chris@10: T5O = VSUB(T5y, T5N); Chris@10: T6H = VADD(T5y, T5N); Chris@10: T6r = VSUB(T6p, T6q); Chris@10: T6w = VSUB(T6s, T6v); Chris@10: T6x = VSUB(T6r, T6w); Chris@10: T6F = VADD(T6w, T6r); Chris@10: { Chris@10: V T65, T6y, T6m, T6z; Chris@10: { Chris@10: V T5Z, T64, T6g, T6l; Chris@10: T5Z = VSUB(T5V, T5Y); Chris@10: T64 = VSUB(T62, T63); Chris@10: T65 = VFMA(LDK(KP831469612), T5Z, VMUL(LDK(KP555570233), T64)); Chris@10: T6y = VFNMS(LDK(KP555570233), T5Z, VMUL(LDK(KP831469612), T64)); Chris@10: T6g = VSUB(T6c, T6f); Chris@10: T6l = VSUB(T6j, T6k); Chris@10: T6m = VFNMS(LDK(KP555570233), T6l, VMUL(LDK(KP831469612), T6g)); Chris@10: T6z = VFMA(LDK(KP555570233), T6g, VMUL(LDK(KP831469612), T6l)); Chris@10: } Chris@10: T6n = VSUB(T65, T6m); Chris@10: T6I = VADD(T6y, T6z); Chris@10: T6A = VSUB(T6y, T6z); Chris@10: T6E = VADD(T65, T6m); Chris@10: } Chris@10: } Chris@10: { Chris@10: V T6o, T6B, T6K, T6L; Chris@10: T6o = VADD(T5O, T6n); Chris@10: T6B = VBYI(VADD(T6x, T6A)); Chris@10: ST(&(x[WS(rs, 54)]), VSUB(T6o, T6B), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 10)]), VADD(T6o, T6B), ms, &(x[0])); Chris@10: T6K = VBYI(VADD(T6F, T6E)); Chris@10: T6L = VADD(T6H, T6I); Chris@10: ST(&(x[WS(rs, 6)]), VADD(T6K, T6L), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 58)]), VSUB(T6L, T6K), ms, &(x[0])); Chris@10: } Chris@10: { Chris@10: V T6C, T6D, T6G, T6J; Chris@10: T6C = VSUB(T5O, T6n); Chris@10: T6D = VBYI(VSUB(T6A, T6x)); Chris@10: ST(&(x[WS(rs, 42)]), VSUB(T6C, T6D), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 22)]), VADD(T6C, T6D), ms, &(x[0])); Chris@10: T6G = VBYI(VSUB(T6E, T6F)); Chris@10: T6J = VSUB(T6H, T6I); Chris@10: ST(&(x[WS(rs, 26)]), VADD(T6G, T6J), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 38)]), VSUB(T6J, T6G), ms, &(x[0])); Chris@10: } Chris@10: } Chris@10: { Chris@10: V T6O, T79, T6Z, T77, T6V, T7a, T72, T76; Chris@10: { Chris@10: V T6M, T6N, T6X, T6Y; Chris@10: T6M = VADD(T5q, T5x); Chris@10: T6N = VADD(T6p, T6q); Chris@10: T6O = VSUB(T6M, T6N); Chris@10: T79 = VADD(T6M, T6N); Chris@10: T6X = VADD(T5F, T5M); Chris@10: T6Y = VADD(T6v, T6s); Chris@10: T6Z = VSUB(T6X, T6Y); Chris@10: T77 = VADD(T6Y, T6X); Chris@10: { Chris@10: V T6R, T70, T6U, T71; Chris@10: { Chris@10: V T6P, T6Q, T6S, T6T; Chris@10: T6P = VADD(T5Y, T5V); Chris@10: T6Q = VADD(T62, T63); Chris@10: T6R = VFMA(LDK(KP980785280), T6P, VMUL(LDK(KP195090322), T6Q)); Chris@10: T70 = VFNMS(LDK(KP195090322), T6P, VMUL(LDK(KP980785280), T6Q)); Chris@10: T6S = VADD(T6f, T6c); Chris@10: T6T = VADD(T6j, T6k); Chris@10: T6U = VFNMS(LDK(KP195090322), T6T, VMUL(LDK(KP980785280), T6S)); Chris@10: T71 = VFMA(LDK(KP195090322), T6S, VMUL(LDK(KP980785280), T6T)); Chris@10: } Chris@10: T6V = VSUB(T6R, T6U); Chris@10: T7a = VADD(T70, T71); Chris@10: T72 = VSUB(T70, T71); Chris@10: T76 = VADD(T6R, T6U); Chris@10: } Chris@10: } Chris@10: { Chris@10: V T6W, T73, T7c, T7d; Chris@10: T6W = VADD(T6O, T6V); Chris@10: T73 = VBYI(VADD(T6Z, T72)); Chris@10: ST(&(x[WS(rs, 50)]), VSUB(T6W, T73), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 14)]), VADD(T6W, T73), ms, &(x[0])); Chris@10: T7c = VBYI(VADD(T77, T76)); Chris@10: T7d = VADD(T79, T7a); Chris@10: ST(&(x[WS(rs, 2)]), VADD(T7c, T7d), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 62)]), VSUB(T7d, T7c), ms, &(x[0])); Chris@10: } Chris@10: { Chris@10: V T74, T75, T78, T7b; Chris@10: T74 = VSUB(T6O, T6V); Chris@10: T75 = VBYI(VSUB(T72, T6Z)); Chris@10: ST(&(x[WS(rs, 46)]), VSUB(T74, T75), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 18)]), VADD(T74, T75), ms, &(x[0])); Chris@10: T78 = VBYI(VSUB(T76, T77)); Chris@10: T7b = VSUB(T79, T7a); Chris@10: ST(&(x[WS(rs, 30)]), VADD(T78, T7b), ms, &(x[0])); Chris@10: ST(&(x[WS(rs, 34)]), VSUB(T7b, T78), ms, &(x[0])); Chris@10: } Chris@10: } Chris@10: { Chris@10: V T4z, T5g, T4R, T59, T4H, T5j, T4O, T55, T4o, T4S, T4K, T4P, T52, T5k, T5c; Chris@10: V T5h; Chris@10: { Chris@10: V T4y, T57, T4v, T58, T4r, T4u; Chris@10: T4y = VADD(T4w, T4x); Chris@10: T57 = VSUB(T4B, T4C); Chris@10: T4r = VFMA(LDK(KP980785280), T4p, VMUL(LDK(KP195090322), T4q)); Chris@10: T4u = VFNMS(LDK(KP195090322), T4t, VMUL(LDK(KP980785280), T4s)); Chris@10: T4v = VADD(T4r, T4u); Chris@10: T58 = VSUB(T4r, T4u); Chris@10: T4z = VSUB(T4v, T4y); Chris@10: T5g = VADD(T57, T58); Chris@10: T4R = VADD(T4y, T4v); Chris@10: T59 = VSUB(T57, T58); Chris@10: } Chris@10: { Chris@10: V T4D, T54, T4G, T53, T4E, T4F; Chris@10: T4D = VADD(T4B, T4C); Chris@10: T54 = VSUB(T4x, T4w); Chris@10: T4E = VFNMS(LDK(KP195090322), T4p, VMUL(LDK(KP980785280), T4q)); Chris@10: T4F = VFMA(LDK(KP195090322), T4s, VMUL(LDK(KP980785280), T4t)); Chris@10: T4G = VADD(T4E, T4F); Chris@10: T53 = VSUB(T4E, T4F); Chris@10: T4H = VSUB(T4D, T4G); Chris@10: T5j = VADD(T54, T53); Chris@10: T4O = VADD(T4D, T4G); Chris@10: T55 = VSUB(T53, T54); Chris@10: } Chris@10: { Chris@10: V T4g, T4I, T4n, T4J; Chris@10: { Chris@10: V T4c, T4f, T4j, T4m; Chris@10: T4c = VADD(T4a, T4b); Chris@10: T4f = VADD(T4d, T4e); Chris@10: T4g = VFNMS(LDK(KP098017140), T4f, VMUL(LDK(KP995184726), T4c)); Chris@10: T4I = VFMA(LDK(KP098017140), T4c, VMUL(LDK(KP995184726), T4f)); Chris@10: T4j = VADD(T4h, T4i); Chris@10: T4m = VADD(T4k, T4l); Chris@10: T4n = VFMA(LDK(KP995184726), T4j, VMUL(LDK(KP098017140), T4m)); Chris@10: T4J = VFNMS(LDK(KP098017140), T4j, VMUL(LDK(KP995184726), T4m)); Chris@10: } Chris@10: T4o = VSUB(T4g, T4n); Chris@10: T4S = VADD(T4I, T4J); Chris@10: T4K = VSUB(T4I, T4J); Chris@10: T4P = VADD(T4g, T4n); Chris@10: } Chris@10: { Chris@10: V T4Y, T5a, T51, T5b; Chris@10: { Chris@10: V T4W, T4X, T4Z, T50; Chris@10: T4W = VSUB(T4a, T4b); Chris@10: T4X = VSUB(T4e, T4d); Chris@10: T4Y = VFNMS(LDK(KP634393284), T4X, VMUL(LDK(KP773010453), T4W)); Chris@10: T5a = VFMA(LDK(KP634393284), T4W, VMUL(LDK(KP773010453), T4X)); Chris@10: T4Z = VSUB(T4h, T4i); Chris@10: T50 = VSUB(T4l, T4k); Chris@10: T51 = VFMA(LDK(KP773010453), T4Z, VMUL(LDK(KP634393284), T50)); Chris@10: T5b = VFNMS(LDK(KP634393284), T4Z, VMUL(LDK(KP773010453), T50)); Chris@10: } Chris@10: T52 = VSUB(T4Y, T51); Chris@10: T5k = VADD(T5a, T5b); Chris@10: T5c = VSUB(T5a, T5b); Chris@10: T5h = VADD(T4Y, T51); Chris@10: } Chris@10: { Chris@10: V T4A, T4L, T5i, T5l; Chris@10: T4A = VBYI(VSUB(T4o, T4z)); Chris@10: T4L = VSUB(T4H, T4K); Chris@10: ST(&(x[WS(rs, 17)]), VADD(T4A, T4L), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 47)]), VSUB(T4L, T4A), ms, &(x[WS(rs, 1)])); Chris@10: T5i = VADD(T5g, T5h); Chris@10: T5l = VBYI(VADD(T5j, T5k)); Chris@10: ST(&(x[WS(rs, 57)]), VSUB(T5i, T5l), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 7)]), VADD(T5i, T5l), ms, &(x[WS(rs, 1)])); Chris@10: } Chris@10: { Chris@10: V T5m, T5n, T4M, T4N; Chris@10: T5m = VSUB(T5g, T5h); Chris@10: T5n = VBYI(VSUB(T5k, T5j)); Chris@10: ST(&(x[WS(rs, 39)]), VSUB(T5m, T5n), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 25)]), VADD(T5m, T5n), ms, &(x[WS(rs, 1)])); Chris@10: T4M = VBYI(VADD(T4z, T4o)); Chris@10: T4N = VADD(T4H, T4K); Chris@10: ST(&(x[WS(rs, 15)]), VADD(T4M, T4N), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 49)]), VSUB(T4N, T4M), ms, &(x[WS(rs, 1)])); Chris@10: } Chris@10: { Chris@10: V T4Q, T4T, T56, T5d; Chris@10: T4Q = VADD(T4O, T4P); Chris@10: T4T = VBYI(VADD(T4R, T4S)); Chris@10: ST(&(x[WS(rs, 63)]), VSUB(T4Q, T4T), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 1)]), VADD(T4Q, T4T), ms, &(x[WS(rs, 1)])); Chris@10: T56 = VBYI(VSUB(T52, T55)); Chris@10: T5d = VSUB(T59, T5c); Chris@10: ST(&(x[WS(rs, 23)]), VADD(T56, T5d), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 41)]), VSUB(T5d, T56), ms, &(x[WS(rs, 1)])); Chris@10: } Chris@10: { Chris@10: V T5e, T5f, T4U, T4V; Chris@10: T5e = VBYI(VADD(T55, T52)); Chris@10: T5f = VADD(T59, T5c); Chris@10: ST(&(x[WS(rs, 9)]), VADD(T5e, T5f), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 55)]), VSUB(T5f, T5e), ms, &(x[WS(rs, 1)])); Chris@10: T4U = VSUB(T4O, T4P); Chris@10: T4V = VBYI(VSUB(T4S, T4R)); Chris@10: ST(&(x[WS(rs, 33)]), VSUB(T4U, T4V), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 31)]), VADD(T4U, T4V), ms, &(x[WS(rs, 1)])); Chris@10: } Chris@10: } Chris@10: { Chris@10: V T1u, T43, T3D, T3V, T3t, T45, T3B, T3K, T3d, T3E, T3w, T3A, T3R, T46, T3Y; Chris@10: V T42; Chris@10: { Chris@10: V TE, T3U, T1t, T3T, T13, T1s; Chris@10: TE = VSUB(Tg, TD); Chris@10: T3U = VADD(T3r, T3k); Chris@10: T13 = VFMA(LDK(KP831469612), TV, VMUL(LDK(KP555570233), T12)); Chris@10: T1s = VFNMS(LDK(KP555570233), T1r, VMUL(LDK(KP831469612), T1k)); Chris@10: T1t = VSUB(T13, T1s); Chris@10: T3T = VADD(T13, T1s); Chris@10: T1u = VSUB(TE, T1t); Chris@10: T43 = VADD(T3U, T3T); Chris@10: T3D = VADD(TE, T1t); Chris@10: T3V = VSUB(T3T, T3U); Chris@10: } Chris@10: { Chris@10: V T3s, T3I, T3h, T3J, T3f, T3g; Chris@10: T3s = VSUB(T3k, T3r); Chris@10: T3I = VADD(Tg, TD); Chris@10: T3f = VFNMS(LDK(KP555570233), TV, VMUL(LDK(KP831469612), T12)); Chris@10: T3g = VFMA(LDK(KP555570233), T1k, VMUL(LDK(KP831469612), T1r)); Chris@10: T3h = VSUB(T3f, T3g); Chris@10: T3J = VADD(T3f, T3g); Chris@10: T3t = VSUB(T3h, T3s); Chris@10: T45 = VADD(T3I, T3J); Chris@10: T3B = VADD(T3s, T3h); Chris@10: T3K = VSUB(T3I, T3J); Chris@10: } Chris@10: { Chris@10: V T2l, T3u, T3c, T3v; Chris@10: { Chris@10: V T29, T2k, T30, T3b; Chris@10: T29 = VSUB(T1R, T28); Chris@10: T2k = VSUB(T2g, T2j); Chris@10: T2l = VFMA(LDK(KP881921264), T29, VMUL(LDK(KP471396736), T2k)); Chris@10: T3u = VFNMS(LDK(KP471396736), T29, VMUL(LDK(KP881921264), T2k)); Chris@10: T30 = VSUB(T2I, T2Z); Chris@10: T3b = VSUB(T37, T3a); Chris@10: T3c = VFNMS(LDK(KP471396736), T3b, VMUL(LDK(KP881921264), T30)); Chris@10: T3v = VFMA(LDK(KP471396736), T30, VMUL(LDK(KP881921264), T3b)); Chris@10: } Chris@10: T3d = VSUB(T2l, T3c); Chris@10: T3E = VADD(T3u, T3v); Chris@10: T3w = VSUB(T3u, T3v); Chris@10: T3A = VADD(T2l, T3c); Chris@10: } Chris@10: { Chris@10: V T3N, T3W, T3Q, T3X; Chris@10: { Chris@10: V T3L, T3M, T3O, T3P; Chris@10: T3L = VADD(T28, T1R); Chris@10: T3M = VADD(T2g, T2j); Chris@10: T3N = VFMA(LDK(KP956940335), T3L, VMUL(LDK(KP290284677), T3M)); Chris@10: T3W = VFNMS(LDK(KP290284677), T3L, VMUL(LDK(KP956940335), T3M)); Chris@10: T3O = VADD(T2Z, T2I); Chris@10: T3P = VADD(T37, T3a); Chris@10: T3Q = VFNMS(LDK(KP290284677), T3P, VMUL(LDK(KP956940335), T3O)); Chris@10: T3X = VFMA(LDK(KP290284677), T3O, VMUL(LDK(KP956940335), T3P)); Chris@10: } Chris@10: T3R = VSUB(T3N, T3Q); Chris@10: T46 = VADD(T3W, T3X); Chris@10: T3Y = VSUB(T3W, T3X); Chris@10: T42 = VADD(T3N, T3Q); Chris@10: } Chris@10: { Chris@10: V T3e, T3x, T44, T47; Chris@10: T3e = VADD(T1u, T3d); Chris@10: T3x = VBYI(VADD(T3t, T3w)); Chris@10: ST(&(x[WS(rs, 53)]), VSUB(T3e, T3x), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 11)]), VADD(T3e, T3x), ms, &(x[WS(rs, 1)])); Chris@10: T44 = VBYI(VSUB(T42, T43)); Chris@10: T47 = VSUB(T45, T46); Chris@10: ST(&(x[WS(rs, 29)]), VADD(T44, T47), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 35)]), VSUB(T47, T44), ms, &(x[WS(rs, 1)])); Chris@10: } Chris@10: { Chris@10: V T48, T49, T3y, T3z; Chris@10: T48 = VBYI(VADD(T43, T42)); Chris@10: T49 = VADD(T45, T46); Chris@10: ST(&(x[WS(rs, 3)]), VADD(T48, T49), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 61)]), VSUB(T49, T48), ms, &(x[WS(rs, 1)])); Chris@10: T3y = VSUB(T1u, T3d); Chris@10: T3z = VBYI(VSUB(T3w, T3t)); Chris@10: ST(&(x[WS(rs, 43)]), VSUB(T3y, T3z), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 21)]), VADD(T3y, T3z), ms, &(x[WS(rs, 1)])); Chris@10: } Chris@10: { Chris@10: V T3C, T3F, T3S, T3Z; Chris@10: T3C = VBYI(VSUB(T3A, T3B)); Chris@10: T3F = VSUB(T3D, T3E); Chris@10: ST(&(x[WS(rs, 27)]), VADD(T3C, T3F), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 37)]), VSUB(T3F, T3C), ms, &(x[WS(rs, 1)])); Chris@10: T3S = VADD(T3K, T3R); Chris@10: T3Z = VBYI(VADD(T3V, T3Y)); Chris@10: ST(&(x[WS(rs, 51)]), VSUB(T3S, T3Z), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 13)]), VADD(T3S, T3Z), ms, &(x[WS(rs, 1)])); Chris@10: } Chris@10: { Chris@10: V T40, T41, T3G, T3H; Chris@10: T40 = VSUB(T3K, T3R); Chris@10: T41 = VBYI(VSUB(T3Y, T3V)); Chris@10: ST(&(x[WS(rs, 45)]), VSUB(T40, T41), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 19)]), VADD(T40, T41), ms, &(x[WS(rs, 1)])); Chris@10: T3G = VBYI(VADD(T3B, T3A)); Chris@10: T3H = VADD(T3D, T3E); Chris@10: ST(&(x[WS(rs, 5)]), VADD(T3G, T3H), ms, &(x[WS(rs, 1)])); Chris@10: ST(&(x[WS(rs, 59)]), VSUB(T3H, T3G), ms, &(x[WS(rs, 1)])); Chris@10: } Chris@10: } Chris@10: } Chris@10: } Chris@10: VLEAVE(); Chris@10: } Chris@10: Chris@10: static const tw_instr twinstr[] = { Chris@10: VTW(0, 1), Chris@10: VTW(0, 2), Chris@10: VTW(0, 3), Chris@10: VTW(0, 4), Chris@10: VTW(0, 5), Chris@10: VTW(0, 6), Chris@10: VTW(0, 7), Chris@10: VTW(0, 8), Chris@10: VTW(0, 9), Chris@10: VTW(0, 10), Chris@10: VTW(0, 11), Chris@10: VTW(0, 12), Chris@10: VTW(0, 13), Chris@10: VTW(0, 14), Chris@10: VTW(0, 15), Chris@10: VTW(0, 16), Chris@10: VTW(0, 17), Chris@10: VTW(0, 18), Chris@10: VTW(0, 19), Chris@10: VTW(0, 20), Chris@10: VTW(0, 21), Chris@10: VTW(0, 22), Chris@10: VTW(0, 23), Chris@10: VTW(0, 24), Chris@10: VTW(0, 25), Chris@10: VTW(0, 26), Chris@10: VTW(0, 27), Chris@10: VTW(0, 28), Chris@10: VTW(0, 29), Chris@10: VTW(0, 30), Chris@10: VTW(0, 31), Chris@10: VTW(0, 32), Chris@10: VTW(0, 33), Chris@10: VTW(0, 34), Chris@10: VTW(0, 35), Chris@10: VTW(0, 36), Chris@10: VTW(0, 37), Chris@10: VTW(0, 38), Chris@10: VTW(0, 39), Chris@10: VTW(0, 40), Chris@10: VTW(0, 41), Chris@10: VTW(0, 42), Chris@10: VTW(0, 43), Chris@10: VTW(0, 44), Chris@10: VTW(0, 45), Chris@10: VTW(0, 46), Chris@10: VTW(0, 47), Chris@10: VTW(0, 48), Chris@10: VTW(0, 49), Chris@10: VTW(0, 50), Chris@10: VTW(0, 51), Chris@10: VTW(0, 52), Chris@10: VTW(0, 53), Chris@10: VTW(0, 54), Chris@10: VTW(0, 55), Chris@10: VTW(0, 56), Chris@10: VTW(0, 57), Chris@10: VTW(0, 58), Chris@10: VTW(0, 59), Chris@10: VTW(0, 60), Chris@10: VTW(0, 61), Chris@10: VTW(0, 62), Chris@10: VTW(0, 63), Chris@10: {TW_NEXT, VL, 0} Chris@10: }; Chris@10: Chris@10: static const ct_desc desc = { 64, XSIMD_STRING("t1bv_64"), twinstr, &GENUS, {467, 198, 52, 0}, 0, 0, 0 }; Chris@10: Chris@10: void XSIMD(codelet_t1bv_64) (planner *p) { Chris@10: X(kdft_dit_register) (p, t1bv_64, &desc); Chris@10: } Chris@10: #endif /* HAVE_FMA */