annotate src/fftw-3.3.8/dft/simd/common/n2sv_4.c @ 83:ae30d91d2ffe

Replace these with versions built using an older toolset (so as to avoid ABI compatibilities when linking on Ubuntu 14.04 for packaging purposes)
author Chris Cannam
date Fri, 07 Feb 2020 11:51:13 +0000
parents d0c2a83c1364
children
rev   line source
Chris@82 1 /*
Chris@82 2 * Copyright (c) 2003, 2007-14 Matteo Frigo
Chris@82 3 * Copyright (c) 2003, 2007-14 Massachusetts Institute of Technology
Chris@82 4 *
Chris@82 5 * This program is free software; you can redistribute it and/or modify
Chris@82 6 * it under the terms of the GNU General Public License as published by
Chris@82 7 * the Free Software Foundation; either version 2 of the License, or
Chris@82 8 * (at your option) any later version.
Chris@82 9 *
Chris@82 10 * This program is distributed in the hope that it will be useful,
Chris@82 11 * but WITHOUT ANY WARRANTY; without even the implied warranty of
Chris@82 12 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
Chris@82 13 * GNU General Public License for more details.
Chris@82 14 *
Chris@82 15 * You should have received a copy of the GNU General Public License
Chris@82 16 * along with this program; if not, write to the Free Software
Chris@82 17 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
Chris@82 18 *
Chris@82 19 */
Chris@82 20
Chris@82 21 /* This file was automatically generated --- DO NOT EDIT */
Chris@82 22 /* Generated on Thu May 24 08:05:19 EDT 2018 */
Chris@82 23
Chris@82 24 #include "dft/codelet-dft.h"
Chris@82 25
Chris@82 26 #if defined(ARCH_PREFERS_FMA) || defined(ISA_EXTENSION_PREFERS_FMA)
Chris@82 27
Chris@82 28 /* Generated by: ../../../genfft/gen_notw.native -fma -simd -compact -variables 4 -pipeline-latency 8 -n 4 -name n2sv_4 -with-ostride 1 -include dft/simd/n2s.h -store-multiple 4 */
Chris@82 29
Chris@82 30 /*
Chris@82 31 * This function contains 16 FP additions, 0 FP multiplications,
Chris@82 32 * (or, 16 additions, 0 multiplications, 0 fused multiply/add),
Chris@82 33 * 17 stack variables, 0 constants, and 18 memory accesses
Chris@82 34 */
Chris@82 35 #include "dft/simd/n2s.h"
Chris@82 36
Chris@82 37 static void n2sv_4(const R *ri, const R *ii, R *ro, R *io, stride is, stride os, INT v, INT ivs, INT ovs)
Chris@82 38 {
Chris@82 39 {
Chris@82 40 INT i;
Chris@82 41 for (i = v; i > 0; i = i - (2 * VL), ri = ri + ((2 * VL) * ivs), ii = ii + ((2 * VL) * ivs), ro = ro + ((2 * VL) * ovs), io = io + ((2 * VL) * ovs), MAKE_VOLATILE_STRIDE(16, is), MAKE_VOLATILE_STRIDE(16, os)) {
Chris@82 42 V T3, Tb, T9, Tf, T6, Ta, Te, Tg;
Chris@82 43 {
Chris@82 44 V T1, T2, T7, T8;
Chris@82 45 T1 = LD(&(ri[0]), ivs, &(ri[0]));
Chris@82 46 T2 = LD(&(ri[WS(is, 2)]), ivs, &(ri[0]));
Chris@82 47 T3 = VADD(T1, T2);
Chris@82 48 Tb = VSUB(T1, T2);
Chris@82 49 T7 = LD(&(ii[0]), ivs, &(ii[0]));
Chris@82 50 T8 = LD(&(ii[WS(is, 2)]), ivs, &(ii[0]));
Chris@82 51 T9 = VSUB(T7, T8);
Chris@82 52 Tf = VADD(T7, T8);
Chris@82 53 }
Chris@82 54 {
Chris@82 55 V T4, T5, Tc, Td;
Chris@82 56 T4 = LD(&(ri[WS(is, 1)]), ivs, &(ri[WS(is, 1)]));
Chris@82 57 T5 = LD(&(ri[WS(is, 3)]), ivs, &(ri[WS(is, 1)]));
Chris@82 58 T6 = VADD(T4, T5);
Chris@82 59 Ta = VSUB(T4, T5);
Chris@82 60 Tc = LD(&(ii[WS(is, 1)]), ivs, &(ii[WS(is, 1)]));
Chris@82 61 Td = LD(&(ii[WS(is, 3)]), ivs, &(ii[WS(is, 1)]));
Chris@82 62 Te = VSUB(Tc, Td);
Chris@82 63 Tg = VADD(Tc, Td);
Chris@82 64 }
Chris@82 65 {
Chris@82 66 V Th, Ti, Tj, Tk;
Chris@82 67 Th = VSUB(T3, T6);
Chris@82 68 STM4(&(ro[2]), Th, ovs, &(ro[0]));
Chris@82 69 Ti = VSUB(Tf, Tg);
Chris@82 70 STM4(&(io[2]), Ti, ovs, &(io[0]));
Chris@82 71 Tj = VADD(T3, T6);
Chris@82 72 STM4(&(ro[0]), Tj, ovs, &(ro[0]));
Chris@82 73 Tk = VADD(Tf, Tg);
Chris@82 74 STM4(&(io[0]), Tk, ovs, &(io[0]));
Chris@82 75 {
Chris@82 76 V Tl, Tm, Tn, To;
Chris@82 77 Tl = VSUB(T9, Ta);
Chris@82 78 STM4(&(io[1]), Tl, ovs, &(io[1]));
Chris@82 79 Tm = VADD(Tb, Te);
Chris@82 80 STM4(&(ro[1]), Tm, ovs, &(ro[1]));
Chris@82 81 Tn = VADD(Ta, T9);
Chris@82 82 STM4(&(io[3]), Tn, ovs, &(io[1]));
Chris@82 83 STN4(&(io[0]), Tk, Tl, Ti, Tn, ovs);
Chris@82 84 To = VSUB(Tb, Te);
Chris@82 85 STM4(&(ro[3]), To, ovs, &(ro[1]));
Chris@82 86 STN4(&(ro[0]), Tj, Tm, Th, To, ovs);
Chris@82 87 }
Chris@82 88 }
Chris@82 89 }
Chris@82 90 }
Chris@82 91 VLEAVE();
Chris@82 92 }
Chris@82 93
Chris@82 94 static const kdft_desc desc = { 4, XSIMD_STRING("n2sv_4"), {16, 0, 0, 0}, &GENUS, 0, 1, 0, 0 };
Chris@82 95
Chris@82 96 void XSIMD(codelet_n2sv_4) (planner *p) {
Chris@82 97 X(kdft_register) (p, n2sv_4, &desc);
Chris@82 98 }
Chris@82 99
Chris@82 100 #else
Chris@82 101
Chris@82 102 /* Generated by: ../../../genfft/gen_notw.native -simd -compact -variables 4 -pipeline-latency 8 -n 4 -name n2sv_4 -with-ostride 1 -include dft/simd/n2s.h -store-multiple 4 */
Chris@82 103
Chris@82 104 /*
Chris@82 105 * This function contains 16 FP additions, 0 FP multiplications,
Chris@82 106 * (or, 16 additions, 0 multiplications, 0 fused multiply/add),
Chris@82 107 * 17 stack variables, 0 constants, and 18 memory accesses
Chris@82 108 */
Chris@82 109 #include "dft/simd/n2s.h"
Chris@82 110
Chris@82 111 static void n2sv_4(const R *ri, const R *ii, R *ro, R *io, stride is, stride os, INT v, INT ivs, INT ovs)
Chris@82 112 {
Chris@82 113 {
Chris@82 114 INT i;
Chris@82 115 for (i = v; i > 0; i = i - (2 * VL), ri = ri + ((2 * VL) * ivs), ii = ii + ((2 * VL) * ivs), ro = ro + ((2 * VL) * ovs), io = io + ((2 * VL) * ovs), MAKE_VOLATILE_STRIDE(16, is), MAKE_VOLATILE_STRIDE(16, os)) {
Chris@82 116 V T3, Tb, T9, Tf, T6, Ta, Te, Tg;
Chris@82 117 {
Chris@82 118 V T1, T2, T7, T8;
Chris@82 119 T1 = LD(&(ri[0]), ivs, &(ri[0]));
Chris@82 120 T2 = LD(&(ri[WS(is, 2)]), ivs, &(ri[0]));
Chris@82 121 T3 = VADD(T1, T2);
Chris@82 122 Tb = VSUB(T1, T2);
Chris@82 123 T7 = LD(&(ii[0]), ivs, &(ii[0]));
Chris@82 124 T8 = LD(&(ii[WS(is, 2)]), ivs, &(ii[0]));
Chris@82 125 T9 = VSUB(T7, T8);
Chris@82 126 Tf = VADD(T7, T8);
Chris@82 127 }
Chris@82 128 {
Chris@82 129 V T4, T5, Tc, Td;
Chris@82 130 T4 = LD(&(ri[WS(is, 1)]), ivs, &(ri[WS(is, 1)]));
Chris@82 131 T5 = LD(&(ri[WS(is, 3)]), ivs, &(ri[WS(is, 1)]));
Chris@82 132 T6 = VADD(T4, T5);
Chris@82 133 Ta = VSUB(T4, T5);
Chris@82 134 Tc = LD(&(ii[WS(is, 1)]), ivs, &(ii[WS(is, 1)]));
Chris@82 135 Td = LD(&(ii[WS(is, 3)]), ivs, &(ii[WS(is, 1)]));
Chris@82 136 Te = VSUB(Tc, Td);
Chris@82 137 Tg = VADD(Tc, Td);
Chris@82 138 }
Chris@82 139 {
Chris@82 140 V Th, Ti, Tj, Tk;
Chris@82 141 Th = VSUB(T3, T6);
Chris@82 142 STM4(&(ro[2]), Th, ovs, &(ro[0]));
Chris@82 143 Ti = VSUB(Tf, Tg);
Chris@82 144 STM4(&(io[2]), Ti, ovs, &(io[0]));
Chris@82 145 Tj = VADD(T3, T6);
Chris@82 146 STM4(&(ro[0]), Tj, ovs, &(ro[0]));
Chris@82 147 Tk = VADD(Tf, Tg);
Chris@82 148 STM4(&(io[0]), Tk, ovs, &(io[0]));
Chris@82 149 {
Chris@82 150 V Tl, Tm, Tn, To;
Chris@82 151 Tl = VSUB(T9, Ta);
Chris@82 152 STM4(&(io[1]), Tl, ovs, &(io[1]));
Chris@82 153 Tm = VADD(Tb, Te);
Chris@82 154 STM4(&(ro[1]), Tm, ovs, &(ro[1]));
Chris@82 155 Tn = VADD(Ta, T9);
Chris@82 156 STM4(&(io[3]), Tn, ovs, &(io[1]));
Chris@82 157 STN4(&(io[0]), Tk, Tl, Ti, Tn, ovs);
Chris@82 158 To = VSUB(Tb, Te);
Chris@82 159 STM4(&(ro[3]), To, ovs, &(ro[1]));
Chris@82 160 STN4(&(ro[0]), Tj, Tm, Th, To, ovs);
Chris@82 161 }
Chris@82 162 }
Chris@82 163 }
Chris@82 164 }
Chris@82 165 VLEAVE();
Chris@82 166 }
Chris@82 167
Chris@82 168 static const kdft_desc desc = { 4, XSIMD_STRING("n2sv_4"), {16, 0, 0, 0}, &GENUS, 0, 1, 0, 0 };
Chris@82 169
Chris@82 170 void XSIMD(codelet_n2sv_4) (planner *p) {
Chris@82 171 X(kdft_register) (p, n2sv_4, &desc);
Chris@82 172 }
Chris@82 173
Chris@82 174 #endif