Loop Id: 1153 | Module: bench | Source: t2fv_4.c:97-114 [...] | Coverage: 34.22% |
---|
Loop Id: 1153 | Module: bench | Source: t2fv_4.c:97-114 [...] | Coverage: 34.22% |
---|
0x104da0 MOV 0x18(%RCX),%RSI [3] |
0x104da4 VMOVAPD (%RAX),%XMM5 [2] |
0x104da8 INC %R8 |
0x104dab ADD $0x60,%RDX |
0x104daf VMOVAPD (%RAX,%RSI,8),%XMM0 [2] |
0x104db4 MOV 0x10(%RCX),%RSI [3] |
0x104db8 VPERMILPD $0x1,%XMM0,%XMM3 |
0x104dbe VMOVAPD (%RAX,%RSI,8),%XMM1 [2] |
0x104dc3 MOV 0x8(%RCX),%RSI [3] |
0x104dc7 VMULPD -0x20(%RDX),%XMM0,%XMM0 [4] |
0x104dcc LEA (%RAX,%RSI,8),%RSI |
0x104dd0 VMOVAPD (%RSI),%XMM2 [1] |
0x104dd4 VFNMADD132PD -0x10(%RDX),%XMM0,%XMM3 [4] |
0x104dda VPERMILPD $0x1,%XMM1,%XMM0 |
0x104de0 VMULPD -0x40(%RDX),%XMM1,%XMM1 [4] |
0x104de5 VFNMADD132PD -0x30(%RDX),%XMM1,%XMM0 [4] |
0x104deb VPERMILPD $0x1,%XMM2,%XMM1 |
0x104df1 VMULPD -0x60(%RDX),%XMM2,%XMM2 [4] |
0x104df6 VSUBPD %XMM0,%XMM5,%XMM4 |
0x104dfa VADDPD %XMM5,%XMM0,%XMM0 |
0x104dfe VFNMADD132PD -0x50(%RDX),%XMM2,%XMM1 [4] |
0x104e04 VSUBPD %XMM3,%XMM1,%XMM2 |
0x104e08 VADDPD %XMM3,%XMM1,%XMM1 |
0x104e0c VXORPD %XMM2,%XMM7,%XMM2 |
0x104e10 VPERMILPD $0x1,%XMM2,%XMM2 |
0x104e16 VSUBPD %XMM2,%XMM4,%XMM6 |
0x104e1a VADDPD %XMM2,%XMM4,%XMM4 |
0x104e1e VSUBPD %XMM1,%XMM0,%XMM2 |
0x104e22 VADDPD %XMM1,%XMM0,%XMM0 |
0x104e26 VMOVAPD %XMM6,(%RSI) [1] |
0x104e2a MOV 0x18(%RCX),%RSI [3] |
0x104e2e VMOVAPD %XMM4,(%RAX,%RSI,8) [2] |
0x104e33 MOV 0x10(%RCX),%RSI [3] |
0x104e37 ADD %RDI,%RCX |
0x104e3a VMOVAPD %XMM2,(%RAX,%RSI,8) [2] |
0x104e3f VMOVAPD %XMM0,(%RAX) [2] |
0x104e43 ADD %R10,%RAX |
0x104e46 CMP %R8,%R9 |
0x104e49 JNE 104da0 |
/usr/lib/gcc/x86_64-pc-linux-gnu/14.2.1/include/emmintrin.h: 251 - 251 |
-------------------------------------------------------------------------------- |
/usr/lib/gcc/x86_64-pc-linux-gnu/14.2.1/include/emmintrin.h: 263 - 263 |
-------------------------------------------------------------------------------- |
/usr/lib/gcc/x86_64-pc-linux-gnu/14.2.1/include/emmintrin.h: 275 - 275 |
-------------------------------------------------------------------------------- |
/usr/lib/gcc/x86_64-pc-linux-gnu/14.2.1/include/emmintrin.h: 355 - 355 |
-------------------------------------------------------------------------------- |
/usr/lib/gcc/x86_64-pc-linux-gnu/14.2.1/include/emmintrin.h: 953 - 953 |
-------------------------------------------------------------------------------- |
/home/fmusial/FFTW_Benchmarks/fftw-3.3.10-gcc-sse-128/dft/simd/sse2/../../../simd-support/simd-sse2.h: 113 - 120 |
-------------------------------------------------------------------------------- |
113: return *(const V *)x; |
[...] |
120: *(V *)x = v; |
/home/fmusial/FFTW_Benchmarks/fftw-3.3.10-gcc-sse-128/dft/simd/sse2/../common/t2fv_4.c: 97 - 114 |
-------------------------------------------------------------------------------- |
97: for (m = mb, W = W + (mb * ((TWVL / VL) * 6)); m < me; m = m + VL, x = x + (VL * ms), W = W + (TWVL * 6), MAKE_VOLATILE_STRIDE(4, rs)) { |
98: V T1, T8, T3, T6, T7, T2, T5; |
99: T1 = LD(&(x[0]), ms, &(x[0])); |
100: T7 = LD(&(x[WS(rs, 3)]), ms, &(x[WS(rs, 1)])); |
101: T8 = BYTWJ(&(W[TWVL * 4]), T7); |
102: T2 = LD(&(x[WS(rs, 2)]), ms, &(x[0])); |
103: T3 = BYTWJ(&(W[TWVL * 2]), T2); |
104: T5 = LD(&(x[WS(rs, 1)]), ms, &(x[WS(rs, 1)])); |
[...] |
111: ST(&(x[WS(rs, 3)]), VADD(T4, T9), ms, &(x[WS(rs, 1)])); |
112: Ta = VADD(T1, T3); |
113: Tb = VADD(T6, T8); |
114: ST(&(x[WS(rs, 2)]), VSUB(Ta, Tb), ms, &(x[0])); |
Coverage (%) | Name | Source Location | Module |
---|---|---|---|
►99.78+ | apply#0x120c50 | dftw-direct.c:51 | bench |
○ | doit | fftw-bench.c:274 | bench |
○ | speed | bench | |
○ | bench_main | bench | |
○ | __libc_init_first | libc.so.6 | |
○ | __libc_start_main | libc.so.6 | |
○ | _start | hook.c:185 | bench |
min | med | avg | max |
---|---|---|---|
Percentile Index | 10 | 20 | 30 | 40 | 50 | 60 | 70 | 80 | 90 | 100 |
---|---|---|---|---|---|---|---|---|---|---|
Value |
min | med | avg | max |
---|---|---|---|
Percentile Index | 10 | 20 | 30 | 40 | 50 | 60 | 70 | 80 | 90 | 100 |
---|---|---|---|---|---|---|---|---|---|---|
Value |
Path / |
Metric | Value |
---|---|
CQA speedup if no scalar integer | 1.31 |
CQA speedup if FP arith vectorized | 1.69 |
CQA speedup if fully vectorized | 4.00 |
CQA speedup if no inter-iteration dependency | NA |
CQA speedup if next bottleneck killed | 1.27 |
Bottlenecks | micro-operation queue, |
Function | t2fv_4 |
Source | emmintrin.h:251-251,emmintrin.h:263-263,emmintrin.h:275-275,emmintrin.h:355-355,emmintrin.h:953-953,simd-sse2.h:113-113,simd-sse2.h:120-120,t2fv_4.c:97-104,t2fv_4.c:111-114 |
Source loop unroll info | not unrolled or unrolled with no peel/tail loop |
Source loop unroll confidence level | max |
Unroll/vectorization loop type | NA |
Unroll factor | NA |
CQA cycles | 9.50 |
CQA cycles if no scalar integer | 7.25 |
CQA cycles if FP arith vectorized | 5.63 |
CQA cycles if fully vectorized | 2.38 |
Front-end cycles | 9.50 |
P0 cycles | 7.50 |
P1 cycles | 7.50 |
P2 cycles | 7.50 |
P3 cycles | 7.50 |
P4 cycles | 4.00 |
P5 cycles | 5.00 |
P6 cycles | 5.00 |
P7 cycles | 4.00 |
DIV/SQRT cycles | 0.00 |
Inter-iter dependencies cycles | 1 |
FE+BE cycles (UFS) | NA |
Stall cycles (UFS) | NA |
Nb insns | 39.00 |
Nb uops | 38.00 |
Nb loads | 15.00 |
Nb stores | 4.00 |
Nb stack references | 0.00 |
FLOP/cycle | 3.58 |
Nb FLOP add-sub | 16.00 |
Nb FLOP mul | 6.00 |
Nb FLOP fma | 6.00 |
Nb FLOP div | 0.00 |
Nb FLOP rcp | 0.00 |
Nb FLOP sqrt | 0.00 |
Nb FLOP rsqrt | 0.00 |
Bytes/cycle | 27.79 |
Bytes prefetched | 0.00 |
Bytes loaded | 200.00 |
Bytes stored | 64.00 |
Stride 0 | 0.00 |
Stride 1 | 1.00 |
Stride n | 0.00 |
Stride unknown | 3.00 |
Stride indirect | 0.00 |
Vectorization ratio all | 100.00 |
Vectorization ratio load | 100.00 |
Vectorization ratio store | 100.00 |
Vectorization ratio mul | 100.00 |
Vectorization ratio add_sub | 100.00 |
Vectorization ratio fma | 100.00 |
Vectorization ratio div_sqrt | NA |
Vectorization ratio other | 100.00 |
Vector-efficiency ratio all | 25.00 |
Vector-efficiency ratio load | 25.00 |
Vector-efficiency ratio store | 25.00 |
Vector-efficiency ratio mul | 25.00 |
Vector-efficiency ratio add_sub | 25.00 |
Vector-efficiency ratio fma | 25.00 |
Vector-efficiency ratio div_sqrt | NA |
Vector-efficiency ratio other | 25.00 |
Metric | Value |
---|---|
CQA speedup if no scalar integer | 1.31 |
CQA speedup if FP arith vectorized | 1.69 |
CQA speedup if fully vectorized | 4.00 |
CQA speedup if no inter-iteration dependency | NA |
CQA speedup if next bottleneck killed | 1.27 |
Bottlenecks | micro-operation queue, |
Function | t2fv_4 |
Source | emmintrin.h:251-251,emmintrin.h:263-263,emmintrin.h:275-275,emmintrin.h:355-355,emmintrin.h:953-953,simd-sse2.h:113-113,simd-sse2.h:120-120,t2fv_4.c:97-104,t2fv_4.c:111-114 |
Source loop unroll info | not unrolled or unrolled with no peel/tail loop |
Source loop unroll confidence level | max |
Unroll/vectorization loop type | NA |
Unroll factor | NA |
CQA cycles | 9.50 |
CQA cycles if no scalar integer | 7.25 |
CQA cycles if FP arith vectorized | 5.63 |
CQA cycles if fully vectorized | 2.38 |
Front-end cycles | 9.50 |
P0 cycles | 7.50 |
P1 cycles | 7.50 |
P2 cycles | 7.50 |
P3 cycles | 7.50 |
P4 cycles | 4.00 |
P5 cycles | 5.00 |
P6 cycles | 5.00 |
P7 cycles | 4.00 |
DIV/SQRT cycles | 0.00 |
Inter-iter dependencies cycles | 1 |
FE+BE cycles (UFS) | NA |
Stall cycles (UFS) | NA |
Nb insns | 39.00 |
Nb uops | 38.00 |
Nb loads | 15.00 |
Nb stores | 4.00 |
Nb stack references | 0.00 |
FLOP/cycle | 3.58 |
Nb FLOP add-sub | 16.00 |
Nb FLOP mul | 6.00 |
Nb FLOP fma | 6.00 |
Nb FLOP div | 0.00 |
Nb FLOP rcp | 0.00 |
Nb FLOP sqrt | 0.00 |
Nb FLOP rsqrt | 0.00 |
Bytes/cycle | 27.79 |
Bytes prefetched | 0.00 |
Bytes loaded | 200.00 |
Bytes stored | 64.00 |
Stride 0 | 0.00 |
Stride 1 | 1.00 |
Stride n | 0.00 |
Stride unknown | 3.00 |
Stride indirect | 0.00 |
Vectorization ratio all | 100.00 |
Vectorization ratio load | 100.00 |
Vectorization ratio store | 100.00 |
Vectorization ratio mul | 100.00 |
Vectorization ratio add_sub | 100.00 |
Vectorization ratio fma | 100.00 |
Vectorization ratio div_sqrt | NA |
Vectorization ratio other | 100.00 |
Vector-efficiency ratio all | 25.00 |
Vector-efficiency ratio load | 25.00 |
Vector-efficiency ratio store | 25.00 |
Vector-efficiency ratio mul | 25.00 |
Vector-efficiency ratio add_sub | 25.00 |
Vector-efficiency ratio fma | 25.00 |
Vector-efficiency ratio div_sqrt | NA |
Vector-efficiency ratio other | 25.00 |
Path / |
Function | t2fv_4 |
Source file and lines | t2fv_4.c:97-114 |
Module | bench |
nb instructions | 39 |
nb uops | 38 |
loop length | 175 |
used x86 registers | 8 |
used mmx registers | 0 |
used xmm registers | 8 |
used ymm registers | 0 |
used zmm registers | 0 |
nb stack references | 0 |
ADD-SUB / MUL ratio | 2.67 |
micro-operation queue | 9.50 cycles |
front end | 9.50 cycles |
P0 | P1 | P2 | P3 | P4 | P5 | P6 | P7 | |
---|---|---|---|---|---|---|---|---|
uops | 7.50 | 7.50 | 7.50 | 7.50 | 4.00 | 5.00 | 5.00 | 4.00 |
cycles | 7.50 | 7.50 | 7.50 | 7.50 | 4.00 | 5.00 | 5.00 | 4.00 |
Cycles executing div or sqrt instructions | NA |
Longest recurrence chain latency (RecMII) | 1.00 |
Front-end | 9.50 |
Dispatch | 7.50 |
Data deps. | 1.00 |
Overall L1 | 9.50 |
all | 100% |
load | 100% |
store | 100% |
mul | 100% |
add-sub | 100% |
fma | 100% |
div/sqrt | NA (no div/sqrt vectorizable/vectorized instructions) |
other | 100% |
all | 25% |
load | 25% |
store | 25% |
mul | 25% |
add-sub | 25% |
fma | 25% |
div/sqrt | NA (no div/sqrt vectorizable/vectorized instructions) |
other | 25% |
Instruction | Nb FU | P0 | P1 | P2 | P3 | P4 | P5 | P6 | P7 | Latency | Recip. throughput | Vectorization |
---|---|---|---|---|---|---|---|---|---|---|---|---|
MOV 0x18(%RCX),%RSI | 1 | 0 | 0 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 4-5 | 0.50 | N/A |
VMOVAPD (%RAX),%XMM5 | 1 | 0 | 0 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 4-5 | 0.50 | vect (25.0%) |
INC %R8 | 1 | 0.25 | 0.25 | 0 | 0 | 0 | 0.25 | 0.25 | 0 | 1 | 0.25 | N/A |
ADD $0x60,%RDX | 1 | 0.25 | 0.25 | 0 | 0 | 0 | 0.25 | 0.25 | 0 | 1 | 0.25 | N/A |
VMOVAPD (%RAX,%RSI,8),%XMM0 | 1 | 0 | 0 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 4-5 | 0.50 | vect (25.0%) |
MOV 0x10(%RCX),%RSI | 1 | 0 | 0 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 4-5 | 0.50 | N/A |
VPERMILPD $0x1,%XMM0,%XMM3 | 1 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 1 | 1 | vect (25.0%) |
VMOVAPD (%RAX,%RSI,8),%XMM1 | 1 | 0 | 0 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 4-5 | 0.50 | vect (25.0%) |
MOV 0x8(%RCX),%RSI | 1 | 0 | 0 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 4-5 | 0.50 | N/A |
VMULPD -0x20(%RDX),%XMM0,%XMM0 | 1 | 0.50 | 0.50 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 4 | 0.50 | vect (25.0%) |
LEA (%RAX,%RSI,8),%RSI | 1 | 0 | 0.50 | 0 | 0 | 0 | 0.50 | 0 | 0 | 1 | 0.50 | N/A |
VMOVAPD (%RSI),%XMM2 | 1 | 0 | 0 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 4-5 | 0.50 | vect (25.0%) |
VFNMADD132PD -0x10(%RDX),%XMM0,%XMM3 | 1 | 0.50 | 0.50 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 4 | 0.50 | vect (25.0%) |
VPERMILPD $0x1,%XMM1,%XMM0 | 1 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 1 | 1 | vect (25.0%) |
VMULPD -0x40(%RDX),%XMM1,%XMM1 | 1 | 0.50 | 0.50 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 4 | 0.50 | vect (25.0%) |
VFNMADD132PD -0x30(%RDX),%XMM1,%XMM0 | 1 | 0.50 | 0.50 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 4 | 0.50 | vect (25.0%) |
VPERMILPD $0x1,%XMM2,%XMM1 | 1 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 1 | 1 | vect (25.0%) |
VMULPD -0x60(%RDX),%XMM2,%XMM2 | 1 | 0.50 | 0.50 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 4 | 0.50 | vect (25.0%) |
VSUBPD %XMM0,%XMM5,%XMM4 | 1 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 0 | 0 | 4 | 0.50 | vect (25.0%) |
VADDPD %XMM5,%XMM0,%XMM0 | 1 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 0 | 0 | 4 | 0.50 | vect (25.0%) |
VFNMADD132PD -0x50(%RDX),%XMM2,%XMM1 | 1 | 0.50 | 0.50 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 4 | 0.50 | vect (25.0%) |
VSUBPD %XMM3,%XMM1,%XMM2 | 1 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 0 | 0 | 4 | 0.50 | vect (25.0%) |
VADDPD %XMM3,%XMM1,%XMM1 | 1 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 0 | 0 | 4 | 0.50 | vect (25.0%) |
VXORPD %XMM2,%XMM7,%XMM2 | 1 | 0.33 | 0.33 | 0 | 0 | 0 | 0.33 | 0 | 0 | 1 | 0.33 | vect (25.0%) |
VPERMILPD $0x1,%XMM2,%XMM2 | 1 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 1 | 1 | vect (25.0%) |
VSUBPD %XMM2,%XMM4,%XMM6 | 1 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 0 | 0 | 4 | 0.50 | vect (25.0%) |
VADDPD %XMM2,%XMM4,%XMM4 | 1 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 0 | 0 | 4 | 0.50 | vect (25.0%) |
VSUBPD %XMM1,%XMM0,%XMM2 | 1 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 0 | 0 | 4 | 0.50 | vect (25.0%) |
VADDPD %XMM1,%XMM0,%XMM0 | 1 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 0 | 0 | 4 | 0.50 | vect (25.0%) |
VMOVAPD %XMM6,(%RSI) | 1 | 0 | 0 | 0.33 | 0.33 | 1 | 0 | 0 | 0.33 | 3 | 1 | vect (25.0%) |
MOV 0x18(%RCX),%RSI | 1 | 0 | 0 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 4-5 | 0.50 | N/A |
VMOVAPD %XMM4,(%RAX,%RSI,8) | 1 | 0 | 0 | 0.33 | 0.33 | 1 | 0 | 0 | 0.33 | 3 | 1 | vect (25.0%) |
MOV 0x10(%RCX),%RSI | 1 | 0 | 0 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 4-5 | 0.50 | N/A |
ADD %RDI,%RCX | 1 | 0.25 | 0.25 | 0 | 0 | 0 | 0.25 | 0.25 | 0 | 1 | 0.25 | N/A |
VMOVAPD %XMM2,(%RAX,%RSI,8) | 1 | 0 | 0 | 0.33 | 0.33 | 1 | 0 | 0 | 0.33 | 3 | 1 | vect (25.0%) |
VMOVAPD %XMM0,(%RAX) | 1 | 0 | 0 | 0.33 | 0.33 | 1 | 0 | 0 | 0.33 | 3 | 1 | vect (25.0%) |
ADD %R10,%RAX | 1 | 0.25 | 0.25 | 0 | 0 | 0 | 0.25 | 0.25 | 0 | 1 | 0.25 | N/A |
CMP %R8,%R9 | 1 | 0.25 | 0.25 | 0 | 0 | 0 | 0.25 | 0.25 | 0 | 1 | 0.25 | N/A |
JNE 104da0 <t2fv_4+0x40> | 1 | 0.50 | 0 | 0 | 0 | 0 | 0 | 0.50 | 0 | 0 | 0.50-1 | N/A |
Function | t2fv_4 |
Source file and lines | t2fv_4.c:97-114 |
Module | bench |
nb instructions | 39 |
nb uops | 38 |
loop length | 175 |
used x86 registers | 8 |
used mmx registers | 0 |
used xmm registers | 8 |
used ymm registers | 0 |
used zmm registers | 0 |
nb stack references | 0 |
ADD-SUB / MUL ratio | 2.67 |
micro-operation queue | 9.50 cycles |
front end | 9.50 cycles |
P0 | P1 | P2 | P3 | P4 | P5 | P6 | P7 | |
---|---|---|---|---|---|---|---|---|
uops | 7.50 | 7.50 | 7.50 | 7.50 | 4.00 | 5.00 | 5.00 | 4.00 |
cycles | 7.50 | 7.50 | 7.50 | 7.50 | 4.00 | 5.00 | 5.00 | 4.00 |
Cycles executing div or sqrt instructions | NA |
Longest recurrence chain latency (RecMII) | 1.00 |
Front-end | 9.50 |
Dispatch | 7.50 |
Data deps. | 1.00 |
Overall L1 | 9.50 |
all | 100% |
load | 100% |
store | 100% |
mul | 100% |
add-sub | 100% |
fma | 100% |
div/sqrt | NA (no div/sqrt vectorizable/vectorized instructions) |
other | 100% |
all | 25% |
load | 25% |
store | 25% |
mul | 25% |
add-sub | 25% |
fma | 25% |
div/sqrt | NA (no div/sqrt vectorizable/vectorized instructions) |
other | 25% |
Instruction | Nb FU | P0 | P1 | P2 | P3 | P4 | P5 | P6 | P7 | Latency | Recip. throughput | Vectorization |
---|---|---|---|---|---|---|---|---|---|---|---|---|
MOV 0x18(%RCX),%RSI | 1 | 0 | 0 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 4-5 | 0.50 | N/A |
VMOVAPD (%RAX),%XMM5 | 1 | 0 | 0 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 4-5 | 0.50 | vect (25.0%) |
INC %R8 | 1 | 0.25 | 0.25 | 0 | 0 | 0 | 0.25 | 0.25 | 0 | 1 | 0.25 | N/A |
ADD $0x60,%RDX | 1 | 0.25 | 0.25 | 0 | 0 | 0 | 0.25 | 0.25 | 0 | 1 | 0.25 | N/A |
VMOVAPD (%RAX,%RSI,8),%XMM0 | 1 | 0 | 0 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 4-5 | 0.50 | vect (25.0%) |
MOV 0x10(%RCX),%RSI | 1 | 0 | 0 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 4-5 | 0.50 | N/A |
VPERMILPD $0x1,%XMM0,%XMM3 | 1 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 1 | 1 | vect (25.0%) |
VMOVAPD (%RAX,%RSI,8),%XMM1 | 1 | 0 | 0 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 4-5 | 0.50 | vect (25.0%) |
MOV 0x8(%RCX),%RSI | 1 | 0 | 0 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 4-5 | 0.50 | N/A |
VMULPD -0x20(%RDX),%XMM0,%XMM0 | 1 | 0.50 | 0.50 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 4 | 0.50 | vect (25.0%) |
LEA (%RAX,%RSI,8),%RSI | 1 | 0 | 0.50 | 0 | 0 | 0 | 0.50 | 0 | 0 | 1 | 0.50 | N/A |
VMOVAPD (%RSI),%XMM2 | 1 | 0 | 0 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 4-5 | 0.50 | vect (25.0%) |
VFNMADD132PD -0x10(%RDX),%XMM0,%XMM3 | 1 | 0.50 | 0.50 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 4 | 0.50 | vect (25.0%) |
VPERMILPD $0x1,%XMM1,%XMM0 | 1 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 1 | 1 | vect (25.0%) |
VMULPD -0x40(%RDX),%XMM1,%XMM1 | 1 | 0.50 | 0.50 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 4 | 0.50 | vect (25.0%) |
VFNMADD132PD -0x30(%RDX),%XMM1,%XMM0 | 1 | 0.50 | 0.50 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 4 | 0.50 | vect (25.0%) |
VPERMILPD $0x1,%XMM2,%XMM1 | 1 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 1 | 1 | vect (25.0%) |
VMULPD -0x60(%RDX),%XMM2,%XMM2 | 1 | 0.50 | 0.50 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 4 | 0.50 | vect (25.0%) |
VSUBPD %XMM0,%XMM5,%XMM4 | 1 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 0 | 0 | 4 | 0.50 | vect (25.0%) |
VADDPD %XMM5,%XMM0,%XMM0 | 1 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 0 | 0 | 4 | 0.50 | vect (25.0%) |
VFNMADD132PD -0x50(%RDX),%XMM2,%XMM1 | 1 | 0.50 | 0.50 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 4 | 0.50 | vect (25.0%) |
VSUBPD %XMM3,%XMM1,%XMM2 | 1 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 0 | 0 | 4 | 0.50 | vect (25.0%) |
VADDPD %XMM3,%XMM1,%XMM1 | 1 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 0 | 0 | 4 | 0.50 | vect (25.0%) |
VXORPD %XMM2,%XMM7,%XMM2 | 1 | 0.33 | 0.33 | 0 | 0 | 0 | 0.33 | 0 | 0 | 1 | 0.33 | vect (25.0%) |
VPERMILPD $0x1,%XMM2,%XMM2 | 1 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 1 | 1 | vect (25.0%) |
VSUBPD %XMM2,%XMM4,%XMM6 | 1 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 0 | 0 | 4 | 0.50 | vect (25.0%) |
VADDPD %XMM2,%XMM4,%XMM4 | 1 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 0 | 0 | 4 | 0.50 | vect (25.0%) |
VSUBPD %XMM1,%XMM0,%XMM2 | 1 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 0 | 0 | 4 | 0.50 | vect (25.0%) |
VADDPD %XMM1,%XMM0,%XMM0 | 1 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 0 | 0 | 4 | 0.50 | vect (25.0%) |
VMOVAPD %XMM6,(%RSI) | 1 | 0 | 0 | 0.33 | 0.33 | 1 | 0 | 0 | 0.33 | 3 | 1 | vect (25.0%) |
MOV 0x18(%RCX),%RSI | 1 | 0 | 0 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 4-5 | 0.50 | N/A |
VMOVAPD %XMM4,(%RAX,%RSI,8) | 1 | 0 | 0 | 0.33 | 0.33 | 1 | 0 | 0 | 0.33 | 3 | 1 | vect (25.0%) |
MOV 0x10(%RCX),%RSI | 1 | 0 | 0 | 0.50 | 0.50 | 0 | 0 | 0 | 0 | 4-5 | 0.50 | N/A |
ADD %RDI,%RCX | 1 | 0.25 | 0.25 | 0 | 0 | 0 | 0.25 | 0.25 | 0 | 1 | 0.25 | N/A |
VMOVAPD %XMM2,(%RAX,%RSI,8) | 1 | 0 | 0 | 0.33 | 0.33 | 1 | 0 | 0 | 0.33 | 3 | 1 | vect (25.0%) |
VMOVAPD %XMM0,(%RAX) | 1 | 0 | 0 | 0.33 | 0.33 | 1 | 0 | 0 | 0.33 | 3 | 1 | vect (25.0%) |
ADD %R10,%RAX | 1 | 0.25 | 0.25 | 0 | 0 | 0 | 0.25 | 0.25 | 0 | 1 | 0.25 | N/A |
CMP %R8,%R9 | 1 | 0.25 | 0.25 | 0 | 0 | 0 | 0.25 | 0.25 | 0 | 1 | 0.25 | N/A |
JNE 104da0 <t2fv_4+0x40> | 1 | 0.50 | 0 | 0 | 0 | 0 | 0 | 0.50 | 0 | 0 | 0.50-1 | N/A |