OMP: pid 25509 tid 25509 thread 0 bound to OS proc set {0}
OMP: pid 25509 tid 25914 thread 1 bound to OS proc set {1}
OMP: pid 25509 tid 25925 thread 12 bound to OS proc set {12}
OMP: pid 25509 tid 25921 thread 8 bound to OS proc set {8}
OMP: pid 25509 tid 25924 thread 11 bound to OS proc set {11}
OMP: pid 25509 tid 25928 thread 15 bound to OS proc set {15}
OMP: pid 25509 tid 25919 thread 6 bound to OS proc set {6}
OMP: pid 25509 tid 25929 thread 16 bound to OS proc set {16}
OMP: pid 25509 tid 25963 thread 50 bound to OS proc set {50}
OMP: pid 25509 tid 25961 thread 48 bound to OS proc set {48}
OMP: pid 25509 tid 25965 thread 52 bound to OS proc set {52}
OMP: pid 25509 tid 25962 thread 49 bound to OS proc set {49}
OMP: pid 25509 tid 25964 thread 51 bound to OS proc set {51}
OMP: pid 25509 tid 25977 thread 64 bound to OS proc set {64}
OMP: pid 25509 tid 25980 thread 67 bound to OS proc set {67}
OMP: pid 25509 tid 25981 thread 68 bound to OS proc set {68}
OMP: pid 25509 tid 25979 thread 66 bound to OS proc set {66}
OMP: pid 25509 tid 25978 thread 65 bound to OS proc set {65}
OMP: pid 25509 tid 25968 thread 55 bound to OS proc set {55}
OMP: pid 25509 tid 25967 thread 54 bound to OS proc set {54}
OMP: pid 25509 tid 25915 thread 2 bound to OS proc set {2}
OMP: pid 25509 tid 25966 thread 53 bound to OS proc set {53}
OMP: pid 25509 tid 25920 thread 7 bound to OS proc set {7}
OMP: pid 25509 tid 25941 thread 28 bound to OS proc set {28}
OMP: pid 25509 tid 25923 thread 10 bound to OS proc set {10}
OMP: pid 25509 tid 25933 thread 20 bound to OS proc set {20}
OMP: pid 25509 tid 25945 thread 32 bound to OS proc set {32}
OMP: pid 25509 tid 25989 thread 76 bound to OS proc set {76}
OMP: pid 25509 tid 25988 thread 75 bound to OS proc set {75}
OMP: pid 25509 tid 25992 thread 79 bound to OS proc set {79}
OMP: pid 25509 tid 25984 thread 71 bound to OS proc set {71}
OMP: pid 25509 tid 25937 thread 24 bound to OS proc set {24}
OMP: pid 25509 tid 25957 thread 44 bound to OS proc set {44}
OMP: pid 25509 tid 25953 thread 40 bound to OS proc set {40}
OMP: pid 25509 tid 25960 thread 47 bound to OS proc set {47}
OMP: pid 25509 tid 25949 thread 36 bound to OS proc set {36}
OMP: pid 25509 tid 25959 thread 46 bound to OS proc set {46}
OMP: pid 25509 tid 25991 thread 78 bound to OS proc set {78}
OMP: pid 25509 tid 25993 thread 80 bound to OS proc set {80}
OMP: pid 25509 tid 25997 thread 84 bound to OS proc set {84}
OMP: pid 25509 tid 25996 thread 83 bound to OS proc set {83}
OMP: pid 25509 tid 25970 thread 57 bound to OS proc set {57}
OMP: pid 25509 tid 25973 thread 60 bound to OS proc set {60}
OMP: pid 25509 tid 25972 thread 59 bound to OS proc set {59}
OMP: pid 25509 tid 25969 thread 56 bound to OS proc set {56}
OMP: pid 25509 tid 25971 thread 58 bound to OS proc set {58}
OMP: pid 25509 tid 25974 thread 61 bound to OS proc set {61}
OMP: pid 25509 tid 25975 thread 62 bound to OS proc set {62}
OMP: pid 25509 tid 25976 thread 63 bound to OS proc set {63}
OMP: pid 25509 tid 25956 thread 43 bound to OS proc set {43}
OMP: pid 25509 tid 25955 thread 42 bound to OS proc set {42}
OMP: pid 25509 tid 25987 thread 74 bound to OS proc set {74}
OMP: pid 25509 tid 25995 thread 82 bound to OS proc set {82}
OMP: pid 25509 tid 25944 thread 31 bound to OS proc set {31}
OMP: pid 25509 tid 25985 thread 72 bound to OS proc set {72}
OMP: pid 25509 tid 25927 thread 14 bound to OS proc set {14}
OMP: pid 25509 tid 25948 thread 35 bound to OS proc set {35}
OMP: pid 25509 tid 25916 thread 3 bound to OS proc set {3}
OMP: pid 25509 tid 25947 thread 34 bound to OS proc set {34}
OMP: pid 25509 tid 25917 thread 4 bound to OS proc set {4}
OMP: pid 25509 tid 25918 thread 5 bound to OS proc set {5}
OMP: pid 25509 tid 25954 thread 41 bound to OS proc set {41}
OMP: pid 25509 tid 25983 thread 70 bound to OS proc set {70}
OMP: pid 25509 tid 25946 thread 33 bound to OS proc set {33}
OMP: pid 25509 tid 25943 thread 30 bound to OS proc set {30}
OMP: pid 25509 tid 25931 thread 18 bound to OS proc set {18}
OMP: pid 25509 tid 25932 thread 19 bound to OS proc set {19}
OMP: pid 25509 tid 26001 thread 88 bound to OS proc set {88}
OMP: pid 25509 tid 26005 thread 92 bound to OS proc set {92}
OMP: pid 25509 tid 25958 thread 45 bound to OS proc set {45}
OMP: pid 25509 tid 25952 thread 39 bound to OS proc set {39}
OMP: pid 25509 tid 25951 thread 38 bound to OS proc set {38}
OMP: pid 25509 tid 25994 thread 81 bound to OS proc set {81}
OMP: pid 25509 tid 25998 thread 85 bound to OS proc set {85}
OMP: pid 25509 tid 25922 thread 9 bound to OS proc set {9}
OMP: pid 25509 tid 26008 thread 95 bound to OS proc set {95}
OMP: pid 25509 tid 25926 thread 13 bound to OS proc set {13}
OMP: pid 25509 tid 25936 thread 23 bound to OS proc set {23}
OMP: pid 25509 tid 26000 thread 87 bound to OS proc set {87}
OMP: pid 25509 tid 25990 thread 77 bound to OS proc set {77}
OMP: pid 25509 tid 26004 thread 91 bound to OS proc set {91}
OMP: pid 25509 tid 25940 thread 27 bound to OS proc set {27}
OMP: pid 25509 tid 25930 thread 17 bound to OS proc set {17}
OMP: pid 25509 tid 25986 thread 73 bound to OS proc set {73}
OMP: pid 25509 tid 25982 thread 69 bound to OS proc set {69}
OMP: pid 25509 tid 25942 thread 29 bound to OS proc set {29}
OMP: pid 25509 tid 26007 thread 94 bound to OS proc set {94}
OMP: pid 25509 tid 26006 thread 93 bound to OS proc set {93}
OMP: pid 25509 tid 25935 thread 22 bound to OS proc set {22}
OMP: pid 25509 tid 25939 thread 26 bound to OS proc set {26}
OMP: pid 25509 tid 25938 thread 25 bound to OS proc set {25}
OMP: pid 25509 tid 25950 thread 37 bound to OS proc set {37}
OMP: pid 25509 tid 25934 thread 21 bound to OS proc set {21}
OMP: pid 25509 tid 26003 thread 90 bound to OS proc set {90}
OMP: pid 25509 tid 26002 thread 89 bound to OS proc set {89}
OMP: pid 25509 tid 25999 thread 86 bound to OS proc set {86}
-------------------------------------------------------------
STREAM version $Revision: 5.10 $
-------------------------------------------------------------
This system uses 8 bytes per array element.
-------------------------------------------------------------
Array size = 860160000 (elements), Offset = 0 (elements)
Memory per array = 6562.5 MiB (= 6.4 GiB).
Total memory required = 19687.5 MiB (= 19.2 GiB).
Each kernel will be executed 100 times.
The *best* time for each kernel (excluding the first iteration)
will be used to compute the reported bandwidth.
-------------------------------------------------------------
Number of Threads requested = 96
Number of Threads counted = 96
-------------------------------------------------------------
Your clock granularity/precision appears to be 1 microseconds.
Each test below will take on the order of 37689 microseconds.
(= 37689 clock ticks)
Increase the size of the arrays if this shows that
you are not getting at least 20 clock ticks per test.
-------------------------------------------------------------
WARNING -- The above is only a rough guideline.
For best results, please be sure you know the
precision of your system timer.
-------------------------------------------------------------
Function Best Rate MB/s Avg time Min time Max time
Copy: 370579.1 0.037226 0.037138 0.037398
Scale: inf 0.000000 0.000000 0.000001
Add: inf 0.000000 0.000000 0.000001
Triad: inf 0.000000 0.000000 0.000001
-------------------------------------------------------------
-------------------------------------------------------------
Your experiment path is /beegfs/hackathon/users/eoseret/qaas_runs_test/ins01.benchmarkcenter.megware.com/177-374-0012/stream-copy/run/oneview_runs/compilers/aocc_10/oneview_results_1773740672/tools/lprof_run_0
To display your profiling results:
#####################################################################################################################################################################################################################################################
# LEVEL | REPORT | COMMAND #
#####################################################################################################################################################################################################################################################
# Functions | Cluster-wide | maqao lprof -df xp=/beegfs/hackathon/users/eoseret/qaas_runs_test/ins01.benchmarkcenter.megware.com/177-374-0012/stream-copy/run/oneview_runs/compilers/aocc_10/oneview_results_1773740672/tools/lprof_run_0 #
# Functions | Per-node | maqao lprof -df -dn xp=/beegfs/hackathon/users/eoseret/qaas_runs_test/ins01.benchmarkcenter.megware.com/177-374-0012/stream-copy/run/oneview_runs/compilers/aocc_10/oneview_results_1773740672/tools/lprof_run_0 #
# Functions | Per-process | maqao lprof -df -dp xp=/beegfs/hackathon/users/eoseret/qaas_runs_test/ins01.benchmarkcenter.megware.com/177-374-0012/stream-copy/run/oneview_runs/compilers/aocc_10/oneview_results_1773740672/tools/lprof_run_0 #
# Functions | Per-thread | maqao lprof -df -dt xp=/beegfs/hackathon/users/eoseret/qaas_runs_test/ins01.benchmarkcenter.megware.com/177-374-0012/stream-copy/run/oneview_runs/compilers/aocc_10/oneview_results_1773740672/tools/lprof_run_0 #
# Loops | Cluster-wide | maqao lprof -dl xp=/beegfs/hackathon/users/eoseret/qaas_runs_test/ins01.benchmarkcenter.megware.com/177-374-0012/stream-copy/run/oneview_runs/compilers/aocc_10/oneview_results_1773740672/tools/lprof_run_0 #
# Loops | Per-node | maqao lprof -dl -dn xp=/beegfs/hackathon/users/eoseret/qaas_runs_test/ins01.benchmarkcenter.megware.com/177-374-0012/stream-copy/run/oneview_runs/compilers/aocc_10/oneview_results_1773740672/tools/lprof_run_0 #
# Loops | Per-process | maqao lprof -dl -dp xp=/beegfs/hackathon/users/eoseret/qaas_runs_test/ins01.benchmarkcenter.megware.com/177-374-0012/stream-copy/run/oneview_runs/compilers/aocc_10/oneview_results_1773740672/tools/lprof_run_0 #
# Loops | Per-thread | maqao lprof -dl -dt xp=/beegfs/hackathon/users/eoseret/qaas_runs_test/ins01.benchmarkcenter.megware.com/177-374-0012/stream-copy/run/oneview_runs/compilers/aocc_10/oneview_results_1773740672/tools/lprof_run_0 #
#####################################################################################################################################################################################################################################################