Intel(R) VTune(TM) Profiler Self Check Utility Copyright (C) 2009 Intel Corporation. All rights reserved. Build Number: 624757 Ignored warnings: ['To profile kernel modules during the session, make sure they are available in the /lib/modules/kernel_version/ location.', 'To enable hardware event-based sampling, PRODUCT_LEGAL_SHORT_NAME has disabled the NMI watchdog timer. The watchdog timer will be re-enabled after collection completes.'] Check of files: Ok ================================================================================ Context values: Command line: /opt/intel/oneapi/vtune/2023.0.0/bin64/amplxe-runss --context-value-list Stdout: targetOS: Linux OS: Linux OSBuildNumber: 0 OSBitness: 64 RootPrivileges: false isPtraceScopeLimited: false isCATSupportedByCPU: true isL3CATAvailable: true L3CATDetails: COS=16;ways=11 isL2CATAvailable: false isL3MonitoringSupportedByCPU: true LLCSize: 37486592 cacheMonitoringUpscalingFactor: 106496 isL3CacheOccupancyAvailable: true isL3TotalBWAvailable: true isL3LocalBWAvailable: true isTSXAvailable: false isPTAvailable: true isHTEnabled: true fpgaOnBoard: None omniPathOnBoard: None genArchOnBoard: 0 pciClassParts: 0x1:0,1;0x2:0;0x3:1;0x6:0,1;0x8:0; tidValuesForIO: 0x1d8;0x1f0;0x1f8 populatedIoParts: 0;1 populatedIoUnits: 0;1;3 populatedTidValuesForIO: 0x1d8;0x1f0 isSGXAvailable: false LinuxRelease: 4.18.0-348.23.1.el8_5.x86_64 is3DXPPresent: false is3DXP2LMMode: false is3DXPAppDirectMode: false IsNUMANodeWithoutCPUsPresent: false Hypervisor: None PerfmonVersion: 4 isMaxDRAMBandwidthMeasurementSupported: true preferedGpuAdapter: none isEHFIAvailable: false isPtraceAvailable: true i915Status: MissingDriver isFtraceAvailable: ftraceAccessError,debugfsNotAccessible isMdfEtwAvailable: false isCSwitchAvailable: no isGpuBusynessAvailable: unsupportedHardware isGpuWaitAvailable: no isFunctionTracingAvailable: no isIowaitTracingAvailable: no isVSyncAvailable: no HypervisorType: None isDeviceOrCredentialGuardEnabled: false isSEPDriverAvailable: false isPAXDriverLoaded: false platformType: 127 CPU_NAME: Intel(R) Xeon(R) Processor code named Cascadelake PMU: cascadelake_server referenceFrequency: 2400000000 isPStateAvailable: true isVTSSPPDriverAvailable: false isNMIWatchDogTimerRunning: true LinuxPerfCredentials: User LinuxPerfCapabilities: breakpoint:raw;cpu:raw,format,events,ldlat,frontend;cstate_core:raw,format,events;cstate_pkg:raw,format,events;intel_bts:raw;intel_pt:raw,format;kprobe:raw,format;msr:raw,format,events;power:raw,format,events;software:raw;tracepoint:raw;uncore_cha:26,raw,format;uncore_iio:6,raw,format;uncore_iio_free_running:6,raw,format,events;uncore_imc:6,raw,format,events;uncore_irp:6,raw,format;uncore_m2m:2,raw,format;uncore_m2pcie:3,raw,format;uncore_m3upi:3,raw,format;uncore_pcu:raw,format;uncore_ubox:raw,format;uncore_upi:3,raw,format;uprobe:raw,format LinuxPerfStackCapabilities: fp,dwarf,lbr areKernelPtrsRestricted: yes LinuxPerfMuxIntervalMs: 1 isPerfPCIeMappingAvailable: true isAOCLAvailable: false isTPSSAvailable: true isPytraceAvailable: true forceShowInlines: false isEnergyCollectionSupported: true isSocwatchDriverLoaded: false isCPUSupportedBySocwatch: true isCpuThrottlingAvailable: false isIPMWatchReady: false osCountersCollectorAvailability: dstatNotAvailable l0LoaderStatus: LibNotFound l0DevicesAvailable: false l0VPUDevicesAvailable: false l0GPUDevicesAvailable: false Getting context values: OK ================================================================================ Check driver: isSEPDriverAvailable: false isPAXDriverLoaded: false Command line: lsmod Is SEP in lsmod: False The SEP driver is not available. ================================================================================ SEP version: Command line: /opt/intel/oneapi/vtune/2023.0.0/bin64/sep -version Stdout: Sampling Enabling Product Version: 5.38 Beta built on Dec 6 2022 03:53:32 SEP Driver Version: PAX Driver Version: Platform type: 127 CPU name: Intel(R) Xeon(R) Processor code named Cascadelake PMU: cascadelake_server Stderr: Error retrieving SEP driver version Error retrieving PAX driver version Check driver with sep -version: Fail ================================================================================ HW event-based analysis (counting mode)... Command line: /opt/intel/oneapi/vtune/2023.0.0/bin64/vtune -collect performance-snapshot -r /tmp/vtune-tmp-resteam/self-checker-2024.05.07_09.00.52/result_ps -data-limit 0 -finalization-mode none -source-search-dir /opt/intel/oneapi/vtune/2023.0.0/samples/en/C++/matrix/src -- /opt/intel/oneapi/vtune/2023.0.0/samples/en/C++/matrix/matrix Stdout: Addr of buf1 = 0x7f5420709010 Offs of buf1 = 0x7f5420709180 Addr of buf2 = 0x7f541e708010 Offs of buf2 = 0x7f541e7081c0 Addr of buf3 = 0x7f541c707010 Offs of buf3 = 0x7f541c707100 Addr of buf4 = 0x7f541a706010 Offs of buf4 = 0x7f541a706140 Threads #: 16 Pthreads Matrix size: 2048 Using multiply kernel: multiply1 Execution time = 2.390 seconds Stderr: vtune: Collection started. To stop the collection, either press CTRL-C or enter from another console window: vtune -r /tmp/vtune-tmp-resteam/self-checker-2024.05.07_09.00.52/result_ps -command stop. vtune: Collection stopped. vtune: Using result path `/tmp/vtune-tmp-resteam/self-checker-2024.05.07_09.00.52/result_ps' vtune: Executing actions 0 % vtune: Executing actions 100 % vtune: Executing actions 100 % done HW event-based analysis (counting mode) (Perf) Example of analysis types: Performance Snapshot Collection: Ok -------------------------------------------------------------------------------- Running finalization... Command line: /opt/intel/oneapi/vtune/2023.0.0/bin64/vtune -finalize -r /tmp/vtune-tmp-resteam/self-checker-2024.05.07_09.00.52/result_ps Stderr: vtune: Using result path `/tmp/vtune-tmp-resteam/self-checker-2024.05.07_09.00.52/result_ps' vtune: Executing actions 0 % vtune: Executing actions 0 % Finalizing results vtune: Executing actions 0 % Finalizing the result vtune: Executing actions 0 % Clearing the database vtune: Executing actions 14 % Clearing the database vtune: Executing actions 14 % Loading raw data to the database vtune: Executing actions 14 % Loading 'systemcollector-918381-srv-sopt1.sc' fil vtune: Executing actions 25 % Loading 'systemcollector-918381-srv-sopt1.sc' fil vtune: Executing actions 25 % Loading '918391.stat.perf' file vtune: Executing actions 25 % Updating precomputed scalar metrics vtune: Executing actions 28 % Updating precomputed scalar metrics vtune: Executing actions 28 % Processing profile metrics and debug information vtune: Executing actions 39 % Processing profile metrics and debug information vtune: Executing actions 39 % Setting data model parameters vtune: Executing actions 39 % Resolving module symbols vtune: Executing actions 39 % Resolving thread name information vtune: Executing actions 43 % Resolving thread name information vtune: Executing actions 43 % Resolving call target names for dynamic code vtune: Executing actions 48 % Resolving call target names for dynamic code vtune: Executing actions 48 % Resolving interrupt name information vtune: Executing actions 53 % Resolving interrupt name information vtune: Executing actions 53 % Processing profile metrics and debug information vtune: Executing actions 56 % Processing profile metrics and debug information vtune: Executing actions 57 % Processing profile metrics and debug information vtune: Executing actions 58 % Processing profile metrics and debug information vtune: Executing actions 59 % Processing profile metrics and debug information vtune: Executing actions 60 % Processing profile metrics and debug information vtune: Executing actions 60 % Setting data model parameters vtune: Executing actions 60 % Precomputing frequently used data vtune: Executing actions 60 % Precomputing frequently used data vtune: Executing actions 62 % Precomputing frequently used data vtune: Executing actions 63 % Precomputing frequently used data vtune: Executing actions 64 % Precomputing frequently used data vtune: Executing actions 65 % Precomputing frequently used data vtune: Executing actions 66 % Precomputing frequently used data vtune: Executing actions 67 % Precomputing frequently used data vtune: Executing actions 68 % Precomputing frequently used data vtune: Executing actions 69 % Precomputing frequently used data vtune: Executing actions 71 % Precomputing frequently used data vtune: Executing actions 72 % Precomputing frequently used data vtune: Executing actions 72 % Updating precomputed scalar metrics vtune: Executing actions 75 % Updating precomputed scalar metrics vtune: Executing actions 75 % Discarding redundant overtime data vtune: Executing actions 78 % Discarding redundant overtime data vtune: Executing actions 78 % Saving the result vtune: Executing actions 82 % Saving the result vtune: Executing actions 85 % Saving the result vtune: Executing actions 99 % Saving the result vtune: Executing actions 100 % Saving the result vtune: Executing actions 100 % done Finalization: Ok -------------------------------------------------------------------------------- Command line: /opt/intel/oneapi/vtune/2023.0.0/bin64/vtune -R summary -r /tmp/vtune-tmp-resteam/self-checker-2024.05.07_09.00.52/result_ps Stdout: Elapsed Time: 2.410s IPC: 0.596 | The IPC may be too low. This could be caused by issues such as memory | stalls, instruction starvation, branch misprediction or long latency | instructions. Explore the other hardware-related metrics to identify what | is causing low IPC. | DP GFLOPS: 7.129 Average CPU Frequency: 3.100 GHz Logical Core Utilization: 16.1% (15.438 out of 96) | The metric value is low, which may signal a poor logical CPU cores | utilization. Consider improving physical core utilization as the first step | and then look at opportunities to utilize logical cores, which in some cases | can improve processor throughput and overall performance of multi-threaded | applications. | Physical Core Utilization: 31.9% (15.320 out of 48) | The metric value is low, which may signal a poor physical CPU cores | utilization caused by: | - load imbalance | - threading runtime overhead | - contended synchronization | - thread/process underutilization | - incorrect affinity that utilizes logical cores instead of physical | cores | Run the HPC Performance Characterization analysis to estimate the | efficiency of MPI and OpenMP parallelism or run the Locks and Waits | analysis to identify parallel bottlenecks for other parallel runtimes. | Microarchitecture Usage: 15.1% of Pipeline Slots | You code efficiency on this platform is too low. | | Possible cause: memory stalls, instruction starvation, branch misprediction | or long latency instructions. | | Next steps: Run Microarchitecture Exploration analysis to identify the cause | of the low microarchitecture usage efficiency. | Retiring: 15.1% of Pipeline Slots Front-End Bound: 0.1% of Pipeline Slots Bad Speculation: 0.1% of Pipeline Slots Back-End Bound: 84.7% of Pipeline Slots | A significant portion of pipeline slots are remaining empty. When | operations take too long in the back-end, they introduce bubbles in the | pipeline that ultimately cause fewer pipeline slots containing useful | work to be retired per cycle than the machine is capable to support. This | opportunity cost results in slower execution. Long-latency operations | like divides and memory operations can cause this, as can too many | operations being directed to a single execution port (for example, more | multiply operations arriving in the back-end per cycle than the execution | unit can support). | Memory Bound: 67.6% of Pipeline Slots | The metric value is high. This can indicate that the significant | fraction of execution pipeline slots could be stalled due to demand | memory load and stores. Use Memory Access analysis to have the metric | breakdown by memory hierarchy, memory bandwidth information, | correlation by memory objects. | L1 Bound: 0.1% of Clockticks FB Full: 100.0% of Clockticks L2 Bound: 0.0% of Clockticks L3 Bound: 49.1% of Clockticks | This metric shows how often CPU was stalled on L3 cache, or | contended with a sibling Core. Avoiding cache misses (L2 | misses/L3 hits) improves the latency and increases performance. | L3 Latency: 100.0% of Clockticks | This metric shows a fraction of cycles with demand load | accesses that hit the L3 cache under unloaded scenarios | (possibly L3 latency limited). Avoiding private cache misses | (i.e. L2 misses/L3 hits) will improve the latency, reduce | contention with sibling physical cores and increase | performance. Note the value of this node may overlap with its | siblings. | DRAM Bound: 13.5% of Clockticks | This metric shows how often CPU was stalled on the main memory | (DRAM). Caching typically improves the latency and increases | performance. | Memory Bandwidth: 86.9% of Clockticks | Issue: A significant fraction of cycles was stalled due to | approaching bandwidth limits of the main memory (DRAM). | | Tips: Improve data accesses to reduce cacheline transfers | from/to memory using these possible techniques: | - Consume all bytes of each cacheline before it is | evicted (for example, reorder structure elements and | split non-hot ones). | - Merge compute-limited and bandwidth-limited loops. | - Use NUMA optimizations on a multi-socket system. | | Note: software prefetches do not help a bandwidth-limited | application. | Memory Latency: 12.3% of Clockticks | Issue: A significant fraction of cycles was stalled due to | the latency of the main memory (DRAM). | | Tips: Improve data accesses or interleave them with compute | using such possible techniques as data layout re-structuring | or software prefetches (through the compiler). | Local DRAM: 25.0% of Clockticks | The number of CPU stalls on loads from the local memory | exceeds the threshold. Consider caching data to improve | the latency and increase the performance. | Remote DRAM: 4.3% of Clockticks Remote Cache: 9.1% of Clockticks | The number of CPU stalls on loads from the remote cache | exceeds the threshold. This is often caused by non- | optimal NUMA memory allocations. | Store Bound: 0.0% of Clockticks Core Bound: 17.1% of Pipeline Slots | This metric represents how much Core non-memory issues were of a | bottleneck. Shortage in hardware compute resources, or dependencies | software's instructions are both categorized under Core Bound. Hence | it may indicate the machine ran out of an OOO resources, certain | execution units are overloaded or dependencies in program's data- or | instruction- flow are limiting the performance (e.g. FP-chained long- | latency arithmetic operations). | Memory Bound: 67.6% of Pipeline Slots | The metric value is high. This can indicate that the significant fraction of | execution pipeline slots could be stalled due to demand memory load and | stores. Use Memory Access analysis to have the metric breakdown by memory | hierarchy, memory bandwidth information, correlation by memory objects. | Cache Bound: 49.3% of Clockticks | A significant proportion of cycles are being spent on data fetches from | caches. Check Memory Access analysis to see if accesses to L2 or L3 | caches are problematic and consider applying the same performance tuning | as you would for a cache-missing workload. This may include reducing the | data working set size, improving data access locality, blocking or | partitioning the working set to fit in the lower cache levels, or | exploiting hardware prefetchers. Consider using software prefetchers, but | note that they can interfere with normal loads, increase latency, and | increase pressure on the memory system. This metric includes coherence | penalties for shared data. Check Microarchitecture Exploration analysis | to see if contested accesses or data sharing are indicated as likely | issues. | DRAM Bound: 13.5% of Clockticks | The metric value is high. This indicates that a significant fraction of | cycles could be stalled on the main memory (DRAM) because of demand loads | or stores. | | The code is memory bandwidth bound, which means that there are a | significant fraction of cycles during which the bandwidth limits of the | main memory are being reached and the code could stall. Review the | Bandwidth Utilization Histogram to estimate the scale of the issue. | Improve data accesses to reduce cacheline transfers from/to memory using | these possible techniques: 1) consume all bytes of each cacheline before | it is evicted (for example, reorder structure elements and split non-hot | ones); 2) merge compute-limited and bandwidth-limited loops; 3) use NUMA | optimizations on a multi-socket system. | | The code is latency bound, which means that there are a significant | fraction of cycles during which the code could be stalled due to main | memory latency. Consider optimizing data layout or using software | prefetches through the compiler to improve cache reuse and to reduce the | data fetched from the main memory. | NUMA: % of Remote Accesses: 29.0% | A significant amount of DRAM loads were serviced from remote DRAM. | Wherever possible, try to consistently use data on the same core, or at | least the same package, as it was allocated on. | Vectorization: 0.3% of Packed FP Operations | A significant fraction of floating point arithmetic instructions are scalar. | This indicates that the code was not fully vectorized. Use Intel Advisor to | see possible reasons why the code was not vectorized. | Instruction Mix SP FLOPs: 0.0% of uOps Packed: 0.0% from SP FP 128-bit: 0.0% from SP FP 256-bit: 0.0% from SP FP 512-bit: 0.0% from SP FP Scalar: 0.0% from SP FP DP FLOPs: 24.8% of uOps Packed: 0.3% from DP FP 128-bit: 0.3% from DP FP 256-bit: 0.0% from DP FP 512-bit: 0.0% from DP FP Scalar: 99.7% from DP FP | A significant fraction of floating point arithmetic instructions | are scalar. This indicates that the code was not fully | vectorized. Use Intel Advisor to see possible reasons why the | code was not vectorized. | x87 FLOPs: 0.0% of uOps Non-FP: 75.2% of uOps FP Arith/Mem Rd Instr. Ratio: 0.997 FP Arith/Mem Wr Instr. Ratio: 1.989 Collection and Platform Info Application Command Line: /opt/intel/oneapi/vtune/2023.0.0/samples/en/C++/matrix/matrix Operating System: 4.18.0-348.23.1.el8_5.x86_64 Red Hat Enterprise Linux release 8.2 (Ootpa) Computer Name: srv-sopt1 Result Size: 3,6 MB Collection start time: 07:00:54 07/05/2024 UTC Collection stop time: 07:00:57 07/05/2024 UTC Collector Type: Driverless Perf per-process counting CPU Name: Intel(R) Xeon(R) Processor code named Cascadelake Frequency: 2.400 GHz Logical CPU Count: 96 LLC size: 37,5 MB Cache Allocation Technology Level 2 capability: not detected Level 3 capability: available Recommendations: Hotspots: Start with Hotspots analysis to understand the efficiency of your algorithm. | Use Hotspots analysis to identify the most time consuming functions. | Drill down to see the time spent on every line of code. Memory Access: The Memory Bound metric is high (67.6%). A significant fraction of execution pipeline slots could be stalled due to demand memory load and stores. | Use Memory Access analysis to measure metrics that can identify memory | access issues. Threading: There is poor utilization of logical CPU cores (16.1%) in your application. | Use Threading to explore more opportunities to increase parallelism in | your application. HPC Performance Characterization: Vectorization (0.3%) is low. A significant fraction of floating point arithmetic instructions are scalar. This indicates that the code was not fully vectorized. Use Intel Advisor to see possible reasons why the code was not vectorized. | Use HPC Performance Characterization analysis to examine the performance | of compute-intensive applications. Understand CPU/GPU utilization and get | information about OpenMP efficiency, memory access, and vectorization. If you want to skip descriptions of detected performance issues in the report, enter: vtune -report summary -report-knob show-issues=false -r . Alternatively, you may view the report in the csv format: vtune -report -format=csv. Stderr: vtune: Using result path `/tmp/vtune-tmp-resteam/self-checker-2024.05.07_09.00.52/result_ps' vtune: Executing actions 0 % vtune: Executing actions 0 % Finalizing results vtune: Executing actions 50 % Finalizing results vtune: Executing actions 50 % Generating a report vtune: Executing actions 50 % Setting data model parameters vtune: Executing actions 75 % Setting data model parameters vtune: Executing actions 75 % Generating a report vtune: Executing actions 100 % Generating a report vtune: Executing actions 100 % done Report: Ok ================================================================================ Instrumentation based analysis check... Command line: /opt/intel/oneapi/vtune/2023.0.0/bin64/vtune -collect hotspots -r /tmp/vtune-tmp-resteam/self-checker-2024.05.07_09.00.52/result_tpss -data-limit 0 -finalization-mode none -source-search-dir /opt/intel/oneapi/vtune/2023.0.0/samples/en/C++/matrix/src -- /opt/intel/oneapi/vtune/2023.0.0/samples/en/C++/matrix/matrix Stdout: Addr of buf1 = 0x7f49165c1010 Offs of buf1 = 0x7f49165c1180 Addr of buf2 = 0x7f49145c0010 Offs of buf2 = 0x7f49145c01c0 Addr of buf3 = 0x7f49125bf010 Offs of buf3 = 0x7f49125bf100 Addr of buf4 = 0x7f49105be010 Offs of buf4 = 0x7f49105be140 Threads #: 16 Pthreads Matrix size: 2048 Using multiply kernel: multiply1 Execution time = 2.302 seconds Stderr: vtune: Warning: Only user space will be profiled due to credentials lack. Consider changing /proc/sys/kernel/perf_event_paranoid file for enabling kernel space profiling. vtune: Collection started. To stop the collection, either press CTRL-C or enter from another console window: vtune -r /tmp/vtune-tmp-resteam/self-checker-2024.05.07_09.00.52/result_tpss -command stop. vtune: Collection stopped. vtune: Using result path `/tmp/vtune-tmp-resteam/self-checker-2024.05.07_09.00.52/result_tpss' vtune: Executing actions 0 % vtune: Executing actions 100 % vtune: Executing actions 100 % done Instrumentation based analysis check Example of analysis types: Hotspots and Threading with user-mode sampling Collection: Ok vtune: Warning: Only user space will be profiled due to credentials lack. Consider changing /proc/sys/kernel/perf_event_paranoid file for enabling kernel space profiling. -------------------------------------------------------------------------------- Running finalization... Command line: /opt/intel/oneapi/vtune/2023.0.0/bin64/vtune -finalize -r /tmp/vtune-tmp-resteam/self-checker-2024.05.07_09.00.52/result_tpss Stderr: vtune: Using result path `/tmp/vtune-tmp-resteam/self-checker-2024.05.07_09.00.52/result_tpss' vtune: Executing actions 0 % vtune: Executing actions 0 % Finalizing results vtune: Executing actions 0 % Finalizing the result vtune: Executing actions 0 % Clearing the database vtune: Executing actions 14 % Clearing the database vtune: Executing actions 14 % Loading raw data to the database vtune: Executing actions 14 % Loading 'systemcollector-918505-srv-sopt1.sc' fil vtune: Executing actions 25 % Loading 'systemcollector-918505-srv-sopt1.sc' fil vtune: Executing actions 25 % Loading '918515.stat.perf' file vtune: Executing actions 25 % Loading '918505-918515.0.trace' file vtune: Executing actions 25 % Updating precomputed scalar metrics vtune: Executing actions 28 % Updating precomputed scalar metrics vtune: Executing actions 28 % Processing profile metrics and debug information vtune: Executing actions 39 % Processing profile metrics and debug information vtune: Executing actions 39 % Setting data model parameters vtune: Executing actions 39 % Resolving module symbols vtune: Executing actions 39 % Resolving information for `matrix' vtune: Executing actions 39 % Resolving information for `libpthread.so.0' vtune: Warning: Cannot locate debugging information for file `/lib64/libpthread.so.0'. vtune: Executing actions 39 % Resolving information for `libc.so.6' vtune: Executing actions 41 % Resolving information for `libc.so.6' vtune: Executing actions 43 % Resolving information for `libc.so.6' vtune: Warning: Cannot locate debugging information for file `/lib64/libc.so.6'. vtune: Executing actions 45 % Resolving information for `libc.so.6' vtune: Executing actions 45 % Resolving bottom user stack information vtune: Executing actions 46 % Resolving bottom user stack information vtune: Executing actions 46 % Resolving thread name information vtune: Executing actions 47 % Resolving thread name information vtune: Executing actions 48 % Resolving thread name information vtune: Executing actions 48 % Resolving call target names for dynamic code vtune: Executing actions 49 % Resolving call target names for dynamic code vtune: Executing actions 49 % Resolving interrupt name information vtune: Executing actions 53 % Resolving interrupt name information vtune: Executing actions 53 % Processing profile metrics and debug information vtune: Executing actions 56 % Processing profile metrics and debug information vtune: Executing actions 57 % Processing profile metrics and debug information vtune: Executing actions 58 % Processing profile metrics and debug information vtune: Executing actions 60 % Processing profile metrics and debug information vtune: Executing actions 60 % Setting data model parameters vtune: Executing actions 60 % Precomputing frequently used data vtune: Executing actions 60 % Precomputing frequently used data vtune: Executing actions 63 % Precomputing frequently used data vtune: Executing actions 66 % Precomputing frequently used data vtune: Executing actions 69 % Precomputing frequently used data vtune: Executing actions 72 % Precomputing frequently used data vtune: Executing actions 72 % Updating precomputed scalar metrics vtune: Executing actions 75 % Updating precomputed scalar metrics vtune: Executing actions 75 % Discarding redundant overtime data vtune: Executing actions 78 % Discarding redundant overtime data vtune: Executing actions 78 % Saving the result vtune: Executing actions 82 % Saving the result vtune: Executing actions 85 % Saving the result vtune: Executing actions 99 % Saving the result vtune: Executing actions 100 % Saving the result vtune: Executing actions 100 % done Finalization: Ok -------------------------------------------------------------------------------- Command line: /opt/intel/oneapi/vtune/2023.0.0/bin64/vtune -limit 5 -format csv -csv-delimiter comma -report hotspots -group-by function -r /tmp/vtune-tmp-resteam/self-checker-2024.05.07_09.00.52/result_tpss Stdout: Function,CPU Time,CPU Time:Effective Time,CPU Time:Spin Time,CPU Time:Overhead Time,Module,Function (Full),Source File,Start Address multiply1,36.309990,36.309990,0.0,0.0,matrix,multiply1,multiply.c,0x401550 init_arr,0.010000,0.010000,0.0,0.0,matrix,init_arr,matrix.c,0x400d4f munmap,0.010000,0.010000,0.0,0.0,libc.so.6,munmap,[Unknown],0xf77d0 Stderr: vtune: Using result path `/tmp/vtune-tmp-resteam/self-checker-2024.05.07_09.00.52/result_tpss' vtune: Executing actions 0 % vtune: Executing actions 0 % Finalizing results vtune: Executing actions 50 % Finalizing results vtune: Executing actions 50 % Generating a report vtune: Executing actions 50 % Setting data model parameters vtune: Executing actions 75 % Setting data model parameters vtune: Executing actions 75 % Generating a report vtune: Executing actions 100 % Generating a report vtune: Executing actions 100 % done Report: Ok ================================================================================ HW event-based analysis check... Command line: /opt/intel/oneapi/vtune/2023.0.0/bin64/vtune -collect hotspots -knob sampling-mode=hw -r /tmp/vtune-tmp-resteam/self-checker-2024.05.07_09.00.52/result_ah -data-limit 0 -finalization-mode none -source-search-dir /opt/intel/oneapi/vtune/2023.0.0/samples/en/C++/matrix/src -- /opt/intel/oneapi/vtune/2023.0.0/samples/en/C++/matrix/matrix Stdout: Addr of buf1 = 0x7fbfcb361010 Offs of buf1 = 0x7fbfcb361180 Addr of buf2 = 0x7fbfc9360010 Offs of buf2 = 0x7fbfc93601c0 Addr of buf3 = 0x7fbfc735f010 Offs of buf3 = 0x7fbfc735f100 Addr of buf4 = 0x7fbfc535e010 Offs of buf4 = 0x7fbfc535e140 Threads #: 16 Pthreads Matrix size: 2048 Using multiply kernel: multiply1 Execution time = 1.924 seconds Stderr: vtune: Warning: Access to /proc/kallsyms file is limited. Consider changing /proc/sys/kernel/kptr_restrict to 0 to enable resolution of OS kernel and kernel module symbols. vtune: Warning: To analyze modules at the kernel level in this configuration, load the Intel sampling driver and set an unlimited (0) value for the Stack size option (if you require stack collection). Alternatively, enable access to kernel-mode monitoring by setting the /proc/sys/kernel/perf_event_paranoid value to 1 or lower. vtune: Warning: To profile kernel modules during the session, make sure they are available in the /lib/modules/kernel_version/ location. vtune: Warning: Consider reducing possible collection overhead by setting the /proc/sys/kernel/perf_event_paranoid value to 0 (or less). vtune: Collection started. To stop the collection, either press CTRL-C or enter from another console window: vtune -r /tmp/vtune-tmp-resteam/self-checker-2024.05.07_09.00.52/result_ah -command stop. vtune: Collection stopped. vtune: Using result path `/tmp/vtune-tmp-resteam/self-checker-2024.05.07_09.00.52/result_ah' vtune: Executing actions 0 % vtune: Executing actions 100 % vtune: Executing actions 100 % done HW event-based analysis check (Perf) Example of analysis types: Hotspots with HW event-based sampling, HPC Performance Characterization, etc. Collection: Ok vtune: Warning: Access to /proc/kallsyms file is limited. Consider changing /proc/sys/kernel/kptr_restrict to 0 to enable resolution of OS kernel and kernel module symbols. vtune: Warning: To analyze modules at the kernel level in this configuration, load the Intel sampling driver and set an unlimited (0) value for the Stack size option (if you require stack collection). Alternatively, enable access to kernel-mode monitoring by setting the /proc/sys/kernel/perf_event_paranoid value to 1 or lower. vtune: Warning: To profile kernel modules during the session, make sure they are available in the /lib/modules/kernel_version/ location. vtune: Warning: Consider reducing possible collection overhead by setting the /proc/sys/kernel/perf_event_paranoid value to 0 (or less). -------------------------------------------------------------------------------- Running finalization... Command line: /opt/intel/oneapi/vtune/2023.0.0/bin64/vtune -finalize -r /tmp/vtune-tmp-resteam/self-checker-2024.05.07_09.00.52/result_ah Stderr: vtune: Using result path `/tmp/vtune-tmp-resteam/self-checker-2024.05.07_09.00.52/result_ah' vtune: Executing actions 0 % vtune: Executing actions 0 % Finalizing results vtune: Executing actions 0 % Finalizing the result vtune: Executing actions 0 % Clearing the database vtune: Executing actions 14 % Clearing the database vtune: Executing actions 14 % Loading raw data to the database vtune: Executing actions 14 % Loading 'systemcollector-918637-srv-sopt1.sc' fil vtune: Executing actions 25 % Loading 'systemcollector-918637-srv-sopt1.sc' fil vtune: Executing actions 25 % Loading '918646.perf' file vtune: Executing actions 25 % Updating precomputed scalar metrics vtune: Executing actions 28 % Updating precomputed scalar metrics vtune: Executing actions 28 % Processing profile metrics and debug information vtune: Executing actions 39 % Processing profile metrics and debug information vtune: Executing actions 39 % Setting data model parameters vtune: Executing actions 39 % Resolving module symbols vtune: Executing actions 39 % Resolving information for dangling locations vtune: Executing actions 39 % Resolving information for `matrix' vtune: Executing actions 41 % Resolving information for `matrix' vtune: Executing actions 44 % Resolving information for `matrix' vtune: Executing actions 44 % Resolving bottom user stack information vtune: Executing actions 45 % Resolving bottom user stack information vtune: Executing actions 46 % Resolving bottom user stack information vtune: Executing actions 46 % Resolving thread name information vtune: Executing actions 47 % Resolving thread name information vtune: Executing actions 48 % Resolving thread name information vtune: Executing actions 48 % Resolving call target names for dynamic code vtune: Executing actions 49 % Resolving call target names for dynamic code vtune: Executing actions 49 % Resolving interrupt name information vtune: Executing actions 53 % Resolving interrupt name information vtune: Executing actions 53 % Processing profile metrics and debug information vtune: Executing actions 54 % Processing profile metrics and debug information vtune: Executing actions 55 % Processing profile metrics and debug information vtune: Executing actions 56 % Processing profile metrics and debug information vtune: Executing actions 57 % Processing profile metrics and debug information vtune: Executing actions 58 % Processing profile metrics and debug information vtune: Executing actions 60 % Processing profile metrics and debug information vtune: Executing actions 60 % Setting data model parameters vtune: Executing actions 60 % Precomputing frequently used data vtune: Executing actions 60 % Precomputing frequently used data vtune: Executing actions 62 % Precomputing frequently used data vtune: Executing actions 63 % Precomputing frequently used data vtune: Executing actions 64 % Precomputing frequently used data vtune: Executing actions 65 % Precomputing frequently used data vtune: Executing actions 66 % Precomputing frequently used data vtune: Executing actions 67 % Precomputing frequently used data vtune: Executing actions 68 % Precomputing frequently used data vtune: Executing actions 69 % Precomputing frequently used data vtune: Executing actions 71 % Precomputing frequently used data vtune: Executing actions 72 % Precomputing frequently used data vtune: Executing actions 72 % Updating precomputed scalar metrics vtune: Executing actions 75 % Updating precomputed scalar metrics vtune: Executing actions 75 % Discarding redundant overtime data vtune: Executing actions 78 % Discarding redundant overtime data vtune: Executing actions 78 % Saving the result vtune: Executing actions 82 % Saving the result vtune: Executing actions 85 % Saving the result vtune: Executing actions 99 % Saving the result vtune: Executing actions 100 % Saving the result vtune: Executing actions 100 % done Finalization: Ok -------------------------------------------------------------------------------- Command line: /opt/intel/oneapi/vtune/2023.0.0/bin64/vtune -limit 5 -format csv -csv-delimiter comma -report hotspots -group-by function -r /tmp/vtune-tmp-resteam/self-checker-2024.05.07_09.00.52/result_ah Stdout: Function,CPU Time,CPU Time:Effective Time,CPU Time:Spin Time,CPU Time:Overhead Time,Instructions Retired,Microarchitecture Usage(%),Microarchitecture Usage:Microarchitecture Usage(%),Microarchitecture Usage:CPI Rate,Module,Function (Full),Source File,Start Address multiply1,27.245012,27.245012,0.0,0.0,64980000000,17.7,17.7,1.301385,matrix,multiply1,multiply.c,0x401550 [Outside any known module],0.030000,0.030000,0.0,0.0,0,41.7,41.7,,[Unknown],[Outside any known module],[Unknown],0x0 init_arr,0.0,0.0,0.0,0.0,12000000,0.0,0.0,1.000000,matrix,init_arr,matrix.c,0x400d4f init_arr,0.0,0.0,0.0,0.0,12000000,0.0,0.0,0.0,matrix,init_arr,matrix.c,0x400bab Stderr: vtune: Using result path `/tmp/vtune-tmp-resteam/self-checker-2024.05.07_09.00.52/result_ah' vtune: Executing actions 0 % vtune: Executing actions 0 % Finalizing results vtune: Executing actions 50 % Finalizing results vtune: Executing actions 50 % Generating a report vtune: Executing actions 50 % Setting data model parameters vtune: Executing actions 75 % Setting data model parameters vtune: Executing actions 75 % Generating a report vtune: Executing actions 100 % Generating a report vtune: Executing actions 100 % done Report: Ok ================================================================================ HW event-based analysis check... Command line: /opt/intel/oneapi/vtune/2023.0.0/bin64/vtune -collect uarch-exploration -r /tmp/vtune-tmp-resteam/self-checker-2024.05.07_09.00.52/result_ge -data-limit 0 -finalization-mode none -source-search-dir /opt/intel/oneapi/vtune/2023.0.0/samples/en/C++/matrix/src -- /opt/intel/oneapi/vtune/2023.0.0/samples/en/C++/matrix/matrix Stderr: vtune: Error: This analysis requires one of these actions: a) Install Intel Sampling Drivers. b) Configure driverless collection with Perf system-wide profiling. To enable Perf system-wide profiling, set /proc/sys/kernel/perf_event_paranoid to 0 or set up Perf tool capabilities. vtune: Warning: Access to /proc/kallsyms file is limited. Consider changing /proc/sys/kernel/kptr_restrict to 0 to enable resolution of OS kernel and kernel module symbols. Cannot find 'runsa.options' by path: /tmp/vtune-tmp-resteam/self-checker-2024.05.07_09.00.52/result_ge/config/runsa.options HW event-based analysis check Example of analysis types: Microarchitecture Exploration Collection: Fail vtune: Error: This analysis requires one of these actions: a) Install Intel Sampling Drivers. b) Configure driverless collection with Perf system-wide profiling. To enable Perf system-wide profiling, set /proc/sys/kernel/perf_event_paranoid to 0 or set up Perf tool capabilities. vtune: Warning: Access to /proc/kallsyms file is limited. Consider changing /proc/sys/kernel/kptr_restrict to 0 to enable resolution of OS kernel and kernel module symbols. ================================================================================ HW event-based analysis with uncore events... Command line: /opt/intel/oneapi/vtune/2023.0.0/bin64/vtune -collect memory-access -r /tmp/vtune-tmp-resteam/self-checker-2024.05.07_09.00.52/result_ma -data-limit 0 -finalization-mode none -source-search-dir /opt/intel/oneapi/vtune/2023.0.0/samples/en/C++/matrix/src -- /opt/intel/oneapi/vtune/2023.0.0/samples/en/C++/matrix/matrix Stderr: vtune: Error: Cannot collect memory bandwidth data. Make sure the sampling driver is installed and enabled on your system. See the Sampling Drivers help topic for more details. Note that memory bandwidth collection is not possible if you are profiling inside a virtualized environment. Cannot find 'runsa.options' by path: /tmp/vtune-tmp-resteam/self-checker-2024.05.07_09.00.52/result_ma/config/runsa.options HW event-based analysis with uncore events Example of analysis types: Memory Access Collection: Fail vtune: Error: Cannot collect memory bandwidth data. Make sure the sampling driver is installed and enabled on your system. See the Sampling Drivers help topic for more details. Note that memory bandwidth collection is not possible if you are profiling inside a virtualized environment. ================================================================================ HW event-based analysis with stacks... Command line: /opt/intel/oneapi/vtune/2023.0.0/bin64/vtune -collect hotspots -knob sampling-mode=hw -knob enable-stack-collection=true -r /tmp/vtune-tmp-resteam/self-checker-2024.05.07_09.00.52/result_ah_with_stacks -data-limit 0 -finalization-mode none -source-search-dir /opt/intel/oneapi/vtune/2023.0.0/samples/en/C++/matrix/src -- /opt/intel/oneapi/vtune/2023.0.0/samples/en/C++/matrix/matrix Stdout: Addr of buf1 = 0x7f32e2a6b010 Offs of buf1 = 0x7f32e2a6b180 Addr of buf2 = 0x7f32e0a6a010 Offs of buf2 = 0x7f32e0a6a1c0 Addr of buf3 = 0x7f32dea69010 Offs of buf3 = 0x7f32dea69100 Addr of buf4 = 0x7f32dca68010 Offs of buf4 = 0x7f32dca68140 Threads #: 16 Pthreads Matrix size: 2048 Using multiply kernel: multiply1 Execution time = 2.032 seconds Stderr: vtune: Warning: Access to /proc/kallsyms file is limited. Consider changing /proc/sys/kernel/kptr_restrict to 0 to enable resolution of OS kernel and kernel module symbols. vtune: Warning: To analyze modules at the kernel level in this configuration, load the Intel sampling driver and set an unlimited (0) value for the Stack size option (if you require stack collection). Alternatively, enable access to kernel-mode monitoring by setting the /proc/sys/kernel/perf_event_paranoid value to 1 or lower. vtune: Warning: To profile kernel modules during the session, make sure they are available in the /lib/modules/kernel_version/ location. vtune: Warning: Consider reducing possible collection overhead by setting the /proc/sys/kernel/perf_event_paranoid value to 0 (or less). vtune: Collection started. To stop the collection, either press CTRL-C or enter from another console window: vtune -r /tmp/vtune-tmp-resteam/self-checker-2024.05.07_09.00.52/result_ah_with_stacks -command stop. vtune: Collection stopped. vtune: Using result path `/tmp/vtune-tmp-resteam/self-checker-2024.05.07_09.00.52/result_ah_with_stacks' vtune: Executing actions 0 % vtune: Executing actions 100 % vtune: Executing actions 100 % done HW event-based analysis with stacks (Perf) Example of analysis types: Hotspots with HW event-based sampling and call stacks Collection: Ok vtune: Warning: Access to /proc/kallsyms file is limited. Consider changing /proc/sys/kernel/kptr_restrict to 0 to enable resolution of OS kernel and kernel module symbols. vtune: Warning: To analyze modules at the kernel level in this configuration, load the Intel sampling driver and set an unlimited (0) value for the Stack size option (if you require stack collection). Alternatively, enable access to kernel-mode monitoring by setting the /proc/sys/kernel/perf_event_paranoid value to 1 or lower. vtune: Warning: To profile kernel modules during the session, make sure they are available in the /lib/modules/kernel_version/ location. vtune: Warning: Consider reducing possible collection overhead by setting the /proc/sys/kernel/perf_event_paranoid value to 0 (or less). -------------------------------------------------------------------------------- Running finalization... Command line: /opt/intel/oneapi/vtune/2023.0.0/bin64/vtune -finalize -r /tmp/vtune-tmp-resteam/self-checker-2024.05.07_09.00.52/result_ah_with_stacks Stderr: vtune: Using result path `/tmp/vtune-tmp-resteam/self-checker-2024.05.07_09.00.52/result_ah_with_stacks' vtune: Executing actions 0 % vtune: Executing actions 0 % Finalizing results vtune: Executing actions 0 % Finalizing the result vtune: Executing actions 0 % Clearing the database vtune: Executing actions 14 % Clearing the database vtune: Executing actions 14 % Loading raw data to the database vtune: Executing actions 14 % Loading 'systemcollector-918936-srv-sopt1.sc' fil vtune: Executing actions 25 % Loading 'systemcollector-918936-srv-sopt1.sc' fil vtune: Executing actions 25 % Loading '918946.perf' file vtune: Executing actions 25 % Updating precomputed scalar metrics vtune: Executing actions 28 % Updating precomputed scalar metrics vtune: Executing actions 28 % Processing profile metrics and debug information vtune: Executing actions 39 % Processing profile metrics and debug information vtune: Executing actions 39 % Setting data model parameters vtune: Executing actions 39 % Resolving module symbols vtune: Executing actions 39 % Resolving information for dangling locations vtune: Executing actions 39 % Resolving information for `libpthread-2.28.so' vtune: Executing actions 39 % Resolving information for `matrix' vtune: Warning: Cannot locate debugging information for file `/usr/lib64/libpthread-2.28.so'. vtune: Executing actions 40 % Resolving information for `matrix' vtune: Executing actions 40 % Resolving information for `libc-2.28.so' vtune: Executing actions 42 % Resolving information for `libc-2.28.so' vtune: Executing actions 43 % Resolving information for `libc-2.28.so' vtune: Warning: Cannot locate debugging information for file `/usr/lib64/libc-2.28.so'. vtune: Executing actions 45 % Resolving information for `libc-2.28.so' vtune: Executing actions 45 % Resolving bottom user stack information vtune: Executing actions 46 % Resolving bottom user stack information vtune: Executing actions 46 % Resolving thread name information vtune: Executing actions 47 % Resolving thread name information vtune: Executing actions 48 % Resolving thread name information vtune: Executing actions 48 % Resolving call target names for dynamic code vtune: Executing actions 50 % Resolving call target names for dynamic code vtune: Executing actions 50 % Resolving interrupt name information vtune: Executing actions 53 % Resolving interrupt name information vtune: Executing actions 53 % Processing profile metrics and debug information vtune: Executing actions 54 % Processing profile metrics and debug information vtune: Executing actions 55 % Processing profile metrics and debug information vtune: Executing actions 56 % Processing profile metrics and debug information vtune: Executing actions 57 % Processing profile metrics and debug information vtune: Executing actions 58 % Processing profile metrics and debug information vtune: Executing actions 60 % Processing profile metrics and debug information vtune: Executing actions 60 % Setting data model parameters vtune: Executing actions 60 % Precomputing frequently used data vtune: Executing actions 60 % Precomputing frequently used data vtune: Executing actions 62 % Precomputing frequently used data vtune: Executing actions 63 % Precomputing frequently used data vtune: Executing actions 64 % Precomputing frequently used data vtune: Executing actions 65 % Precomputing frequently used data vtune: Executing actions 66 % Precomputing frequently used data vtune: Executing actions 67 % Precomputing frequently used data vtune: Executing actions 68 % Precomputing frequently used data vtune: Executing actions 69 % Precomputing frequently used data vtune: Executing actions 71 % Precomputing frequently used data vtune: Executing actions 72 % Precomputing frequently used data vtune: Executing actions 72 % Updating precomputed scalar metrics vtune: Executing actions 75 % Updating precomputed scalar metrics vtune: Executing actions 75 % Discarding redundant overtime data vtune: Executing actions 78 % Discarding redundant overtime data vtune: Executing actions 78 % Saving the result vtune: Executing actions 82 % Saving the result vtune: Executing actions 85 % Saving the result vtune: Executing actions 99 % Saving the result vtune: Executing actions 100 % Saving the result vtune: Executing actions 100 % done Finalization: Ok -------------------------------------------------------------------------------- Command line: /opt/intel/oneapi/vtune/2023.0.0/bin64/vtune -limit 5 -format csv -csv-delimiter comma -report hotspots -group-by function -r /tmp/vtune-tmp-resteam/self-checker-2024.05.07_09.00.52/result_ah_with_stacks Stdout: Function,CPU Time,CPU Time:Effective Time,CPU Time:Spin Time,CPU Time:Overhead Time,Instructions Retired,Microarchitecture Usage(%),Microarchitecture Usage:Microarchitecture Usage(%),Microarchitecture Usage:CPI Rate,Module,Function (Full),Source File,Start Address multiply1,30.625014,30.625014,0.0,0.0,68280000000,16.6,16.6,1.391213,matrix,multiply1,multiply.c,0x401550 [Outside any known module],0.025000,0.025000,0.0,0.0,0,0.0,0.0,,[Unknown],[Outside any known module],[Unknown],0x0 init_arr,0.0,0.0,0.0,0.0,12000000,0.0,0.0,0.0,matrix,init_arr,matrix.c,0x400bab init_arr,0.0,0.0,0.0,0.0,12000000,0.0,0.0,1.000000,matrix,init_arr,matrix.c,0x400d4f Stderr: vtune: Using result path `/tmp/vtune-tmp-resteam/self-checker-2024.05.07_09.00.52/result_ah_with_stacks' vtune: Executing actions 0 % vtune: Executing actions 0 % Finalizing results vtune: Executing actions 50 % Finalizing results vtune: Executing actions 50 % Generating a report vtune: Executing actions 50 % Setting data model parameters vtune: Executing actions 75 % Setting data model parameters vtune: Executing actions 75 % Generating a report vtune: Executing actions 100 % Generating a report vtune: Executing actions 100 % done Report: Ok ================================================================================ HW event-based analysis with context switches... Command line: /opt/intel/oneapi/vtune/2023.0.0/bin64/vtune -collect threading -knob sampling-and-waits=hw -knob enable-stack-collection=false -r /tmp/vtune-tmp-resteam/self-checker-2024.05.07_09.00.52/result_th -data-limit 0 -finalization-mode none -source-search-dir /opt/intel/oneapi/vtune/2023.0.0/samples/en/C++/matrix/src -- /opt/intel/oneapi/vtune/2023.0.0/samples/en/C++/matrix/matrix Stderr: vtune: Warning: Access to /proc/kallsyms file is limited. Consider changing /proc/sys/kernel/kptr_restrict to 0 to enable resolution of OS kernel and kernel module symbols. vtune: Warning: To analyze modules at the kernel level in this configuration, load the Intel sampling driver and set an unlimited (0) value for the Stack size option (if you require stack collection). Alternatively, enable access to kernel-mode monitoring by setting the /proc/sys/kernel/perf_event_paranoid value to 1 or lower. vtune: Warning: Cannot collect context switch data in Perf-based driverless mode if the value of /proc/sys/kernel/perf_event_paranoid is greater than 1. vtune: Warning: To profile kernel modules during the session, make sure they are available in the /lib/modules/kernel_version/ location. vtune: Warning: Consider reducing possible collection overhead by setting the /proc/sys/kernel/perf_event_paranoid value to 0 (or less). vtune: Error: amplxe-perf: threads_spec: cpu Using CPUID GenuineIntel-6-55-7 DEBUGINFOD_URLS= Compression enabled, disabling build id collection at the end of the session. nr_threads: 96 thread_masks[0]: 0x55c908a79d00: maps mask[96]: 0 thread_masks[0]: 0x55c908a79d10: affinity mask[96]: 0 thread_masks[1]: 0x55c908a79d20: maps mask[96]: 1 thread_masks[1]: 0x55c908a79d30: affinity mask[96]: 1 thread_masks[2]: 0x55c908a79d40: maps mask[96]: 2 thread_masks[2]: 0x55c908a79d50: affinity mask[96]: 2 thread_masks[3]: 0x55c908a79d60: maps mask[96]: 3 thread_masks[3]: 0x55c908a79d70: affinity mask[96]: 3 thread_masks[4]: 0x55c908a79d80: maps mask[96]: 4 thread_masks[4]: 0x55c908a79d90: affinity mask[96]: 4 thread_masks[5]: 0x55c908a79da0: maps mask[96]: 5 thread_masks[5]: 0x55c908a79db0: affinity mask[96]: 5 thread_masks[6]: 0x55c908a79dc0: maps mask[96]: 6 thread_masks[6]: 0x55c908a79dd0: affinity mask[96]: 6 thread_masks[7]: 0x55c908a79de0: maps mask[96]: 7 thread_masks[7]: 0x55c908a79df0: affinity mask[96]: 7 thread_masks[8]: 0x55c908a79e00: maps mask[96]: 8 thread_masks[8]: 0x55c908a79e10: affinity mask[96]: 8 thread_masks[9]: 0x55c908a79e20: maps mask[96]: 9 thread_masks[9]: 0x55c908a79e30: affinity mask[96]: 9 thread_masks[10]: 0x55c908a79e40: maps mask[96]: 10 thread_masks[10]: 0x55c908a79e50: affinity mask[96]: 10 thread_masks[11]: 0x55c908a79e60: maps mask[96]: 11 thread_masks[11]: 0x55c908a79e70: affinity mask[96]: 11 thread_masks[12]: 0x55c908a79e80: maps mask[96]: 12 thread_masks[12]: 0x55c908a79e90: affinity mask[96]: 12 thread_masks[13]: 0x55c908a79ea0: maps mask[96]: 13 thread_masks[13]: 0x55c908a79eb0: affinity mask[96]: 13 thread_masks[14]: 0x55c908a79ec0: maps mask[96]: 14 thread_masks[14]: 0x55c908a79ed0: affinity mask[96]: 14 thread_masks[15]: 0x55c908a79ee0: maps mask[96]: 15 thread_masks[15]: 0x55c908a79ef0: affinity mask[96]: 15 thread_masks[16]: 0x55c908a79f00: maps mask[96]: 16 thread_masks[16]: 0x55c908a79f10: affinity mask[96]: 16 thread_masks[17]: 0x55c908a79f20: maps mask[96]: 17 thread_masks[17]: 0x55c908a79f30: affinity mask[96]: 17 thread_masks[18]: 0x55c908a79f40: maps mask[96]: 18 thread_masks[18]: 0x55c908a79f50: affinity mask[96]: 18 thread_masks[19]: 0x55c908a79f60: maps mask[96]: 19 thread_masks[19]: 0x55c908a79f70: affinity mask[96]: 19 thread_masks[20]: 0x55c908a79f80: maps mask[96]: 20 thread_masks[20]: 0x55c908a79f90: affinity mask[96]: 20 thread_masks[21]: 0x55c908a79fa0: maps mask[96]: 21 thread_masks[21]: 0x55c908a79fb0: affinity mask[96]: 21 thread_masks[22]: 0x55c908a79fc0: maps mask[96]: 22 thread_masks[22]: 0x55c908a79fd0: affinity mask[96]: 22 thread_masks[23]: 0x55c908a79fe0: maps mask[96]: 23 thread_masks[23]: 0x55c908a79ff0: affinity mask[96]: 23 thread_masks[24]: 0x55c908a7a000: maps mask[96]: 24 thread_masks[24]: 0x55c908a7a010: affinity mask[96]: 24 thread_masks[25]: 0x55c908a7a020: maps mask[96]: 25 thread_masks[25]: 0x55c908a7a030: affinity mask[96]: 25 thread_masks[26]: 0x55c908a7a040: maps mask[96]: 26 thread_masks[26]: 0x55c908a7a050: affinity mask[96]: 26 thread_masks[27]: 0x55c908a7a060: maps mask[96]: 27 thread_masks[27]: 0x55c908a7a070: affinity mask[96]: 27 thread_masks[28]: 0x55c908a7a080: maps mask[96]: 28 thread_masks[28]: 0x55c908a7a090: affinity mask[96]: 28 thread_masks[29]: 0x55c908a7a0a0: maps mask[96]: 29 thread_masks[29]: 0x55c908a7a0b0: affinity mask[96]: 29 thread_masks[30]: 0x55c908a7a0c0: maps mask[96]: 30 thread_masks[30]: 0x55c908a7a0d0: affinity mask[96]: 30 thread_masks[31]: 0x55c908a7a0e0: maps mask[96]: 31 thread_masks[31]: 0x55c908a7a0f0: affinity mask[96]: 31 thread_masks[32]: 0x55c908a7a100: maps mask[96]: 32 thread_masks[32]: 0x55c908a7a110: affinity mask[96]: 32 thread_masks[33]: 0x55c908a7a120: maps mask[96]: 33 thread_masks[33]: 0x55c908a7a130: affinity mask[96]: 33 thread_masks[34]: 0x55c908a7a140: maps mask[96]: 34 thread_masks[34]: 0x55c908a7a150: affinity mask[96]: 34 thread_masks[35]: 0x55c908a7a160: maps mask[96]: 35 thread_masks[35]: 0x55c908a7a170: affinity mask[96]: 35 thread_masks[36]: 0x55c908a7a180: maps mask[96]: 36 thread_masks[36]: 0x55c908a7a190: affinity mask[96]: 36 thread_masks[37]: 0x55c908a7a1a0: maps mask[96]: 37 thread_masks[37]: 0x55c908a7a1b0: affinity mask[96]: 37 thread_masks[38]: 0x55c908a7a1c0: maps mask[96]: 38 thread_masks[38]: 0x55c908a7a1d0: affinity mask[96]: 38 thread_masks[39]: 0x55c908a7a1e0: maps mask[96]: 39 thread_masks[39]: 0x55c908a7a1f0: affinity mask[96]: 39 thread_masks[40]: 0x55c908a7a200: maps mask[96]: 40 thread_masks[40]: 0x55c908a7a210: affinity mask[96]: 40 thread_masks[41]: 0x55c908a7a220: maps mask[96]: 41 thread_masks[41]: 0x55c908a7a230: affinity mask[96]: 41 thread_masks[42]: 0x55c908a7a240: maps mask[96]: 42 thread_masks[42]: 0x55c908a7a250: affinity mask[96]: 42 thread_masks[43]: 0x55c908a7a260: maps mask[96]: 43 thread_masks[43]: 0x55c908a7a270: affinity mask[96]: 43 thread_masks[44]: 0x55c908a7a280: maps mask[96]: 44 thread_masks[44]: 0x55c908a7a290: affinity mask[96]: 44 thread_masks[45]: 0x55c908a7a2a0: maps mask[96]: 45 thread_masks[45]: 0x55c908a7a2b0: affinity mask[96]: 45 thread_masks[46]: 0x55c908a7a2c0: maps mask[96]: 46 thread_masks[46]: 0x55c908a7a2d0: affinity mask[96]: 46 thread_masks[47]: 0x55c908a7a2e0: maps mask[96]: 47 thread_masks[47]: 0x55c908a7a2f0: affinity mask[96]: 47 thread_masks[48]: 0x55c908a7a300: maps mask[96]: 48 thread_masks[48]: 0x55c908a7a310: affinity mask[96]: 48 thread_masks[49]: 0x55c908a7a320: maps mask[96]: 49 thread_masks[49]: 0x55c908a7a330: affinity mask[96]: 49 thread_masks[50]: 0x55c908a7a340: maps mask[96]: 50 thread_masks[50]: 0x55c908a7a350: affinity mask[96]: 50 thread_masks[51]: 0x55c908a7a360: maps mask[96]: 51 thread_masks[51]: 0x55c908a7a370: affinity mask[96]: 51 thread_masks[52]: 0x55c908a7a380: maps mask[96]: 52 thread_masks[52]: 0x55c908a7a390: affinity mask[96]: 52 thread_masks[53]: 0x55c908a7a3a0: maps mask[96]: 53 thread_masks[53]: 0x55c908a7a3b0: affinity mask[96]: 53 thread_masks[54]: 0x55c908a7a3c0: maps mask[96]: 54 thread_masks[54]: 0x55c908a7a3d0: affinity mask[96]: 54 thread_masks[55]: 0x55c908a7a3e0: maps mask[96]: 55 thread_masks[55]: 0x55c908a7a3f0: affinity mask[96]: 55 thread_masks[56]: 0x55c908a7a400: maps mask[96]: 56 thread_masks[56]: 0x55c908a7a410: affinity mask[96]: 56 thread_masks[57]: 0x55c908a7a420: maps mask[96]: 57 thread_masks[57]: 0x55c908a7a430: affinity mask[96]: 57 thread_masks[58]: 0x55c908a7a440: maps mask[96]: 58 thread_masks[58]: 0x55c908a7a450: affinity mask[96]: 58 thread_masks[59]: 0x55c908a7a460: maps mask[96]: 59 thread_masks[59]: 0x55c908a7a470: affinity mask[96]: 59 thread_masks[60]: 0x55c908a7a480: maps mask[96]: 60 thread_masks[60]: 0x55c908a7a490: affinity mask[96]: 60 thread_masks[61]: 0x55c908a7a4a0: maps mask[96]: 61 thread_masks[61]: 0x55c908a7a4b0: affinity mask[96]: 61 thread_masks[62]: 0x55c908a7a4c0: maps mask[96]: 62 thread_masks[62]: 0x55c908a7a4d0: affinity mask[96]: 62 thread_masks[63]: 0x55c908a7a4e0: maps mask[96]: 63 thread_masks[63]: 0x55c908a7a4f0: affinity mask[96]: 63 thread_masks[64]: 0x55c908a7a500: maps mask[96]: 64 thread_masks[64]: 0x55c908a7a510: affinity mask[96]: 64 thread_masks[65]: 0x55c908a7a520: maps mask[96]: 65 thread_masks[65]: 0x55c908a7a530: affinity mask[96]: 65 thread_masks[66]: 0x55c908a7a540: maps mask[96]: 66 thread_masks[66]: 0x55c908a7a550: affinity mask[96]: 66 thread_masks[67]: 0x55c908a7a560: maps mask[96]: 67 thread_masks[67]: 0x55c908a7a570: affinity mask[96]: 67 thread_masks[68]: 0x55c908a7a580: maps mask[96]: 68 thread_masks[68]: 0x55c908a7a590: affinity mask[96]: 68 thread_masks[69]: 0x55c908a7a5a0: maps mask[96]: 69 thread_masks[69]: 0x55c908a7a5b0: affinity mask[96]: 69 thread_masks[70]: 0x55c908a7a5c0: maps mask[96]: 70 thread_masks[70]: 0x55c908a7a5d0: affinity mask[96]: 70 thread_masks[71]: 0x55c908a7a5e0: maps mask[96]: 71 thread_masks[71]: 0x55c908a7a5f0: affinity mask[96]: 71 thread_masks[72]: 0x55c908a7a600: maps mask[96]: 72 thread_masks[72]: 0x55c908a7a610: affinity mask[96]: 72 thread_masks[73]: 0x55c908a7a620: maps mask[96]: 73 thread_masks[73]: 0x55c908a7a630: affinity mask[96]: 73 thread_masks[74]: 0x55c908a7a640: maps mask[96]: 74 thread_masks[74]: 0x55c908a7a650: affinity mask[96]: 74 thread_masks[75]: 0x55c908a7a660: maps mask[96]: 75 thread_masks[75]: 0x55c908a7a670: affinity mask[96]: 75 thread_masks[76]: 0x55c908a7a680: maps mask[96]: 76 thread_masks[76]: 0x55c908a7a690: affinity mask[96]: 76 thread_masks[77]: 0x55c908a7a6a0: maps mask[96]: 77 thread_masks[77]: 0x55c908a7a6b0: affinity mask[96]: 77 thread_masks[78]: 0x55c908a7a6c0: maps mask[96]: 78 thread_masks[78]: 0x55c908a7a6d0: affinity mask[96]: 78 thread_masks[79]: 0x55c908a7a6e0: maps mask[96]: 79 thread_masks[79]: 0x55c908a7a6f0: affinity mask[96]: 79 thread_masks[80]: 0x55c908a7a700: maps mask[96]: 80 thread_masks[80]: 0x55c908a7a710: affinity mask[96]: 80 thread_masks[81]: 0x55c908a7a720: maps mask[96]: 81 thread_masks[81]: 0x55c908a7a730: affinity mask[96]: 81 thread_masks[82]: 0x55c908a7a740: maps mask[96]: 82 thread_masks[82]: 0x55c908a7a750: affinity mask[96]: 82 thread_masks[83]: 0x55c908a7a760: maps mask[96]: 83 thread_masks[83]: 0x55c908a7a770: affinity mask[96]: 83 thread_masks[84]: 0x55c908a7a780: maps mask[96]: 84 thread_masks[84]: 0x55c908a7a790: affinity mask[96]: 84 thread_masks[85]: 0x55c908a7a7a0: maps mask[96]: 85 thread_masks[85]: 0x55c908a7a7b0: affinity mask[96]: 85 thread_masks[86]: 0x55c908a7a7c0: maps mask[96]: 86 thread_masks[86]: 0x55c908a7a7d0: affinity mask[96]: 86 thread_masks[87]: 0x55c908a7a7e0: maps mask[96]: 87 thread_masks[87]: 0x55c908a7a7f0: affinity mask[96]: 87 thread_masks[88]: 0x55c908a7a800: maps mask[96]: 88 thread_masks[88]: 0x55c908a7a810: affinity mask[96]: 88 thread_masks[89]: 0x55c908a7a820: maps mask[96]: 89 thread_masks[89]: 0x55c908a7a830: affinity mask[96]: 89 thread_masks[90]: 0x55c908a7a840: maps mask[96]: 90 thread_masks[90]: 0x55c908a7a850: affinity mask[96]: 90 thread_masks[91]: 0x55c908a7a860: maps mask[96]: 91 thread_masks[91]: 0x55c908a7a870: affinity mask[96]: 91 thread_masks[92]: 0x55c908a7a880: maps mask[96]: 92 thread_masks[92]: 0x55c908a7a890: affinity mask[96]: 92 thread_masks[93]: 0x55c908a7a8a0: maps mask[96]: 93 thread_masks[93]: 0x55c908a7a8b0: affinity mask[96]: 93 thread_masks[94]: 0x55c908a7a8c0: maps mask[96]: 94 thread_masks[94]: 0x55c908a7a8d0: affinity mask[96]: 94 thread_masks[95]: 0x55c908a7a8e0: maps mask[96]: 95 thread_masks[95]: 0x55c908a7a8f0: affinity mask[96]: 95 nr_cblocks: 0 affinity: SYS mmap flush: 1 comp level: 1 Problems creating module maps, continuing anyway... mmap size 528384B Failed to open thread[85] communication pipes vtune: Collection failed. vtune: Internal Error Cannot find 'runsa.options' by path: /tmp/vtune-tmp-resteam/self-checker-2024.05.07_09.00.52/result_th/config/runsa.options HW event-based analysis with context switches Example of analysis types: Threading with HW event-based sampling Collection: Fail vtune: Warning: Access to /proc/kallsyms file is limited. Consider changing /proc/sys/kernel/kptr_restrict to 0 to enable resolution of OS kernel and kernel module symbols. vtune: Warning: To analyze modules at the kernel level in this configuration, load the Intel sampling driver and set an unlimited (0) value for the Stack size option (if you require stack collection). Alternatively, enable access to kernel-mode monitoring by setting the /proc/sys/kernel/perf_event_paranoid value to 1 or lower. vtune: Warning: Cannot collect context switch data in Perf-based driverless mode if the value of /proc/sys/kernel/perf_event_paranoid is greater than 1. vtune: Warning: To profile kernel modules during the session, make sure they are available in the /lib/modules/kernel_version/ location. vtune: Warning: Consider reducing possible collection overhead by setting the /proc/sys/kernel/perf_event_paranoid value to 0 (or less). vtune: Error: amplxe-perf: threads_spec: cpu Using CPUID GenuineIntel-6-55-7 DEBUGINFOD_URLS= Compression enabled, disabling build id collection at the end of the session. nr_threads: 96 thread_masks[0]: 0x55c908a79d00: maps mask[96]: 0 thread_masks[0]: 0x55c908a79d10: affinity mask[96]: 0 thread_masks[1]: 0x55c908a79d20: maps mask[96]: 1 thread_masks[1]: 0x55c908a79d30: affinity mask[96]: 1 thread_masks[2]: 0x55c908a79d40: maps mask[96]: 2 thread_masks[2]: 0x55c908a79d50: affinity mask[96]: 2 thread_masks[3]: 0x55c908a79d60: maps mask[96]: 3 thread_masks[3]: 0x55c908a79d70: affinity mask[96]: 3 thread_masks[4]: 0x55c908a79d80: maps mask[96]: 4 thread_masks[4]: 0x55c908a79d90: affinity mask[96]: 4 thread_masks[5]: 0x55c908a79da0: maps mask[96]: 5 thread_masks[5]: 0x55c908a79db0: affinity mask[96]: 5 thread_masks[6]: 0x55c908a79dc0: maps mask[96]: 6 thread_masks[6]: 0x55c908a79dd0: affinity mask[96]: 6 thread_masks[7]: 0x55c908a79de0: maps mask[96]: 7 thread_masks[7]: 0x55c908a79df0: affinity mask[96]: 7 thread_masks[8]: 0x55c908a79e00: maps mask[96]: 8 thread_masks[8]: 0x55c908a79e10: affinity mask[96]: 8 thread_masks[9]: 0x55c908a79e20: maps mask[96]: 9 thread_masks[9]: 0x55c908a79e30: affinity mask[96]: 9 thread_masks[10]: 0x55c908a79e40: maps mask[96]: 10 thread_masks[10]: 0x55c908a79e50: affinity mask[96]: 10 thread_masks[11]: 0x55c908a79e60: maps mask[96]: 11 thread_masks[11]: 0x55c908a79e70: affinity mask[96]: 11 thread_masks[12]: 0x55c908a79e80: maps mask[96]: 12 thread_masks[12]: 0x55c908a79e90: affinity mask[96]: 12 thread_masks[13]: 0x55c908a79ea0: maps mask[96]: 13 thread_masks[13]: 0x55c908a79eb0: affinity mask[96]: 13 thread_masks[14]: 0x55c908a79ec0: maps mask[96]: 14 thread_masks[14]: 0x55c908a79ed0: affinity mask[96]: 14 thread_masks[15]: 0x55c908a79ee0: maps mask[96]: 15 thread_masks[15]: 0x55c908a79ef0: affinity mask[96]: 15 thread_masks[16]: 0x55c908a79f00: maps mask[96]: 16 thread_masks[16]: 0x55c908a79f10: affinity mask[96]: 16 thread_masks[17]: 0x55c908a79f20: maps mask[96]: 17 thread_masks[17]: 0x55c908a79f30: affinity mask[96]: 17 thread_masks[18]: 0x55c908a79f40: maps mask[96]: 18 thread_masks[18]: 0x55c908a79f50: affinity mask[96]: 18 thread_masks[19]: 0x55c908a79f60: maps mask[96]: 19 thread_masks[19]: 0x55c908a79f70: affinity mask[96]: 19 thread_masks[20]: 0x55c908a79f80: maps mask[96]: 20 thread_masks[20]: 0x55c908a79f90: affinity mask[96]: 20 thread_masks[21]: 0x55c908a79fa0: maps mask[96]: 21 thread_masks[21]: 0x55c908a79fb0: affinity mask[96]: 21 thread_masks[22]: 0x55c908a79fc0: maps mask[96]: 22 thread_masks[22]: 0x55c908a79fd0: affinity mask[96]: 22 thread_masks[23]: 0x55c908a79fe0: maps mask[96]: 23 thread_masks[23]: 0x55c908a79ff0: affinity mask[96]: 23 thread_masks[24]: 0x55c908a7a000: maps mask[96]: 24 thread_masks[24]: 0x55c908a7a010: affinity mask[96]: 24 thread_masks[25]: 0x55c908a7a020: maps mask[96]: 25 thread_masks[25]: 0x55c908a7a030: affinity mask[96]: 25 thread_masks[26]: 0x55c908a7a040: maps mask[96]: 26 thread_masks[26]: 0x55c908a7a050: affinity mask[96]: 26 thread_masks[27]: 0x55c908a7a060: maps mask[96]: 27 thread_masks[27]: 0x55c908a7a070: affinity mask[96]: 27 thread_masks[28]: 0x55c908a7a080: maps mask[96]: 28 thread_masks[28]: 0x55c908a7a090: affinity mask[96]: 28 thread_masks[29]: 0x55c908a7a0a0: maps mask[96]: 29 thread_masks[29]: 0x55c908a7a0b0: affinity mask[96]: 29 thread_masks[30]: 0x55c908a7a0c0: maps mask[96]: 30 thread_masks[30]: 0x55c908a7a0d0: affinity mask[96]: 30 thread_masks[31]: 0x55c908a7a0e0: maps mask[96]: 31 thread_masks[31]: 0x55c908a7a0f0: affinity mask[96]: 31 thread_masks[32]: 0x55c908a7a100: maps mask[96]: 32 thread_masks[32]: 0x55c908a7a110: affinity mask[96]: 32 thread_masks[33]: 0x55c908a7a120: maps mask[96]: 33 thread_masks[33]: 0x55c908a7a130: affinity mask[96]: 33 thread_masks[34]: 0x55c908a7a140: maps mask[96]: 34 thread_masks[34]: 0x55c908a7a150: affinity mask[96]: 34 thread_masks[35]: 0x55c908a7a160: maps mask[96]: 35 thread_masks[35]: 0x55c908a7a170: affinity mask[96]: 35 thread_masks[36]: 0x55c908a7a180: maps mask[96]: 36 thread_masks[36]: 0x55c908a7a190: affinity mask[96]: 36 thread_masks[37]: 0x55c908a7a1a0: maps mask[96]: 37 thread_masks[37]: 0x55c908a7a1b0: affinity mask[96]: 37 thread_masks[38]: 0x55c908a7a1c0: maps mask[96]: 38 thread_masks[38]: 0x55c908a7a1d0: affinity mask[96]: 38 thread_masks[39]: 0x55c908a7a1e0: maps mask[96]: 39 thread_masks[39]: 0x55c908a7a1f0: affinity mask[96]: 39 thread_masks[40]: 0x55c908a7a200: maps mask[96]: 40 thread_masks[40]: 0x55c908a7a210: affinity mask[96]: 40 thread_masks[41]: 0x55c908a7a220: maps mask[96]: 41 thread_masks[41]: 0x55c908a7a230: affinity mask[96]: 41 thread_masks[42]: 0x55c908a7a240: maps mask[96]: 42 thread_masks[42]: 0x55c908a7a250: affinity mask[96]: 42 thread_masks[43]: 0x55c908a7a260: maps mask[96]: 43 thread_masks[43]: 0x55c908a7a270: affinity mask[96]: 43 thread_masks[44]: 0x55c908a7a280: maps mask[96]: 44 thread_masks[44]: 0x55c908a7a290: affinity mask[96]: 44 thread_masks[45]: 0x55c908a7a2a0: maps mask[96]: 45 thread_masks[45]: 0x55c908a7a2b0: affinity mask[96]: 45 thread_masks[46]: 0x55c908a7a2c0: maps mask[96]: 46 thread_masks[46]: 0x55c908a7a2d0: affinity mask[96]: 46 thread_masks[47]: 0x55c908a7a2e0: maps mask[96]: 47 thread_masks[47]: 0x55c908a7a2f0: affinity mask[96]: 47 thread_masks[48]: 0x55c908a7a300: maps mask[96]: 48 thread_masks[48]: 0x55c908a7a310: affinity mask[96]: 48 thread_masks[49]: 0x55c908a7a320: maps mask[96]: 49 thread_masks[49]: 0x55c908a7a330: affinity mask[96]: 49 thread_masks[50]: 0x55c908a7a340: maps mask[96]: 50 thread_masks[50]: 0x55c908a7a350: affinity mask[96]: 50 thread_masks[51]: 0x55c908a7a360: maps mask[96]: 51 thread_masks[51]: 0x55c908a7a370: affinity mask[96]: 51 thread_masks[52]: 0x55c908a7a380: maps mask[96]: 52 thread_masks[52]: 0x55c908a7a390: affinity mask[96]: 52 thread_masks[53]: 0x55c908a7a3a0: maps mask[96]: 53 thread_masks[53]: 0x55c908a7a3b0: affinity mask[96]: 53 thread_masks[54]: 0x55c908a7a3c0: maps mask[96]: 54 thread_masks[54]: 0x55c908a7a3d0: affinity mask[96]: 54 thread_masks[55]: 0x55c908a7a3e0: maps mask[96]: 55 thread_masks[55]: 0x55c908a7a3f0: affinity mask[96]: 55 thread_masks[56]: 0x55c908a7a400: maps mask[96]: 56 thread_masks[56]: 0x55c908a7a410: affinity mask[96]: 56 thread_masks[57]: 0x55c908a7a420: maps mask[96]: 57 thread_masks[57]: 0x55c908a7a430: affinity mask[96]: 57 thread_masks[58]: 0x55c908a7a440: maps mask[96]: 58 thread_masks[58]: 0x55c908a7a450: affinity mask[96]: 58 thread_masks[59]: 0x55c908a7a460: maps mask[96]: 59 thread_masks[59]: 0x55c908a7a470: affinity mask[96]: 59 thread_masks[60]: 0x55c908a7a480: maps mask[96]: 60 thread_masks[60]: 0x55c908a7a490: affinity mask[96]: 60 thread_masks[61]: 0x55c908a7a4a0: maps mask[96]: 61 thread_masks[61]: 0x55c908a7a4b0: affinity mask[96]: 61 thread_masks[62]: 0x55c908a7a4c0: maps mask[96]: 62 thread_masks[62]: 0x55c908a7a4d0: affinity mask[96]: 62 thread_masks[63]: 0x55c908a7a4e0: maps mask[96]: 63 thread_masks[63]: 0x55c908a7a4f0: affinity mask[96]: 63 thread_masks[64]: 0x55c908a7a500: maps mask[96]: 64 thread_masks[64]: 0x55c908a7a510: affinity mask[96]: 64 thread_masks[65]: 0x55c908a7a520: maps mask[96]: 65 thread_masks[65]: 0x55c908a7a530: affinity mask[96]: 65 thread_masks[66]: 0x55c908a7a540: maps mask[96]: 66 thread_masks[66]: 0x55c908a7a550: affinity mask[96]: 66 thread_masks[67]: 0x55c908a7a560: maps mask[96]: 67 thread_masks[67]: 0x55c908a7a570: affinity mask[96]: 67 thread_masks[68]: 0x55c908a7a580: maps mask[96]: 68 thread_masks[68]: 0x55c908a7a590: affinity mask[96]: 68 thread_masks[69]: 0x55c908a7a5a0: maps mask[96]: 69 thread_masks[69]: 0x55c908a7a5b0: affinity mask[96]: 69 thread_masks[70]: 0x55c908a7a5c0: maps mask[96]: 70 thread_masks[70]: 0x55c908a7a5d0: affinity mask[96]: 70 thread_masks[71]: 0x55c908a7a5e0: maps mask[96]: 71 thread_masks[71]: 0x55c908a7a5f0: affinity mask[96]: 71 thread_masks[72]: 0x55c908a7a600: maps mask[96]: 72 thread_masks[72]: 0x55c908a7a610: affinity mask[96]: 72 thread_masks[73]: 0x55c908a7a620: maps mask[96]: 73 thread_masks[73]: 0x55c908a7a630: affinity mask[96]: 73 thread_masks[74]: 0x55c908a7a640: maps mask[96]: 74 thread_masks[74]: 0x55c908a7a650: affinity mask[96]: 74 thread_masks[75]: 0x55c908a7a660: maps mask[96]: 75 thread_masks[75]: 0x55c908a7a670: affinity mask[96]: 75 thread_masks[76]: 0x55c908a7a680: maps mask[96]: 76 thread_masks[76]: 0x55c908a7a690: affinity mask[96]: 76 thread_masks[77]: 0x55c908a7a6a0: maps mask[96]: 77 thread_masks[77]: 0x55c908a7a6b0: affinity mask[96]: 77 thread_masks[78]: 0x55c908a7a6c0: maps mask[96]: 78 thread_masks[78]: 0x55c908a7a6d0: affinity mask[96]: 78 thread_masks[79]: 0x55c908a7a6e0: maps mask[96]: 79 thread_masks[79]: 0x55c908a7a6f0: affinity mask[96]: 79 thread_masks[80]: 0x55c908a7a700: maps mask[96]: 80 thread_masks[80]: 0x55c908a7a710: affinity mask[96]: 80 thread_masks[81]: 0x55c908a7a720: maps mask[96]: 81 thread_masks[81]: 0x55c908a7a730: affinity mask[96]: 81 thread_masks[82]: 0x55c908a7a740: maps mask[96]: 82 thread_masks[82]: 0x55c908a7a750: affinity mask[96]: 82 thread_masks[83]: 0x55c908a7a760: maps mask[96]: 83 thread_masks[83]: 0x55c908a7a770: affinity mask[96]: 83 thread_masks[84]: 0x55c908a7a780: maps mask[96]: 84 thread_masks[84]: 0x55c908a7a790: affinity mask[96]: 84 thread_masks[85]: 0x55c908a7a7a0: maps mask[96]: 85 thread_masks[85]: 0x55c908a7a7b0: affinity mask[96]: 85 thread_masks[86]: 0x55c908a7a7c0: maps mask[96]: 86 thread_masks[86]: 0x55c908a7a7d0: affinity mask[96]: 86 thread_masks[87]: 0x55c908a7a7e0: maps mask[96]: 87 thread_masks[87]: 0x55c908a7a7f0: affinity mask[96]: 87 thread_masks[88]: 0x55c908a7a800: maps mask[96]: 88 thread_masks[88]: 0x55c908a7a810: affinity mask[96]: 88 thread_masks[89]: 0x55c908a7a820: maps mask[96]: 89 thread_masks[89]: 0x55c908a7a830: affinity mask[96]: 89 thread_masks[90]: 0x55c908a7a840: maps mask[96]: 90 thread_masks[90]: 0x55c908a7a850: affinity mask[96]: 90 thread_masks[91]: 0x55c908a7a860: maps mask[96]: 91 thread_masks[91]: 0x55c908a7a870: affinity mask[96]: 91 thread_masks[92]: 0x55c908a7a880: maps mask[96]: 92 thread_masks[92]: 0x55c908a7a890: affinity mask[96]: 92 thread_masks[93]: 0x55c908a7a8a0: maps mask[96]: 93 thread_masks[93]: 0x55c908a7a8b0: affinity mask[96]: 93 thread_masks[94]: 0x55c908a7a8c0: maps mask[96]: 94 thread_masks[94]: 0x55c908a7a8d0: affinity mask[96]: 94 thread_masks[95]: 0x55c908a7a8e0: maps mask[96]: 95 thread_masks[95]: 0x55c908a7a8f0: affinity mask[96]: 95 nr_cblocks: 0 affinity: SYS mmap flush: 1 comp level: 1 Problems creating module maps, continuing anyway... mmap size 528384B Failed to open thread[85] communication pipes Getting available devices: Command line: sycl-ls Exception: [Errno 2] No such file or directory: 'sycl-ls': 'sycl-ls' Checking DPC++ application as prerequisite for GPU analyses... Setting envirnoment variable: SYCL_DEVICE_FILTER=opencl:gpu Command line: /opt/intel/oneapi/vtune/2023.0.0/bin64/self_check_apps/matrix.dpcpp/matrix.dpcpp Stderr: /opt/intel/oneapi/vtune/2023.0.0/bin64/self_check_apps/matrix.dpcpp/matrix.dpcpp: error while loading shared libraries: libsycl.so.5: cannot open shared object file: No such file or directory Setting envirnoment variable: SYCL_DEVICE_FILTER=level_zero:gpu Command line: /opt/intel/oneapi/vtune/2023.0.0/bin64/self_check_apps/matrix.dpcpp/matrix.dpcpp Stderr: /opt/intel/oneapi/vtune/2023.0.0/bin64/self_check_apps/matrix.dpcpp/matrix.dpcpp: error while loading shared libraries: libsycl.so.5: cannot open shared object file: No such file or directory Setting envirnoment variable: ZES_ENABLE_SYSMAN=1 Command line: /opt/intel/oneapi/vtune/2023.0.0/bin64/self_check_apps/matrix.dpcpp/matrix.dpcpp Stderr: /opt/intel/oneapi/vtune/2023.0.0/bin64/self_check_apps/matrix.dpcpp/matrix.dpcpp: error while loading shared libraries: libsycl.so.5: cannot open shared object file: No such file or directory Checking DPC++ application as prerequisite for GPU analyses: Fail Unable to run DPC++ application on GPU connected to this system. If you are using an Intel GPU and want to verify profiling support for DPC++ applications, check these requirements: * Install Intel(R) GPU driver. * Install Intel(R) Level Zero GPU runtime. * Install Intel(R) oneAPI DPC++ Runtime and set the environment. The check observed a product failure on your system. Review errors in the output above to fix a problem or contact Intel technical support. The system is ready for the following analyses: * Performance Snapshot * Hotspots and Threading with user-mode sampling * Hotspots with HW event-based sampling, HPC Performance Characterization, etc. * Hotspots with HW event-based sampling and call stacks The following analyses have failed on the system: * Microarchitecture Exploration * Memory Access * Threading with HW event-based sampling * GPU Compute/Media Hotspots (characterization mode) * GPU Compute/Media Hotspots (source analysis mode) Log location: /tmp/vtune-tmp-resteam/self-checker-2024.05.07_09.00.52/log.txt