- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Hi all !
I'm new using Intel MKL routines. I have the following function that multiplies two dense matrices:
I compile with the followin flags
g++ -std=c++11 -O3 -I/opt/intel/oneapi/2022/mkl/latest/include c laicoMult.cpp -L/opt/intel/oneapi/2022/mkl/latest/lib/intel64 -Wl,-rpath,/opt/intel/oneapi/mkl/2024.0/lib/intel64 -lmkl_intel_lp64 -lmkl_sequential -lmkl_core -lpthread -lm -ldl -fopenmp -march=native -fopt-info-vec -ffast-math -ftree-vectorize -DARMA_DONT_USE_WRAPPER -o laicoMult
At the moment I'm using a Intel(R) Core(TM) i7-10700 CPU @ 2.90GHz processor with 16GB DDR4.
The code is pretty fast but I'd like to know if there is something I could do to make this code faster.
Can anyone help me ?
Thanks in advance !
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
yes, you could link against threaded version of MKL instead of sequential (-lmkl_sequential ) once. Please check with MKL Linker Adviser ( https://www.intel.com/content/www/us/en/developer/tools/oneapi/onemkl-link-line-advisor.html ) how to link against OpenMP or TBB threaded runtimes. You also might check MKL Developer Guide where you could find out many of such examples.
Link Copied
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
yes, you could link against threaded version of MKL instead of sequential (-lmkl_sequential ) once. Please check with MKL Linker Adviser ( https://www.intel.com/content/www/us/en/developer/tools/oneapi/onemkl-link-line-advisor.html ) how to link against OpenMP or TBB threaded runtimes. You also might check MKL Developer Guide where you could find out many of such examples.

- Subscribe to RSS Feed
- Mark Topic as New
- Mark Topic as Read
- Float this Topic for Current User
- Bookmark
- Subscribe
- Printer Friendly Page