<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Run more docker containters with Inter-optimized-tensorflow on One 8 physical core 16cores Cpu in Intel® Optimized AI Frameworks</title>
    <link>https://community.intel.com/t5/Intel-Optimized-AI-Frameworks/Run-more-docker-containters-with-Inter-optimized-tensorflow-on/m-p/1304180#M303</link>
    <description>&lt;P&gt;Hi!&lt;/P&gt;
&lt;P&gt;Thanks for you advice&amp;nbsp;&lt;/P&gt;
&lt;P&gt;I should share more details&lt;/P&gt;
&lt;P&gt;my Inter-optimized-tensorflow containter &lt;STRONG&gt;Environment variables is&amp;nbsp;&lt;BR /&gt;ENV OMP_NUM_THREADS='4'&lt;BR /&gt;ENV KMP_BLOCKTIME='1'&lt;BR /&gt;ENV KMP_AFFINITY=granularity=fine,verbose,compact,1,0&lt;BR /&gt;&lt;/STRONG&gt;&lt;/P&gt;
&lt;P&gt;&lt;STRONG&gt;i run the CMD &lt;SPAN&gt;docker run -itd --cpuset-cpus=7, 8, 9 , 10&lt;/SPAN&gt;&lt;/STRONG&gt;&lt;/P&gt;
&lt;P&gt;also i set tf.config&amp;nbsp;&amp;nbsp;&lt;SPAN&gt;intra_/inter_op_parallelism_threads =4, 2&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;this is the verbose when i run one containter:&lt;/P&gt;
&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="image.png" style="width: 679px;"&gt;&lt;img src="https://community.intel.com/t5/image/serverpage/image-id/18600iA62CBA9B25D74D32/image-size/large/is-moderation-mode/true?v=v2&amp;amp;px=999&amp;amp;whitelist-exif-data=Orientation%2CResolution%2COriginalDefaultFinalSize%2CCopyright" role="button" title="image.png" alt="image.png" /&gt;&lt;/span&gt;&lt;/P&gt;
&lt;P&gt;the train phase cost time is 23s, it is very fast!&lt;/P&gt;
&lt;P&gt;when I set OMP_NUM_THREADS = '8', and other param is fixed,&amp;nbsp; I find the train speed is very slow. it set 4 the train speed is fast.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;but when i run two containters:(the another is run cpu1,2,3,4)&lt;/P&gt;
&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="OosakiKaNa_0-1628232167130.png" style="width: 400px;"&gt;&lt;img src="https://community.intel.com/t5/image/serverpage/image-id/18603iBA9E7E81D29F6457/image-size/medium/is-moderation-mode/true?v=v2&amp;amp;px=400&amp;amp;whitelist-exif-data=Orientation%2CResolution%2COriginalDefaultFinalSize%2CCopyright" role="button" title="OosakiKaNa_0-1628232167130.png" alt="OosakiKaNa_0-1628232167130.png" /&gt;&lt;/span&gt;&lt;/P&gt;
&lt;P&gt;you can find the train phase cost time is increasing,&amp;nbsp; i dont know why&lt;/P&gt;
&lt;P&gt;and this is my host Htop status&lt;/P&gt;
&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="OosakiKaNa_1-1628232241797.png" style="width: 400px;"&gt;&lt;img src="https://community.intel.com/t5/image/serverpage/image-id/18604i59128D460D4C742A/image-size/medium/is-moderation-mode/true?v=v2&amp;amp;px=400&amp;amp;whitelist-exif-data=Orientation%2CResolution%2COriginalDefaultFinalSize%2CCopyright" role="button" title="OosakiKaNa_1-1628232241797.png" alt="OosakiKaNa_1-1628232241797.png" /&gt;&lt;/span&gt;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;thanks.&lt;/P&gt;</description>
    <pubDate>Fri, 06 Aug 2021 06:46:13 GMT</pubDate>
    <dc:creator>OosakiKaNa</dc:creator>
    <dc:date>2021-08-06T06:46:13Z</dc:date>
    <item>
      <title>Run more docker containters with Inter-optimized-tensorflow on One 8 physical core 16cores Cpu</title>
      <link>https://community.intel.com/t5/Intel-Optimized-AI-Frameworks/Run-more-docker-containters-with-Inter-optimized-tensorflow-on/m-p/1303280#M294</link>
      <description>hello, I find the inter-optimized-tensorflow has the great increasing on train phase.
but i want to run 3 docker containters in 8 physical core 16cores Cpu, i set every containter with 4 logical core
how i set the param intra_/inter_op_parallelism_threads and OMP_NUM_THREADS?
when one containter runs, the train time cost 17s every epoch, but when i run 3 containters, in every containter the train time cost 50s/epoch.
by the way i set intra_/inter_op_parallelism_threads =2, OMP_NUM_THREADS= 2 ,KMP_BLOCKTIME=1 in containter.
please tell me why?</description>
      <pubDate>Tue, 03 Aug 2021 11:59:59 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-Optimized-AI-Frameworks/Run-more-docker-containters-with-Inter-optimized-tensorflow-on/m-p/1303280#M294</guid>
      <dc:creator>OosakiKaNa</dc:creator>
      <dc:date>2021-08-03T11:59:59Z</dc:date>
    </item>
    <item>
      <title>Re: Run more docker containters with Inter-optimized-tensorflow on One 8 physical core 16cores Cpu</title>
      <link>https://community.intel.com/t5/Intel-Optimized-AI-Frameworks/Run-more-docker-containters-with-Inter-optimized-tensorflow-on/m-p/1303435#M295</link>
      <description>&lt;P&gt;Hello&amp;nbsp;OosakiKaNa,&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Thank you for posting on the Intel® communities.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;To better assist you, we have moved your question to another forum.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Regards,&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Adrian M.&lt;/P&gt;
&lt;P&gt;Intel Customer Support Technician &lt;/P&gt;</description>
      <pubDate>Tue, 03 Aug 2021 22:07:05 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-Optimized-AI-Frameworks/Run-more-docker-containters-with-Inter-optimized-tensorflow-on/m-p/1303435#M295</guid>
      <dc:creator>AdrianM_Intel</dc:creator>
      <dc:date>2021-08-03T22:07:05Z</dc:date>
    </item>
    <item>
      <title>Re:Run more docker containters with Inter-optimized-tensorflow on One 8 physical core 16cores Cpu</title>
      <link>https://community.intel.com/t5/Intel-Optimized-AI-Frameworks/Run-more-docker-containters-with-Inter-optimized-tensorflow-on/m-p/1303542#M297</link>
      <description>&lt;P&gt;Hi,&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;Could you please share the following details:&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;1) Docker images you used?&lt;/P&gt;&lt;P&gt;2) Complete steps to reproduce the issue including the commands you used&lt;/P&gt;&lt;P&gt;3) Intel tensorflow version used&lt;/P&gt;&lt;P&gt;4) OS details&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;Thanks&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;/P&gt;&lt;BR /&gt;</description>
      <pubDate>Wed, 04 Aug 2021 06:43:54 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-Optimized-AI-Frameworks/Run-more-docker-containters-with-Inter-optimized-tensorflow-on/m-p/1303542#M297</guid>
      <dc:creator>AthiraM_Intel</dc:creator>
      <dc:date>2021-08-04T06:43:54Z</dc:date>
    </item>
    <item>
      <title>Re: Re:Run more docker containters with Inter-optimized-tensorflow on One 8 physical core 16cores Cp</title>
      <link>https://community.intel.com/t5/Intel-Optimized-AI-Frameworks/Run-more-docker-containters-with-Inter-optimized-tensorflow-on/m-p/1303550#M298</link>
      <description>&lt;P&gt;docker images: intel/intel-optimized-tensorflow:2.2.0-centos-8-mpich-horovod&lt;/P&gt;
&lt;P&gt;my os: centos8&lt;/P&gt;
&lt;P&gt;docker run -itd --cpuset-cpus=1,2,3,4 -v /home/liangliang/nfscontent/:/tf/tft/output tft:v1&lt;/P&gt;
&lt;P&gt;tft:v1 is my program iamge&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;thanks&lt;/P&gt;</description>
      <pubDate>Wed, 04 Aug 2021 07:08:21 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-Optimized-AI-Frameworks/Run-more-docker-containters-with-Inter-optimized-tensorflow-on/m-p/1303550#M298</guid>
      <dc:creator>OosakiKaNa</dc:creator>
      <dc:date>2021-08-04T07:08:21Z</dc:date>
    </item>
    <item>
      <title>Re: Run more docker containters with Inter-optimized-tensorflow on One 8 physical core 16cores Cpu</title>
      <link>https://community.intel.com/t5/Intel-Optimized-AI-Frameworks/Run-more-docker-containters-with-Inter-optimized-tensorflow-on/m-p/1304156#M302</link>
      <description>&lt;P&gt;Hi,&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Thanks for sharing the details.&lt;/P&gt;
&lt;P&gt;Could you please share the log file by enabling KMP_AFFINITY verbose.&lt;/P&gt;
&lt;P&gt;ie, KMP_AFFINITY=verbose&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Please find the below link for more information:&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&lt;A href="https://software.intel.com/content/www/us/en/develop/documentation/cpp-compiler-developer-guide-and-reference/top/optimization-and-programming-guide/openmp-support/openmp-library-support/thread-affinity-interface-linux-and-windows.html" target="_blank" rel="noopener"&gt;https://software.intel.com/content/www/us/en/develop/documentation/cpp-compiler-developer-guide-and-reference/top/optimization-and-programming-guide/openmp-support/openmp-library-support/thread-affinity-interface-linux-and-windows.html&lt;/A&gt;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Also you can try by increasing the OMP_NUM_THREADS , set OMP_NUM_THREADS = 8 and check whether there is any improvement?&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Thanks.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Fri, 06 Aug 2021 05:49:11 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-Optimized-AI-Frameworks/Run-more-docker-containters-with-Inter-optimized-tensorflow-on/m-p/1304156#M302</guid>
      <dc:creator>AthiraM_Intel</dc:creator>
      <dc:date>2021-08-06T05:49:11Z</dc:date>
    </item>
    <item>
      <title>Re: Run more docker containters with Inter-optimized-tensorflow on One 8 physical core 16cores Cpu</title>
      <link>https://community.intel.com/t5/Intel-Optimized-AI-Frameworks/Run-more-docker-containters-with-Inter-optimized-tensorflow-on/m-p/1304180#M303</link>
      <description>&lt;P&gt;Hi!&lt;/P&gt;
&lt;P&gt;Thanks for you advice&amp;nbsp;&lt;/P&gt;
&lt;P&gt;I should share more details&lt;/P&gt;
&lt;P&gt;my Inter-optimized-tensorflow containter &lt;STRONG&gt;Environment variables is&amp;nbsp;&lt;BR /&gt;ENV OMP_NUM_THREADS='4'&lt;BR /&gt;ENV KMP_BLOCKTIME='1'&lt;BR /&gt;ENV KMP_AFFINITY=granularity=fine,verbose,compact,1,0&lt;BR /&gt;&lt;/STRONG&gt;&lt;/P&gt;
&lt;P&gt;&lt;STRONG&gt;i run the CMD &lt;SPAN&gt;docker run -itd --cpuset-cpus=7, 8, 9 , 10&lt;/SPAN&gt;&lt;/STRONG&gt;&lt;/P&gt;
&lt;P&gt;also i set tf.config&amp;nbsp;&amp;nbsp;&lt;SPAN&gt;intra_/inter_op_parallelism_threads =4, 2&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;this is the verbose when i run one containter:&lt;/P&gt;
&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="image.png" style="width: 679px;"&gt;&lt;img src="https://community.intel.com/t5/image/serverpage/image-id/18600iA62CBA9B25D74D32/image-size/large/is-moderation-mode/true?v=v2&amp;amp;px=999&amp;amp;whitelist-exif-data=Orientation%2CResolution%2COriginalDefaultFinalSize%2CCopyright" role="button" title="image.png" alt="image.png" /&gt;&lt;/span&gt;&lt;/P&gt;
&lt;P&gt;the train phase cost time is 23s, it is very fast!&lt;/P&gt;
&lt;P&gt;when I set OMP_NUM_THREADS = '8', and other param is fixed,&amp;nbsp; I find the train speed is very slow. it set 4 the train speed is fast.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;but when i run two containters:(the another is run cpu1,2,3,4)&lt;/P&gt;
&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="OosakiKaNa_0-1628232167130.png" style="width: 400px;"&gt;&lt;img src="https://community.intel.com/t5/image/serverpage/image-id/18603iBA9E7E81D29F6457/image-size/medium/is-moderation-mode/true?v=v2&amp;amp;px=400&amp;amp;whitelist-exif-data=Orientation%2CResolution%2COriginalDefaultFinalSize%2CCopyright" role="button" title="OosakiKaNa_0-1628232167130.png" alt="OosakiKaNa_0-1628232167130.png" /&gt;&lt;/span&gt;&lt;/P&gt;
&lt;P&gt;you can find the train phase cost time is increasing,&amp;nbsp; i dont know why&lt;/P&gt;
&lt;P&gt;and this is my host Htop status&lt;/P&gt;
&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="OosakiKaNa_1-1628232241797.png" style="width: 400px;"&gt;&lt;img src="https://community.intel.com/t5/image/serverpage/image-id/18604i59128D460D4C742A/image-size/medium/is-moderation-mode/true?v=v2&amp;amp;px=400&amp;amp;whitelist-exif-data=Orientation%2CResolution%2COriginalDefaultFinalSize%2CCopyright" role="button" title="OosakiKaNa_1-1628232241797.png" alt="OosakiKaNa_1-1628232241797.png" /&gt;&lt;/span&gt;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;thanks.&lt;/P&gt;</description>
      <pubDate>Fri, 06 Aug 2021 06:46:13 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-Optimized-AI-Frameworks/Run-more-docker-containters-with-Inter-optimized-tensorflow-on/m-p/1304180#M303</guid>
      <dc:creator>OosakiKaNa</dc:creator>
      <dc:date>2021-08-06T06:46:13Z</dc:date>
    </item>
    <item>
      <title>Re:Run more docker containters with Inter-optimized-tensorflow on One 8 physical core 16cores Cpu</title>
      <link>https://community.intel.com/t5/Intel-Optimized-AI-Frameworks/Run-more-docker-containters-with-Inter-optimized-tensorflow-on/m-p/1306258#M304</link>
      <description>&lt;P&gt;Hi,&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;We are checking on your issue. Could you please share the sample reproducer and complete steps to try out the same from our end.&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;Thanks&lt;/P&gt;&lt;BR /&gt;</description>
      <pubDate>Thu, 12 Aug 2021 12:47:40 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-Optimized-AI-Frameworks/Run-more-docker-containters-with-Inter-optimized-tensorflow-on/m-p/1306258#M304</guid>
      <dc:creator>AthiraM_Intel</dc:creator>
      <dc:date>2021-08-12T12:47:40Z</dc:date>
    </item>
    <item>
      <title>Re: Re:Run more docker containters with Inter-optimized-tensorflow on One 8 physical core 16cores Cp</title>
      <link>https://community.intel.com/t5/Intel-Optimized-AI-Frameworks/Run-more-docker-containters-with-Inter-optimized-tensorflow-on/m-p/1306269#M305</link>
      <description>&lt;P&gt;Hi!&lt;/P&gt;
&lt;P&gt;What should I do? send you my program and dataset?&lt;BR /&gt;I dont know how to do, please tell me&lt;BR /&gt;Thanks&lt;/P&gt;</description>
      <pubDate>Thu, 12 Aug 2021 13:19:52 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-Optimized-AI-Frameworks/Run-more-docker-containters-with-Inter-optimized-tensorflow-on/m-p/1306269#M305</guid>
      <dc:creator>OosakiKaNa</dc:creator>
      <dc:date>2021-08-12T13:19:52Z</dc:date>
    </item>
    <item>
      <title>Re:Run more docker containters with Inter-optimized-tensorflow on One 8 physical core 16cores Cpu</title>
      <link>https://community.intel.com/t5/Intel-Optimized-AI-Frameworks/Run-more-docker-containters-with-Inter-optimized-tensorflow-on/m-p/1307420#M308</link>
      <description>&lt;P&gt;Hi,&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;Yes, you can share your sample reproducer and commands used. Regarding this we will contact you through private message shortly.&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;Thanks&lt;/P&gt;&lt;BR /&gt;</description>
      <pubDate>Tue, 17 Aug 2021 10:01:16 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-Optimized-AI-Frameworks/Run-more-docker-containters-with-Inter-optimized-tensorflow-on/m-p/1307420#M308</guid>
      <dc:creator>AthiraM_Intel</dc:creator>
      <dc:date>2021-08-17T10:01:16Z</dc:date>
    </item>
    <item>
      <title>Re: Re:Run more docker containters with Inter-optimized-tensorflow on One 8 physical core 16cores Cp</title>
      <link>https://community.intel.com/t5/Intel-Optimized-AI-Frameworks/Run-more-docker-containters-with-Inter-optimized-tensorflow-on/m-p/1309408#M312</link>
      <description>&lt;DIV&gt;Hi&lt;/DIV&gt;
&lt;DIV&gt;I am sorry to reply you for a so long time&lt;/DIV&gt;
&lt;DIV&gt;My company doesn't let me share the Program and Data&lt;/DIV&gt;
&lt;DIV&gt;&amp;nbsp;&lt;/DIV&gt;
&lt;DIV&gt;Actually， I have gived up at the Issue，I think maybe it's Hardware Limitation，So It can't solve this problem with Software Setting.&lt;/DIV&gt;
&lt;DIV&gt;The Model is not so complex, it just have 220K parameter，The data is just a excel file with 10K row and 13 columns.&lt;/DIV&gt;
&lt;DIV&gt;The model source code is&amp;nbsp;&lt;A href="https://github.com/google-research/google-research/tree/master/tft" target="_blank"&gt;https://github.com/google-research/google-research/tree/master/tft&lt;/A&gt;&lt;/DIV&gt;
&lt;DIV&gt;but this code is not for running in the docker Containter.&lt;/DIV&gt;
&lt;DIV&gt;I run the model with inter-optimized-inter 2.2.0&amp;nbsp; but i doesn't using the Tensorflow2 property&lt;/DIV&gt;
&lt;DIV&gt;I import tensorflow.compat.v1 as tf so I think maybe use tf2.0 can bring some advancement&lt;/DIV&gt;
&lt;DIV&gt;But recently I can't do the experiment with this setting, if i have time i will try. and i will contact you.&lt;/DIV&gt;
&lt;DIV&gt;So the issue maybe is over&lt;/DIV&gt;
&lt;DIV&gt;My English is poor, sorry.&amp;nbsp;&lt;/DIV&gt;
&lt;DIV&gt;Thanks for your help !&amp;nbsp;&lt;/DIV&gt;</description>
      <pubDate>Tue, 24 Aug 2021 13:59:02 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-Optimized-AI-Frameworks/Run-more-docker-containters-with-Inter-optimized-tensorflow-on/m-p/1309408#M312</guid>
      <dc:creator>OosakiKaNa</dc:creator>
      <dc:date>2021-08-24T13:59:02Z</dc:date>
    </item>
    <item>
      <title>Re: Run more docker containters with Inter-optimized-tensorflow on One 8 physical core 16cores Cpu</title>
      <link>https://community.intel.com/t5/Intel-Optimized-AI-Frameworks/Run-more-docker-containters-with-Inter-optimized-tensorflow-on/m-p/1309880#M314</link>
      <description>&lt;P&gt;Hi&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;From the KMP verbose log, you could see 8 threads bound to cpu 7-10 when you&amp;nbsp;&lt;SPAN&gt;set OMP_NUM_THREADS = '4'.&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;If you have hyperthreading on, each thread could use 1 hyper thread because number of hyper threading is 8 in this case.&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;However, when you set&amp;nbsp;set OMP_NUM_THREADS = '8', you will have 16 threads to compete 8 hyper threads. the performance will be impacted.&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;For the two container case, do you run your workloads on a system with&lt;/SPAN&gt;2 sockets&lt;SPAN style="font-family: inherit;"&gt;?&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;If yes, you might need to use numactl to make all threads within a container to run on one socket instead of two sockets to reduce some NUMA remote access issue.&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;regards&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Wed, 25 Aug 2021 23:15:23 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-Optimized-AI-Frameworks/Run-more-docker-containters-with-Inter-optimized-tensorflow-on/m-p/1309880#M314</guid>
      <dc:creator>Louie_T_Intel</dc:creator>
      <dc:date>2021-08-25T23:15:23Z</dc:date>
    </item>
    <item>
      <title>Re: Run more docker containters with Inter-optimized-tensorflow on One 8 physical core 16cores Cpu</title>
      <link>https://community.intel.com/t5/Intel-Optimized-AI-Frameworks/Run-more-docker-containters-with-Inter-optimized-tensorflow-on/m-p/1309896#M315</link>
      <description>&lt;P&gt;Hi~&lt;/P&gt;
&lt;P&gt;Thanks for your reply&lt;/P&gt;
&lt;P&gt;I don't run my workloads on a system with 2 sockets&amp;nbsp;&lt;/P&gt;
&lt;P&gt;This is My computer cpu information&lt;/P&gt;
&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="OosakiKaNa_0-1629939364467.png" style="width: 400px;"&gt;&lt;img src="https://community.intel.com/t5/image/serverpage/image-id/19012iDB5A27D9CEB77D25/image-size/medium/is-moderation-mode/true?v=v2&amp;amp;px=400&amp;amp;whitelist-exif-data=Orientation%2CResolution%2COriginalDefaultFinalSize%2CCopyright" role="button" title="OosakiKaNa_0-1629939364467.png" alt="OosakiKaNa_0-1629939364467.png" /&gt;&lt;/span&gt;&lt;/P&gt;
&lt;P&gt;But tomorrow my company buy 10 computers with Gold 6248R 2sockets 24C/48T&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Actually I use k8s manage my model at 29 computers, so Do you know how can I&amp;nbsp;&lt;SPAN&gt;make all threads within a container to run on one socket instead of two sockets with k8s setting?&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;My English is poor, sorry.&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;Regards&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Thu, 26 Aug 2021 01:10:00 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-Optimized-AI-Frameworks/Run-more-docker-containters-with-Inter-optimized-tensorflow-on/m-p/1309896#M315</guid>
      <dc:creator>OosakiKaNa</dc:creator>
      <dc:date>2021-08-26T01:10:00Z</dc:date>
    </item>
    <item>
      <title>Re: Run more docker containters with Inter-optimized-tensorflow on One 8 physical core 16cores Cpu</title>
      <link>https://community.intel.com/t5/Intel-Optimized-AI-Frameworks/Run-more-docker-containters-with-Inter-optimized-tensorflow-on/m-p/1311633#M319</link>
      <description>&lt;P&gt;Hi,&lt;/P&gt;
&lt;P&gt;&amp;nbsp; To simplify the description, we use physical cores in this topic.&lt;/P&gt;
&lt;P&gt;&amp;nbsp; I think in your case, set the same cores numbers to each container, but the containers share some cores in same time. So, the performance is reduced to 1/3 of one container.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;To resolve this issue, please assign different cores to different containers. Like:&amp;nbsp;&amp;nbsp;&lt;/P&gt;
&lt;PRE class="highlight"&gt;&lt;CODE&gt;docker run &lt;SPAN class="nt"&gt;-it&lt;/SPAN&gt; &lt;SPAN class="nt"&gt;--cpus&lt;/SPAN&gt;&lt;SPAN class="o"&gt;=&lt;/SPAN&gt;&lt;SPAN class="s2"&gt;"1,2"&lt;/SPAN&gt; ubuntu /bin/bash&lt;BR /&gt;docker run &lt;SPAN class="nt"&gt;-it&lt;/SPAN&gt; &lt;SPAN class="nt"&gt;--cpus&lt;/SPAN&gt;&lt;SPAN class="o"&gt;=&lt;/SPAN&gt;&lt;SPAN class="s2"&gt;"3,4"&lt;/SPAN&gt; ubuntu /bin/bash&lt;BR /&gt;docker run &lt;SPAN class="nt"&gt;-it&lt;/SPAN&gt; &lt;SPAN class="nt"&gt;--cpus&lt;/SPAN&gt;&lt;SPAN class="o"&gt;=&lt;/SPAN&gt;&lt;SPAN class="s2"&gt;"5,6"&lt;/SPAN&gt; ubuntu /bin/bash&lt;BR /&gt;&lt;/CODE&gt;&lt;/PRE&gt;
&lt;P&gt;Refer to:&amp;nbsp;&lt;A href="https://docs.docker.com/config/containers/resource_constraints/" target="_blank"&gt;https://docs.docker.com/config/containers/resource_constraints/&lt;/A&gt;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Thank you!&lt;/P&gt;</description>
      <pubDate>Thu, 02 Sep 2021 04:39:14 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-Optimized-AI-Frameworks/Run-more-docker-containters-with-Inter-optimized-tensorflow-on/m-p/1311633#M319</guid>
      <dc:creator>Jianyu_Z_Intel</dc:creator>
      <dc:date>2021-09-02T04:39:14Z</dc:date>
    </item>
    <item>
      <title>Re: Run more docker containters with Inter-optimized-tensorflow on One 8 physical core 16cores Cpu</title>
      <link>https://community.intel.com/t5/Intel-Optimized-AI-Frameworks/Run-more-docker-containters-with-Inter-optimized-tensorflow-on/m-p/1311647#M320</link>
      <description>&lt;P&gt;Hi!&lt;/P&gt;
&lt;P&gt;Thanks for your reply&lt;/P&gt;
&lt;P&gt;Please take a look on my reply at &lt;SPAN class="local-date"&gt;‎08-05-2021&lt;/SPAN&gt;&lt;SPAN&gt;&amp;nbsp;&lt;/SPAN&gt;&lt;SPAN class="local-time"&gt;11:46 PM&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;I run the two docker containter on cpu 7,8,9,10 and 1,2,3,4&lt;/P&gt;
&lt;P&gt;My computer RAM is 32GB, I set they run different cpu, but the issue is still &lt;SPAN&gt;exists&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;My English is poor, sorry&lt;/P&gt;
&lt;P&gt;Regards&lt;/P&gt;</description>
      <pubDate>Thu, 02 Sep 2021 06:01:04 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-Optimized-AI-Frameworks/Run-more-docker-containters-with-Inter-optimized-tensorflow-on/m-p/1311647#M320</guid>
      <dc:creator>OosakiKaNa</dc:creator>
      <dc:date>2021-09-02T06:01:04Z</dc:date>
    </item>
    <item>
      <title>Re: Run more docker containters with Inter-optimized-tensorflow on One 8 physical core 16cores Cpu</title>
      <link>https://community.intel.com/t5/Intel-Optimized-AI-Frameworks/Run-more-docker-containters-with-Inter-optimized-tensorflow-on/m-p/1311648#M321</link>
      <description>&lt;P&gt;Hi,&lt;/P&gt;
&lt;P&gt;&amp;nbsp; Don't warry! I fully understand your words.&amp;nbsp; &lt;LI-EMOJI id="lia_slightly-smiling-face" title=":slightly_smiling_face:"&gt;&lt;/LI-EMOJI&gt;&lt;/P&gt;
&lt;P&gt;&amp;nbsp; In your CPU, there are 8 cores.&amp;nbsp;The cores 0-7 are the index of them.&lt;/P&gt;
&lt;P&gt;&amp;nbsp; Index 8 and index 0 are same core in fact.&lt;/P&gt;
&lt;P&gt;&amp;nbsp; In your case:&amp;nbsp;&lt;SPAN&gt;cpu 7,8,9,10 and 1,2,3,4&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&lt;SPAN style="font-family: inherit;"&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; 1,9 &amp;amp; 2, 10, they are same cores in fact.&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;&amp;nbsp; That means they share 2 cores (1(9), 2(10)). That will impact the performance.&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;&amp;nbsp; If you want to use 4 cores per container, please use 0-3, 4-7.&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;&amp;nbsp; Avoid to assign one core to more than one container.&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;&amp;nbsp; Thank you!&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;&amp;nbsp; In my example:&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;&amp;nbsp;&amp;nbsp;&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Thu, 02 Sep 2021 06:15:56 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-Optimized-AI-Frameworks/Run-more-docker-containters-with-Inter-optimized-tensorflow-on/m-p/1311648#M321</guid>
      <dc:creator>Jianyu_Z_Intel</dc:creator>
      <dc:date>2021-09-02T06:15:56Z</dc:date>
    </item>
    <item>
      <title>Re: Run more docker containters with Inter-optimized-tensorflow on One 8 physical core 16cores Cpu</title>
      <link>https://community.intel.com/t5/Intel-Optimized-AI-Frameworks/Run-more-docker-containters-with-Inter-optimized-tensorflow-on/m-p/1311650#M322</link>
      <description>&lt;P&gt;Hi!&lt;/P&gt;
&lt;P&gt;Thanks for your so fast reply&lt;/P&gt;
&lt;P&gt;I will do the experiment with this setting&lt;/P&gt;
&lt;P&gt;But Actually I manage my model on 34 computers with k8s, the k8s control the docker containter with Cgroups, It can't assign physical core(maybe can't, at now i don't know this)&lt;/P&gt;
&lt;P&gt;&lt;BR /&gt;So if this issue is about the cpu share(means hardware issue), it can't solve by software setting(I guess).&lt;/P&gt;
&lt;P&gt;I just want to know what brings this problem.&lt;/P&gt;
&lt;P&gt;Thanks&lt;/P&gt;</description>
      <pubDate>Thu, 02 Sep 2021 06:25:39 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-Optimized-AI-Frameworks/Run-more-docker-containters-with-Inter-optimized-tensorflow-on/m-p/1311650#M322</guid>
      <dc:creator>OosakiKaNa</dc:creator>
      <dc:date>2021-09-02T06:25:39Z</dc:date>
    </item>
    <item>
      <title>Re: Run more docker containters with Inter-optimized-tensorflow on One 8 physical core 16cores Cpu</title>
      <link>https://community.intel.com/t5/Intel-Optimized-AI-Frameworks/Run-more-docker-containters-with-Inter-optimized-tensorflow-on/m-p/1311693#M323</link>
      <description>&lt;P&gt;Hi,&lt;/P&gt;
&lt;P&gt;&amp;nbsp; In K8S case, Intel provide the solution for CPU pinning:&amp;nbsp;&lt;SPAN&gt;CPU Manager for&amp;nbsp;&lt;/SPAN&gt;&lt;SPAN&gt;Kubernetes* (also called CMK).&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;&amp;nbsp; Here is the guide for it.&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;&amp;nbsp;&amp;nbsp;&lt;A href="https://builders.intel.com/docs/networkbuilders/cpu-pin-and-isolation-in-kubernetes-app-note.pdf" target="_blank"&gt;https://builders.intel.com/docs/networkbuilders/cpu-pin-and-isolation-in-kubernetes-app-note.pdf&lt;/A&gt;&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;&amp;nbsp; &lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;&amp;nbsp; If you have more question about CMK, please create new issue for CMK in Intel Community.&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;&amp;nbsp; Good luck!&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;&amp;nbsp; Thank you!&amp;nbsp;&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;&amp;nbsp; &amp;nbsp;&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Thu, 02 Sep 2021 09:39:41 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-Optimized-AI-Frameworks/Run-more-docker-containters-with-Inter-optimized-tensorflow-on/m-p/1311693#M323</guid>
      <dc:creator>Jianyu_Z_Intel</dc:creator>
      <dc:date>2021-09-02T09:39:41Z</dc:date>
    </item>
    <item>
      <title>Re: Run more docker containters with Inter-optimized-tensorflow on One 8 physical core 16cores Cpu</title>
      <link>https://community.intel.com/t5/Intel-Optimized-AI-Frameworks/Run-more-docker-containters-with-Inter-optimized-tensorflow-on/m-p/1311706#M324</link>
      <description>&lt;P&gt;Hi!&lt;/P&gt;
&lt;P&gt;From your reply I know the reason cause the Issue and the tools to solve it&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Thus, the issue is over!&lt;/P&gt;
&lt;P&gt;Thank you and community's everyone!&lt;/P&gt;
&lt;P&gt;Thank Intel!&lt;/P&gt;</description>
      <pubDate>Thu, 02 Sep 2021 10:37:54 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-Optimized-AI-Frameworks/Run-more-docker-containters-with-Inter-optimized-tensorflow-on/m-p/1311706#M324</guid>
      <dc:creator>OosakiKaNa</dc:creator>
      <dc:date>2021-09-02T10:37:54Z</dc:date>
    </item>
    <item>
      <title>Re: Run more docker containters with Inter-optimized-tensorflow on One 8 physical core 16cores Cpu</title>
      <link>https://community.intel.com/t5/Intel-Optimized-AI-Frameworks/Run-more-docker-containters-with-Inter-optimized-tensorflow-on/m-p/1311871#M326</link>
      <description>&lt;P&gt;Hi,&lt;/P&gt;
&lt;P&gt;&amp;nbsp; It's our pleasure!&amp;nbsp; &lt;LI-EMOJI id="lia_slightly-smiling-face" title=":slightly_smiling_face:"&gt;&lt;/LI-EMOJI&gt;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;Thank your support!&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Fri, 03 Sep 2021 02:03:40 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-Optimized-AI-Frameworks/Run-more-docker-containters-with-Inter-optimized-tensorflow-on/m-p/1311871#M326</guid>
      <dc:creator>Jianyu_Z_Intel</dc:creator>
      <dc:date>2021-09-03T02:03:40Z</dc:date>
    </item>
  </channel>
</rss>

