Difference between revisions of "CSIT/csit-perf-env-tuning-ubuntu1604-obsolete"

From fd.io
Jump to: navigation, search
Line 37: Line 37:
 
* Jira CSIT-428, https://gerrit.fd.io/r/#/c/3244/
 
* Jira CSIT-428, https://gerrit.fd.io/r/#/c/3244/
  
csit resources/tools/testbed-setup/playbooks/files
+
 
$ more cpufrequtils
+
  $ csit resources/tools/testbed-setup/playbooks/files
GOVERNOR="performance"
+
  $ more cpufrequtils
 +
  GOVERNOR="performance"
  
 
   - name: Set cpufrequtils defaults
 
   - name: Set cpufrequtils defaults
 
     copy: src=files/cpufrequtils dest=/etc/default/cpufrequtils owner=root group=root mode=0644
 
     copy: src=files/cpufrequtils dest=/etc/default/cpufrequtils owner=root group=root mode=0644
  
$ more irqbalance
+
  $ more irqbalance
#Configuration for the irqbalance daemon
+
  #Configuration for the irqbalance daemon
  
#Should irqbalance be enabled?
+
  #Should irqbalance be enabled?
ENABLED="0"
+
  ENABLED="0"
#Balance the IRQs only once?
+
  #Balance the IRQs only once?
ONESHOT="0"
+
  ONESHOT="0"
  
 
   - name: Disable IRQ load balancing
 
   - name: Disable IRQ load balancing
 
     copy: src=files/irqbalance dest=/etc/default/irqbalance owner=root group=root mode=0644
 
     copy: src=files/irqbalance dest=/etc/default/irqbalance owner=root group=root mode=0644
  
# sysctl settings
+
==== sysctl settings ====
  
a. not using hugepages setting in GRUB_CMDLINE_LINUX, e.g. GRUB_CMDLINE_LINUX="default_hugepagesz=1GB hugepagesz=1G hugepages=64"
+
* not using hugepages setting in GRUB_CMDLINE_LINUX, e.g. GRUB_CMDLINE_LINUX="default_hugepagesz=1GB hugepagesz=1G hugepages=64"
b. using sysctl instead to set additional related parameters
+
* using sysctl instead to set additional related parameters
c. with ubuntu 14.04.3
+
* with ubuntu 14.04.3
     - mk: we did have hugepages setting there, no? was it in GRUB_CMDLINE_LINUX - set manually using 90-csit.conf, but not in ansible script ########
+
     Hugepages were applied by VPP via 80-vpp.conf (1024 of 2M hugepages). For the vhost measurements we dynamically allocate additional hugepages during the vhost tests. This aproach leads to huge fragmentation of memory space and caused issues to testbeds.
    - pm: we did have manually applied setting that comes from 80-vpp.conf (1024 of 2M hugepages). We dynamically allocate more during the vhost testing. This aproach leads to huge fragmentation of memory space and caused issues to testbeds.
+
* with ubuntu 16.04.1:
d. with ubuntu 16.04.1:
+
 
     - ansible .yaml:
 
     - ansible .yaml:
 
           - name: copy sysctl file
 
           - name: copy sysctl file
Line 119: Line 119:
 
     kernel.watchdog_cpumask=0,18
 
     kernel.watchdog_cpumask=0,18
  
e. Jira CSIT-429, https://gerrit.fd.io/r/#/c/3245/
+
==== Host CFS optimizations (QEMU+VPP) ====
 
+
# Host CFS optimizations (QEMU+VPP)
+
  
 
Applying CFS scheduler tuning on all Qemu vcpu worker threads (those are handling testpmd - pmd threads) and VPP
 
Applying CFS scheduler tuning on all Qemu vcpu worker threads (those are handling testpmd - pmd threads) and VPP
PMD worker threads. List of VPP PMD threads can be obtained either from
+
PMD worker threads. List of VPP PMD threads can be obtained either from:
"cat /proc/`pidof vpp`/task/*/stat | awk '{print $1" "$2" "$39}'" (prefixed by vpp_wk) or via VPP cli command "show threads".
+
 
 +
  "cat /proc/`pidof vpp`/task/*/stat | awk '{print $1" "$2" "$39}'"
  
    $ chrt -r -p 1 <worker_pid>
+
  $ chrt -r -p 1 <worker_pid>
  
  
# Host IRQ optimizations
+
==== Host IRQ optimizations ====
  
 
Changing the default pinning of IRQ to core 0. (Same does apply on both guest and host OS)
 
Changing the default pinning of IRQ to core 0. (Same does apply on both guest and host OS)
  
    $ for l in `ls /proc/irq`; do echo 1 | sudo tee /proc/irq/$l/smp_affinity; done
+
  $ for l in `ls /proc/irq`; do echo 1 | sudo tee /proc/irq/$l/smp_affinity; done
  
  
# Host RCU optimizations
+
==== Host RCU optimizations ====
  
 
Changing the default pinning of RCU to core 0. (Same does apply on both guest and host OS)
 
Changing the default pinning of RCU to core 0. (Same does apply on both guest and host OS)
  
    $ for i in `pgrep rcu[^c]` ; do sudo taskset -pc 0 $i ; done
+
  $ for i in `pgrep rcu[^c]` ; do sudo taskset -pc 0 $i ; done
  
  
# Host Writeback affiniy
+
==== Host Writeback affinity ====
  
 
Changing the default pinning of writebacks to core 0. (Same does apply on both guest and host OS)
 
Changing the default pinning of writebacks to core 0. (Same does apply on both guest and host OS)
  
    $ echo 1 | sudo tee /sys/bus/workqueue/devices/writeback/cpumask
+
  $ echo 1 | sudo tee /sys/bus/workqueue/devices/writeback/cpumask
  
 
For more information please follow: https://www.kernel.org/doc/Documentation/kernel-per-CPU-kthreads.txt
 
For more information please follow: https://www.kernel.org/doc/Documentation/kernel-per-CPU-kthreads.txt

Revision as of 09:24, 6 February 2017

DRAFT - work in progress as part of CSIT rls1701 report

https://wiki.fd.io/view/CSIT/csit-perf-env-tuning-ubuntu1604

We have upgraded the CSIT performance testbeds, and used this to apply kernel configuration changes that should address some of the issues we observed during perf tests in csit1609, mainly related to interactions with qemu in vhost tests.

Kernel boot parameters (grub)

Following kernel boot parameters are used in CSIT performance testbeds

All grub command line parameters are applied during installation using CSIT ansible scripts

  • csit / resources/tools/testbed-setup/playbooks/01-host-setup.yaml
    • line=GRUB_CMDLINE_LINUX="\"isolcpus=Template:Isolcpus nohz_full=Template:Nohz rcu_nocbs=Template:Rcu intel_pstate=disable\""
    • sample of generated grub config line: GRUB_CMDLINE_LINUX="isolcpus=1-17,19-35 intel_pstate=disable nohz_full=1-17,19-35 rcu_nocbs=1-17,19-35"

Changes applied during upgrade from ubuntu 14.04.03 to ubuntu 16.04.1

  • in ubuntu 14.04.3
    • sample of generated grub config line: GRUB_CMDLINE_LINUX="isolcpus=1-17,19-35 intel_pstate=disable"
  • in ubuntu 16.04.1
    • sample of generated grub config line: GRUB_CMDLINE_LINUX="isolcpus=1-17,19-35 intel_pstate=disable nohz_full=1-17,19-35 rcu_nocbs=1-17,19-35"
  • Jira CSIT-428, https://gerrit.fd.io/r/#/c/3244/


 $ csit resources/tools/testbed-setup/playbooks/files
 $ more cpufrequtils
 GOVERNOR="performance"
 - name: Set cpufrequtils defaults
   copy: src=files/cpufrequtils dest=/etc/default/cpufrequtils owner=root group=root mode=0644
 $ more irqbalance
 #Configuration for the irqbalance daemon
 #Should irqbalance be enabled?
 ENABLED="0"
 #Balance the IRQs only once?
 ONESHOT="0"
 - name: Disable IRQ load balancing
   copy: src=files/irqbalance dest=/etc/default/irqbalance owner=root group=root mode=0644

sysctl settings

  • not using hugepages setting in GRUB_CMDLINE_LINUX, e.g. GRUB_CMDLINE_LINUX="default_hugepagesz=1GB hugepagesz=1G hugepages=64"
  • using sysctl instead to set additional related parameters
  • with ubuntu 14.04.3
   Hugepages were applied by VPP via 80-vpp.conf (1024 of 2M hugepages). For the vhost measurements we dynamically allocate additional hugepages during the vhost tests. This aproach leads to huge fragmentation of memory space and caused issues to testbeds.
  • with ubuntu 16.04.1:
   - ansible .yaml:
         - name: copy sysctl file
           template: src=files/90-csit dest=/etc/sysctl.d/90-csit.conf owner=root group=root mode=644
   - systcl file content - csit resources/tools/testbed-setup/playbooks/files/90-csit
   # change the minimum size of the hugepage pool.
   vm.nr_hugepages=4096
   
   # maximum number of memory map areas a process
   # may have. memory map areas are used as a side-effect of calling
   # malloc, directly by mmap and mprotect, and also when loading shared
   # libraries.
   # while most applications need less than a thousand maps, certain
   # programs, particularly malloc debuggers, may consume lots of them,
   # e.g., up to one or two maps per allocation.
   # must be greater than or equal to (2 * vm.nr_hugepages).
   vm.max_map_count=200000
   
   # hugetlb_shm_group contains group id that is allowed to create sysv
   # shared memory segment using hugetlb page.
   vm.hugetlb_shm_group=0
   
   # this control is used to define how aggressive the kernel will swap
   # memory pages.  higher values will increase agressiveness, lower values
   # decrease the amount of swap.  a value of 0 instructs the kernel not to
   # initiate swap until the amount of free and file-backed pages is less
   # than the high water mark in a zone.
   vm.swappiness=0
   
   # shared memory max must be greator or equal to the total size of hugepages.
   # for 2mb pages, totalhugepagesize = vm.nr_hugepages * 2 * 1024 * 1024
   # if the existing kernel.shmmax setting  (cat /sys/proc/kernel/shmmax)
   # is greater than the calculated totalhugepagesize then set this parameter
   # to current shmmax value.
   kernel.shmmax=8589934592
   
   # this option can be used to select the type of process address
   # space randomization that is used in the system, for architectures
   # that support this feature.
   # 0 - turn the process address space randomization off.  this is the
   #     default for architectures that do not support this feature anyways,
   #     and kernels that are booted with the "norandmaps" parameter.
   kernel.randomize_va_space=0
   
   # this value can be used to control on which cpus the watchdog may run.
   # the default cpumask is all possible cores, but if no_hz_full is
   # enabled in the kernel config, and cores are specified with the
   # nohz_full= boot argument, those cores are excluded by default.
   # offline cores can be included in this mask, and if the core is later
   # brought online, the watchdog will be started based on the mask value.
   #
   # typically this value would only be touched in the nohz_full case
   # to re-enable cores that by default were not running the watchdog,
   # if a kernel lockup was suspected on those cores.
   kernel.watchdog_cpumask=0,18

Host CFS optimizations (QEMU+VPP)

Applying CFS scheduler tuning on all Qemu vcpu worker threads (those are handling testpmd - pmd threads) and VPP PMD worker threads. List of VPP PMD threads can be obtained either from:

 "cat /proc/`pidof vpp`/task/*/stat | awk '{print $1" "$2" "$39}'"
 $ chrt -r -p 1 <worker_pid>


Host IRQ optimizations

Changing the default pinning of IRQ to core 0. (Same does apply on both guest and host OS)

 $ for l in `ls /proc/irq`; do echo 1 | sudo tee /proc/irq/$l/smp_affinity; done


Host RCU optimizations

Changing the default pinning of RCU to core 0. (Same does apply on both guest and host OS)

 $ for i in `pgrep rcu[^c]` ; do sudo taskset -pc 0 $i ; done


Host Writeback affinity

Changing the default pinning of writebacks to core 0. (Same does apply on both guest and host OS)

 $ echo 1 | sudo tee /sys/bus/workqueue/devices/writeback/cpumask

For more information please follow: https://www.kernel.org/doc/Documentation/kernel-per-CPU-kthreads.txt

--- END