Changes

Jump to navigation Jump to search

Performance Optimization

5,475 bytes added, 02:56, 11 September 2021
<span style="color: red;">'''Note: When hugepages are configured this portion of memory is taken away from the Host. This means the Host will no longer be able to use it. Keep this in mind.'''</span>
=== Debain Debian (Ubuntu/Mint/Lubuntu/PopOS/etc) ===First check if Linux isn't already using Hugepages with:<code>cat /proc/meminfo | grep Huge</code>.If the output resembles the following:<syntaxhighlight lang=text>AnonHugePages: 2048 kBShmemHugePages: 0 kBHugePages_Total: 0HugePages_Free: 0HugePages_Rsvd: 0HugePages_Surp: 0Hugepagesize: 2048 kBHugetlb: 0 kB</syntaxhighlight>then Hugepages aren't enabled. To enable Hugepages first check <code>/etc/sysctl.conf</code> for the following entries:<syntaxhighlight lang=text>vm.nr_hugepages=vm.hugetlb_shm_group=</syntaxhighlight>If they don't exist they can be appended to the end of the file. The general rule of thumb is 1 Hugepage for every 2MB of of RAM to be assigned to the VM:<syntaxhighlight lang=text>vm.nr_hugepages=8192vm.hugetlb_shm_group=48</syntaxhighlight>The above example will set aside approx 16GB. After saving the changes reboot the system. Now to verify the changes rerun: <code>cat /proc/meminfo | grep Huge</code>The output should resemble the following:<syntaxhighlight lang=text>AnonHugePages: 0 kBShmemHugePages: 0 kBHugePages_Total: 8192HugePages_Free: 8192HugePages_Rsvd: 0HugePages_Surp: 0Hugepagesize: 2048 kBHugetlb: 16777216 kB</syntaxhighlight>Hugepages are now enabled. It's also recommended by RedHat to disable the old transparent hugepages. This can be done with:<syntaxhighlight lang=text>echo 'never' > /sys/kernel/mm/transparent_hugepage/defragecho 'never' > /sys/kernel/mm/transparent_hugepage/enabled</syntaxhighlight>Now restart the computer. === Assigning Hugepages to VM === To make the VM use Hugepages enter the VM's .XML file and add <code><memoryBacking><hugepages/></memoryBacking></code> to the memory section:<syntaxhighlight lang=text>...<memory unit='KiB'>16777216</memory> <currentMemory unit='KiB'>16777216</currentMemory> <memoryBacking> <hugepages/> </memoryBacking>...</syntaxhighlight>Hugepages are now enabled.
== hyperv ==
Hyperv has a number of variables that can change the way in which the VM interacts with system resources. A few variables that can help preserve resources for the system are by appending:
<syntaxhighlight lang=text>
<vpindex state='on'/>
<runtime state='on'/>
<synic state='on'/>
<stimer state='on'/>
</syntaxhighlight>
to the hyperv section of the VM's .XML file.
<syntaxhighlight lang=text>
...
<hyperv>
<related state='on'/>
<vapic state='on'/>
<spinlocks state='on' retries='8191'/>
<vpindex state='on'/>
<runtime state='on'/>
<synic state='on'/>
<stimer state='on'/>
...
</syntaxhighlight>
This edit doesn't help the VM perform better as much as it helps preserve system resources if the plan is to run multiple simultaneous instances.
== vcpupin ==
vcpupin is the process where-in each vCPU assigned to the VM is tied to a physical core/thread. Configuring this has the most profound impact when dealing with a system that has multiple NUMA Nodes because it forces requests to memory to stay on one node without having to cross Intel's QPI links or AMD's Infinity Fabric. This also helps by tying the vCPUs to the Node that is directly connected to the GPU.
 
<span style="color: red;">'''NOTE: In many cases a function in the BIOS known as Memory Interleave will obfuscate the NUMA Nodes making the system treat multiple sockets or multiple dies as one UMA(Uniform Memory Access) Node. This is a problem for multiple reasons. To fix this find Memory Interleave in your BIOS and set it from Auto -> Channel. If Memory Interleave does not have these options on a single socket system then chances are the system only operates in UMA mode. In such case this is fine.'''</span>
 
=== Identifying CPU Affinity ===
Their are a couple of options available to determine what PCI_e device and CPU threads are connected to which NUMA Node. On Debian <code>lscpu</code> & <code>lspci -vnn</code> can both be used to determine which node a PCI_e device is connected to and which threads belong to that node.
 
Another option is lstopo (the <code>hwloc</code> package). This application provides a GUI overview of the cores/threads and what PCI_e device(s) are connected to which.
 
=== Assigning CPU Affinity ===
To tie cores/threads to a VM open the VM's .XML file and find the <code>vcpu placement</code> section. Beneath this you will append <code>cputune</code> and modify it's configuration:
<syntaxhighlight lang=text>
<vcpu placement='static'>16</vcpu>
<cputune>
<vcpupin vcpu='0' cpuset='0'/>
<vcpupin vcpu='1' cpuset='1'/>
<vcpupin vcpu='2' cpuset='2'/>
<vcpupin vcpu='3' cpuset='3'/>
<vcpupin vcpu='4' cpuset='8'/>
<vcpupin vcpu='5' cpuset='9'/>
<vcpupin vcpu='6' cpuset='10'/>
<vcpupin vcpu='7' cpuset='11'/>
</cputune>
</syntaxhighlight>
CPU tune works on a per thread basis. If NUMA Node 0 or NUMA Node 1 were threads 0-3,8-11 then you would assign these to the VM's vCPU's with the above example.
 
=== Verifying CPU Affinity ===
To verify if the CPU threads have been successfully pinned to the VM the <code>virsh vcpupin name-of-vm</code> command can be used. The output for a VM with 8 vCPUs configured as above would look like this:
<syntaxhighlight lang=text>
VCPU CPU Affinity
----------------------
0 0
1 1
2 2
3 3
4 8
5 9
6 10
7 11
</syntaxhighlight>
In the event that your system only uses UMA(Uniform Memory Access) the impact of pinning threads is definitely less pronounced but some benefit can still be gained.
Anonymous user

Navigation menu