1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358
|
.. SPDX-License-Identifier: BSD-3-Clause
Copyright 2016 Red Hat, Inc.
PVP reference benchmark setup using testpmd
===========================================
This guide lists the steps required to setup a PVP benchmark using testpmd as
a simple forwarder between NICs and Vhost interfaces. The goal of this setup
is to have a reference PVP benchmark without using external vSwitches (OVS,
VPP, ...) to make it easier to obtain reproducible results and to facilitate
continuous integration testing.
The guide covers two ways of launching the VM, either by directly calling the
QEMU command line, or by relying on libvirt. It has been tested with DPDK
v16.11 using RHEL7 for both host and guest.
Setup overview
--------------
.. _figure_pvp_2nics:
.. figure:: img/pvp_2nics.*
PVP setup using 2 NICs
In this diagram, each red arrow represents one logical core. This use case
requires 6 dedicated logical cores. A forwarding configuration with a single
NIC is also possible, requiring 3 logical cores.
Host setup
----------
In this setup, we isolate 6 cores (from CPU2 to CPU7) on the same NUMA
node. Two cores are assigned to the VM vCPUs running testpmd and four are
assigned to testpmd on the host.
Host tuning
~~~~~~~~~~~
#. On BIOS, disable turbo-boost and hyper-threads.
#. Append these options to Kernel command line:
.. code-block:: console
intel_pstate=disable mce=ignore_ce default_hugepagesz=1G hugepagesz=1G hugepages=6 isolcpus=2-7 nohz_full=2-7 iommu=pt intel_iommu=on
#. Disable hyper-threads at runtime if necessary or if BIOS is not accessible:
.. code-block:: console
cat /sys/devices/system/cpu/cpu*[0-9]/topology/thread_siblings_list \
| sort | uniq \
| awk -F, '{system("echo 0 > /sys/devices/system/cpu/cpu"$2"/online")}'
#. Disable NMIs:
.. code-block:: console
echo 0 > /proc/sys/kernel/nmi_watchdog
#. Exclude isolated CPUs from the writeback cpumask:
.. code-block:: console
echo ffffff03 > /sys/bus/workqueue/devices/writeback/cpumask
#. Isolate CPUs from IRQs:
.. code-block:: console
clear_mask=0xfc #Isolate CPU2 to CPU7 from IRQs
for i in /proc/irq/*/smp_affinity
do
echo "obase=16;$(( 0x$(cat $i) & ~$clear_mask ))" | bc > $i
done
Qemu build
~~~~~~~~~~
Build Qemu:
.. code-block:: console
git clone git://git.qemu.org/qemu.git
cd qemu
mkdir bin
cd bin
../configure --target-list=x86_64-softmmu
make
DPDK build
~~~~~~~~~~
See :doc:`../linux_gsg/build_dpdk` for details.
Testpmd launch
~~~~~~~~~~~~~~
#. Assign NICs to DPDK:
.. code-block:: console
modprobe vfio-pci
usertools/dpdk-devbind -b vfio-pci 0000:11:00.0 0000:11:00.1
.. Note::
The Sandy Bridge family seems to have some IOMMU limitations giving poor
performance results. To achieve good performance on these machines
consider using UIO instead.
#. Launch the testpmd application:
.. code-block:: console
<build_dir>/app/dpdk-testpmd -l 0,2,3,4,5 --socket-mem=1024 -n 4 \
--vdev 'net_vhost0,iface=/tmp/vhost-user1' \
--vdev 'net_vhost1,iface=/tmp/vhost-user2' -- \
--portmask=f -i --rxq=1 --txq=1 \
--nb-cores=4 --forward-mode=io
With this command, isolated CPUs 2 to 5 will be used as lcores for PMD threads.
#. In testpmd interactive mode, set the portlist to obtain the correct port
chaining:
.. code-block:: console
set portlist 0,2,1,3
start
VM launch
~~~~~~~~~
The VM may be launched either by calling QEMU directly, or by using libvirt.
Qemu way
^^^^^^^^
Launch QEMU with two Virtio-net devices paired to the vhost-user sockets
created by testpmd. Below example uses default Virtio-net options, but options
may be specified, for example to disable mergeable buffers or indirect
descriptors.
.. code-block:: console
<QEMU path>/bin/x86_64-softmmu/qemu-system-x86_64 \
-enable-kvm -cpu host -m 3072 -smp 3 \
-chardev socket,id=char0,path=/tmp/vhost-user1 \
-netdev type=vhost-user,id=mynet1,chardev=char0,vhostforce \
-device virtio-net-pci,netdev=mynet1,mac=52:54:00:02:d9:01,addr=0x10 \
-chardev socket,id=char1,path=/tmp/vhost-user2 \
-netdev type=vhost-user,id=mynet2,chardev=char1,vhostforce \
-device virtio-net-pci,netdev=mynet2,mac=52:54:00:02:d9:02,addr=0x11 \
-object memory-backend-file,id=mem,size=3072M,mem-path=/dev/hugepages,share=on \
-numa node,memdev=mem -mem-prealloc \
-net user,hostfwd=tcp::1002$1-:22 -net nic \
-qmp unix:/tmp/qmp.socket,server,nowait \
-monitor stdio <vm_image>.qcow2
You can use this `qmp-vcpu-pin <https://patchwork.kernel.org/patch/9361617/>`_
script to pin vCPUs.
It can be used as follows, for example to pin 3 vCPUs to CPUs 1, 6 and 7,
where isolated CPUs 6 and 7 will be used as lcores for Virtio PMDs:
.. code-block:: console
export PYTHONPATH=$PYTHONPATH:<QEMU path>/scripts/qmp
./qmp-vcpu-pin -s /tmp/qmp.socket 1 6 7
Libvirt way
^^^^^^^^^^^
Some initial steps are required for libvirt to be able to connect to testpmd's
sockets.
First, SELinux policy needs to be set to permissive, since testpmd is
generally run as root (note, as reboot is required):
.. code-block:: console
cat /etc/selinux/config
# This file controls the state of SELinux on the system.
# SELINUX= can take one of these three values:
# enforcing - SELinux security policy is enforced.
# permissive - SELinux prints warnings instead of enforcing.
# disabled - No SELinux policy is loaded.
SELINUX=permissive
# SELINUXTYPE= can take one of three two values:
# targeted - Targeted processes are protected,
# minimum - Modification of targeted policy.
# Only selected processes are protected.
# mls - Multi Level Security protection.
SELINUXTYPE=targeted
Also, Qemu needs to be run as root, which has to be specified in
``/etc/libvirt/qemu.conf``:
.. code-block:: console
user = "root"
Once the domain created, the following snippet is an extract of he most
important information (hugepages, vCPU pinning, Virtio PCI devices):
.. code-block:: xml
<domain type='kvm'>
<memory unit='KiB'>3145728</memory>
<currentMemory unit='KiB'>3145728</currentMemory>
<memoryBacking>
<hugepages>
<page size='1048576' unit='KiB' nodeset='0'/>
</hugepages>
<locked/>
</memoryBacking>
<vcpu placement='static'>3</vcpu>
<cputune>
<vcpupin vcpu='0' cpuset='1'/>
<vcpupin vcpu='1' cpuset='6'/>
<vcpupin vcpu='2' cpuset='7'/>
<emulatorpin cpuset='0'/>
</cputune>
<numatune>
<memory mode='strict' nodeset='0'/>
</numatune>
<os>
<type arch='x86_64' machine='pc-i440fx-rhel7.0.0'>hvm</type>
<boot dev='hd'/>
</os>
<cpu mode='host-passthrough'>
<topology sockets='1' cores='3' threads='1'/>
<numa>
<cell id='0' cpus='0-2' memory='3145728' unit='KiB' memAccess='shared'/>
</numa>
</cpu>
<devices>
<interface type='vhostuser'>
<mac address='56:48:4f:53:54:01'/>
<source type='unix' path='/tmp/vhost-user1' mode='client'/>
<model type='virtio'/>
<driver name='vhost' rx_queue_size='256' />
<address type='pci' domain='0x0000' bus='0x00' slot='0x10' function='0x0'/>
</interface>
<interface type='vhostuser'>
<mac address='56:48:4f:53:54:02'/>
<source type='unix' path='/tmp/vhost-user2' mode='client'/>
<model type='virtio'/>
<driver name='vhost' rx_queue_size='256' />
<address type='pci' domain='0x0000' bus='0x00' slot='0x11' function='0x0'/>
</interface>
</devices>
</domain>
Guest setup
-----------
Guest tuning
~~~~~~~~~~~~
#. Append these options to the Kernel command line:
.. code-block:: console
default_hugepagesz=1G hugepagesz=1G hugepages=1 intel_iommu=on iommu=pt isolcpus=1,2 nohz_full=1,2
#. Disable NMIs:
.. code-block:: console
echo 0 > /proc/sys/kernel/nmi_watchdog
#. Exclude isolated CPU1 and CPU2 from the writeback cpumask:
.. code-block:: console
echo 1 > /sys/bus/workqueue/devices/writeback/cpumask
#. Isolate CPUs from IRQs:
.. code-block:: console
clear_mask=0x6 #Isolate CPU1 and CPU2 from IRQs
for i in /proc/irq/*/smp_affinity
do
echo "obase=16;$(( 0x$(cat $i) & ~$clear_mask ))" | bc > $i
done
DPDK build
~~~~~~~~~~
See :doc:`../linux_gsg/build_dpdk` for details.
Testpmd launch
~~~~~~~~~~~~~~
Probe vfio module without iommu:
.. code-block:: console
modprobe -r vfio_iommu_type1
modprobe -r vfio
modprobe vfio enable_unsafe_noiommu_mode=1
cat /sys/module/vfio/parameters/enable_unsafe_noiommu_mode
modprobe vfio-pci
Bind the virtio-net devices to DPDK:
.. code-block:: console
usertools/dpdk-devbind.py -b vfio-pci 0000:00:10.0 0000:00:11.0
Start testpmd:
.. code-block:: console
<build_dir>/app/dpdk-testpmd -l 0,1,2 --socket-mem 1024 -n 4 \
--proc-type auto --file-prefix pg -- \
--portmask=3 --forward-mode=macswap --port-topology=chained \
--disable-rss -i --rxq=1 --txq=1 \
--rxd=256 --txd=256 --nb-cores=2 --auto-start
Results template
----------------
Below template should be used when sharing results:
.. code-block:: none
Traffic Generator: <Test equipment (e.g. IXIA, Moongen, ...)>
Acceptable Loss: <n>%
Validation run time: <n>min
Host DPDK version/commit: <version, SHA-1>
Guest DPDK version/commit: <version, SHA-1>
Patches applied: <link to patchwork>
QEMU version/commit: <version>
Virtio features: <features (e.g. mrg_rxbuf='off', leave empty if default)>
CPU: <CPU model>, <CPU frequency>
NIC: <NIC model>
Result: <n> Mpps
|