Single-core CPU performance cannot meet the requirement of processing NIC interruptions incurred with the increase of network I/O bandwidth. NIC multi-queue enables multiple CPUs to process ECS NIC interruptions, thereby improving PPS and I/O performance.
The ECS described in this section is assumed to comply with the requirements on specifications and virtualization type.
NIC multi-queue can be enabled on an ECS only when the ECS specifications, virtualization type, and image OS meet the requirements described in this section.
If the number of NIC queues is greater than 1, NIC multi-queue is supported.
Run the uname -r command to obtain the kernel version. If the kernel version is earlier than 2.6.35, contact customer service to upgrade the kernel.
Image |
Support of NIC Multi-Queue |
NIC Multi-Queue Enabled by Default |
---|---|---|
Ubuntu 14.04/16.04/18.04/20.04 server 64bit |
Yes |
Yes |
OpenSUSE 42.2/15.* 64bit |
Yes |
Yes |
SUSE Enterprise 12 SP1/SP2 64bit |
Yes |
Yes |
CentOS 6.8/6.9/7.*/8.* 64bit |
Yes |
Yes |
Debian 8.0.0/8.8.0/8.9.0/9.0.0/10.0.0/10.2.0 64bit |
Yes |
Yes |
Fedora 24/25/30 64bit |
Yes |
Yes |
EulerOS 2.2/2.3/2.5 64bit |
Yes |
Yes |
OS |
Image |
Status |
---|---|---|
Windows |
Windows Server 2008 Web R2 64-bit |
Supported using private images |
Windows Server 2008 R2 Standard/DataCenter/Enterprise 64bit |
Supported using private images |
|
Windows Server 2012 R2 Standard/DataCenter 64bit |
Supported using private images |
|
Windows Server 2016 Standard/DataCenter 64bit |
Supported using private images |
|
Linux |
Ubuntu 14.04/16.04 server 64bit |
Supported |
OpenSUSE 42.2 64bit |
Supported |
|
SUSE Enterprise 12 SP1/SP2 64bit |
Supported |
|
CentOS 6.8/6.9/7.0/7.1/7.2/7.3/7.4/7.5/7.6 64bit |
Supported |
|
Debian 8.0.0/8.8.0/8.9.0/9.0.0 64bit |
Supported |
|
Fedora 24/25 64bit |
Supported |
|
EulerOS 2.2 64bit |
Supported |
For details, see "Registering an Image File as a Private Image" in Image Management Service User Guide.
Windows OSs have not commercially supported NIC multi-queue. If you enable NIC multi-queue in a Windows image, starting an ECS created using such an image may be slow.
Use one of the following methods to set the NIC multi-queue attribute:
Method 3: Add hw_vif_multiqueue_enabled to an image through the API.
The value of X-Auth-Token is the token obtained in step 1.
The value of Content-Type is application/openstack-images-v2.1-json-patch.
The request URI is in the following format:
PATCH /v2/images/{image_id}
[ { "op":"add", "path":"/hw_vif_multiqueue_enabled", "value": "true" } ]
Figure 1 shows an example request body for modifying the NIC multi-queue attribute.
KVM Windows ECSs use private images to support NIC multi-queue. For details, see "How Do I Set NIC Multi-queue Feature of an Image?" in Image Management Service User Guide.
This section uses a Linux ECS running CentOS 7.4 as an example to describe how to enable NIC multi-queue.
An example is provided as follows:
[root@localhost ~]# ethtool -l eth0 #View the number of queues used by NIC eth0. Channel parameters for eth0: Pre-set maximums: RX: 0 TX: 0 Other: 0 Combined: 4 #Indicates that a maximum of four queues can be enabled for the NIC. Current hardware settings: RX: 0 TX: 0 Other: 0 Combined: 1 #Indicates that one queue has been enabled. [root@localhost ~]# ethtool -L eth0 combined 4 #Enable four queues on NIC eth0.
Enabling irqbalance allows the system to automatically allocate NIC interrupts, improving network performance. If the improved network performance still fails to meet your requirements, manually configure interrupt affinity on the ECS.
To do so, perform the following operations:
Configure the following script so that one ECS vCPU serves the interrupt requests initialized by one queue. One queue corresponds to one interrupt, and one interrupt binds to one vCPU.
#!/bin/bash service irqbalance stop eth_dirs=$(ls -d /sys/class/net/eth*) if [ $? -ne 0 ];then echo "Failed to find eth* , sleep 30" >> $ecs_network_log sleep 30 eth_dirs=$(ls -d /sys/class/net/eth*) fi for eth in $eth_dirs do cur_eth=$(basename $eth) cpu_count=`cat /proc/cpuinfo| grep "processor"| wc -l` virtio_name=$(ls -l /sys/class/net/"$cur_eth"/device/driver/ | grep pci |awk {'print $9'}) affinity_cpu=0 virtio_input="$virtio_name""-input" irqs_in=$(grep "$virtio_input" /proc/interrupts | awk -F ":" '{print $1}') for irq in ${irqs_in[*]} do echo $((affinity_cpu%cpu_count)) > /proc/irq/"$irq"/smp_affinity_list affinity_cpu=$[affinity_cpu+2] done affinity_cpu=1 virtio_output="$virtio_name""-output" irqs_out=$(grep "$virtio_output" /proc/interrupts | awk -F ":" '{print $1}') for irq in ${irqs_out[*]} do echo $((affinity_cpu%cpu_count)) > /proc/irq/"$irq"/smp_affinity_list affinity_cpu=$[affinity_cpu+2] done done
XPS allows the system with NIC multi-queue enabled to select a queue by vCPU when sending a data packet.
#!/bin/bash # enable XPS feature cpu_count=$(grep -c processor /proc/cpuinfo) dec2hex(){ echo $(printf "%x" $1) } eth_dirs=$(ls -d /sys/class/net/eth*) if [ $? -ne 0 ];then echo "Failed to find eth* , sleep 30" >> $ecs_network_log sleep 30 eth_dirs=$(ls -d /sys/class/net/eth*) fi for eth in $eth_dirs do cpu_id=1 cur_eth=$(basename $eth) cur_q_num=$(ethtool -l $cur_eth | grep -iA5 current | grep -i combined | awk {'print $2'}) for((i=0;i<cur_q_num;i++)) do if [ $i -eq $cpu_count ];then cpu_id=1 fi xps_file="/sys/class/net/${cur_eth}/queues/tx-$i/xps_cpus" rps_file="/sys/class/net/${cur_eth}/queues/rx-$i/rps_cpus" cpuset=$(dec2hex "$cpu_id") echo $cpuset > $xps_file echo $cpuset > $rps_file let cpu_id=cpu_id*2 done done