1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
|
.. This work is licensed under a Creative Commons Attribution 4.0 International License.
.. http://creativecommons.org/licenses/by/4.0
Fast Live Migration
===================
The NFV project requires fast live migration. The specific requirement is
total live migration time < 2Sec, while keeping the VM down time < 10ms when
running DPDK L2 forwarding workload.
We measured the baseline data of migrating an idle 8GiB guest running a DPDK L2
forwarding work load and observed that the total live migration time was 2271ms
while the VM downtime was 26ms. Both of these two indicators failed to satisfy
the requirements.
Current Challenges
------------------
The following 4 features have been developed over the years to make the live
migration process faster.
+ XBZRLE:
Helps to reduce the network traffic by just sending the
compressed data.
+ RDMA:
Uses a specific NIC to increase the efficiency of data
transmission.
+ Multi thread compression:
Compresses the data before transmission.
+ Auto convergence:
Reduces the data rate of dirty pages.
Tests show none of the above features can satisfy the requirement of NFV.
XBZRLE and Multi thread compression do the compression entirely in software and
they are not fast enough in a 10Gbps network environment. RDMA is not flexible
because it has to transport all the guest memory to the destination without zero
page optimization. Auto convergence is not appropriate for NFV because it will
impact guest’s performance.
So we need to find other ways for optimization.
Optimizations
-------------------------
a. Delay non-emergency operations
By profiling, it was discovered that some of the cleanup operations during
the stop and copy stage are the main reason for the long VM down time. The
cleanup operation includes stopping the dirty page logging, which is a time
consuming operation. By deferring these operations until the data transmission
is completed the VM down time is reduced to about 5-7ms.
b. Optimize zero page checking
Currently QEMU uses the SSE2 instruction to optimize the zero pages
checking. The SSE2 instruction can process 16 bytes per instruction.
By using the AVX2 instruction, we can process 32 bytes per instruction.
Testing shows that using AVX2 can speed up the zero pages checking process
by about 25%.
c. Remove unnecessary context synchronization.
The CPU context was being synchronized twice during live migration. Removing
this unnecessary synchronization shortened the VM downtime by about 100us.
Test Environment
----------------
The source and destination host have the same hardware and OS:
::
Host: HSW-EP
CPU: Intel(R) Xeon(R) CPU E5-2699 v3 @ 2.30GHz
RAM: 64G
OS: RHEL 7.1
Kernel: 4.2
QEMU v2.4.0
Ethernet controller: Intel Corporation Ethernet Controller 10-Gigabit
X540-AT2 (rev 01)
Vhost-user with OVS/DPDK as backend:
::
The goal is to connect guests' virtio-net devices having vhost-user backend to OVS dpdkvhostuser
ports and be able to run any kind of network traffic between them.
Installation of OVS and DPDK:
::
Using vsperf,installing the OVS and DPDk. Prepare the directories
.. code:: bash
mkdir -p /var/run/openvswitch
mount -t hugetlbfs -o pagesize=2048k none /dev/hugepages
Load Kernel modules
.. code:: bash
modprobe openvswitch
For OVS setup, clean the environment
.. code:: bash
rm -f /usr/local/var/run/openvswitch/vhost-user*
rm -f /usr/local/etc/openvswitch/conf.db
Start database server
.. code:: bash
ovsdb-tool create /usr/local/etc/openvswitch/conf.db $VSPERF/src/ovs/ovs/vswitchd/vswitch.ovsschema
ovsdb-server --remote=punix:$DB_SOCK --remote=db:Open_vSwitch,Open_vSwitch,manager_options --pidfile --detach
Start OVS
.. code:: bash
ovs-vsctl --no-wait init
ovs-vsctl --no-wait set Open_vSwitch . other_config:dpdk-lcore-mask=0xf
ovs-vsctl --no-wait set Open_vSwitch . other_config:dpdk-socket-mem=1024
ovs-vsctl --no-wait set Open_vSwitch . other_config:dpdk-init=true
Configure the bridge
.. code:: bash
ovs-vsctl add-br ovsbr0 -- set bridge ovsbr0 datapath_type=netdev
ovs-vsctl add-port ovsbr0 vhost-user1 -- set Interface vhost-user1 type=dpdkvhostuser
ovs-vsctl add-port ovsbr0 vhost-user2 -- set Interface vhost-user2 type=dpdkvhostuser
QEMU parameters:
::
qemu-system-x86_64 -enable-kvm -cpu host -smp 2
-chardev socket,id=char1,path=/usr/local/var/run/openvswitch/vhost-user1
-netdev type=vhost-user,id=net1,chardev=char1,vhostforce \
-device virtio-net-pci,netdev=net1,mac=52:54:00:12:34:56 \
-chardev socket,id=char2,path=/usr/local/var/run/openvswitch/vhost-user2\
-netdev type=vhost-user,id=net2,chardev=char2,vhostforce \
-device virtio-net-pci,netdev=net2,mac=54:54:00:12:34:56 -m 1024 -mem-path /dev/hugepages \
-mem-prealloc -realtime mlock=on -monitor unix:/tmp/qmp-sock-src,server,nowait \
-balloon virtio -drive file=/root/guest1.qcow2 -vnc :1 &
Run the standby qemu with -incoming tcp:-incoming tcp:${incoming_ip}:${migrate_port}:${migrate_port}
For local live migration
.. code:: bash
incoming ip=0
For peer -peer live migration
.. code:: bash
incoming ip=dest_host
Network connection
.. figure:: lmnetwork.jpg
:align: center
:alt: live migration network connection
:figwidth: 80%
Commands for performing live migration:
::
.. code:: bash
echo "migrate_set_speed 0" |nc -U /tmp/qmp-sock-src
echo "migrate_set_downtime 0.10" |nc -U /tmp/qmp-sock-src
echo "migrate -d tcp:0:4444" |nc -U /tmp/qmp-sock-src
#Wait till livemigration completed
echo "info migrate" | nc -U /tmp/qmp-sock-src
Test Result
-----------
The down time is set to 10ms when doing the test. We use pktgen to send the
packages to guest, the package size is 64 bytes, and the line rate is 2013
Mbps.
a. Total live migration time
The total live migration time before and after optimization is shown in the
chart below. For an idle guest, we can reduce the total live migration time
from 2070ms to 401ms. For a guest running the DPDK L2 forwarding workload,
the total live migration time is reduced from 2271ms to 654ms.
.. figure:: lmtotaltime.jpg
:align: center
:alt: total live migration time
b. VM downtime
The VM down time before and after optimization is shown in the chart below.
For an idle guest, we can reduce the VM down time from 29ms to 9ms. For a guest
running the DPDK L2 forwarding workload, the VM down time is reduced from 26ms to
5ms.
.. figure:: lmdowntime.jpg
:align: center
:alt: vm downtime
:figwidth: 80%
|