【DPDK】虚拟机开发环境配置

正文 本文介绍的步骤基本适用于dpdk 1.7.0 - dpdk 2.0.0 各版本。只是setup.sh显示的菜单有一些小的不同;同样的,也适用于ubuntu更高版本(已在ubuntu 12.04+及14.04上验证过)

1. 所用系统与软件版本
系统:Ubuntu 12.04.3 LTS 64位, CentOS Linux release 7.0.1406 64位
dpdk: 1.7.0,http://www.dpdk.org/browse/dpdk/refs/。

dpdk 1.7.1 经过试验,发现在这两个系统上都有问题, 运行各示例程序都有以下错误
EAL: Error reading from file descriptor

这个bug已经由dpdk的开发人员修复,patch内容如下:

diff --gita/lib/librte_eal/linuxapp/igb_uio/igb_uio.cb/lib/librte_eal/linuxapp/igb_uio/igb_uio.c
index d1ca26e..c46a00f 100644
--- a/lib/librte_eal/linuxapp/igb_uio/igb_uio.c
+++b/lib/librte_eal/linuxapp/igb_uio/igb_uio.c
@@ -505,14 +505,11 @@igbuio_pci_probe(struct pci_dev *dev, const struct pci_device_id *id)
}
/* fall back to INTX */
case RTE_INTR_MODE_LEGACY:
-if (pci_intx_mask_supported(dev)) {
-dev_dbg(&dev->dev, "using INTX");
-udev->info.irq_flags = IRQF_SHARED;
-udev->info.irq = dev->irq;
-udev->mode = RTE_INTR_MODE_LEGACY;
-break;
-}
-dev_notice(&dev->dev, "PCI INTX mask not supported\n");
+dev_dbg(&dev->dev, "using INTX");
+udev->info.irq_flags = IRQF_SHARED;
+udev->info.irq = dev->irq;
+udev->mode = RTE_INTR_MODE_LEGACY;
+break;
/* fall back to no IRQ */
case RTE_INTR_MODE_NONE:
udev->mode = RTE_INTR_MODE_NONE;


在虚拟机里使用时,打上以上的补丁,或手工修改文件后重新编译即可。

2. 虚拟机配置
虚拟机软件:VMWare WorkStation 10.0.1 build-1379776
CPU: 2个CPU, 每个CPU2个核心
内存: 1GB+
网卡:intel网卡*2,用于dpdk试验;另一块网卡用于和宿主系统进行通信


3. Ubuntu 12.04上的配置
3.1 准备 需要安装gcc及其他一些小工具等,默认都有了,没有的话运行sudo apt-get install装一下。dkdk的一些脚本用到了python,也装一下。

3.2 通过setup脚本进行配置 首先运行su切换到root权限,root没有开的话使用sudo passwd roo,来开一下。
dpdk提供了一个方便的配置脚本:/tools/setup.sh,通过它可以方便地配置环境。


1)设置环境变量,这里是linux 64位的配置
export RTE_SDK=
exportRTE_TARGET=x86_64-native-linuxapp-gcc


2)运行setup.sh,显示如下
------------------------------------------------------------------------------
RTE_SDK exported as /home/hack/dpdk-1.7.0
------------------------------------------------------------------------------
----------------------------------------------------------
Step1: Select the DPDK environment to build
----------------------------------------------------------
[1] i686-native-linuxapp-gcc
[2] i686-native-linuxapp-icc
[3] x86_64-ivshmem-linuxapp-gcc
[4] x86_64-ivshmem-linuxapp-icc
[5] x86_64-native-bsdapp-gcc
[6] x86_64-native-linuxapp-gcc
[7] x86_64-native-linuxapp-icc

----------------------------------------------------------
Step2: Setup linuxapp environment
----------------------------------------------------------
[8] Insert IGB UIO module
[9] Insert VFIO module
[10] Insert KNI module
[11] Setup hugepage mappings for non-NUMAsystems
[12] Setup hugepage mappings for NUMAsystems
[13] Display current Ethernet devicesettings
[14] Bind Ethernet device to IGB UIO module
[15] Bind Ethernet device to VFIO module
[16] Setup VFIO permissions

----------------------------------------------------------
Step3: Run test application for linuxapp environment
----------------------------------------------------------
[17] Run test application($RTE_TARGET/app/test)
[18] Run testpmd application in interactivemode ($RTE_TARGET/app/testpmd)

【【DPDK】虚拟机开发环境配置】----------------------------------------------------------
Step4: Other tools
----------------------------------------------------------
[19] List hugepage info from /proc/meminfo

----------------------------------------------------------
Step5: Uninstall and system cleanup
----------------------------------------------------------
[20] Uninstall all targets
[21] Unbind NICs from IGB UIO driver
[22] Remove IGB UIO module
[23] Remove VFIO module
[24] Remove KNI module
[25] Remove hugepage mappings

[26] Exit Script
复制代码
选择6,进行编译

3)选择8, 插入igb_uio模块

4)选择11,配置大页内存(非NUMA),选择后会提示你选择页数,输入64,128什么的即可

Removing currently reserved hugepages
Unmounting /mnt/huge and removing directory

Input the number of 2MB pages
Example: to have 128MB of hugepages available, enter '64' to
reserve 64 * 2MB pages
Number of pages: 128

选择19,可以确认一下大页内存的配置:

AnonHugePages:0 kB
HugePages_Total:128
HugePages_Free:128
HugePages_Rsvd:0
HugePages_Surp:0
Hugepagesize:2048 kB

5)选择14, 绑定dpdk要使用的网卡

Network devices using DPDK-compatibledriver
============================================


Network devices using kernel driver
===================================
0000:02:01.0 '82545EM Gigabit EthernetController (Copper)' if=eth0 drv=e1000 unused=igb_uio *Active*
0000:02:06.0 '82545EM Gigabit EthernetController (Copper)' if=eth1 drv=e1000 unused=igb_uio
0000:02:07.0 '82545EM Gigabit EthernetController (Copper)' if=eth2 drv=e1000 unused=igb_uio

Other network devices
=====================


Enter PCI address of device to bind to IGBUIO driver: 0000:02:06.0

绑定好后,选择13,可以查看当前的网卡配置:

Network devices using DPDK-compatibledriver
============================================
0000:02:06.0 '82545EM Gigabit EthernetController (Copper)' drv=igb_uio unused=e1000
0000:02:07.0 '82545EM Gigabit EthernetController (Copper)' drv=igb_uio unused=e1000

Network devices using kernel driver
===================================
0000:02:01.0 '82545EM Gigabit EthernetController (Copper)' if=eth0 drv=e1000 unused=igb_uio *Active*

Other network devices
=====================


6)选择18, 运行testpmd测试程序

注意,运行这个测试程序,虚拟机最好提供2个网卡用于dpdk。

Enter hex bitmask of cores to execute testpmd app on
Example: to execute app on cores 0 to 7, enter 0xff
bitmask: f
如果没问题,按回车后会出现以下输出:

Launching app
EAL: Detected lcore 0 as core 0 on socket 0
EAL: Detected lcore 1 as core 1 on socket 0
EAL: Detected lcore 2 as core 0 on socket 0
EAL: Detected lcore 3 as core 1 on socket 0
EAL: Support maximum 64 logical core(s) byconfiguration.
EAL: Detected 4 lcore(s)
EAL: Setting up memory...
EAL: Ask a virtual area of 0xf000000 bytes
EAL: Virtual area found at 0x7fe828000000(size = 0xf000000)
EAL: Ask a virtual area of 0x200000 bytes
EAL: Virtual area found at 0x7fe827c00000(size = 0x200000)
EAL: Ask a virtual area of 0x200000 bytes
EAL: Virtual area found at 0x7fe827800000(size = 0x200000)
EAL: Ask a virtual area of 0x800000 bytes
EAL: Virtual area found at 0x7fe826e00000(size = 0x800000)
EAL: Ask a virtual area of 0x400000 bytes
EAL: Virtual area found at 0x7fe826800000(size = 0x400000)
EAL: Requesting 128 pages of size 2MB fromsocket 0
EAL: TSC frequency is ~3292453 KHz
EAL: Master core 0 is ready (tid=37c79800)
EAL: Core 3 is ready (tid=24ffc700)
EAL: Core 2 is ready (tid=257fd700)
EAL: Core 1 is ready (tid=25ffe700)
EAL: PCI device 0000:02:01.0 on NUMA socket-1
EAL:probe driver: 8086:100f rte_em_pmd
EAL:0000:02:01.0 not managed by UIO driver, skipping
EAL: PCI device 0000:02:06.0 on NUMA socket-1
EAL:probe driver: 8086:100f rte_em_pmd
EAL:PCI memory mapped at 0x7fe837c23000
EAL:PCI memory mapped at 0x7fe837c13000
EAL: PCI device 0000:02:07.0 on NUMA socket-1
EAL:probe driver: 8086:100f rte_em_pmd
EAL:PCI memory mapped at 0x7fe837bf3000
EAL:PCI memory mapped at 0x7fe837be3000
Interactive-mode selected
Configuring Port 0 (socket 0)
Port 0: 00:0C:29:14:50:CE
Configuring Port 1 (socket 0)
Port 1: 00:0C:29:14:50:D8
Checking link statuses...
Port 0 Link Up - speed 1000 Mbps -full-duplex
Port 1 Link Up - speed 1000 Mbps -full-duplex
Done
testpmd>

输入start,开始包转发

testpmd> start
iopacket forwarding - CRC stripping disabled - packets/burst=32
nbforwarding cores=1 - nb forwarding ports=2
RXqueues=1 - RX desc=128 - RX free threshold=0
RXthreshold registers: pthresh=8 hthresh=8 wthresh=0
TXqueues=1 - TX desc=512 - TX free threshold=0
TXthreshold registers: pthresh=32 hthresh=0 wthresh=0
TXRS bit threshold=0 - TXQ flags=0x0

输入stop,停止包转发,这时会显示统计信息


testpmd> stop
Telling cores to stop...
Waiting for lcores to finish...

---------------------- Forward statistics for port 0----------------------
RX-packets: 5544832RX-dropped: 0RX-total: 5544832
TX-packets: 5544832TX-dropped: 0TX-total: 5544832
----------------------------------------------------------------------------

---------------------- Forward statistics for port 1----------------------
RX-packets: 5544832RX-dropped: 0RX-total: 5544832
TX-packets: 5544832TX-dropped: 0TX-total: 5544832
----------------------------------------------------------------------------

+++++++++++++++ Accumulated forward statistics for allports+++++++++++++++
RX-packets: 11089664RX-dropped: 0RX-total: 11089664
TX-packets: 11089664TX-dropped: 0TX-total: 11089664
++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++

Done.

3.3 通过命令配置 最好切换到root权限。

1)编译dpdk

进入dpdk主目录,输入

make install T=x86_64-native-linuxapp-gcc
进行编译

2)配置大页内存(非NUMA)

echo 128 >/sys/kernel/mm/hugepages/hugepages-2048kB/nr_hugepages
mkdir /mnt/huge
mount -t hugetlbfs nodev /mnt/huge
可以用以下命令查看大页内存状态:

cat /proc/meminfo | grep Huge

3)安装igb_uio驱动

modprobe uio
insmod x86_64-native-linuxapp-gcc/kmod/igb_uio.ko

4)绑定网卡

先看一下当前网卡的状态

./tools/dpdk_nic_bind.py --status

Network devices using DPDK-compatibledriver
============================================


Network devices using kernel driver
===================================
0000:02:01.0 '82545EM Gigabit EthernetController (Copper)' if=eth0 drv=e1000 unused=igb_uio *Active*

Other network devices
=====================
0000:02:06.0 '82545EM Gigabit EthernetController (Copper)' unused=e1000,igb_uio
0000:02:07.0 '82545EM Gigabit Ethernet Controller(Copper)' unused=e1000,igb_uio

进行绑定:

./tools/dpdk_nic_bind.py -b igb_uio0000:02:06.0
./tools/dpdk_nic_bind.py -b igb_uio0000:02:07.0
如果网卡有接口名,如eth1, eth2, 也可以在-b igb_uio后面使用接口名,而不使用pci地址。

5)运行testpmd测试程序

./x86_64-native-linuxapp-gcc/app/testpmd -c0x3 -n 2 -- -i

6)编译运行其他示例程序

/examples下面有很多示例程序,这些程序在dpdk编译时,没有被编译。这里以编译helloworld为例,首先要设置环境变量:

export RTE_SDK=
exportRTE_TARGET=x86_64-native-linuxapp-gcc
之后进入/examples/helloworld,运行make,成功会生成build目录,其中有编译好的helloworld程序。


4. CentOS 7.0上的配置
4.1 准备 安装CentOS虚拟机时,如果选择minimal安装,还需要安装其下的基本开发工具集(含gcc,python等)

另外,dpdk提供的dpdk_nic_bind.py脚本中会调用到lspci命令,这个默认没有安装,运行以下命令安装(不安装此工具则无法绑定网卡):

yum install pciutils
ifconfig默认也没有安装,如果想用它,应运行:

yum install net-tools
在CentOS上,要绑定给dpdk使用的网卡在绑定前,可能是活动的(active),应将其禁用,否则无法绑定。禁用的一种方式是运行:

ifconfig eno33554984 down
eno33554984是接口名,如同eth0一样。

在CentOS上使用setup.sh和通过命令编译和配置dpdk的过程与Ubuntu一样,这里就从略了。



5. CPU和网卡队列绑定

以l3fwd为例,启动命令如下:

./l3fwd -c fffffffffff -n 4 -- -p 0xf -P--config="(0,0,1)(0,1,2)(1,0,8)(1,1,9)(2,0,16)(2,1,17)(3,0,24)(3,1,25)"

参数说明:
-c标识绑定的core
-ncpu中numa的数量
-p绑定的网卡
-q一个网卡上收包队列的数量

该硬件设备有四个CPU socket,使四个网口绑定到四个不同的CPU socket上,每个网口起两个队列,绑定两个CPU的核。注意:如果PCI插槽对应不同的socket,会导致测试性能降低。

也可以使用l3fwd进行一个网口的自发自收,命令:l3fwd -cff -n 4 -- -p0x1 -P --config=“(0,0,1),(0,1,2),(0,2,3),(0,3,4)”




文章来源
http://www.cnblogs.com/zzqcn/p/4024205.html



    推荐阅读