基於DPDK抓包的Suricata安裝部署

2023-03-22 18:01:47

一、背景

Suricata支援網路卡線上抓包和離線讀取PCAP包兩種形式的抓包:

  • 離線抓包天然具有速度慢、非實時的特點
  • 線上捕獲封包又包括常規網路卡抓包、PF_RING和DPDK的方式

由於專案分光的流量較大, 軟體自帶的抓包方式並不能滿足需求,因此採用了基於DPDK的Suricata線上捕獲網路卡封包的方式。

 

二、 伺服器設定與對應軟體版本

作業系統:Centos7.6 1810
核心版本:3.10.0-1160.88.1.el7.x86_64
網路卡資訊:Intel X722 萬兆光口乙太網卡
DPDK版本:dpdk-19.11.14
Suricata版本:DPDK_Suricata-4.1.4

 

三、部署安裝DPDK

基於DPDK抓包的Suricata版本只更新到4.1.4,因此對DPDK版本有要求,經過測試推薦使用 DPDK-19.11.14

DPDK官網下載地址:http://fast.dpdk.org/rel/dpdk-19.11.14.tar.gz

1、安裝dpdk-19.11.14需要作業系統核心版本大於3.2,如果版本過低,可以通過以下方式升級

1 [root@ids-dpdk ~]# cat /etc/redhat-release
2 CentOS Linux release 7.6.1810 (Core)
3 [root@ids-dpdk ~]# uname -r #檢視一下系統版本
4 3.10.0-957.el7.x86_64
5 [root@ids-dpdk ~]# rpm -qa kernel #通過rpm命令檢視我所安裝的核心版本
6 kernel-3.10.0-957.el7.x86_64
7 [root@ids-dpdk ~]# ls /usr/src/kernels/ #檢視有沒有相應的核心開發包
8 空 #如果該目錄下沒有系統核心原始碼,執行以下操作
9 [root@ids-dpdk ~]# yum install kernel-devel #安裝核心標頭檔案後
10 [root@ids-dpdk ~]# ls /usr/src/kernels/
11 3.10.0-1160.88.1.el7.x86_64
12 #兩個版本號不一致,進行升級並重啟
13 [root@ids-dpdk ~]# yum -y update kernel kernel-devel
14 [root@ids-dpdk ~]# reboot
15 #再此檢視,版本號一致,問題解決
16 [root@ids-dpdk ~]# ls /usr/src/kernels/
17 3.10.0-1160.88.1.el7.x86_64
18 [root@ids-dpdk ~]# uname -r
19 3.10.0-1160.88.1.el7.x86_64

 

2、安裝依賴包

sudo yum install -y gcc make
sudo yum install -y libpcap libpcap-devel
sudo yum install -y numactl numactl-devel
sudo yum install -y pciutils

 

3、從dpdk官網下載dpdk壓縮包並解壓

將dpdk壓縮包下載到/home目錄下並解壓
wget http://fast.dpdk.org/rel/dpdk-19.11.14.tar.gz
tar -zxvf dpdk-19.11.14.tar.gz

 

4、DPDK編譯和網路卡繫結

1. 設定環境變數,命令列執行:
[root@ids-dpdk ~]# export RTE_SDK='/home/dpdk-19.11.14'
[root@ids-dpdk ~]# export RTE_TARGET=x86_64-native-linuxapp-gcc
#(對於64位元機用這個命令,對於32位元機用i686-native-linuxapp-gcc)

2. 檢視環境變數是否設定好:
[root@ids-dpdk ~]# env |grep RTE
RTE_SDK=/home/dpdk-stable-19.11.14
RTE_TARGET=x86_64-native-linuxapp-gcc

3. 關閉要繫結的網路卡,否則繫結dpdk時不成功
ifconfig  ens1f0 down

4. 進入到dpdk-19.11.14/usertools目錄下
cd /home/dpdk-19.11.14/usertools
執行./dpdk-setup.sh

會輸出一列可選操作:

------------------------------------------------------------------------------
 RTE_SDK exported as /home/dpdk-stable-19.11.14
------------------------------------------------------------------------------
----------------------------------------------------------
 Step 1: Select the DPDK environment to build
----------------------------------------------------------
[1] arm64-armada-linuxapp-gcc
[2] arm64-armada-linux-gcc
[3] arm64-armv8a-linuxapp-clang
[4] arm64-armv8a-linuxapp-gcc
[5] arm64-armv8a-linux-clang
[6] arm64-armv8a-linux-gcc
[7] arm64-bluefield-linuxapp-gcc
[8] arm64-bluefield-linux-gcc
[9] arm64-dpaa-linuxapp-gcc
[10] arm64-dpaa-linux-gcc
[11] arm64-emag-linuxapp-gcc
[12] arm64-emag-linux-gcc
[13] arm64-graviton2-linuxapp-gcc
[14] arm64-graviton2-linux-gcc
[15] arm64-n1sdp-linuxapp-gcc
[16] arm64-n1sdp-linux-gcc
[17] arm64-octeontx2-linuxapp-gcc
[18] arm64-octeontx2-linux-gcc
[19] arm64-stingray-linuxapp-gcc
[20] arm64-stingray-linux-gcc
[21] arm64-thunderx2-linuxapp-gcc
[22] arm64-thunderx2-linux-gcc
[23] arm64-thunderx-linuxapp-gcc
[24] arm64-thunderx-linux-gcc
[25] arm64-xgene1-linuxapp-gcc
[26] arm64-xgene1-linux-gcc
[27] arm-armv7a-linuxapp-gcc
[28] arm-armv7a-linux-gcc
[29] graviton2
[30] i686-native-linuxapp-gcc
[31] i686-native-linuxapp-icc
[32] i686-native-linux-gcc
[33] i686-native-linux-icc
[34] ppc_64-power8-linuxapp-gcc
[35] ppc_64-power8-linux-gcc
[36] x86_64-native-bsdapp-clang
[37] x86_64-native-bsdapp-gcc
[38] x86_64-native-freebsd-clang
[39] x86_64-native-freebsd-gcc
[40] x86_64-native-linuxapp-clang
[41] x86_64-native-linuxapp-gcc
[42] x86_64-native-linuxapp-icc
[43] x86_64-native-linux-clang
[44] x86_64-native-linux-gcc
[45] x86_64-native-linux-icc
[46] x86_x32-native-linuxapp-gcc
[47] x86_x32-native-linux-gcc

----------------------------------------------------------
 Step 2: Setup linux environment
----------------------------------------------------------
[48] Insert IGB UIO module
[49] Insert VFIO module
[50] Insert KNI module
[51] Setup hugepage mappings for non-NUMA systems
[52] Setup hugepage mappings for NUMA systems
[53] Display current Ethernet/Baseband/Crypto device settings
[54] Bind Ethernet/Baseband/Crypto device to IGB UIO module
[55] Bind Ethernet/Baseband/Crypto device to VFIO module
[56] Setup VFIO permissions

----------------------------------------------------------
 Step 3: Run test application for linux environment
----------------------------------------------------------
[57] Run test application ($RTE_TARGET/app/test)
[58] Run testpmd application in interactive mode ($RTE_TARGET/app/testpmd)

----------------------------------------------------------
 Step 4: Other tools
----------------------------------------------------------
[59] List hugepage info from /proc/meminfo

----------------------------------------------------------
 Step 5: Uninstall and system cleanup
----------------------------------------------------------
[60] Unbind devices from IGB UIO or VFIO driver
[61] Remove IGB UIO module
[62] Remove VFIO module
[63] Remove KNI module
[64] Remove hugepage mappings

[65] Exit Script

Option: 

在最下面的 Option: 處輸入 41
會選擇適合x86_64機器的gcc編譯器,如果是其他架構的機器,需要選擇對應的其他編譯器
編譯中,稍等片刻......
編譯完成後會出現
Build complete [x86_64-native-linuxapp-gcc]
Installation cannot run with T defined and DESTDIR undefined
------------------------------------------------------------------------------
 RTE_TARGET exported as x86_64-native-linuxapp-gcc
------------------------------------------------------------------------------

Press enter to continue ...

按下確認鍵,繼續選擇其他操作:
Option: 48
載入igb uio模組
Unloading any existing DPDK UIO module
Loading DPDK UIO module

Press enter to continue ...

按下確認鍵,繼續選擇其他操作:
Option: 52
設定大頁
Removing currently reserved hugepages
Unmounting /mnt/huge and removing directory

  Input the number of 2048kB hugepages for each node
  Example: to have 128MB of hugepages available per node in a 2MB huge page system,
  enter '64' to reserve 64 * 2MB pages on each node
Number of pages for node0: 1024(在此輸入1024)

Reserving hugepages
Creating /mnt/huge and mounting as hugetlbfs

Press enter to continue ...

按下確認鍵,繼續選擇其他操作:
Option: 53
列出所有網路卡
Network devices using kernel driver
===================================
0000:17:00.0 'MT27800 Family [ConnectX-5] 1017' if=ens5f0 drv=mlx5_core unused=igb_uio *Active*
0000:17:00.1 'MT27800 Family [ConnectX-5] 1017' if=ens5f1 drv=mlx5_core unused=igb_uio *Active*
0000:33:00.0 'Ethernet Connection X722 for 10GbE SFP+ 0dda' if=ens1f0 drv=i40e unused=igb_uio
0000:33:00.1 'Ethernet Connection X722 for 10GbE SFP+ 0dda' if=ens1f1 drv=i40e unused=igb_uio

No 'Baseband' devices detected
==============================

No 'Crypto' devices detected
============================

No 'Eventdev' devices detected
==============================

No 'Mempool' devices detected
=============================

No 'Compress' devices detected
==============================

No 'Misc (rawdev)' devices detected
===================================

Press enter to continue ...

按下確認鍵,繼續選擇其他操作:
Option: 54
繫結網路卡
Network devices using kernel driver
===================================
0000:17:00.0 'MT27800 Family [ConnectX-5] 1017' if=ens5f0 drv=mlx5_core unused=igb_uio *Active*
0000:17:00.1 'MT27800 Family [ConnectX-5] 1017' if=ens5f1 drv=mlx5_core unused=igb_uio *Active*
0000:33:00.0 'Ethernet Connection X722 for 10GbE SFP+ 0dda' if=ens1f0 drv=i40e unused=igb_uio
0000:33:00.1 'Ethernet Connection X722 for 10GbE SFP+ 0dda' if=ens1f1 drv=i40e unused=igb_uio

No 'Baseband' devices detected
==============================

No 'Crypto' devices detected
============================

No 'Eventdev' devices detected
==============================

No 'Mempool' devices detected
=============================

No 'Compress' devices detected
==============================

No 'Misc (rawdev)' devices detected
===================================

Enter PCI address of device to bind to IGB UIO driver: 0000:33:00.0(在此輸入網路卡pci號,就是上邊列出來的網路卡資訊)

成功後會輸出:
ok

Press enter to continue ...

按下確認鍵,繼續選擇其他操作:
Option: 53
檢視網路卡繫結情況
Network devices using DPDK-compatible driver
============================================
0000:33:00.0 'Ethernet Connection X722 for 10GbE SFP+ 0dda' drv=igb_uio unused=i40e

Network devices using kernel driver
===================================
0000:17:00.0 'MT27800 Family [ConnectX-5] 1017' if=ens5f0 drv=mlx5_core unused=igb_uio  *Active*
0000:17:00.1 'MT27800 Family [ConnectX-5] 1017' if=ens5f1 drv=mlx5_core unused=igb_uio  *Active*
0000:33:00.1 'Ethernet Connection X722 for 10GbE SFP+ 0dda' if=enp51s0f1 drv=i40e unused=igb_uio 

No 'Baseband' devices detected
==============================

No 'Crypto' devices detected
============================

No 'Eventdev' devices detected
==============================

No 'Mempool' devices detected
=============================

No 'Compress' devices detected
==============================

No 'Misc (rawdev)' devices detected
===================================

Press enter to continue ...

按下確認鍵,繼續選擇其他操作:
Option: 57
進行簡單測試:
  Enter hex bitmask of cores to execute test app on
  Example: to execute app on cores 0 to 7, enter 0xff
bitmask: 0xff(在此輸入 0xff )
Launching app
EAL: Detected 40 lcore(s)
EAL: Detected 2 NUMA nodes
EAL: Multi-process socket /var/run/dpdk/rte/mp_socket
EAL: Selected IOVA mode 'PA'
EAL: No available hugepages reported in hugepages-1048576kB
EAL: Probing VFIO support...
EAL: PCI device 0000:00:04.0 on NUMA socket 0
EAL:   probe driver: 8086:2021 rawdev_ioat
EAL: PCI device 0000:00:04.1 on NUMA socket 0
EAL:   probe driver: 8086:2021 rawdev_ioat
EAL: PCI device 0000:00:04.2 on NUMA socket 0
EAL:   probe driver: 8086:2021 rawdev_ioat
EAL: PCI device 0000:00:04.3 on NUMA socket 0
EAL:   probe driver: 8086:2021 rawdev_ioat
EAL: PCI device 0000:00:04.4 on NUMA socket 0
EAL:   probe driver: 8086:2021 rawdev_ioat
EAL: PCI device 0000:00:04.5 on NUMA socket 0
EAL:   probe driver: 8086:2021 rawdev_ioat
EAL: PCI device 0000:00:04.6 on NUMA socket 0
EAL:   probe driver: 8086:2021 rawdev_ioat
EAL: PCI device 0000:00:04.7 on NUMA socket 0
EAL:   probe driver: 8086:2021 rawdev_ioat
EAL: PCI device 0000:3d:00.0 on NUMA socket 0
EAL:   probe driver: 8086:37d3 net_i40e
i40e_GLQF_reg_init(): i40e device 0000:3d:00.0 changed global register [0x002689a0]. original: 0x0000002a, new: 0x00000029 
i40e_GLQF_reg_init(): i40e device 0000:3d:00.0 changed global register [0x00268ca4]. original: 0x00002826, new: 0x00009420 
EAL: PCI device 0000:3d:00.1 on NUMA socket 0
EAL:   probe driver: 8086:37d3 net_i40e
EAL: PCI device 0000:80:04.0 on NUMA socket 1
EAL:   probe driver: 8086:2021 rawdev_ioat
EAL: PCI device 0000:80:04.1 on NUMA socket 1
EAL:   probe driver: 8086:2021 rawdev_ioat
EAL: PCI device 0000:80:04.2 on NUMA socket 1
EAL:   probe driver: 8086:2021 rawdev_ioat
EAL: PCI device 0000:80:04.3 on NUMA socket 1
EAL:   probe driver: 8086:2021 rawdev_ioat
EAL: PCI device 0000:80:04.4 on NUMA socket 1
EAL:   probe driver: 8086:2021 rawdev_ioat
EAL: PCI device 0000:80:04.5 on NUMA socket 1
EAL:   probe driver: 8086:2021 rawdev_ioat
EAL: PCI device 0000:80:04.6 on NUMA socket 1
EAL:   probe driver: 8086:2021 rawdev_ioat
EAL: PCI device 0000:80:04.7 on NUMA socket 1
EAL:   probe driver: 8086:2021 rawdev_ioat
APP: HPET is not enabled, using TSC as default timer
輸出以上內容表示安裝dpdk成功
RTE>>quit(在此輸入quit)

Press enter to continue ....

按下確認鍵,繼續選擇其他操作:
Option: 58
進行抓包測試
  Enter hex bitmask of cores to execute testpmd app on
  Example: to execute app on cores 0 to 7, enter 0xff
bitmask: 7(輸入 7)
Launching app
EAL: Detected 40 lcore(s)
EAL: Detected 2 NUMA nodes
EAL: Multi-process socket /var/run/dpdk/rte/mp_socket
EAL: Selected IOVA mode 'PA'
EAL: No available hugepages reported in hugepages-1048576kB
EAL: Probing VFIO support...
EAL: PCI device 0000:00:04.0 on NUMA socket 0
EAL:   probe driver: 8086:2021 rawdev_ioat
EAL: PCI device 0000:00:04.1 on NUMA socket 0
EAL:   probe driver: 8086:2021 rawdev_ioat
EAL: PCI device 0000:00:04.2 on NUMA socket 0
EAL:   probe driver: 8086:2021 rawdev_ioat
EAL: PCI device 0000:00:04.3 on NUMA socket 0
EAL:   probe driver: 8086:2021 rawdev_ioat
EAL: PCI device 0000:00:04.4 on NUMA socket 0
EAL:   probe driver: 8086:2021 rawdev_ioat
EAL: PCI device 0000:00:04.5 on NUMA socket 0
EAL:   probe driver: 8086:2021 rawdev_ioat
EAL: PCI device 0000:00:04.6 on NUMA socket 0
EAL:   probe driver: 8086:2021 rawdev_ioat
EAL: PCI device 0000:00:04.7 on NUMA socket 0
EAL:   probe driver: 8086:2021 rawdev_ioat
EAL: PCI device 0000:3d:00.0 on NUMA socket 0
EAL:   probe driver: 8086:37d3 net_i40e
EAL: PCI device 0000:3d:00.1 on NUMA socket 0
EAL:   probe driver: 8086:37d3 net_i40e
EAL: PCI device 0000:80:04.0 on NUMA socket 1
EAL:   probe driver: 8086:2021 rawdev_ioat
EAL: PCI device 0000:80:04.1 on NUMA socket 1
EAL:   probe driver: 8086:2021 rawdev_ioat
EAL: PCI device 0000:80:04.2 on NUMA socket 1
EAL:   probe driver: 8086:2021 rawdev_ioat
EAL: PCI device 0000:80:04.3 on NUMA socket 1
EAL:   probe driver: 8086:2021 rawdev_ioat
EAL: PCI device 0000:80:04.4 on NUMA socket 1
EAL:   probe driver: 8086:2021 rawdev_ioat
EAL: PCI device 0000:80:04.5 on NUMA socket 1
EAL:   probe driver: 8086:2021 rawdev_ioat
EAL: PCI device 0000:80:04.6 on NUMA socket 1
EAL:   probe driver: 8086:2021 rawdev_ioat
EAL: PCI device 0000:80:04.7 on NUMA socket 1
EAL:   probe driver: 8086:2021 rawdev_ioat
Interactive-mode selected
testpmd: create a new mbuf pool <mbuf_pool_socket_0>: n=163456, size=2176, socket=0
testpmd: preferred mempool ops selected: ring_mp_mc

Warning! port-topology=paired and odd forward ports number, the last port will pair with itself.

Configuring Port 0 (socket 0)
Port 0: F0:10:90:7E:FF:03
Checking link statuses...
Done
testpmd> start (輸入start開始抓包)
io packet forwarding - ports=1 - cores=1 - streams=1 - NUMA support enabled, MP allocation mode: native
Logical Core 1 (socket 0) forwards packets on 1 streams:
  RX P=0/Q=0 (socket 0) -> TX P=0/Q=0 (socket 0) peer=02:00:00:00:00:00

  io packet forwarding packets/burst=32
  nb forwarding cores=1 - nb forwarding ports=1
  port 0: RX queue number: 1 Tx queue number: 1
    Rx offloads=0x0 Tx offloads=0x10000
    RX queue: 0
      RX desc=256 - RX free threshold=32
      RX threshold registers: pthresh=0 hthresh=0  wthresh=0
      RX Offloads=0x0
    TX queue: 0
      TX desc=256 - TX free threshold=32
      TX threshold registers: pthresh=32 hthresh=0  wthresh=0
      TX offloads=0x10000 - TX RS bit threshold=32
testpmd> stop (輸入stop停止抓包)
Telling cores to stop...
Waiting for lcores to finish...

  ---------------------- Forward statistics for port 0  ----------------------
  RX-packets: 2685              RX-dropped: 0             RX-total: 2685
  TX-packets: 2686              TX-dropped: 0             TX-total: 2686
  ----------------------------------------------------------------------------

  +++++++++++++++ Accumulated forward statistics for all ports+++++++++++++++
  RX-packets: 2685              RX-dropped: 0             RX-total: 2685
  TX-packets: 2686              TX-dropped: 0             TX-total: 2686
  ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++

Done.
testpmd> quit(在此輸入quit)

Stopping port 0...
Stopping ports...
Done

Shutting down port 0...
Closing ports...
Done

Bye...

Press enter to continue ...

按下確認鍵,繼續選擇其他操作:
Option: 65
退出安裝程式,至此,dpdk安裝成功。

 

5、DPDK_Suricata編譯與安裝

1、安裝依賴
[root@ids-dpdk ~]# yum -y install gcc libpcap-devel pcre-devel libyaml-devel file-devel \
  zlib-devel jansson-devel nss-devel libcap-ng-devel libnet-devel tar make openssl openssl-devel \
  libnetfilter_queue-devel lua-devel PyYAML libmaxminddb-devel rustc cargo librdkafka-devel \
  lz4-devel libxml2 autoconf

2、下載DPDK_Suricata安裝包
##目前DPDK_Suricata專案作者已經到4.1.4,只是目錄仍沿用4.1.1,直接克隆專案程式碼即可
[root@ids-dpdk ~]# git clone https://github.com/vipinpv85/DPDK_SURICATA-4_1_1

3、編譯並安裝DPDK_Suricata
進入DPDK_Suricata目錄下:
[root@ids-dpdk ~]# cd /home/DPDK_SURICATA-4_1_1/suricata-4.1.4
構建支援DPDK的設定指令碼:
[root@ids-dpdk suricata-4.1.4]# autoconf
使用dpdk進行設定:
[root@ids-dpdk suricata-4.1.4]# ./configure --prefix=/usr --sysconfdir=/etc --localstatedir=/var --enable-dpdk  
編譯DPDK_Suricata:
[root@ids-dpdk suricata-4.1.4]# make
安裝DPDK_Suricata:
[root@ids-dpdk suricata-4.1.4]# make install
更新規則檔案:
[root@ids-dpdk suricata-4.1.4]# make install-full
安裝相關組態檔:
[root@ids-dpdk suricata-4.1.4]# make install-conf

 

 6、測試執行抓包

修改組態檔suricata.yaml

[root@ids-dpdk ~]# vim /etc/suricata/suricata.yaml
搜尋 DPDK關鍵詞,修改input-output-map和interface兩處引數值
將input-output-map修改成:input-output-map: ["0"]
將interface修改成:- interface: 0
修改完成後執行 :wq 儲存並退出

 

 

測試執行:

[root@ids-dpdk ~]# suricata
如果報以下錯誤,是庫檔案未找到的原因
/usr/bin/suricata: error while loading shared libraries: libhtp.so.2: cannot open shared object file: No such file or directory
執行命令手動建立到libhtp.so.2的軟連結
[root@ids-dpdk ~]# ln -s /usr/lib/libhtp.so.2 /lib64/libhtp.so.2
然後執行命令開始執行dpdk抓包測試
[root@ids-dpdk ~]# suricata --dpdk
22/3/2023 -- 15:36:51 - <Notice> -  --26. (protocol-ff) 

22/3/2023 -- 15:36:51 - <Notice> -  --27. (protocol-ff) 

22/3/2023 -- 15:36:51 - <Notice> -  --28. (protocol-ff) 

22/3/2023 -- 15:36:51 - <Notice> -  --29. (protocol-ff) 

22/3/2023 -- 15:36:51 - <Notice> -  --30. (protocol-ff) 

22/3/2023 -- 15:36:51 - <Notice> -  --31. (protocol-ff) 

22/3/2023 -- 15:36:51 - <Notice> -  addr_dst_match4_cnt 1 addr_src_match4_cnt 1 addr_dst_match6_cnt i1 addr_src_match6_cnt 1
22/3/2023 -- 15:36:51 - <Notice> -  IPV4 
22/3/2023 -- 15:36:51 - <Notice> -  0:ffffffff 
22/3/2023 -- 15:36:51 - <Notice> -  0:ffffffff 
22/3/2023 -- 15:36:51 - <Notice> - -----------------------
22/3/2023 -- 15:36:51 - <Notice> -  IPV6 
22/3/2023 -- 15:36:51 - <Notice> -  0-0-0-0:ffffffff-ffffffff-ffffffff-ffffffff 
22/3/2023 -- 15:36:51 - <Notice> -  0-0-0-0:ffffffff-ffffffff-ffffffff-ffffffff 
22/3/2023 -- 15:36:51 - <Notice> - -----------------------
22/3/2023 -- 15:36:51 - <Notice> -  Source Port 
22/3/2023 -- 15:36:51 - <Notice> -  port:port2 (0:ffff)
22/3/2023 -- 15:36:51 - <Notice> -  Destiantion Port 
22/3/2023 -- 15:36:51 - <Notice> -  port:port2 (0:ffff)
22/3/2023 -- 15:36:51 - <Notice> -  prio 3 
22/3/2023 -- 15:36:51 - <Notice> -  Port 0 RX-q (4) hence trying RSS
22/3/2023 -- 15:36:51 - <Notice> -  rss_hf 3ef8, rss_key_len 0

22/3/2023 -- 15:36:52 - <Notice> - all 4 packet processing threads, 4 management threads initialized, engine started.
至此,DPDK_Suricata開啟DPDK抓包

 

7、DPDK指令碼補充

由於伺服器開關機會導致DPDK繫結的網路卡會被預設解綁,為簡化重新機械的繫結工作,通過shell指令碼實現自動化DPDK繫結網路卡。

編輯/etc/profile,設定環境變數

在done和unset i之間新增環境變數,執行:wq儲存並退出,然後執行source使之生效

[root@ids-dpdk ~]# vim /etc/profile
新增內容:

export RTE_SDK='/home/dpdk-stable-19.11.14'
export RTE_TARGET=x86_64-native-linuxapp-gcc

[root@ids-dpdk ~]# source /etc/profile

 

建立DPDK繫結指令碼dpdk-bind.sh,並賦予執行許可權

[root@ids-dpdk usertools]# vim dpdk-bind.sh
[root@ids-dpdk usertools]# chmod -R 775 dpdk-bind.sh

 

dpdk-bind.sh指令碼內容如下,根據自己需要修改網路卡名稱和路徑資訊

#!/bin/sh

#網路卡名
uio=ens1f0

#需要繫結的驅動型別igb_uio或者vfio-pci 
pci_type=igb_uio

#掛載驅動
modprobe uio
insmod /home/dpdk-stable-19.11.14/x86_64-native-linuxapp-gcc/kmod/igb_uio.ko

#關閉網路卡
ifconfig $uio down

#繫結網路卡到igb_uio
python /home/dpdk-stable-19.11.14/usertools/dpdk-devbind.py --bind=$pci_type $uio

 

設定大頁儲存,將DPDK繫結指令碼加入開機啟動項中,編輯/etc/rc.local,根據自己需求修改路徑等資訊

#!/bin/bash
# THIS FILE IS ADDED FOR COMPATIBILITY PURPOSES
#
# It is highly advisable to create own systemd services or udev rules
# to run scripts during boot instead of using this file.
#
# In contrast to previous versions due to parallel execution during boot
# this script will NOT be run after all other services.
#
# Please note that you must run 'chmod +x /etc/rc.d/rc.local' to ensure
# that this script will be executed during boot.

touch /var/lock/subsys/local

#設定大頁儲存
echo 1024 > /sys/kernel/mm/hugepages/hugepages-2048kB/nr_hugepages
mkdir /mnt/huge > /dev/null 2>&1
mount -t hugetlbfs nodev /mnt/huge

#網路卡繫結DPDK
sh /home/dpdk-stable-19.11.14/usertools/dpdk-bind.sh > /dev/null 2>&1

編輯完成後,執行:wq儲存並退出