Hi all, I also tested master and stable/1710 release using 1G hugepage. Below is the test result: For stable/1710, it could work well using 1G hugepage. For master, it failed to start up when using 1G hugepage, just like below.
Because 1G hugepage support is required for some performance tests, and virtio-user in DPDK-based container. Please help to take a look at this issue. Thanks a lot, Hongjun From: vpp-dev-boun...@lists.fd.io [mailto:vpp-dev-boun...@lists.fd.io] On Behalf Of Samuel Eliáš Sent: Monday, January 22, 2018 11:30 PM To: vpp-dev@lists.fd.io Subject: [vpp-dev] "ftruncate: Invalid argument" in VPP startup Hello vpp-dev, I've encountered an issue when trying to run VPP using default configuration on baremetal (512GB RAM, 36 phys cores, Ubuntu16.04). It appears to be related to dpdk memory allocation, and only occurs on VPP 18.01, not on 17.10. Just wondering if anyone's seen this before, and/or whether I should go bother the dpdk folks instead. I understand dpdk version was bumped to 17.11 in this release, so that's one potential cause. $ sudo /usr/bin/vpp unix interactive vlib_plugin_early_init:356: plugin path /usr/lib/vpp_plugins load_one_plugin:184: Loaded plugin: acl_plugin.so (Access Control Lists) ... more plugins ... load_one_plugin:63: Loaded plugin: /usr/lib/vpp_api_test_plugins/lb_test_plugin.so clib_sysfs_read: open `/sys/devices/system/node/node0/hugepages/hugepages-2048kB/free_hugepages': No such file or directory clib_sysfs_read: open `/sys/devices/system/node/node1/hugepages/hugepages-2048kB/free_hugepages': No such file or directory dpdk_bind_devices_to_uio:758: Unsupported PCI device 0x8086:0x0435 found at PCI address 0000:08:00.0 vlib_pci_bind_to_uio: Skipping PCI device 0000:0a:00.0 as host interface eth0 is up dpdk_bind_devices_to_uio:758: Unsupported PCI device 0x8086:0x0435 found at PCI address 0000:84:00.0 dpdk_config:1240: EAL init args: -c 1 -n 4 --huge-dir /run/vpp/hugepages --file-prefix vpp -b 0000:0a:00.0 --master-lcore 0 --socket-mem 64,64 EAL: VFIO support initialized DPDK physical memory layout: Segment 0: IOVA:0x100000000, len:1073741824, virt:0x7efbc0000000, socket_id:0, hugepage_sz:1073741824, nchannel:0, nrank:0 Segment 1: IOVA:0x4080000000, len:1073741824, virt:0x7ee480000000, socket_id:1, hugepage_sz:1073741824, nchannel:0, nrank:0 clib_mem_vm_ext_alloc: ftruncate: Invalid argument dpdk_buffer_pool_create: failed to allocate mempool on socket 0 Memory info: $ cat /proc/meminfo | tail -n 8 HugePages_Total: 128 HugePages_Free: 128 HugePages_Rsvd: 0 HugePages_Surp: 0 Hugepagesize: 1048576 kB DirectMap4k: 224680 kB DirectMap2M: 3862528 kB DirectMap1G: 534773760 kB Thanks, - Sam
_______________________________________________ vpp-dev mailing list vpp-dev@lists.fd.io https://lists.fd.io/mailman/listinfo/vpp-dev