Hi Adarsh, Would you be able to try the following?
In startup config for now, could you add the following change in the unix section of the vpp startup. # nodaemon interactive Also, I see you have the line gid vpp. I am not sure what that does. Most of the configs I see do not have that. You could try removing that line. Then run VPP manually like this so we might get more error information. # cat /proc/meminfo – Be sure you still have some hugepages, sometimes if dpdk crashes it doesn’t return them. # service vpp stop # ps -eaf | grep vpp – Make sure there are no other VPP processes still running # /usr/bin/vpp -c /etc/vpp/startup.conf Then send us the results from the entire startup. Thanks, John From: <vpp-dev-boun...@lists.fd.io> on behalf of adarsh m via vpp-dev <vpp-dev@lists.fd.io> Reply-To: adarsh m <addi.ada...@yahoo.in> Date: Thursday, February 1, 2018 at 8:43 AM To: "Damjan Marion (damarion)" <damar...@cisco.com>, adarsh m via vpp-dev <vpp-dev@lists.fd.io> Subject: Re: [vpp-dev] Error when trying to add interface to vpp on ARM server. Hi, After removing socket-mem now vpp is stable but when we try to connect through CLI connection is refused. ubuntu@vasily:~$ sudo service vpp status ● vpp.service - vector packet processing engine Loaded: loaded (/lib/systemd/system/vpp.service; enabled; vendor preset: enabled) Active: active (running) since Thu 2018-02-01 21:42:23 CST; 1s ago Process: 49787 ExecStopPost=/bin/rm -f /dev/shm/db /dev/shm/global_vm /dev/shm/vpe-api (code=exited, status=0/SUCCESS) Process: 49793 ExecStartPre=/sbin/modprobe uio_pci_generic (code=exited, status=0/SUCCESS) Process: 49790 ExecStartPre=/bin/rm -f /dev/shm/db /dev/shm/global_vm /dev/shm/vpe-api (code=exited, status=0/SUCCESS) Main PID: 49796 (vpp) CGroup: /system.slice/vpp.service └─49796 /usr/bin/vpp -c /etc/vpp/startup.conf Feb 01 19:37:16 vasily /usr/bin/vpp[43454]: load_one_plugin:63: Loaded plugin: /usr/lib/vpp_api_test_plugins/acl_test_plugin.so Feb 01 19:37:16 vasily /usr/bin/vpp[43454]: load_one_plugin:63: Loaded plugin: /usr/lib/vpp_api_test_plugins/kubeproxy_test_plugin.so Feb 01 19:37:16 vasily /usr/bin/vpp[43454]: load_one_plugin:63: Loaded plugin: /usr/lib/vpp_api_test_plugins/pppoe_test_plugin.so Feb 01 19:37:16 vasily /usr/bin/vpp[43454]: load_one_plugin:63: Loaded plugin: /usr/lib/vpp_api_test_plugins/ioam_vxlan_gpe_test_plugin.so Feb 01 19:37:16 vasily /usr/bin/vpp[43454]: load_one_plugin:63: Loaded plugin: /usr/lib/vpp_api_test_plugins/dpdk_test_plugin.so Feb 01 19:37:16 vasily /usr/bin/vpp[43454]: load_one_plugin:63: Loaded plugin: /usr/lib/vpp_api_test_plugins/udp_ping_test_plugin.so Feb 01 19:37:16 vasily vpp[43454]: /usr/bin/vpp[43454]: dpdk_config:1240: EAL init args: -c 1 -n 4 --huge-dir /run/vpp/hugepages --file-prefix vpp -w 0002:f9:00.0 --master-lcore 0 --socket-mem 1024, Feb 01 19:37:16 vasily /usr/bin/vpp[43454]: dpdk_config:1240: EAL init args: -c 1 -n 4 --huge-dir /run/vpp/hugepages --file-prefix vpp -w 0002:f9:00.0 --master-lcore 0 --socket-mem 1024,0,0,0 Feb 01 19:37:16 vasily vpp[43454]: EAL: VFIO support initialized Feb 01 19:37:16 vasily vnet[43454]: EAL: VFIO support initialized ubuntu@vasily:~$ sudo vppctl show int clib_socket_init: connect (fd 3, '/run/vpp/cli.sock'): Connection refused ubuntu@vasily:~$ Startup.conf : unix { nodaemon log /tmp/vpp.log full-coredump cli-listen /run/vpp/cli.sock gid vpp } api-trace { ## This stanza controls binary API tracing. Unless there is a very strong reason, ## please leave this feature enabled. on ## Additional parameters: ## ## To set the number of binary API trace records in the circular buffer, configure nitems ## ## nitems <nnn> ## ## To save the api message table decode tables, configure a filename. Results in /tmp/<filename> ## Very handy for understanding api message changes between versions, identifying missing ## plugins, and so forth. ## ## save-api-table <filename> } api-segment { gid vpp } cpu { ## In the VPP there is one main thread and optionally the user can create worker(s) ## The main thread and worker thread(s) can be pinned to CPU core(s) manually or automatically ## Manual pinning of thread(s) to CPU core(s) ## Set logical CPU core where main thread runs # main-core 1 ## Set logical CPU core(s) where worker threads are running # corelist-workers 2-3,18-19 ## Automatic pinning of thread(s) to CPU core(s) ## Sets number of CPU core(s) to be skipped (1 ... N-1) ## Skipped CPU core(s) are not used for pinning main thread and working thread(s). ## The main thread is automatically pinned to the first available CPU core and worker(s) ## are pinned to next free CPU core(s) after core assigned to main thread # skip-cores 4 ## Specify a number of workers to be created ## Workers are pinned to N consecutive CPU cores while skipping "skip-cores" CPU core(s) ## and main thread's CPU core # workers 2 ## Set scheduling policy and priority of main and worker threads ## Scheduling policy options are: other (SCHED_OTHER), batch (SCHED_BATCH) ## idle (SCHED_IDLE), fifo (SCHED_FIFO), rr (SCHED_RR) # scheduler-policy fifo ## Scheduling priority is used only for "real-time policies (fifo and rr), ## and has to be in the range of priorities supported for a particular policy # scheduler-priority 50 } dpdk { dev 0002:f9:00.0 ## Change default settings for all intefaces # dev default { ## Number of receive queues, enables RSS ## Default is 1 # num-rx-queues 3 ## Number of transmit queues, Default is equal ## to number of worker threads or 1 if no workers treads # num-tx-queues 3 ## Number of descriptors in transmit and receive rings ## increasing or reducing number can impact performance ## Default is 1024 for both rx and tx # num-rx-desc 512 # num-tx-desc 512 ## VLAN strip offload mode for interface ## Default is off # vlan-strip-offload on # } ## Whitelist specific interface by specifying PCI address # dev 0000:02:00.0 ## Whitelist specific interface by specifying PCI address and in ## addition specify custom parameters for this interface # dev 0000:02:00.1 { # num-rx-queues 2 # } ## Specify bonded interface and its slaves via PCI addresses ## ## Bonded interface in XOR load balance mode (mode 2) with L3 and L4 headers # vdev eth_bond0,mode=2,slave=0000:02:00.0,slave=0000:03:00.0,xmit_policy=l34 # vdev eth_bond1,mode=2,slave=0000:02:00.1,slave=0000:03:00.1,xmit_policy=l34 ## ## Bonded interface in Active-Back up mode (mode 1) # vdev eth_bond0,mode=1,slave=0000:02:00.0,slave=0000:03:00.0 # vdev eth_bond1,mode=1,slave=0000:02:00.1,slave=0000:03:00.1 ## Change UIO driver used by VPP, Options are: igb_uio, vfio-pci, ## uio_pci_generic or auto (default) # uio-driver vfio-pci ## Disable mutli-segment buffers, improves performance but ## disables Jumbo MTU support # no-multi-seg ## Increase number of buffers allocated, needed only in scenarios with ## large number of interfaces and worker threads. Value is per CPU socket. ## Default is 16384 # num-mbufs 128000 num-mbufs 4095 ## Change hugepages allocation per-socket, needed only if there is need for ## larger number of mbufs. Default is 256M on each detected CPU socket # socket-mem 2048,2048 ## Disables UDP / TCP TX checksum offload. Typically needed for use ## faster vector PMDs (together with no-multi-seg) # no-tx-checksum-offload } # Adjusting the plugin path depending on where the VPP plugins are: #plugins #{ # path /home/bms/vpp/build-root/install-vpp-native/vpp/lib64/vpp_plugins #} # Alternate syntax to choose plugin path #plugin_path /home/bms/vpp/build-root/install-vpp-native/vpp/lib64/vpp_plugins On Thursday 1 February 2018, 7:00:18 PM IST, adarsh m via vpp-dev <vpp-dev@lists.fd.io> wrote: Support 2 channel 32core processor(ARM64) On Thursday 1 February 2018, 6:47:31 PM IST, Damjan Marion (damarion) <damar...@cisco.com> wrote: You also added socket-mem which is pretty bad idea, try without. If that doesn't help then you will need to run VPP form console and possibly use gdb to collect more details. Which ARM board is that? On 1 Feb 2018, at 14:15, adarsh m <addi.ada...@yahoo.in<mailto:addi.ada...@yahoo.in>> wrote: Hi, This is on ARM board. and yes i have modified the specific pcie address to added in startup.conf dpdk { socket-mem 1024 dev 0002:f9:00.0 } On Thursday 1 February 2018, 5:20:59 PM IST, Damjan Marion (damarion) <damar...@cisco.com<mailto:damar...@cisco.com>> wrote: Please keep mailing list in CC. Those lines doesn't show that anything is wrong... Is this 4 socket computer? Have you modified startup.conf? On 1 Feb 2018, at 12:40, adarsh m <addi.ada...@yahoo.in<mailto:addi.ada...@yahoo.in>> wrote: Hi, Very sorry pls check the complete one. Feb 01 19:37:16 vasily /usr/bin/vpp[43454]: load_one_plugin:63: Loaded plugin: /usr/lib/vpp_api_test_plugins/acl_test_plugin.so Feb 01 19:37:16 vasily /usr/bin/vpp[43454]: load_one_plugin:63: Loaded plugin: /usr/lib/vpp_api_test_plugins/kubeproxy_test_plugin.so Feb 01 19:37:16 vasily /usr/bin/vpp[43454]: load_one_plugin:63: Loaded plugin: /usr/lib/vpp_api_test_plugins/pppoe_test_plugin.so Feb 01 19:37:16 vasily /usr/bin/vpp[43454]: load_one_plugin:63: Loaded plugin: /usr/lib/vpp_api_test_plugins/ioam_vxlan_gpe_test_plugin.so Feb 01 19:37:16 vasily /usr/bin/vpp[43454]: load_one_plugin:63: Loaded plugin: /usr/lib/vpp_api_test_plugins/dpdk_test_plugin.so Feb 01 19:37:16 vasily /usr/bin/vpp[43454]: load_one_plugin:63: Loaded plugin: /usr/lib/vpp_api_test_plugins/udp_ping_test_plugin.so Feb 01 19:37:16 vasily vpp[43454]: /usr/bin/vpp[43454]: dpdk_config:1240: EAL init args: -c 1 -n 4 --huge-dir /run/vpp/hugepages --file-prefix vpp -w 0002:f9:00.0 --master-lcore 0 --socket-mem 1024, Feb 01 19:37:16 vasily /usr/bin/vpp[43454]: dpdk_config:1240: EAL init args: -c 1 -n 4 --huge-dir /run/vpp/hugepages --file-prefix vpp -w 0002:f9:00.0 --master-lcore 0 --socket-mem 1024,0,0,0 Feb 01 19:37:16 vasily vpp[43454]: EAL: VFIO support initialized Feb 01 19:37:16 vasily vnet[43454]: EAL: VFIO support initialized On Thursday 1 February 2018, 4:48:35 PM IST, Damjan Marion (damarion) <damar...@cisco.com<mailto:damar...@cisco.com>> wrote: Unfortunately log you provided is incomplete and truncated so I cannot help much.... On 1 Feb 2018, at 11:59, adarsh m <addi.ada...@yahoo.in<mailto:addi.ada...@yahoo.in>> wrote: Hi, I checked hugepage and it was 0, so i freed this and increased it to 5120 ubuntu@vasily:~$ sudo -i root@vasily:~# echo 5120 > /proc/sys/vm/nr_hugepages now the previous error is not occurring when i start but VPP is not stable, it'll become dead after a few sceonds from start. Logs : ubuntu@vasily:~$ grep HugePages_Free /proc/meminfo HugePages_Free: 5120 ubuntu@vasily:~$ ubuntu@vasily:~$ ubuntu@vasily:~$ ubuntu@vasily:~$ sudo service vpp status ● vpp.service - vector packet processing engine Loaded: loaded (/lib/systemd/system/vpp.service; enabled; vendor preset: enabled) Active: inactive (dead) since Thu 2018-02-01 18:50:46 CST; 5min ago Process: 42736 ExecStopPost=/bin/rm -f /dev/shm/db /dev/shm/global_vm /dev/shm/vpe-api (code=exited, status=0/SUCCESS) Process: 42731 ExecStart=/usr/bin/vpp -c /etc/vpp/startup.conf (code=exited, status=0/SUCCESS) Process: 42728 ExecStartPre=/sbin/modprobe uio_pci_generic (code=exited, status=0/SUCCESS) Process: 42726 ExecStartPre=/bin/rm -f /dev/shm/db /dev/shm/global_vm /dev/shm/vpe-api (code=exited, status=0/SUCCESS) Main PID: 42731 (code=exited, status=0/SUCCESS) Feb 01 18:50:46 vasily systemd[1]: vpp.service: Service hold-off time over, scheduling restart. Feb 01 18:50:46 vasily systemd[1]: Stopped vector packet processing engine. Feb 01 18:50:46 vasily systemd[1]: vpp.service: Start request repeated too quickly. Feb 01 18:50:46 vasily systemd[1]: Failed to start vector packet processing engine. Feb 01 18:56:12 vasily systemd[1]: Stopped vector packet processing engine. ubuntu@vasily:~$ ubuntu@vasily:~$ ubuntu@vasily:~$ ubuntu@vasily:~$ sudo service vpp start ubuntu@vasily:~$ ubuntu@vasily:~$ ubuntu@vasily:~$ sudo service vpp status ● vpp.service - vector packet processing engine Loaded: loaded (/lib/systemd/system/vpp.service; enabled; vendor preset: enabled) Active: active (running) since Thu 2018-02-01 18:56:49 CST; 298ms ago Process: 42857 ExecStopPost=/bin/rm -f /dev/shm/db /dev/shm/global_vm /dev/shm/vpe-api (code=exited, status=0/SUCCESS) Process: 42863 ExecStartPre=/sbin/modprobe uio_pci_generic (code=exited, status=0/SUCCESS) Process: 42860 ExecStartPre=/bin/rm -f /dev/shm/db /dev/shm/global_vm /dev/shm/vpe-api (code=exited, status=0/SUCCESS) Main PID: 42866 (vpp) CGroup: /system.slice/vpp.service └─42866 /usr/bin/vpp -c /etc/vpp/startup.conf Feb 01 18:56:49 vasily /usr/bin/vpp[42866]: load_one_plugin:63: Loaded plugin: /usr/lib/vpp_api_test_plugins/acl_test_plugin.so Feb 01 18:56:49 vasily /usr/bin/vpp[42866]: load_one_plugin:63: Loaded plugin: /usr/lib/vpp_api_test_plugins/kubeproxy_test_plugin.s Feb 01 18:56:49 vasily /usr/bin/vpp[42866]: load_one_plugin:63: Loaded plugin: /usr/lib/vpp_api_test_plugins/pppoe_test_plugin.so Feb 01 18:56:49 vasily /usr/bin/vpp[42866]: load_one_plugin:63: Loaded plugin: /usr/lib/vpp_api_test_plugins/ioam_vxlan_gpe_test_plu Feb 01 18:56:49 vasily /usr/bin/vpp[42866]: load_one_plugin:63: Loaded plugin: /usr/lib/vpp_api_test_plugins/dpdk_test_plugin.so Feb 01 18:56:49 vasily /usr/bin/vpp[42866]: load_one_plugin:63: Loaded plugin: /usr/lib/vpp_api_test_plugins/udp_ping_test_plugin.so Feb 01 18:56:49 vasily vpp[42866]: /usr/bin/vpp[42866]: dpdk_config:1240: EAL init args: -c 1 -n 4 --huge-dir /run/vpp/hugepages --f Feb 01 18:56:49 vasily /usr/bin/vpp[42866]: dpdk_config:1240: EAL init args: -c 1 -n 4 --huge-dir /run/vpp/hugepages --file-prefix v Feb 01 18:56:49 vasily vpp[42866]: EAL: VFIO support initialized Feb 01 18:56:49 vasily vnet[42866]: EAL: VFIO support initialized ...skipping... ● vpp.service - vector packet processing engine Loaded: loaded (/lib/systemd/system/vpp.service; enabled; vendor preset: enabled) Active: active (running) since Thu 2018-02-01 18:56:49 CST; 298ms ago Process: 42857 ExecStopPost=/bin/rm -f /dev/shm/db /dev/shm/global_vm /dev/shm/vpe-api (code=exited, status=0/SUCCESS) Process: 42863 ExecStartPre=/sbin/modprobe uio_pci_generic (code=exited, status=0/SUCCESS) Process: 42860 ExecStartPre=/bin/rm -f /dev/shm/db /dev/shm/global_vm /dev/shm/vpe-api (code=exited, status=0/SUCCESS) Main PID: 42866 (vpp) CGroup: /system.slice/vpp.service └─42866 /usr/bin/vpp -c /etc/vpp/startup.conf Feb 01 18:56:49 vasily /usr/bin/vpp[42866]: load_one_plugin:63: Loaded plugin: /usr/lib/vpp_api_test_plugins/acl_test_plugin.so Feb 01 18:56:49 vasily /usr/bin/vpp[42866]: load_one_plugin:63: Loaded plugin: /usr/lib/vpp_api_test_plugins/kubeproxy_test_plugin.s Feb 01 18:56:49 vasily /usr/bin/vpp[42866]: load_one_plugin:63: Loaded plugin: /usr/lib/vpp_api_test_plugins/pppoe_test_plugin.so Feb 01 18:56:49 vasily /usr/bin/vpp[42866]: load_one_plugin:63: Loaded plugin: /usr/lib/vpp_api_test_plugins/ioam_vxlan_gpe_test_plu Feb 01 18:56:49 vasily /usr/bin/vpp[42866]: load_one_plugin:63: Loaded plugin: /usr/lib/vpp_api_test_plugins/dpdk_test_plugin.so Feb 01 18:56:49 vasily /usr/bin/vpp[42866]: load_one_plugin:63: Loaded plugin: /usr/lib/vpp_api_test_plugins/udp_ping_test_plugin.so Feb 01 18:56:49 vasily vpp[42866]: /usr/bin/vpp[42866]: dpdk_config:1240: EAL init args: -c 1 -n 4 --huge-dir /run/vpp/hugepages --f Feb 01 18:56:49 vasily /usr/bin/vpp[42866]: dpdk_config:1240: EAL init args: -c 1 -n 4 --huge-dir /run/vpp/hugepages --file-prefix v Feb 01 18:56:49 vasily vpp[42866]: EAL: VFIO support initialized Feb 01 18:56:49 vasily vnet[42866]: EAL: VFIO support initialized ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ubuntu@vasily:~$ grep HugePages_Free /proc/meminfo HugePages_Free: 2335 ubuntu@vasily:~$ sudo service vpp status ● vpp.service - vector packet processing engine Loaded: loaded (/lib/systemd/system/vpp.service; enabled; vendor preset: enabled) Active: inactive (dead) since Thu 2018-02-01 18:56:56 CST; 63ms ago Process: 42917 ExecStopPost=/bin/rm -f /dev/shm/db /dev/shm/global_vm /dev/shm/vpe-api (code=exited, status=0/SUCCESS) Process: 42914 ExecStart=/usr/bin/vpp -c /etc/vpp/startup.conf (code=exited, status=0/SUCCESS) Process: 42911 ExecStartPre=/sbin/modprobe uio_pci_generic (code=exited, status=0/SUCCESS) Process: 42908 ExecStartPre=/bin/rm -f /dev/shm/db /dev/shm/global_vm /dev/shm/vpe-api (code=exited, status=0/SUCCESS) Main PID: 42914 (code=exited, status=0/SUCCESS) Feb 01 18:56:56 vasily systemd[1]: vpp.service: Service hold-off time over, scheduling restart. Feb 01 18:56:56 vasily systemd[1]: Stopped vector packet processing engine. Feb 01 18:56:56 vasily systemd[1]: vpp.service: Start request repeated too quickly. Feb 01 18:56:56 vasily systemd[1]: Failed to start vector packet processing engine. ubuntu@vasily:~$ grep HugePages_Free /proc/meminfo HugePages_Free: 5120 ubuntu@vasily:~$ On Wednesday 31 January 2018, 9:30:19 PM IST, Damjan Marion (damarion) <damar...@cisco.com<mailto:damar...@cisco.com>> wrote: On 31 Jan 2018, at 10:34, adarsh m via vpp-dev <vpp-dev@lists.fd.io<mailto:vpp-dev@lists.fd.io>> wrote: Hi, Pls check i am trying to bring up vpp with interface on ARM server but facing issue while doing so, pls let me know if there is any existing issue or method to correct this issue. ubuntu@vasily:~$ sudo service vpp status [sudo] password for ubuntu: ● vpp.service - vector packet processing engine Loaded: loaded (/lib/systemd/system/vpp.service; enabled; vendor preset: enab Active: active (running) since Mon 2018-01-29 22:07:02 CST; 19h ago Process: 2461 ExecStartPre=/sbin/modprobe uio_pci_generic (code=exited, status Process: 2453 ExecStartPre=/bin/rm -f /dev/shm/db /dev/shm/global_vm /dev/shm/ Main PID: 2472 (vpp_main) CGroup: /system.slice/vpp.service └─2472 /usr/bin/vpp -c /etc/vpp/startup.conf Jan 29 22:07:05 vasily vnet[2472]: dpdk_pool_create: failed to create dpdk_mbuf_ Jan 29 22:07:05 vasily vnet[2472]: dpdk_buffer_pool_create:573: WARNING: Failed Jan 29 22:07:05 vasily vnet[2472]: unix_physmem_region_iommu_register: ioctl (VF Jan 29 22:07:05 vasily vnet[2472]: dpdk_pool_create: failed to create dpdk_mbuf_ Jan 29 22:07:05 vasily vnet[2472]: dpdk_buffer_pool_create:573: WARNING: Failed Jan 29 22:07:05 vasily vnet[2472]: unix_physmem_region_iommu_register: ioctl (VF Jan 29 22:07:05 vasily vnet[2472]: dpdk_pool_create: failed to create dpdk_mbuf_ Jan 29 22:07:05 vasily vnet[2472]: dpdk_buffer_pool_create:573: WARNING: Failed Jan 29 22:07:05 vasily vnet[2472]: dpdk_ipsec_process:1011: not enough DPDK cryp Jan 29 22:07:05 vasily vnet[2472]: dpdk_lib_init:221: DPDK drivers found no port Looks like hugepages issue. can you show full log? What you pasted above is truncated... _______________________________________________ vpp-dev mailing list vpp-dev@lists.fd.io<mailto:vpp-dev@lists.fd.io> https://lists.fd.io/mailman/listinfo/vpp-dev _______________________________________________ vpp-dev mailing list vpp-dev@lists.fd.io<mailto:vpp-dev@lists.fd.io> https://lists.fd.io/mailman/listinfo/vpp-dev _______________________________________________ vpp-dev mailing list vpp-dev@lists.fd.io<mailto:vpp-dev@lists.fd.io> https://lists.fd.io/mailman/listinfo/vpp-dev _______________________________________________ vpp-dev mailing list vpp-dev@lists.fd.io<mailto:vpp-dev@lists.fd.io> https://lists.fd.io/mailman/listinfo/vpp-dev _______________________________________________ vpp-dev mailing list vpp-dev@lists.fd.io<mailto:vpp-dev@lists.fd.io> https://lists.fd.io/mailman/listinfo/vpp-dev