Hi Adarsh,

Would you be able to try the following?

In startup config for now, could you add the following change in the unix 
section of the vpp startup.

  # nodaemon
  interactive

Also, I see you have the line gid vpp. I am not sure what that does. Most of 
the configs I see do not have that. You could try removing that line.

Then run VPP manually like this so we might get more error information.

# cat /proc/meminfo – Be sure you still have some hugepages, sometimes if dpdk 
crashes it doesn’t return them.
# service vpp stop
# ps -eaf | grep vpp – Make sure there are no other VPP processes still running
# /usr/bin/vpp -c /etc/vpp/startup.conf

Then send us the results from the entire startup.

Thanks,

John


From: <vpp-dev-boun...@lists.fd.io> on behalf of adarsh m via vpp-dev 
<vpp-dev@lists.fd.io>
Reply-To: adarsh m <addi.ada...@yahoo.in>
Date: Thursday, February 1, 2018 at 8:43 AM
To: "Damjan Marion (damarion)" <damar...@cisco.com>, adarsh m via vpp-dev 
<vpp-dev@lists.fd.io>
Subject: Re: [vpp-dev] Error when trying to add interface to vpp on ARM server.

Hi,

After removing socket-mem now vpp is stable but when we try to connect through 
CLI connection is refused.


ubuntu@vasily:~$ sudo service vpp status
● vpp.service - vector packet processing engine
   Loaded: loaded (/lib/systemd/system/vpp.service; enabled; vendor preset: 
enabled)
   Active: active (running) since Thu 2018-02-01 21:42:23 CST; 1s ago
  Process: 49787 ExecStopPost=/bin/rm -f /dev/shm/db /dev/shm/global_vm 
/dev/shm/vpe-api (code=exited, status=0/SUCCESS)
  Process: 49793 ExecStartPre=/sbin/modprobe uio_pci_generic (code=exited, 
status=0/SUCCESS)
  Process: 49790 ExecStartPre=/bin/rm -f /dev/shm/db /dev/shm/global_vm 
/dev/shm/vpe-api (code=exited, status=0/SUCCESS)
 Main PID: 49796 (vpp)
   CGroup: /system.slice/vpp.service
           └─49796 /usr/bin/vpp -c /etc/vpp/startup.conf
Feb 01 19:37:16 vasily /usr/bin/vpp[43454]: load_one_plugin:63: Loaded plugin: 
/usr/lib/vpp_api_test_plugins/acl_test_plugin.so
Feb 01 19:37:16 vasily /usr/bin/vpp[43454]: load_one_plugin:63: Loaded plugin: 
/usr/lib/vpp_api_test_plugins/kubeproxy_test_plugin.so
Feb 01 19:37:16 vasily /usr/bin/vpp[43454]: load_one_plugin:63: Loaded plugin: 
/usr/lib/vpp_api_test_plugins/pppoe_test_plugin.so
Feb 01 19:37:16 vasily /usr/bin/vpp[43454]: load_one_plugin:63: Loaded plugin: 
/usr/lib/vpp_api_test_plugins/ioam_vxlan_gpe_test_plugin.so
Feb 01 19:37:16 vasily /usr/bin/vpp[43454]: load_one_plugin:63: Loaded plugin: 
/usr/lib/vpp_api_test_plugins/dpdk_test_plugin.so
Feb 01 19:37:16 vasily /usr/bin/vpp[43454]: load_one_plugin:63: Loaded plugin: 
/usr/lib/vpp_api_test_plugins/udp_ping_test_plugin.so
Feb 01 19:37:16 vasily vpp[43454]: /usr/bin/vpp[43454]: dpdk_config:1240: EAL 
init args: -c 1 -n 4 --huge-dir /run/vpp/hugepages --file-prefix vpp -w 
0002:f9:00.0 --master-lcore 0 --socket-mem 1024,
Feb 01 19:37:16 vasily /usr/bin/vpp[43454]: dpdk_config:1240: EAL init args: -c 
1 -n 4 --huge-dir /run/vpp/hugepages --file-prefix vpp -w 0002:f9:00.0 
--master-lcore 0 --socket-mem 1024,0,0,0
Feb 01 19:37:16 vasily vpp[43454]: EAL: VFIO support initialized
Feb 01 19:37:16 vasily vnet[43454]: EAL: VFIO support initialized
ubuntu@vasily:~$ sudo vppctl show int
clib_socket_init: connect (fd 3, '/run/vpp/cli.sock'): Connection refused
ubuntu@vasily:~$

Startup.conf :
unix {
  nodaemon
  log /tmp/vpp.log
  full-coredump
  cli-listen /run/vpp/cli.sock
  gid vpp
}

api-trace {
## This stanza controls binary API tracing. Unless there is a very strong 
reason,
## please leave this feature enabled.
  on
## Additional parameters:
##
## To set the number of binary API trace records in the circular buffer, 
configure nitems
##
## nitems <nnn>
##
## To save the api message table decode tables, configure a filename. Results 
in /tmp/<filename>
## Very handy for understanding api message changes between versions, 
identifying missing
## plugins, and so forth.
##
## save-api-table <filename>
}

api-segment {
  gid vpp
}

cpu {
    ## In the VPP there is one main thread and optionally the user can create 
worker(s)
    ## The main thread and worker thread(s) can be pinned to CPU core(s) 
manually or automatically

    ## Manual pinning of thread(s) to CPU core(s)

    ## Set logical CPU core where main thread runs
    # main-core 1

    ## Set logical CPU core(s) where worker threads are running
    # corelist-workers 2-3,18-19

    ## Automatic pinning of thread(s) to CPU core(s)

    ## Sets number of CPU core(s) to be skipped (1 ... N-1)
    ## Skipped CPU core(s) are not used for pinning main thread and working 
thread(s).
    ## The main thread is automatically pinned to the first available CPU core 
and worker(s)
    ## are pinned to next free CPU core(s) after core assigned to main thread
    # skip-cores 4

    ## Specify a number of workers to be created
    ## Workers are pinned to N consecutive CPU cores while skipping 
"skip-cores" CPU core(s)
    ## and main thread's CPU core
    # workers 2

    ## Set scheduling policy and priority of main and worker threads

    ## Scheduling policy options are: other (SCHED_OTHER), batch (SCHED_BATCH)
    ## idle (SCHED_IDLE), fifo (SCHED_FIFO), rr (SCHED_RR)
    # scheduler-policy fifo

    ## Scheduling priority is used only for "real-time policies (fifo and rr),
    ## and has to be in the range of priorities supported for a particular 
policy
    # scheduler-priority 50
}

 dpdk {

       dev 0002:f9:00.0
    ## Change default settings for all intefaces
    # dev default {
        ## Number of receive queues, enables RSS
        ## Default is 1
        # num-rx-queues 3

        ## Number of transmit queues, Default is equal
        ## to number of worker threads or 1 if no workers treads
        # num-tx-queues 3

        ## Number of descriptors in transmit and receive rings
        ## increasing or reducing number can impact performance
        ## Default is 1024 for both rx and tx
        # num-rx-desc 512
        # num-tx-desc 512

        ## VLAN strip offload mode for interface
        ## Default is off
        # vlan-strip-offload on
    # }

    ## Whitelist specific interface by specifying PCI address
    # dev 0000:02:00.0

    ## Whitelist specific interface by specifying PCI address and in
    ## addition specify custom parameters for this interface
    # dev 0000:02:00.1 {
    #    num-rx-queues 2
    # }

    ## Specify bonded interface and its slaves via PCI addresses
    ##
        ## Bonded interface in XOR load balance mode (mode 2) with L3 and L4 
headers
    # vdev 
eth_bond0,mode=2,slave=0000:02:00.0,slave=0000:03:00.0,xmit_policy=l34
    # vdev 
eth_bond1,mode=2,slave=0000:02:00.1,slave=0000:03:00.1,xmit_policy=l34
    ##
    ## Bonded interface in Active-Back up mode (mode 1)
    # vdev eth_bond0,mode=1,slave=0000:02:00.0,slave=0000:03:00.0
    # vdev eth_bond1,mode=1,slave=0000:02:00.1,slave=0000:03:00.1

    ## Change UIO driver used by VPP, Options are: igb_uio, vfio-pci,
    ## uio_pci_generic or auto (default)
    # uio-driver vfio-pci

    ## Disable mutli-segment buffers, improves performance but
    ## disables Jumbo MTU support
    # no-multi-seg

    ## Increase number of buffers allocated, needed only in scenarios with
    ## large number of interfaces and worker threads. Value is per CPU socket.
    ## Default is 16384
    # num-mbufs 128000
        num-mbufs 4095
    ## Change hugepages allocation per-socket, needed only if there is need for
    ## larger number of mbufs. Default is 256M on each detected CPU socket
    # socket-mem 2048,2048

    ## Disables UDP / TCP TX checksum offload. Typically needed for use
    ## faster vector PMDs (together with no-multi-seg)
    # no-tx-checksum-offload
 }

# Adjusting the plugin path depending on where the VPP plugins are:
#plugins
#{
#    path /home/bms/vpp/build-root/install-vpp-native/vpp/lib64/vpp_plugins
#}

# Alternate syntax to choose plugin path
#plugin_path /home/bms/vpp/build-root/install-vpp-native/vpp/lib64/vpp_plugins

On Thursday 1 February 2018, 7:00:18 PM IST, adarsh m via vpp-dev 
<vpp-dev@lists.fd.io> wrote:



Support 2 channel 32core processor(ARM64)


On Thursday 1 February 2018, 6:47:31 PM IST, Damjan Marion (damarion) 
<damar...@cisco.com> wrote:


You also added socket-mem which is pretty bad idea, try without.
If that doesn't help then you will need to run VPP form console and possibly 
use gdb to collect more details.

Which ARM board is that?



On 1 Feb 2018, at 14:15, adarsh m 
<addi.ada...@yahoo.in<mailto:addi.ada...@yahoo.in>> wrote:

Hi,

This is on ARM board. and yes i have modified the specific pcie address to 
added in startup.conf

 dpdk {
       socket-mem 1024
       dev 0002:f9:00.0

}


On Thursday 1 February 2018, 5:20:59 PM IST, Damjan Marion (damarion) 
<damar...@cisco.com<mailto:damar...@cisco.com>> wrote:


Please keep mailing list in CC.

Those lines doesn't show that anything is wrong...

Is this 4 socket computer? Have you modified startup.conf?



On 1 Feb 2018, at 12:40, adarsh m 
<addi.ada...@yahoo.in<mailto:addi.ada...@yahoo.in>> wrote:

Hi,

Very sorry pls check the complete one.

Feb 01 19:37:16 vasily /usr/bin/vpp[43454]: load_one_plugin:63: Loaded plugin: 
/usr/lib/vpp_api_test_plugins/acl_test_plugin.so
Feb 01 19:37:16 vasily /usr/bin/vpp[43454]: load_one_plugin:63: Loaded plugin: 
/usr/lib/vpp_api_test_plugins/kubeproxy_test_plugin.so
Feb 01 19:37:16 vasily /usr/bin/vpp[43454]: load_one_plugin:63: Loaded plugin: 
/usr/lib/vpp_api_test_plugins/pppoe_test_plugin.so
Feb 01 19:37:16 vasily /usr/bin/vpp[43454]: load_one_plugin:63: Loaded plugin: 
/usr/lib/vpp_api_test_plugins/ioam_vxlan_gpe_test_plugin.so
Feb 01 19:37:16 vasily /usr/bin/vpp[43454]: load_one_plugin:63: Loaded plugin: 
/usr/lib/vpp_api_test_plugins/dpdk_test_plugin.so
Feb 01 19:37:16 vasily /usr/bin/vpp[43454]: load_one_plugin:63: Loaded plugin: 
/usr/lib/vpp_api_test_plugins/udp_ping_test_plugin.so
Feb 01 19:37:16 vasily vpp[43454]: /usr/bin/vpp[43454]: dpdk_config:1240: EAL 
init args: -c 1 -n 4 --huge-dir /run/vpp/hugepages --file-prefix vpp -w 
0002:f9:00.0 --master-lcore 0 --socket-mem 1024,
Feb 01 19:37:16 vasily /usr/bin/vpp[43454]: dpdk_config:1240: EAL init args: -c 
1 -n 4 --huge-dir /run/vpp/hugepages --file-prefix vpp -w 0002:f9:00.0 
--master-lcore 0 --socket-mem 1024,0,0,0
Feb 01 19:37:16 vasily vpp[43454]: EAL: VFIO support initialized
Feb 01 19:37:16 vasily vnet[43454]: EAL: VFIO support initialized


On Thursday 1 February 2018, 4:48:35 PM IST, Damjan Marion (damarion) 
<damar...@cisco.com<mailto:damar...@cisco.com>> wrote:



Unfortunately log you provided is incomplete and truncated so I cannot help 
much....


On 1 Feb 2018, at 11:59, adarsh m 
<addi.ada...@yahoo.in<mailto:addi.ada...@yahoo.in>> wrote:

Hi,

I checked hugepage and it was 0, so i freed this and increased it to 5120

ubuntu@vasily:~$ sudo -i
root@vasily:~# echo 5120 > /proc/sys/vm/nr_hugepages

now the previous error is not occurring when i start but VPP is not stable, 
it'll become dead after a few sceonds from start.

Logs :
ubuntu@vasily:~$ grep HugePages_Free /proc/meminfo
HugePages_Free:     5120
ubuntu@vasily:~$
ubuntu@vasily:~$
ubuntu@vasily:~$
ubuntu@vasily:~$ sudo service vpp status
● vpp.service - vector packet processing engine
   Loaded: loaded (/lib/systemd/system/vpp.service; enabled; vendor preset: 
enabled)
   Active: inactive (dead) since Thu 2018-02-01 18:50:46 CST; 5min ago
  Process: 42736 ExecStopPost=/bin/rm -f /dev/shm/db /dev/shm/global_vm 
/dev/shm/vpe-api (code=exited, status=0/SUCCESS)
  Process: 42731 ExecStart=/usr/bin/vpp -c /etc/vpp/startup.conf (code=exited, 
status=0/SUCCESS)
  Process: 42728 ExecStartPre=/sbin/modprobe uio_pci_generic (code=exited, 
status=0/SUCCESS)
  Process: 42726 ExecStartPre=/bin/rm -f /dev/shm/db /dev/shm/global_vm 
/dev/shm/vpe-api (code=exited, status=0/SUCCESS)
 Main PID: 42731 (code=exited, status=0/SUCCESS)

Feb 01 18:50:46 vasily systemd[1]: vpp.service: Service hold-off time over, 
scheduling restart.
Feb 01 18:50:46 vasily systemd[1]: Stopped vector packet processing engine.
Feb 01 18:50:46 vasily systemd[1]: vpp.service: Start request repeated too 
quickly.
Feb 01 18:50:46 vasily systemd[1]: Failed to start vector packet processing 
engine.
Feb 01 18:56:12 vasily systemd[1]: Stopped vector packet processing engine.
ubuntu@vasily:~$
ubuntu@vasily:~$
ubuntu@vasily:~$
ubuntu@vasily:~$ sudo service vpp start
ubuntu@vasily:~$
ubuntu@vasily:~$
ubuntu@vasily:~$ sudo service vpp status
● vpp.service - vector packet processing engine
   Loaded: loaded (/lib/systemd/system/vpp.service; enabled; vendor preset: 
enabled)
   Active: active (running) since Thu 2018-02-01 18:56:49 CST; 298ms ago
  Process: 42857 ExecStopPost=/bin/rm -f /dev/shm/db /dev/shm/global_vm 
/dev/shm/vpe-api (code=exited, status=0/SUCCESS)
  Process: 42863 ExecStartPre=/sbin/modprobe uio_pci_generic (code=exited, 
status=0/SUCCESS)
  Process: 42860 ExecStartPre=/bin/rm -f /dev/shm/db /dev/shm/global_vm 
/dev/shm/vpe-api (code=exited, status=0/SUCCESS)
 Main PID: 42866 (vpp)
   CGroup: /system.slice/vpp.service
           └─42866 /usr/bin/vpp -c /etc/vpp/startup.conf

Feb 01 18:56:49 vasily /usr/bin/vpp[42866]: load_one_plugin:63: Loaded plugin: 
/usr/lib/vpp_api_test_plugins/acl_test_plugin.so
Feb 01 18:56:49 vasily /usr/bin/vpp[42866]: load_one_plugin:63: Loaded plugin: 
/usr/lib/vpp_api_test_plugins/kubeproxy_test_plugin.s
Feb 01 18:56:49 vasily /usr/bin/vpp[42866]: load_one_plugin:63: Loaded plugin: 
/usr/lib/vpp_api_test_plugins/pppoe_test_plugin.so
Feb 01 18:56:49 vasily /usr/bin/vpp[42866]: load_one_plugin:63: Loaded plugin: 
/usr/lib/vpp_api_test_plugins/ioam_vxlan_gpe_test_plu
Feb 01 18:56:49 vasily /usr/bin/vpp[42866]: load_one_plugin:63: Loaded plugin: 
/usr/lib/vpp_api_test_plugins/dpdk_test_plugin.so
Feb 01 18:56:49 vasily /usr/bin/vpp[42866]: load_one_plugin:63: Loaded plugin: 
/usr/lib/vpp_api_test_plugins/udp_ping_test_plugin.so
Feb 01 18:56:49 vasily vpp[42866]: /usr/bin/vpp[42866]: dpdk_config:1240: EAL 
init args: -c 1 -n 4 --huge-dir /run/vpp/hugepages --f
Feb 01 18:56:49 vasily /usr/bin/vpp[42866]: dpdk_config:1240: EAL init args: -c 
1 -n 4 --huge-dir /run/vpp/hugepages --file-prefix v
Feb 01 18:56:49 vasily vpp[42866]: EAL: VFIO support initialized
Feb 01 18:56:49 vasily vnet[42866]: EAL: VFIO support initialized
...skipping...
● vpp.service - vector packet processing engine
   Loaded: loaded (/lib/systemd/system/vpp.service; enabled; vendor preset: 
enabled)
   Active: active (running) since Thu 2018-02-01 18:56:49 CST; 298ms ago
  Process: 42857 ExecStopPost=/bin/rm -f /dev/shm/db /dev/shm/global_vm 
/dev/shm/vpe-api (code=exited, status=0/SUCCESS)
  Process: 42863 ExecStartPre=/sbin/modprobe uio_pci_generic (code=exited, 
status=0/SUCCESS)
  Process: 42860 ExecStartPre=/bin/rm -f /dev/shm/db /dev/shm/global_vm 
/dev/shm/vpe-api (code=exited, status=0/SUCCESS)
 Main PID: 42866 (vpp)
   CGroup: /system.slice/vpp.service
           └─42866 /usr/bin/vpp -c /etc/vpp/startup.conf

Feb 01 18:56:49 vasily /usr/bin/vpp[42866]: load_one_plugin:63: Loaded plugin: 
/usr/lib/vpp_api_test_plugins/acl_test_plugin.so
Feb 01 18:56:49 vasily /usr/bin/vpp[42866]: load_one_plugin:63: Loaded plugin: 
/usr/lib/vpp_api_test_plugins/kubeproxy_test_plugin.s
Feb 01 18:56:49 vasily /usr/bin/vpp[42866]: load_one_plugin:63: Loaded plugin: 
/usr/lib/vpp_api_test_plugins/pppoe_test_plugin.so
Feb 01 18:56:49 vasily /usr/bin/vpp[42866]: load_one_plugin:63: Loaded plugin: 
/usr/lib/vpp_api_test_plugins/ioam_vxlan_gpe_test_plu
Feb 01 18:56:49 vasily /usr/bin/vpp[42866]: load_one_plugin:63: Loaded plugin: 
/usr/lib/vpp_api_test_plugins/dpdk_test_plugin.so
Feb 01 18:56:49 vasily /usr/bin/vpp[42866]: load_one_plugin:63: Loaded plugin: 
/usr/lib/vpp_api_test_plugins/udp_ping_test_plugin.so
Feb 01 18:56:49 vasily vpp[42866]: /usr/bin/vpp[42866]: dpdk_config:1240: EAL 
init args: -c 1 -n 4 --huge-dir /run/vpp/hugepages --f
Feb 01 18:56:49 vasily /usr/bin/vpp[42866]: dpdk_config:1240: EAL init args: -c 
1 -n 4 --huge-dir /run/vpp/hugepages --file-prefix v
Feb 01 18:56:49 vasily vpp[42866]: EAL: VFIO support initialized
Feb 01 18:56:49 vasily vnet[42866]: EAL: VFIO support initialized
~
~
~
~
~
~
~
~
~
~
~
~
~
~
~
~
~
~
~
~
~
~
~
~
~
~
~
~
~

ubuntu@vasily:~$ grep HugePages_Free /proc/meminfo
HugePages_Free:     2335
ubuntu@vasily:~$ sudo service vpp status
● vpp.service - vector packet processing engine
   Loaded: loaded (/lib/systemd/system/vpp.service; enabled; vendor preset: 
enabled)
   Active: inactive (dead) since Thu 2018-02-01 18:56:56 CST; 63ms ago
  Process: 42917 ExecStopPost=/bin/rm -f /dev/shm/db /dev/shm/global_vm 
/dev/shm/vpe-api (code=exited, status=0/SUCCESS)
  Process: 42914 ExecStart=/usr/bin/vpp -c /etc/vpp/startup.conf (code=exited, 
status=0/SUCCESS)
  Process: 42911 ExecStartPre=/sbin/modprobe uio_pci_generic (code=exited, 
status=0/SUCCESS)
  Process: 42908 ExecStartPre=/bin/rm -f /dev/shm/db /dev/shm/global_vm 
/dev/shm/vpe-api (code=exited, status=0/SUCCESS)
 Main PID: 42914 (code=exited, status=0/SUCCESS)

Feb 01 18:56:56 vasily systemd[1]: vpp.service: Service hold-off time over, 
scheduling restart.
Feb 01 18:56:56 vasily systemd[1]: Stopped vector packet processing engine.
Feb 01 18:56:56 vasily systemd[1]: vpp.service: Start request repeated too 
quickly.
Feb 01 18:56:56 vasily systemd[1]: Failed to start vector packet processing 
engine.
ubuntu@vasily:~$ grep HugePages_Free /proc/meminfo
HugePages_Free:     5120
ubuntu@vasily:~$


On Wednesday 31 January 2018, 9:30:19 PM IST, Damjan Marion (damarion) 
<damar...@cisco.com<mailto:damar...@cisco.com>> wrote:



On 31 Jan 2018, at 10:34, adarsh m via vpp-dev 
<vpp-dev@lists.fd.io<mailto:vpp-dev@lists.fd.io>> wrote:

Hi,

Pls check i am trying to bring up vpp with interface on ARM server but facing 
issue while doing so,

pls let me know if there is any existing issue or method to correct this issue.


ubuntu@vasily:~$ sudo service vpp status
[sudo] password for ubuntu:
● vpp.service - vector packet processing engine
   Loaded: loaded (/lib/systemd/system/vpp.service; enabled; vendor preset: enab
   Active: active (running) since Mon 2018-01-29 22:07:02 CST; 19h ago
  Process: 2461 ExecStartPre=/sbin/modprobe uio_pci_generic (code=exited, status
  Process: 2453 ExecStartPre=/bin/rm -f /dev/shm/db /dev/shm/global_vm /dev/shm/
 Main PID: 2472 (vpp_main)
   CGroup: /system.slice/vpp.service
           └─2472 /usr/bin/vpp -c /etc/vpp/startup.conf

Jan 29 22:07:05 vasily vnet[2472]: dpdk_pool_create: failed to create dpdk_mbuf_
Jan 29 22:07:05 vasily vnet[2472]: dpdk_buffer_pool_create:573: WARNING: Failed
Jan 29 22:07:05 vasily vnet[2472]: unix_physmem_region_iommu_register: ioctl (VF
Jan 29 22:07:05 vasily vnet[2472]: dpdk_pool_create: failed to create dpdk_mbuf_
Jan 29 22:07:05 vasily vnet[2472]: dpdk_buffer_pool_create:573: WARNING: Failed
Jan 29 22:07:05 vasily vnet[2472]: unix_physmem_region_iommu_register: ioctl (VF
Jan 29 22:07:05 vasily vnet[2472]: dpdk_pool_create: failed to create dpdk_mbuf_
Jan 29 22:07:05 vasily vnet[2472]: dpdk_buffer_pool_create:573: WARNING: Failed
Jan 29 22:07:05 vasily vnet[2472]: dpdk_ipsec_process:1011: not enough DPDK cryp
Jan 29 22:07:05 vasily vnet[2472]: dpdk_lib_init:221: DPDK drivers found no port


Looks like hugepages issue. can you show full log? What you pasted above is 
truncated...

_______________________________________________
vpp-dev mailing list
vpp-dev@lists.fd.io<mailto:vpp-dev@lists.fd.io>
https://lists.fd.io/mailman/listinfo/vpp-dev

_______________________________________________
vpp-dev mailing list
vpp-dev@lists.fd.io<mailto:vpp-dev@lists.fd.io>
https://lists.fd.io/mailman/listinfo/vpp-dev

_______________________________________________
vpp-dev mailing list
vpp-dev@lists.fd.io<mailto:vpp-dev@lists.fd.io>
https://lists.fd.io/mailman/listinfo/vpp-dev

_______________________________________________
vpp-dev mailing list
vpp-dev@lists.fd.io<mailto:vpp-dev@lists.fd.io>
https://lists.fd.io/mailman/listinfo/vpp-dev
_______________________________________________
vpp-dev mailing list
vpp-dev@lists.fd.io<mailto:vpp-dev@lists.fd.io>
https://lists.fd.io/mailman/listinfo/vpp-dev

Reply via email to