Dear Chris, Whew, that just made my weekend a lot happier.
I'll look into why the relevant patch didn't make it back into 19.08 - it will now! - unfortunately "stuff happens..." Thanks for confirming... Dave -----Original Message----- From: Christian Hopps <cho...@chopps.org> Sent: Friday, June 5, 2020 5:09 PM To: Dave Barach (dbarach) <dbar...@cisco.com> Cc: Christian Hopps <cho...@chopps.org>; vpp-dev <vpp-dev@lists.fd.io> Subject: Re: [vpp-dev] Interesting backtrace in 1908 Bingo. In fact in 19.08 the value is left as 0 which defaults to 15. I took it from 20 down to 15, starting successfully until I reached 15 which then hit the problem (both with the arp path and the other). Thanks for the help finding this! Chris. > On Jun 5, 2020, at 4:52 PM, Dave Barach via lists.fd.io > <dbarach=cisco....@lists.fd.io> wrote: > > Hmmm. That begins to smell like an undetected stack overflow. To test that > theory: s/18/20/ below: > > /* *INDENT-OFF* */ > VLIB_REGISTER_NODE (startup_config_node,static) = { > .function = startup_config_process, > .type = VLIB_NODE_TYPE_PROCESS, > .name = "startup-config-process", > .process_log2_n_stack_bytes = 18, > }; > /* *INDENT-ON* */ > > It's entirely possible that compiling -O0 blows the stack, especially if you > end up 75 miles deep in fib code. > > Dave > > -----Original Message----- > From: Christian Hopps <cho...@chopps.org> > Sent: Friday, June 5, 2020 4:28 PM > To: Dave Barach (dbarach) <dbar...@cisco.com> > Cc: Christian Hopps <cho...@chopps.org>; vpp-dev <vpp-dev@lists.fd.io> > Subject: Re: [vpp-dev] Interesting backtrace in 1908 > > > >> On Jun 5, 2020, at 2:10 PM, Dave Barach via lists.fd.io >> <dbarach=cisco....@lists.fd.io> wrote: >> >> Step 1 is to make the silly-looking sibling recursion in >> vlib_node_add_next_with_slot(...) disappear. I’m on it... >> >> Just to ask, can you repro w/ master/latest? > > I will try and do this. > > In the meantime I moved the arp configs to later in my startup config (this > is actually built by a test script) and immediately hit another sigsegv in > startup. This one is in infra but is going through my code initialization, > but also rooted in startup config processing... It's also in memcpy code, > which is making me suspicious now. > > Again, I've changed "-O2" to "-O0" in the cmake vpp.mk package, when I change > it back to -O2 I do not hit either bug. So I'm now wondering if there is > something wrong with doing this, like do I need to do something else as well? > > What I'm going for is not to have CLIB_DEBUG defined, but still have useful > levels of debugabillity to do RCA on a much (millions of packets) later > problem I have. > > modified build-data/platforms/vpp.mk > @@ -35,13 +35,21 @@ vpp_debug_TAG_CFLAGS = -O0 -DCLIB_DEBUG > $(vpp_common_cflags) vpp_debug_TAG_CXXFLAGS = -O0 -DCLIB_DEBUG > $(vpp_common_cflags) vpp_debug_TAG_LDFLAGS = -O0 -DCLIB_DEBUG > $(vpp_common_cflags) > > -vpp_TAG_CFLAGS = -O2 $(vpp_common_cflags) -vpp_TAG_CXXFLAGS = -O2 > $(vpp_common_cflags) -vpp_TAG_LDFLAGS = -O2 $(vpp_common_cflags) -pie > +vpp_TAG_CFLAGS = -O0 $(vpp_common_cflags) vpp_TAG_CXXFLAGS = -O0 > +$(vpp_common_cflags) vpp_TAG_LDFLAGS = -O0 $(vpp_common_cflags) -pie > > The new backtrace I'm seeing is: > > Thread 1 "vpp_main" received signal SIGSEGV, Segmentation fault. > 0x00007ffff5cbe8b1 in clib_mov16 (dst=0x6d986a570000 <error: Cannot > access memory at address 0x6d986a570000>, src=0x2d8f0700006c2d08 > <error: Cannot access memory at address 0x2d8f0700006c2d08>) at > /var/build/vpp/src/vppinfra/memcpy_sse3.h:56 > > (gdb) bt > #0 0x00007ffff5cbe8b1 in clib_mov16 (dst=0x6d986a570000 <error: > Cannot access memory at address 0x6d986a570000>, > src=0x2d8f0700006c2d08 <error: Cannot access memory at address > 0x2d8f0700006c2d08>) at /var/build/vpp/src/vppinfra/memcpy_sse3.h:56 > #1 0x00007ffff5cbe910 in clib_mov32 (dst=0x7fffb90bea90 "", > src=0x7fffafa01fd0 "iptfs-refill-zpool sa_index %d before %d requested > %d head %d tail %d") at /var/build/vpp/src/vppinfra/memcpy_sse3.h:66 > #2 0x00007ffff5cbe951 in clib_mov64 (dst=0x7fffb90bea90 "", > src=0x7fffafa01fd0 "iptfs-refill-zpool sa_index %d before %d requested > %d head %d tail %d") at /var/build/vpp/src/vppinfra/memcpy_sse3.h:73 > #3 0x00007ffff5cbed5a in clib_memcpy_fast (dst=0x7fffb90bea90, > src=0x7fffafa01fd0, n=5) at > /var/build/vpp/src/vppinfra/memcpy_sse3.h:273 > #4 0x00007ffff5cc5e72 in do_percent (_s=0x7fffb8141a58, > fmt=0x7ffff5dc7734 "%s%c", va=0x7fffb8141bc8) at > /var/build/vpp/src/vppinfra/format.c:341 > #5 0x00007ffff5cc6564 in va_format (s=0x0, fmt=0x7ffff5dc7734 "%s%c", > va=0x7fffb8141bc8) at /var/build/vpp/src/vppinfra/format.c:404 > #6 0x00007ffff5cc6810 in format (s=0x0, fmt=0x7ffff5dc7734 "%s%c") at > /var/build/vpp/src/vppinfra/format.c:428 > #7 0x00007ffff5cace7d in elog_event_type_register (em=0x7ffff656c7a8 > <vlib_global_main+936>, t=0x7fffb9058300) at > /var/build/vpp/src/vppinfra/elog.c:173 > #8 0x00007fffaf9b8ba4 in elog_event_data_inline > (cpu_time=3193258542505306, track=0x7fffb9093f98, type=0x7fffafc08880 > <e>, em=0x7ffff656c7a8 <vlib_global_main+936>) at > /var/build/vpp/src/vppinfra/elog.h:310 > #9 elog_data_inline (track=0x7fffb9093f98, type=0x7fffafc08880 <e>, > em=0x7ffff656c7a8 <vlib_global_main+936>) at > /var/build/vpp/src/vppinfra/elog.h:435 > #10 iptfs_refill_zpool (vm=0x7ffff656c400 <vlib_global_main>, > zpool=0x7fffb915a8c0, sa_index=1, payload_size=1470, put=false, > track=0x7fffb9093f98) at > /var/build/vpp/src/plugins/iptfs/iptfs_zpool.c:134 > #11 0x00007fffaf9b9d3f in iptfs_zpool_alloc (vm=0x7ffff656c400 > <vlib_global_main>, queue_size=768, sa_index=1, payload_size=1470, > put=false, track=0x7fffb9093f98) at > /var/build/vpp/src/plugins/iptfs/iptfs_zpool.c:235 > #12 0x00007fffaf99d73c in iptfs_tfs_data_init (sa_index=1, > conf=0x7fffb91cd7c0) at > /var/build/vpp/src/plugins/iptfs/ipsec_iptfs.c:347 > #13 0x00007fffaf9a0a09 in iptfs_add_del_sa (sa_index=1, > tfs_config=0x7fffb91cd7c0, is_add=1 '\001') at > /var/build/vpp/src/plugins/iptfs/ipsec_iptfs.c:822 > #14 0x00007ffff6fff6a8 in ipsec_sa_add_and_lock (id=3221225472, > spi=1112, proto=IPSEC_PROTOCOL_ESP, crypto_alg=IPSEC_CRYPTO_ALG_NONE, > ck=0x7fffb8144cf0, integ_alg=IPSEC_INTEG_ALG_NONE, ik=0x7fffb8144d80, > flags=(IPSEC_SA_FLAG_USE_ESN | IPSEC_SA_FLAG_IS_TUNNEL), _tfs_type=2 > '\002', tfs_config=0x7fffb91cd7c0, tx_table_id=0, salt=0, > tun_src=0x7fffb8145485, tun_dst=0x7fffb8145495, > sa_out_index=0x7fffb87dc384) at > /var/build/vpp/src/vnet/ipsec/ipsec_sa.c:217 > #15 0x00007ffff6feb7ce in ipsec_add_del_tunnel_if_internal > (vnm=0x7ffff7b47e80 <vnet_main>, args=0x7fffb8145480, > sw_if_index_p=0x0) at /var/build/vpp/src/vnet/ipsec/ipsec_if.c:370 > #16 0x00007ffff6fea044 in ipsec_add_del_tunnel_if_rpc_callback > (a=0x7fffb8145480) at /var/build/vpp/src/vnet/ipsec/ipsec_if.c:210 > #17 0x00007ffff7ba47c9 in vl_api_rpc_call_main_thread_inline > (force_rpc=0 '\000', data_length=600, data=0x7fffb8145480 "\001", > fp=0x7ffff6fea017 <ipsec_add_del_tunnel_if_rpc_callback>) at > /var/build/vpp/src/vlibmemory/vlib_api.c:571 > #18 vl_api_rpc_call_main_thread (fp=0x7ffff6fea017 > <ipsec_add_del_tunnel_if_rpc_callback>, data=0x7fffb8145480 "\001", > data_length=600) at /var/build/vpp/src/vlibmemory/vlib_api.c:602 > #19 0x00007ffff6fea06a in ipsec_add_del_tunnel_if > (args=0x7fffb8145480) at /var/build/vpp/src/vnet/ipsec/ipsec_if.c:216 > #20 0x00007ffff6fde020 in create_ipsec_tunnel_command_fn > (vm=0x7ffff656c400 <vlib_global_main>, input=0x7fffb8146ed0, > cmd=0x7fffb82e29a8) at /var/build/vpp/src/vnet/ipsec/ipsec_cli.c:857 > #21 0x00007ffff62385f2 in vlib_cli_dispatch_sub_commands > (vm=0x7ffff656c400 <vlib_global_main>, cm=0x7ffff656c630 > <vlib_global_main+560>, input=0x7fffb8146ed0, > parent_command_index=772) at /var/build/vpp/src/vlib/cli.c:649 > #22 0x00007ffff62383a3 in vlib_cli_dispatch_sub_commands > (vm=0x7ffff656c400 <vlib_global_main>, cm=0x7ffff656c630 > <vlib_global_main+560>, input=0x7fffb8146ed0, parent_command_index=10) > at /var/build/vpp/src/vlib/cli.c:609 > #23 0x00007ffff62383a3 in vlib_cli_dispatch_sub_commands > (vm=0x7ffff656c400 <vlib_global_main>, cm=0x7ffff656c630 > <vlib_global_main+560>, input=0x7fffb8146ed0, parent_command_index=0) > at /var/build/vpp/src/vlib/cli.c:609 > #24 0x00007ffff62391c3 in vlib_cli_input (vm=0x7ffff656c400 > <vlib_global_main>, input=0x7fffb8146ed0, function=0x0, > function_arg=0) at /var/build/vpp/src/vlib/cli.c:750 > #25 0x00007ffff6344666 in startup_config_process (vm=0x7ffff656c400 > <vlib_global_main>, rt=0x7fffb813e000, f=0x0) at > /var/build/vpp/src/vlib/unix/main.c:367 > #26 0x00007ffff6297fa0 in vlib_process_bootstrap (_a=140736274672448) > at /var/build/vpp/src/vlib/main.c:1472 > #27 0x00007ffff5ce23d8 in clib_calljmp () from > target:/var/build/vpp/build-root/install-vpp-native/vpp/lib/libvppinfr > a.so.19.08.2 > #28 0x00007fffb7a8a7e0 in ?? () > #29 0x00007ffff62981ec in vlib_process_startup (f=0x7fffb7f71c74, > p=0x7fffb813e000, vm=0x7fffb8f498c8) at > /var/build/vpp/src/vlib/main.c:1494 > #30 dispatch_process (vm=0x7fffb87ef9c0, p=0x30, f=0x30, > last_time_stamp=140736297535488) at > /var/build/vpp/src/vlib/main.c:1539 > #31 0x00007ffff629fa4b in vlib_main_or_worker_loop (is_main=1, > vm=0x7ffff656c400 <vlib_global_main>) at > /var/build/vpp/src/vlib/main.c:1914 > #32 vlib_main_loop (vm=0x7ffff656c400 <vlib_global_main>) at > /var/build/vpp/src/vlib/main.c:1931 > #33 0x00007ffff62a6cef in vlib_main (vm=0x7ffff656c400 > <vlib_global_main>, input=0x7fffb7a8bfb0) at > /var/build/vpp/src/vlib/main.c:2148 > #34 0x00007ffff6345ad6 in thread0 (arg=140737326269440) at > /var/build/vpp/src/vlib/unix/main.c:649 > #35 0x00007ffff5ce23d8 in clib_calljmp () from > target:/var/build/vpp/build-root/install-vpp-native/vpp/lib/libvppinfr > a.so.19.08.2 > #36 0x00007fffffffcf60 in ?? () > #37 0x00007ffff6346b02 in vlib_unix_main (argc=69, > argv=0x7fffffffe4e8) at /var/build/vpp/src/vlib/unix/main.c:719 > #38 0x000055555555b801 in main (argc=69, argv=0x7fffffffe4e8) at > /var/build/vpp/src/vpp/vnet/main.c:280 > (gdb) fr 4 > #4 0x00007ffff5cc5e72 in do_percent (_s=0x7fffb8141a58, fmt=0x7ffff5dc7734 > "%s%c", va=0x7fffb8141bc8) at /var/build/vpp/src/vppinfra/format.c:341 > 341 vec_add (s, cstring, len); > > (gdb) p cstring > $13 = 0x7fffafa01fd0 "iptfs-refill-zpool sa_index %d before %d requested %d > head %d tail %d" > (gdb) p len > $14 = 69 > (gdb) p len > $15 = 69 > > # I believe the vec_add has just done a _vec_resize b/c of this: > (gdb) fr 4 > #4 0x00007ffff5cc5e72 in do_percent (_s=0x7fffb8141a58, fmt=0x7ffff5dc7734 > "%s%c", va=0x7fffb8141bc8) at /var/build/vpp/src/vppinfra/format.c:341 > 341 vec_add (s, cstring, len); > (gdb) p _s > $21 = (u8 **) 0x7fffb8141a58 > (gdb) p *_s > $22 = (u8 *) 0x0 > (gdb) p s > $23 = (u8 *) 0x7fffb90bea90 "" > (gdb) p *((vec_header_t *)s - 1) > $24 = {len = 69, dlmalloc_header_offset = 0, vector_data = > 0x7fffb90bea90 ""} > > So everything should be peachy for clib_memcpy_fast, but something goes > horribly wrong when it gets to clib_mov16. > > #1 0x00007ffff5cbe910 in clib_mov32 (dst=0x7fffb90bea90 "", > src=0x7fffafa01fd0 "iptfs-refill-zpool sa_index %d before %d requested %d > head %d tail %d") at /var/build/vpp/src/vppinfra/memcpy_sse3.h:66 > 66 clib_mov16 ((u8 *) dst + 0 * 16, (const u8 *) src + 0 * 16); > (gdb) p dst > $27 = (u8 *) 0x7fffb90bea90 "" > (gdb) p src > $28 = (const u8 *) 0x7fffafa01fd0 "iptfs-refill-zpool sa_index %d before %d > requested %d head %d tail %d" > (gdb) down > #0 0x00007ffff5cbe8b1 in clib_mov16 (dst=0x6d986a570000 <error: Cannot > access memory at address 0x6d986a570000>, src=0x2d8f0700006c2d08 <error: > Cannot access memory at address 0x2d8f0700006c2d08>) at > /var/build/vpp/src/vppinfra/memcpy_sse3.h:56 > 56 { > > > Thanks, > Chris. > >> >> Thanks... Dave >> >> From: vpp-dev@lists.fd.io <vpp-dev@lists.fd.io> On Behalf Of >> Christian Hopps >> Sent: Friday, June 5, 2020 1:29 PM >> To: vpp-dev <vpp-dev@lists.fd.io> >> Cc: Christian Hopps <cho...@chopps.org> >> Subject: [vpp-dev] Interesting backtrace in 1908 >> >> I'm wondering if maybe this SIGSEGV/backtrace might be related to the other >> recently reported problem with the FIB and barrier code? The workers are at >> the barrier when the SIGSEGV happens, but maybe they aren't when they need >> to be earlier on? >> >> In this case I've compiled w/o CLIB_DEBUG set, but with compiler flags set >> to -O0 instead of -O2 (trying to debug another problem that occurs much >> later). >> >> This is being hit (apparently) when my startup config is adding a >> static arp entry (included below the backtrace) >> >> I've sync'd code to the 2 recent commits past 19.08.02 as well as cherry >> picking the fix from Dave for the counter resize issue in the FIB. >> >> I can try and put together a more in depth bug report (or try an RCA it >> myself), but I'm wondering if something might be easily identified from this >> backtrace w/o doing a bunch more work. >> >> Thanks, >> Chris. >> >> (gdb) info thre >> Id Target Id Frame >> * 1 Thread 83.83 "vpp_main" 0x00007ffff5ccbb11 in clib_memcpy_fast >> (dst=0x1400000000000000, src=0x450000000000e239, n=936751150609465344) at >> /var/build/vpp/src/vppinfra/memcpy_sse3.h:187 >> 2 Thread 83.86 "eal-intr-thread" 0x00007ffff59a3bb7 in epoll_wait >> (epfd=epfd@entry=15, events=events@entry=0x7fff9e8dbe10, >> maxevents=maxevents@entry=1, timeout=timeout@entry=-1) at >> ../sysdeps/unix/sysv/linux/epoll_wait.c:30 >> 3 Thread 83.87 "vpp_wk_0" 0x00007ffff6290c90 in >> vlib_worker_thread_barrier_check () at /var/build/vpp/src/vlib/threads.h:430 >> 4 Thread 83.88 "vpp_wk_1" 0x00007ffff6290c9a in >> vlib_worker_thread_barrier_check () at /var/build/vpp/src/vlib/threads.h:430 >> 5 Thread 83.89 "vpp_wk_2" 0x00007ffff6290c9f in >> vlib_worker_thread_barrier_check () at /var/build/vpp/src/vlib/threads.h:430 >> >> (gdb) bt >> #0 0x00007ffff5ccbb11 in clib_memcpy_fast (dst=0x1400000000000000, >> src=0x450000000000e239, n=936751150609465344) at >> /var/build/vpp/src/vppinfra/memcpy_sse3.h:187 >> #1 0x00007ffff5cd49a8 in lookup (v=0x7fffb7c793c8, key=615, op=SET, >> new_value=0x7fffb7c04880, old_value=0x0) at >> /var/build/vpp/src/vppinfra/hash.c:611 >> #2 0x00007ffff5cd6217 in _hash_set3 (v=0x7fffb7c793c8, key=615, >> value=0x7fffb7c04880, old_value=0x0) at >> /var/build/vpp/src/vppinfra/hash.c:840 >> #3 0x00007ffff62b0b28 in vlib_node_add_next_with_slot >> (vm=0x7ffff656c400 <vlib_global_main>, node_index=522, >> next_node_index=615, slot=3) at /var/build/vpp/src/vlib/node.c:241 >> #4 0x00007ffff62b102b in vlib_node_add_next_with_slot >> (vm=0x7ffff656c400 <vlib_global_main>, node_index=521, >> next_node_index=615, slot=3) at /var/build/vpp/src/vlib/node.c:253 >> #5 0x00007ffff62b102b in vlib_node_add_next_with_slot >> (vm=0x7ffff656c400 <vlib_global_main>, node_index=520, >> next_node_index=615, slot=3) at /var/build/vpp/src/vlib/node.c:253 >> #6 0x00007ffff62b102b in vlib_node_add_next_with_slot >> (vm=0x7ffff656c400 <vlib_global_main>, node_index=519, >> next_node_index=615, slot=3) at /var/build/vpp/src/vlib/node.c:253 >> #7 0x00007ffff62b102b in vlib_node_add_next_with_slot >> (vm=0x7ffff656c400 <vlib_global_main>, node_index=274, >> next_node_index=615, slot=3) at /var/build/vpp/src/vlib/node.c:253 >> #8 0x00007ffff62b102b in vlib_node_add_next_with_slot >> (vm=0x7ffff656c400 <vlib_global_main>, node_index=523, >> next_node_index=615, slot=3) at /var/build/vpp/src/vlib/node.c:253 >> #9 0x00007ffff76e120a in vlib_node_add_next (next_node=615, >> node=523, >> vm=0x7ffff656c400 <vlib_global_main>) at >> /var/build/vpp/src/vlib/node_funcs.h:1109 >> #10 adj_nbr_update_rewrite_internal (adj=0x7fffb821e5c0, >> adj_next_index=IP_LOOKUP_NEXT_REWRITE, this_node=523, next_node=615, >> rewrite=0x0) at /var/build/vpp/src/vnet/adj/adj_nbr.c:488 >> #11 0x00007ffff76e0c34 in adj_nbr_update_rewrite (adj_index=2, >> flags=ADJ_NBR_REWRITE_FLAG_COMPLETE, rewrite=0x7fffb89f4d60 "\002") >> at >> /var/build/vpp/src/vnet/adj/adj_nbr.c:314 >> #12 0x00007ffff6f695b3 in arp_mk_complete (ai=2, e=0x7fffb7f116a8) at >> /var/build/vpp/src/vnet/ethernet/arp.c:385 >> #13 0x00007ffff6f696be in arp_mk_complete_walk (ai=2, >> ctx=0x7fffb7f116a8) at /var/build/vpp/src/vnet/ethernet/arp.c:430 >> #14 0x00007ffff76e15c0 in adj_nbr_walk_nh4 (sw_if_index=1, >> addr=0x7fffb7f116ac, cb=0x7ffff6f69696 <arp_mk_complete_walk>, >> ctx=0x7fffb7f116a8) at /var/build/vpp/src/vnet/adj/adj_nbr.c:624 >> #15 0x00007ffff6f6a436 in arp_update_adjacency (vnm=0x7ffff7b47e80 >> <vnet_main>, sw_if_index=1, ai=2) at >> /var/build/vpp/src/vnet/ethernet/arp.c:540 >> #16 0x00007ffff6b30f27 in ethernet_update_adjacency >> (vnm=0x7ffff7b47e80 <vnet_main>, sw_if_index=1, ai=2) at >> /var/build/vpp/src/vnet/ethernet/interface.c:210 >> #17 0x00007ffff7706ceb in vnet_update_adjacency_for_sw_interface >> (vnm=0x7ffff7b47e80 <vnet_main>, sw_if_index=1, ai=2) at >> /var/build/vpp/src/vnet/adj/rewrite.c:187 >> #18 0x00007ffff76e0b18 in adj_nbr_add_or_lock >> (nh_proto=FIB_PROTOCOL_IP4, link_type=VNET_LINK_IP4, >> nh_addr=0x7fffb821f200, sw_if_index=1) at >> /var/build/vpp/src/vnet/adj/adj_nbr.c:252 >> #19 0x00007ffff76bff09 in fib_path_attached_next_hop_get_adj >> (path=0x7fffb821f1e8, link=VNET_LINK_IP4) at >> /var/build/vpp/src/vnet/fib/fib_path.c:668 >> #20 0x00007ffff76bff50 in fib_path_attached_next_hop_set >> (path=0x7fffb821f1e8) at /var/build/vpp/src/vnet/fib/fib_path.c:682 >> #21 0x00007ffff76c44f3 in fib_path_resolve (path_index=18) at >> /var/build/vpp/src/vnet/fib/fib_path.c:1916 >> #22 0x00007ffff76bbaec in fib_path_list_resolve >> (path_list=0x7fffb821eb58) at >> /var/build/vpp/src/vnet/fib/fib_path_list.c:584 >> #23 0x00007ffff76bc154 in fib_path_list_create >> (flags=FIB_PATH_LIST_FLAG_NONE, rpaths=0x7fffb80677b0) at >> /var/build/vpp/src/vnet/fib/fib_path_list.c:751 >> #24 0x00007ffff76b07f1 in fib_entry_src_adj_path_swap >> (src=0x7fffb7eef7d0, entry=0x7fffb87a7310, >> pl_flags=FIB_PATH_LIST_FLAG_NONE, paths=0x7fffb80677b0) at >> /var/build/vpp/src/vnet/fib/fib_entry_src_adj.c:110 >> #25 0x00007ffff76ad83a in fib_entry_src_action_path_swap >> (fib_entry=0x7fffb87a7310, source=FIB_SOURCE_ADJ, >> flags=FIB_ENTRY_FLAG_ATTACHED, rpaths=0x7fffb80677b0) at >> /var/build/vpp/src/vnet/fib/fib_entry_src.c:1660 >> #26 0x00007ffff76a2b3b in fib_entry_create (fib_index=0, >> prefix=0x7fffb7c06c60, source=FIB_SOURCE_ADJ, >> flags=FIB_ENTRY_FLAG_ATTACHED, paths=0x7fffb80677b0) at >> /var/build/vpp/src/vnet/fib/fib_entry.c:747 >> #27 0x00007ffff768921a in fib_table_entry_path_add2 (fib_index=0, >> prefix=0x7fffb7c06c60, source=FIB_SOURCE_ADJ, >> flags=FIB_ENTRY_FLAG_ATTACHED, rpaths=0x7fffb80677b0) at >> /var/build/vpp/src/vnet/fib/fib_table.c:599 >> #28 0x00007ffff768901a in fib_table_entry_path_add (fib_index=0, >> prefix=0x7fffb7c06c60, source=FIB_SOURCE_ADJ, >> flags=FIB_ENTRY_FLAG_ATTACHED, next_hop_proto=DPO_PROTO_IP4, >> next_hop=0x7fffb7c06c64, next_hop_sw_if_index=1, >> next_hop_fib_index=4294967295, next_hop_weight=1, >> next_hop_labels=0x0, >> path_flags=FIB_ROUTE_PATH_FLAG_NONE) at >> /var/build/vpp/src/vnet/fib/fib_table.c:559 >> #29 0x00007ffff6f6a592 in arp_adj_fib_add (e=0x7fffb7f116a8, >> fib_index=0) at /var/build/vpp/src/vnet/ethernet/arp.c:639 >> #30 0x00007ffff6f6bf99 in vnet_arp_set_ip4_over_ethernet_internal >> (vnm=0x7ffff7b47e80 <vnet_main>, args=0x7fffb7c08040) at >> /var/build/vpp/src/vnet/ethernet/arp.c:766 >> #31 0x00007ffff6f7b672 in set_ip4_over_ethernet_rpc_callback >> (a=0x7fffb7c08040) at /var/build/vpp/src/vnet/ethernet/arp.c:2305 >> #32 0x00007ffff7ba47c9 in vl_api_rpc_call_main_thread_inline >> (force_rpc=0 '\000', data_length=20, data=0x7fffb7c08040 "\001", >> fp=0x7ffff6f7b5be <set_ip4_over_ethernet_rpc_callback>) at >> /var/build/vpp/src/vlibmemory/vlib_api.c:571 >> #33 vl_api_rpc_call_main_thread (fp=0x7ffff6f7b5be >> <set_ip4_over_ethernet_rpc_callback>, data=0x7fffb7c08040 "\001", >> data_length=20) at /var/build/vpp/src/vlibmemory/vlib_api.c:602 >> #34 0x00007ffff6f7bff3 in vnet_arp_set_ip4_over_ethernet >> (vnm=0x7ffff7b47e80 <vnet_main>, sw_if_index=1, a=0x7fffb7c086ce, >> flags=IP_NEIGHBOR_FLAG_STATIC) at >> /var/build/vpp/src/vnet/ethernet/arp.c:2393 >> #35 0x00007ffff6f7d4e0 in ip_arp_add_del_command_fn >> (vm=0x7ffff656c400 <vlib_global_main>, input=0x7fffb7c09ed0, >> cmd=0x7fffb7db2c98) at >> /var/build/vpp/src/vnet/ethernet/arp.c:2604 >> #36 0x00007ffff62385f2 in vlib_cli_dispatch_sub_commands >> (vm=0x7ffff656c400 <vlib_global_main>, cm=0x7ffff656c630 >> <vlib_global_main+560>, input=0x7fffb7c09ed0, >> parent_command_index=812) at /var/build/vpp/src/vlib/cli.c:649 >> #37 0x00007ffff62383a3 in vlib_cli_dispatch_sub_commands >> (vm=0x7ffff656c400 <vlib_global_main>, cm=0x7ffff656c630 >> <vlib_global_main+560>, input=0x7fffb7c09ed0, >> parent_command_index=33) at /var/build/vpp/src/vlib/cli.c:609 >> #38 0x00007ffff62383a3 in vlib_cli_dispatch_sub_commands >> (vm=0x7ffff656c400 <vlib_global_main>, cm=0x7ffff656c630 >> <vlib_global_main+560>, input=0x7fffb7c09ed0, parent_command_index=0) >> at /var/build/vpp/src/vlib/cli.c:609 >> #39 0x00007ffff62391c3 in vlib_cli_input (vm=0x7ffff656c400 >> <vlib_global_main>, input=0x7fffb7c09ed0, function=0x0, >> function_arg=0) at /var/build/vpp/src/vlib/cli.c:750 >> #40 0x00007ffff6344666 in startup_config_process (vm=0x7ffff656c400 >> <vlib_global_main>, rt=0x7fffb7c01000, f=0x0) at >> /var/build/vpp/src/vlib/unix/main.c:367 >> #41 0x00007ffff6297fa0 in vlib_process_bootstrap (_a=140736270031680) >> at /var/build/vpp/src/vlib/main.c:1472 >> #42 0x00007ffff5ce23d8 in clib_calljmp () from >> target:/var/build/vpp/build-root/install-vpp-native/vpp/lib/libvppinf >> r >> a.so.19.08.2 >> #43 0x00007fffb761d7e0 in ?? () >> #44 0x00007ffff62981ec in vlib_process_startup (f=0x7fffb7a7f004, >> p=0x7fffb7c01000, vm=0x7fffb850e978) at >> /var/build/vpp/src/vlib/main.c:1494 >> #45 dispatch_process (vm=0x7fffb8237760, p=0x30, f=0x30, >> last_time_stamp=140736282631376) at >> /var/build/vpp/src/vlib/main.c:1539 >> #46 0x00007ffff629fa4b in vlib_main_or_worker_loop (is_main=1, >> vm=0x7ffff656c400 <vlib_global_main>) at >> /var/build/vpp/src/vlib/main.c:1914 >> #47 vlib_main_loop (vm=0x7ffff656c400 <vlib_global_main>) at >> /var/build/vpp/src/vlib/main.c:1931 >> #48 0x00007ffff62a6cef in vlib_main (vm=0x7ffff656c400 >> <vlib_global_main>, input=0x7fffb761efb0) at >> /var/build/vpp/src/vlib/main.c:2148 >> #49 0x00007ffff6345ad6 in thread0 (arg=140737326269440) at >> /var/build/vpp/src/vlib/unix/main.c:649 >> #50 0x00007ffff5ce23d8 in clib_calljmp () from >> target:/var/build/vpp/build-root/install-vpp-native/vpp/lib/libvppinf >> r >> a.so.19.08.2 >> #51 0x00007fffffffcf60 in ?? () >> #52 0x00007ffff6346b02 in vlib_unix_main (argc=69, >> argv=0x7fffffffe4e8) at /var/build/vpp/src/vlib/unix/main.c:719 >> #53 0x000055555555b801 in main (argc=69, argv=0x7fffffffe4e8) at >> /var/build/vpp/src/vpp/vnet/main.c:280 >> >> >> Config: >> >> set interface rx-placement UnknownEthernet0 worker 0 set int state >> UnknownEthernet0 up set int ip address UnknownEthernet0 >> 11.11.11.11/24 >> >> set interface rx-placement UnknownEthernet1 worker 1 set int state >> UnknownEthernet1 up set int ip address UnknownEthernet1 >> 13.13.13.11/24 >> >> set ip arp UnknownEthernet0 11.11.11.253 02:00:0b:00:00:fd static set >> ip arp UnknownEthernet1 13.13.13.12 02:42:0d:0d:0d:0b static >> >> ...no-more-arp so it's not getting passed... >> > >
-=-=-=-=-=-=-=-=-=-=-=- Links: You receive all messages sent to this group. View/Reply Online (#16675): https://lists.fd.io/g/vpp-dev/message/16675 Mute This Topic: https://lists.fd.io/mt/74698032/21656 Group Owner: vpp-dev+ow...@lists.fd.io Unsubscribe: https://lists.fd.io/g/vpp-dev/unsub [arch...@mail-archive.com] -=-=-=-=-=-=-=-=-=-=-=-