X-Git-Url: https://gerrit.fd.io/r/gitweb?a=blobdiff_plain;f=src%2Fvnet%2Fmpls%2Fmpls_tunnel.c;h=b7bcbfd14f764fdb33f92b0d67dc19a0c97cf7d0;hb=097fa66b986f06281f603767d321ab13ab6c88c3;hp=1d85c3bc78c344acc4260b126dff562f86ac6ea4;hpb=bdc0e6b7204ea0211d4f7881497e4306586fb9ef;p=vpp.git diff --git a/src/vnet/mpls/mpls_tunnel.c b/src/vnet/mpls/mpls_tunnel.c index 1d85c3bc78c..b7bcbfd14f7 100644 --- a/src/vnet/mpls/mpls_tunnel.c +++ b/src/vnet/mpls/mpls_tunnel.c @@ -62,7 +62,7 @@ static u8* mpls_tunnel_build_rewrite_i (void) { /* - * passing the adj code a NULL rewirte means 'i don't have one cos + * passing the adj code a NULL rewrite means 'i don't have one cos * t'other end is unresolved'. That's not the case here. For the mpls * tunnel there are just no bytes of encap to apply in the adj. We'll impose * the label stack once we choose a path. So return a zero length rewrite. @@ -165,9 +165,12 @@ mpls_tunnel_mk_lb (mpls_tunnel_t *mt, lb_proto = fib_forw_chain_type_to_dpo_proto(fct); - fib_path_list_walk(mt->mt_path_list, - mpls_tunnel_collect_forwarding, - &ctx); + if (FIB_NODE_INDEX_INVALID != mt->mt_path_list) + { + fib_path_list_walk(mt->mt_path_list, + mpls_tunnel_collect_forwarding, + &ctx); + } if (!dpo_id_is_valid(dpo_lb)) { @@ -239,9 +242,15 @@ mpls_tunnel_stack (adj_index_t ai) mt = mpls_tunnel_get_from_sw_if_index(sw_if_index); - if (NULL == mt) + if (NULL == mt || FIB_NODE_INDEX_INVALID == mt->mt_path_list) return; + if (FIB_NODE_INDEX_INVALID == mt->mt_path_list) + { + adj_nbr_midchain_unstack(ai); + return; + } + /* * while we're stacking the adj, remove the tunnel from the child list * of the path list. this breaks a circular dependency of walk updates @@ -606,7 +615,7 @@ vnet_mpls_tunnel_create (u8 l2_only, vnm = vnet_get_main(); pool_get(mpls_tunnel_pool, mt); - memset (mt, 0, sizeof (*mt)); + clib_memset (mt, 0, sizeof (*mt)); mti = mt - mpls_tunnel_pool; fib_node_init(&mt->mt_node, FIB_NODE_TYPE_MPLS_TUNNEL); mt->mt_path_list = FIB_NODE_INDEX_INVALID; @@ -645,6 +654,7 @@ void vnet_mpls_tunnel_path_add (u32 sw_if_index, fib_route_path_t *rpaths) { + fib_route_path_t *rpath; mpls_tunnel_t *mt; u32 mti; @@ -686,10 +696,13 @@ vnet_mpls_tunnel_path_add (u32 sw_if_index, */ fib_path_ext_list_resolve(&mt->mt_path_exts, mt->mt_path_list); } - fib_path_ext_list_insert(&mt->mt_path_exts, - mt->mt_path_list, - FIB_PATH_EXT_MPLS, - rpaths); + vec_foreach(rpath, rpaths) + { + fib_path_ext_list_insert(&mt->mt_path_exts, + mt->mt_path_list, + FIB_PATH_EXT_MPLS, + rpath); + } mpls_tunnel_restack(mt); } @@ -721,6 +734,7 @@ vnet_mpls_tunnel_path_remove (u32 sw_if_index, old_pl_index = mt->mt_path_list; + fib_path_list_lock(old_pl_index); mt->mt_path_list = fib_path_list_copy_and_path_remove(old_pl_index, FIB_PATH_LIST_FLAG_SHARED, @@ -732,6 +746,7 @@ vnet_mpls_tunnel_path_remove (u32 sw_if_index, if (FIB_NODE_INDEX_INVALID == mt->mt_path_list) { /* no paths left */ + fib_path_list_unlock(old_pl_index); return (0); } else @@ -755,11 +770,24 @@ vnet_mpls_tunnel_path_remove (u32 sw_if_index, mt->mt_path_list); mpls_tunnel_restack(mt); + fib_path_list_unlock(old_pl_index); } return (fib_path_list_get_n_paths(mt->mt_path_list)); } +int +vnet_mpls_tunnel_get_index (u32 sw_if_index) +{ + mpls_tunnel_t *mt; + + mt = mpls_tunnel_get_from_sw_if_index(sw_if_index); + + if (NULL == mt) + return (~0); + + return (mt - mpls_tunnel_pool); +} static clib_error_t * vnet_create_mpls_tunnel_command_fn (vlib_main_t * vm, @@ -773,7 +801,7 @@ vnet_create_mpls_tunnel_command_fn (vlib_main_t * vm, u32 sw_if_index = ~0, payload_proto; clib_error_t *error = NULL; - memset(&rpath, 0, sizeof(rpath)); + clib_memset(&rpath, 0, sizeof(rpath)); payload_proto = DPO_PROTO_MPLS; /* Get a line of input. */ @@ -863,7 +891,7 @@ format_mpls_tunnel (u8 * s, va_list * args) mpls_tunnel_t *mt = va_arg (*args, mpls_tunnel_t *); mpls_tunnel_attribute_t attr; - s = format(s, "mpls_tunnel%d: sw_if_index:%d hw_if_index:%d", + s = format(s, "mpls-tunnel%d: sw_if_index:%d hw_if_index:%d", mt - mpls_tunnel_pool, mt->mt_sw_if_index, mt->mt_hw_if_index);