X-Git-Url: https://gerrit.fd.io/r/gitweb?a=blobdiff_plain;f=src%2Fvnet%2Ffib%2Ffib_path.c;h=2a4e6ab551f0095ab1f34ea9a734fa92929e7476;hb=e2fe09742;hp=f81f41706c520461d82a3cea410e996bc93564b6;hpb=0f26c5a0138ac86d7ebd197c31a09d8d624c35fe;p=vpp.git diff --git a/src/vnet/fib/fib_path.c b/src/vnet/fib/fib_path.c index f81f41706c5..2a4e6ab551f 100644 --- a/src/vnet/fib/fib_path.c +++ b/src/vnet/fib/fib_path.c @@ -21,8 +21,12 @@ #include #include #include -#include +#include #include +#include +#include +#include +#include #include #include @@ -35,6 +39,12 @@ #include #include #include +#include +#include +#include +#include +#include +#include /** * Enurmeration of path types @@ -72,21 +82,32 @@ typedef enum fib_path_type_t_ { * interface receive. */ FIB_PATH_TYPE_INTF_RX, + /** + * Path resolves via a UDP encap object. + */ + FIB_PATH_TYPE_UDP_ENCAP, /** * receive. it's for-us. */ FIB_PATH_TYPE_RECEIVE, /** - * Marker. Add new types before this one, then update it. + * bier-imp. it's via a BIER imposition. + */ + FIB_PATH_TYPE_BIER_IMP, + /** + * bier-fmask. it's via a BIER ECMP-table. + */ + FIB_PATH_TYPE_BIER_TABLE, + /** + * bier-fmask. it's via a BIER f-mask. */ - FIB_PATH_TYPE_LAST = FIB_PATH_TYPE_RECEIVE, + FIB_PATH_TYPE_BIER_FMASK, + /** + * via a DVR. + */ + FIB_PATH_TYPE_DVR, } __attribute__ ((packed)) fib_path_type_t; -/** - * The maximum number of path_types - */ -#define FIB_PATH_TYPE_MAX (FIB_PATH_TYPE_LAST + 1) - #define FIB_PATH_TYPES { \ [FIB_PATH_TYPE_ATTACHED_NEXT_HOP] = "attached-nexthop", \ [FIB_PATH_TYPE_ATTACHED] = "attached", \ @@ -95,12 +116,14 @@ typedef enum fib_path_type_t_ { [FIB_PATH_TYPE_EXCLUSIVE] = "exclusive", \ [FIB_PATH_TYPE_DEAG] = "deag", \ [FIB_PATH_TYPE_INTF_RX] = "intf-rx", \ + [FIB_PATH_TYPE_UDP_ENCAP] = "udp-encap", \ [FIB_PATH_TYPE_RECEIVE] = "receive", \ + [FIB_PATH_TYPE_BIER_IMP] = "bier-imp", \ + [FIB_PATH_TYPE_BIER_TABLE] = "bier-table", \ + [FIB_PATH_TYPE_BIER_FMASK] = "bier-fmask", \ + [FIB_PATH_TYPE_DVR] = "dvr", \ } -#define FOR_EACH_FIB_PATH_TYPE(_item) \ - for (_item = FIB_PATH_TYPE_FIRST; _item <= FIB_PATH_TYPE_LAST; _item++) - /** * Enurmeration of path operational (i.e. derived) attributes */ @@ -193,12 +216,19 @@ typedef struct fib_path_t_ { * next-hop's address. We can't derive this from the address itself * since the address can be all zeros */ - fib_protocol_t fp_nh_proto; + dpo_proto_t fp_nh_proto; /** - * UCMP [unnormalised] weigt + * UCMP [unnormalised] weigth */ - u32 fp_weight; + u8 fp_weight; + + /** + * A path preference. 0 is the best. + * Only paths of the best preference, that are 'up', are considered + * for forwarding. + */ + u8 fp_preference; /** * per-type union of the data required to resolve the path @@ -215,6 +245,10 @@ typedef struct fib_path_t_ { u32 fp_interface; } attached_next_hop; struct { + /** + * The Connected local address + */ + fib_prefix_t fp_connected; /** * The interface */ @@ -238,30 +272,37 @@ typedef struct fib_path_t_ { mpls_eos_bit_t fp_eos; }; } fp_nh; - /** - * The FIB table index in which to find the next-hop. - * This needs to be fixed. We should lookup the adjacencies in - * a separate table of adjacencies, rather than from the FIB. - * Two reasons I can think of: - * - consider: - * int ip addr Gig0 10.0.0.1/24 - * ip route 10.0.0.2/32 via Gig1 192.168.1.2 - * ip route 1.1.1.1/32 via Gig0 10.0.0.2 - * this is perfectly valid. - * Packets addressed to 10.0.0.2 should be sent via Gig1. - * Packets address to 1.1.1.1 should be sent via Gig0. - * when we perform the adj resolution from the FIB for the path - * "via Gig0 10.0.0.2" the lookup will result in the route via Gig1 - * and so we will pick up the adj via Gig1 - which was not what the - * operator wanted. - * - we can only return link-type IPv4 and so not the link-type MPLS. - * more on this in a later commit. - * - * The table ID should only belong to a recursive path and indicate - * which FIB should be used to resolve the next-hop. - */ - fib_node_index_t fp_tbl_id; + union { + /** + * The FIB table index in which to find the next-hop. + */ + fib_node_index_t fp_tbl_id; + /** + * The BIER FIB the fmask is in + */ + index_t fp_bier_fib; + }; } recursive; + struct { + /** + * BIER FMask ID + */ + index_t fp_bier_fmask; + } bier_fmask; + struct { + /** + * The BIER table's ID + */ + bier_table_id_t fp_bier_tbl; + } bier_table; + struct { + /** + * The BIER imposition object + * this is part of the path's key, since the index_t + * of an imposition object is the object's key. + */ + index_t fp_bier_imp; + } bier_imp; struct { /** * The FIB index in which to perfom the next lookup @@ -296,6 +337,24 @@ typedef struct fib_path_t_ { */ u32 fp_interface; } intf_rx; + struct { + /** + * The UDP Encap object this path resolves through + */ + u32 fp_udp_encap_id; + } udp_encap; + struct { + /** + * The UDP Encap object this path resolves through + */ + u32 fp_classify_table_id; + } classify; + struct { + /** + * The interface + */ + u32 fp_interface; + } dvr; }; STRUCT_MARK(path_hash_end); @@ -310,11 +369,21 @@ typedef struct fib_path_t_ { */ fib_path_oper_flags_t fp_oper_flags; - /** - * the resolving via fib. not part of the union, since it it not part - * of the path's hash. - */ - fib_node_index_t fp_via_fib; + union { + /** + * the resolving via fib. not part of the union, since it it not part + * of the path's hash. + */ + fib_node_index_t fp_via_fib; + /** + * the resolving bier-table + */ + index_t fp_via_bier_tbl; + /** + * the resolving bier-fmask + */ + index_t fp_via_bier_fmask; + }; /** * The Data-path objects through which this path resolves for IP. @@ -339,22 +408,22 @@ static const char *fib_path_cfg_attribute_names[] = FIB_PATH_CFG_ATTRIBUTES; */ static fib_path_t *fib_path_pool; +/** + * the logger + */ +vlib_log_class_t fib_path_logger; + /* * Debug macro */ -#ifdef FIB_DEBUG -#define FIB_PATH_DBG(_p, _fmt, _args...) \ -{ \ - u8 *_tmp = NULL; \ - _tmp = fib_path_format(fib_path_get_index(_p), _tmp); \ - clib_warning("path:[%d:%s]:" _fmt, \ - fib_path_get_index(_p), _tmp, \ - ##_args); \ - vec_free(_tmp); \ +#define FIB_PATH_DBG(_p, _fmt, _args...) \ +{ \ + vlib_log_debug (fib_path_logger, \ + "[%U]: " _fmt, \ + format_fib_path, fib_path_get_index(_p), 0, \ + FIB_PATH_FORMAT_FLAGS_ONE_LINE, \ + ##_args); \ } -#else -#define FIB_PATH_DBG(_p, _fmt, _args...) -#endif static fib_path_t * fib_path_get (fib_node_index_t index) @@ -377,24 +446,39 @@ fib_path_get_node (fib_node_index_t index) static fib_path_t* fib_path_from_fib_node (fib_node_t *node) { -#if CLIB_DEBUG > 0 ASSERT(FIB_NODE_TYPE_PATH == node->fn_type); -#endif return ((fib_path_t*)node); } u8 * format_fib_path (u8 * s, va_list * args) { - fib_path_t *path = va_arg (*args, fib_path_t *); + fib_node_index_t path_index = va_arg (*args, fib_node_index_t); + u32 indent = va_arg (*args, u32); + fib_format_path_flags_t flags = va_arg (*args, fib_format_path_flags_t); vnet_main_t * vnm = vnet_get_main(); fib_path_oper_attribute_t oattr; fib_path_cfg_attribute_t cattr; + fib_path_t *path; + const char *eol; + + if (flags & FIB_PATH_FORMAT_FLAGS_ONE_LINE) + { + eol = ""; + } + else + { + eol = "\n"; + } + + path = fib_path_get(path_index); - s = format (s, " index:%d ", fib_path_get_index(path)); + s = format (s, "%Upath:[%d] ", format_white_space, indent, + fib_path_get_index(path)); s = format (s, "pl-index:%d ", path->fp_pl_index); - s = format (s, "%U ", format_fib_protocol, path->fp_nh_proto); + s = format (s, "%U ", format_dpo_proto, path->fp_nh_proto); s = format (s, "weight=%d ", path->fp_weight); + s = format (s, "pref=%d ", path->fp_preference); s = format (s, "%s: ", fib_path_type_names[path->fp_type]); if (FIB_PATH_OPER_FLAG_NONE != path->fp_oper_flags) { s = format(s, " oper-flags:"); @@ -412,7 +496,8 @@ format_fib_path (u8 * s, va_list * args) } } } - s = format(s, "\n "); + if (!(flags & FIB_PATH_FORMAT_FLAGS_ONE_LINE)) + s = format(s, "\n%U", format_white_space, indent+2); switch (path->fp_type) { @@ -440,19 +525,20 @@ format_fib_path (u8 * s, va_list * args) } if (!dpo_id_is_valid(&path->fp_dpo)) { - s = format(s, "\n unresolved"); + s = format(s, "%s%Uunresolved", eol, format_white_space, indent+2); } else { - s = format(s, "\n %U", - format_dpo_id, + s = format(s, "%s%U%U", eol, + format_white_space, indent, + format_dpo_id, &path->fp_dpo, 13); } break; case FIB_PATH_TYPE_ATTACHED: if (path->fp_oper_flags & FIB_PATH_OPER_FLAG_DROP) { - s = format (s, " if_index:%d", path->attached_next_hop.fp_interface); + s = format (s, "if_index:%d", path->attached_next_hop.fp_interface); } else { @@ -465,7 +551,7 @@ format_fib_path (u8 * s, va_list * args) } break; case FIB_PATH_TYPE_RECURSIVE: - if (FIB_PROTOCOL_MPLS == path->fp_nh_proto) + if (DPO_PROTO_MPLS == path->fp_nh_proto) { s = format (s, "via %U %U", format_mpls_unicast_label, @@ -489,55 +575,54 @@ format_fib_path (u8 * s, va_list * args) path->fp_dpo.dpoi_index); break; + case FIB_PATH_TYPE_UDP_ENCAP: + s = format (s, "UDP-encap ID:%d", path->udp_encap.fp_udp_encap_id); + break; + case FIB_PATH_TYPE_BIER_TABLE: + s = format (s, "via bier-table:[%U}", + format_bier_table_id, + &path->bier_table.fp_bier_tbl); + s = format (s, " via-dpo:[%U:%d]", + format_dpo_type, path->fp_dpo.dpoi_type, + path->fp_dpo.dpoi_index); + break; + case FIB_PATH_TYPE_BIER_FMASK: + s = format (s, "via-fmask:%d", path->bier_fmask.fp_bier_fmask); + s = format (s, " via-dpo:[%U:%d]", + format_dpo_type, path->fp_dpo.dpoi_type, + path->fp_dpo.dpoi_index); + break; + case FIB_PATH_TYPE_BIER_IMP: + s = format (s, "via %U", format_bier_imp, + path->bier_imp.fp_bier_imp, 0, BIER_SHOW_BRIEF); + break; + case FIB_PATH_TYPE_DVR: + s = format (s, " %U", + format_vnet_sw_interface_name, + vnm, + vnet_get_sw_interface( + vnm, + path->dvr.fp_interface)); + break; + case FIB_PATH_TYPE_DEAG: + s = format (s, " %sfib-index:%d", + (path->fp_cfg_flags & FIB_PATH_CFG_FLAG_RPF_ID ? "m" : ""), + path->deag.fp_tbl_id); + break; case FIB_PATH_TYPE_RECEIVE: case FIB_PATH_TYPE_INTF_RX: case FIB_PATH_TYPE_SPECIAL: - case FIB_PATH_TYPE_DEAG: case FIB_PATH_TYPE_EXCLUSIVE: if (dpo_id_is_valid(&path->fp_dpo)) { s = format(s, "%U", format_dpo_id, - &path->fp_dpo, 2); + &path->fp_dpo, indent+2); } break; } return (s); } -u8 * -fib_path_format (fib_node_index_t pi, u8 *s) -{ - fib_path_t *path; - - path = fib_path_get(pi); - ASSERT(NULL != path); - - return (format (s, "%U", format_fib_path, path)); -} - -u8 * -fib_path_adj_format (fib_node_index_t pi, - u32 indent, - u8 *s) -{ - fib_path_t *path; - - path = fib_path_get(pi); - ASSERT(NULL != path); - - if (!dpo_id_is_valid(&path->fp_dpo)) - { - s = format(s, " unresolved"); - } - else - { - s = format(s, "%U", format_dpo_id, - &path->fp_dpo, 2); - } - - return (s); -} - /* * fib_path_last_lock_gone * @@ -550,10 +635,18 @@ fib_path_last_lock_gone (fib_node_t *node) ASSERT(0); } -static const adj_index_t +static fib_path_t* fib_path_attached_next_hop_get_adj (fib_path_t *path, - vnet_link_t link) + vnet_link_t link, + dpo_id_t *dpo) { + fib_node_index_t fib_path_index; + fib_protocol_t nh_proto; + adj_index_t ai; + + fib_path_index = fib_path_get_index(path); + nh_proto = dpo_proto_to_fib(path->fp_nh_proto); + if (vnet_sw_interface_is_p2p(vnet_get_main(), path->attached_next_hop.fp_interface)) { @@ -563,33 +656,34 @@ fib_path_attached_next_hop_get_adj (fib_path_t *path, * the subnet address (the attached route) links to the * auto-adj (see below), we want that adj here too. */ - return (adj_nbr_add_or_lock(path->fp_nh_proto, - link, - &zero_addr, - path->attached_next_hop.fp_interface)); + ai = adj_nbr_add_or_lock(nh_proto, link, &zero_addr, + path->attached_next_hop.fp_interface); } else { - return (adj_nbr_add_or_lock(path->fp_nh_proto, - link, - &path->attached_next_hop.fp_nh, - path->attached_next_hop.fp_interface)); + ai = adj_nbr_add_or_lock(nh_proto, link, + &path->attached_next_hop.fp_nh, + path->attached_next_hop.fp_interface); } + + dpo_set(dpo, DPO_ADJACENCY, vnet_link_to_dpo_proto(link), ai); + adj_unlock(ai); + + return (fib_path_get(fib_path_index)); } static void fib_path_attached_next_hop_set (fib_path_t *path) { /* - * resolve directly via the adjacnecy discribed by the + * resolve directly via the adjacency discribed by the * interface and next-hop */ - dpo_set(&path->fp_dpo, - DPO_ADJACENCY, - fib_proto_to_dpo(path->fp_nh_proto), - fib_path_attached_next_hop_get_adj( - path, - fib_proto_to_link(path->fp_nh_proto))); + path = fib_path_attached_next_hop_get_adj(path, + dpo_proto_to_link(path->fp_nh_proto), + &path->fp_dpo); + + ASSERT(dpo_is_adj(&path->fp_dpo)); /* * become a child of the adjacency so we receive updates @@ -599,14 +693,55 @@ fib_path_attached_next_hop_set (fib_path_t *path) FIB_NODE_TYPE_PATH, fib_path_get_index(path)); - if (!vnet_sw_interface_is_admin_up(vnet_get_main(), - path->attached_next_hop.fp_interface) || + if (!vnet_sw_interface_is_up(vnet_get_main(), + path->attached_next_hop.fp_interface) || !adj_is_up(path->fp_dpo.dpoi_index)) { path->fp_oper_flags &= ~FIB_PATH_OPER_FLAG_RESOLVED; } } +static void +fib_path_attached_get_adj (fib_path_t *path, + vnet_link_t link, + dpo_id_t *dpo) +{ + fib_protocol_t nh_proto; + + nh_proto = dpo_proto_to_fib(path->fp_nh_proto); + + if (vnet_sw_interface_is_p2p(vnet_get_main(), + path->attached.fp_interface)) + { + /* + * point-2-point interfaces do not require a glean, since + * there is nothing to ARP. Install a rewrite/nbr adj instead + */ + adj_index_t ai; + + ai = adj_nbr_add_or_lock(nh_proto, link, &zero_addr, + path->attached.fp_interface); + + dpo_set(dpo, DPO_ADJACENCY, vnet_link_to_dpo_proto(link), ai); + adj_unlock(ai); + } + else if (vnet_sw_interface_is_nbma(vnet_get_main(), + path->attached.fp_interface)) + { + dpo_copy(dpo, drop_dpo_get(path->fp_nh_proto)); + } + else + { + adj_index_t ai; + + ai = adj_glean_add_or_lock(nh_proto, link, + path->attached.fp_interface, + &path->attached.fp_connected); + dpo_set(dpo, DPO_ADJACENCY_GLEAN, vnet_link_to_dpo_proto(link), ai); + adj_unlock(ai); + } +} + /* * create of update the paths recursive adj */ @@ -637,7 +772,7 @@ fib_path_recursive_adj_update (fib_path_t *path, if (path->fp_oper_flags & FIB_PATH_OPER_FLAG_RECURSIVE_LOOP) { path->fp_oper_flags &= ~FIB_PATH_OPER_FLAG_RESOLVED; - dpo_copy(&via_dpo, drop_dpo_get(fib_proto_to_dpo(path->fp_nh_proto))); + dpo_copy(&via_dpo, drop_dpo_get(path->fp_nh_proto)); } else if (path->fp_cfg_flags & FIB_PATH_CFG_FLAG_RESOLVE_HOST) { @@ -655,7 +790,7 @@ fib_path_recursive_adj_update (fib_path_t *path, if (fib_entry_get_best_source(path->fp_via_fib) >= FIB_SOURCE_RR) { path->fp_oper_flags &= ~FIB_PATH_OPER_FLAG_RESOLVED; - dpo_copy(&via_dpo, drop_dpo_get(fib_proto_to_dpo(path->fp_nh_proto))); + dpo_copy(&via_dpo, drop_dpo_get(path->fp_nh_proto)); /* * PIC edge trigger. let the load-balance maps know @@ -672,7 +807,7 @@ fib_path_recursive_adj_update (fib_path_t *path, if (!(FIB_ENTRY_FLAG_ATTACHED & fib_entry_get_flags(path->fp_via_fib))) { path->fp_oper_flags &= ~FIB_PATH_OPER_FLAG_RESOLVED; - dpo_copy(&via_dpo, drop_dpo_get(fib_proto_to_dpo(path->fp_nh_proto))); + dpo_copy(&via_dpo, drop_dpo_get(path->fp_nh_proto)); /* * PIC edge trigger. let the load-balance maps know @@ -686,7 +821,7 @@ fib_path_recursive_adj_update (fib_path_t *path, if (!fib_entry_is_resolved(path->fp_via_fib)) { path->fp_oper_flags &= ~FIB_PATH_OPER_FLAG_RESOLVED; - dpo_copy(&via_dpo, drop_dpo_get(fib_proto_to_dpo(path->fp_nh_proto))); + dpo_copy(&via_dpo, drop_dpo_get(path->fp_nh_proto)); /* * PIC edge trigger. let the load-balance maps know @@ -694,18 +829,46 @@ fib_path_recursive_adj_update (fib_path_t *path, load_balance_map_path_state_change(fib_path_get_index(path)); } + /* + * If this path is contributing a drop, then it's not resolved + */ + if (dpo_is_drop(&via_dpo) || load_balance_is_drop(&via_dpo)) + { + path->fp_oper_flags &= ~FIB_PATH_OPER_FLAG_RESOLVED; + } + /* * update the path's contributed DPO */ dpo_copy(dpo, &via_dpo); - FIB_PATH_DBG(path, "recursive update: %U", - fib_get_lookup_main(path->fp_nh_proto), - &path->fp_dpo, 2); + FIB_PATH_DBG(path, "recursive update:"); dpo_reset(&via_dpo); } +/* + * re-evaulate the forwarding state for a via fmask path + */ +static void +fib_path_bier_fmask_update (fib_path_t *path, + dpo_id_t *dpo) +{ + bier_fmask_contribute_forwarding(path->bier_fmask.fp_bier_fmask, dpo); + + /* + * if we are stakcing on the drop, then the path is not resolved + */ + if (dpo_is_drop(dpo)) + { + path->fp_oper_flags &= ~FIB_PATH_OPER_FLAG_RESOLVED; + } + else + { + path->fp_oper_flags |= FIB_PATH_OPER_FLAG_RESOLVED; + } +} + /* * fib_path_is_permanent_drop * @@ -740,22 +903,35 @@ fib_path_unresolve (fib_path_t *path) case FIB_PATH_TYPE_RECURSIVE: if (FIB_NODE_INDEX_INVALID != path->fp_via_fib) { - fib_prefix_t pfx; - - fib_entry_get_prefix(path->fp_via_fib, &pfx); fib_entry_child_remove(path->fp_via_fib, path->fp_sibling); - fib_table_entry_special_remove(path->recursive.fp_tbl_id, - &pfx, + fib_table_entry_special_remove(path->recursive.fp_tbl_id, + fib_entry_get_prefix(path->fp_via_fib), FIB_SOURCE_RR); + fib_table_unlock(path->recursive.fp_tbl_id, + dpo_proto_to_fib(path->fp_nh_proto), + FIB_SOURCE_RR); path->fp_via_fib = FIB_NODE_INDEX_INVALID; } break; + case FIB_PATH_TYPE_BIER_FMASK: + bier_fmask_child_remove(path->fp_via_bier_fmask, + path->fp_sibling); + break; + case FIB_PATH_TYPE_BIER_IMP: + bier_imp_unlock(path->fp_dpo.dpoi_index); + break; + case FIB_PATH_TYPE_BIER_TABLE: + bier_table_ecmp_unlock(path->fp_via_bier_tbl); + break; case FIB_PATH_TYPE_ATTACHED_NEXT_HOP: case FIB_PATH_TYPE_ATTACHED: - adj_child_remove(path->fp_dpo.dpoi_index, - path->fp_sibling); - adj_unlock(path->fp_dpo.dpoi_index); + if (dpo_is_adj(&path->fp_dpo)) + adj_child_remove(path->fp_dpo.dpoi_index, + path->fp_sibling); + break; + case FIB_PATH_TYPE_UDP_ENCAP: + udp_encap_unlock(path->fp_dpo.dpoi_index); break; case FIB_PATH_TYPE_EXCLUSIVE: dpo_reset(&path->exclusive.fp_ex_dpo); @@ -764,6 +940,7 @@ fib_path_unresolve (fib_path_t *path) case FIB_PATH_TYPE_RECEIVE: case FIB_PATH_TYPE_INTF_RX: case FIB_PATH_TYPE_DEAG: + case FIB_PATH_TYPE_DVR: /* * these hold only the path's DPO, which is reset below. */ @@ -783,13 +960,8 @@ fib_path_unresolve (fib_path_t *path) static fib_forward_chain_type_t fib_path_to_chain_type (const fib_path_t *path) { - switch (path->fp_nh_proto) + if (DPO_PROTO_MPLS == path->fp_nh_proto) { - case FIB_PROTOCOL_IP4: - return (FIB_FORW_CHAIN_TYPE_UNICAST_IP4); - case FIB_PROTOCOL_IP6: - return (FIB_FORW_CHAIN_TYPE_UNICAST_IP6); - case FIB_PROTOCOL_MPLS: if (FIB_PATH_TYPE_RECURSIVE == path->fp_type && MPLS_EOS == path->recursive.fp_nh.fp_eos) { @@ -797,10 +969,13 @@ fib_path_to_chain_type (const fib_path_t *path) } else { - return (FIB_FORW_CHAIN_TYPE_MPLS_EOS); + return (FIB_FORW_CHAIN_TYPE_MPLS_NON_EOS); } } - return (FIB_FORW_CHAIN_TYPE_UNICAST_IP4); + else + { + return (fib_forw_chain_type_from_dpo_proto(path->fp_nh_proto)); + } } /* @@ -816,6 +991,9 @@ fib_path_back_walk_notify (fib_node_t *node, path = fib_path_from_fib_node(node); + FIB_PATH_DBG(path, "bw:%U", + format_fib_node_bw_reason, ctx->fnbw_reason); + switch (path->fp_type) { case FIB_PATH_TYPE_RECURSIVE: @@ -847,6 +1025,30 @@ fib_path_back_walk_notify (fib_node_t *node, return (FIB_NODE_BACK_WALK_CONTINUE); } break; + case FIB_PATH_TYPE_BIER_FMASK: + if (FIB_NODE_BW_REASON_FLAG_EVALUATE & ctx->fnbw_reason) + { + /* + * update to use the BIER fmask's new forwading + */ + fib_path_bier_fmask_update(path, &path->fp_dpo); + } + if ((FIB_NODE_BW_REASON_FLAG_ADJ_UPDATE & ctx->fnbw_reason) || + (FIB_NODE_BW_REASON_FLAG_ADJ_DOWN & ctx->fnbw_reason)) + { + /* + * ADJ updates (complete<->incomplete) do not need to propagate to + * recursive entries. + * The only reason its needed as far back as here, is that the adj + * and the incomplete adj are a different DPO type, so the LBs need + * to re-stack. + * If this walk was quashed in the fib_entry, then any non-fib_path + * children (like tunnels that collapse out the LB when they stack) + * would not see the update. + */ + return (FIB_NODE_BACK_WALK_CONTINUE); + } + break; case FIB_PATH_TYPE_ATTACHED_NEXT_HOP: /* FIXME comment @@ -898,27 +1100,22 @@ FIXME comment * restack the DPO to pick up the correct DPO sub-type */ uword if_is_up; - adj_index_t ai; - if_is_up = vnet_sw_interface_is_admin_up( + if_is_up = vnet_sw_interface_is_up( vnet_get_main(), path->attached_next_hop.fp_interface); - ai = fib_path_attached_next_hop_get_adj( - path, - fib_proto_to_link(path->fp_nh_proto)); + path = fib_path_attached_next_hop_get_adj( + path, + dpo_proto_to_link(path->fp_nh_proto), + &path->fp_dpo); path->fp_oper_flags &= ~FIB_PATH_OPER_FLAG_RESOLVED; - if (if_is_up && adj_is_up(ai)) + if (if_is_up && adj_is_up(path->fp_dpo.dpoi_index)) { path->fp_oper_flags |= FIB_PATH_OPER_FLAG_RESOLVED; } - dpo_set(&path->fp_dpo, DPO_ADJACENCY, - fib_proto_to_dpo(path->fp_nh_proto), - ai); - adj_unlock(ai); - if (!if_is_up) { /* @@ -947,6 +1144,7 @@ FIXME comment } break; case FIB_PATH_TYPE_ATTACHED: + case FIB_PATH_TYPE_DVR: /* * FIXME; this could schedule a lower priority walk, since attached * routes are not usually in ECMP configurations so the backwalk to @@ -966,6 +1164,33 @@ FIXME comment path->fp_oper_flags |= FIB_PATH_OPER_FLAG_DROP; } break; + case FIB_PATH_TYPE_UDP_ENCAP: + { + dpo_id_t via_dpo = DPO_INVALID; + + /* + * hope for the best - clear if restrictions apply. + */ + path->fp_oper_flags |= FIB_PATH_OPER_FLAG_RESOLVED; + + udp_encap_contribute_forwarding(path->udp_encap.fp_udp_encap_id, + path->fp_nh_proto, + &via_dpo); + /* + * If this path is contributing a drop, then it's not resolved + */ + if (dpo_is_drop(&via_dpo) || load_balance_is_drop(&via_dpo)) + { + path->fp_oper_flags &= ~FIB_PATH_OPER_FLAG_RESOLVED; + } + + /* + * update the path's contributed DPO + */ + dpo_copy(&path->fp_dpo, &via_dpo); + dpo_reset(&via_dpo); + break; + } case FIB_PATH_TYPE_INTF_RX: ASSERT(0); case FIB_PATH_TYPE_DEAG: @@ -975,6 +1200,8 @@ FIXME comment case FIB_PATH_TYPE_SPECIAL: case FIB_PATH_TYPE_RECEIVE: case FIB_PATH_TYPE_EXCLUSIVE: + case FIB_PATH_TYPE_BIER_TABLE: + case FIB_PATH_TYPE_BIER_IMP: /* * these path types have no parents. so to be * walked from one is unexpected. @@ -1015,6 +1242,8 @@ fib_path_route_flags_to_cfg_flags (const fib_route_path_t *rpath) { fib_path_cfg_flags_t cfg_flags = FIB_PATH_CFG_FLAG_NONE; + if (rpath->frp_flags & FIB_ROUTE_PATH_POP_PW_CW) + cfg_flags |= FIB_PATH_CFG_FLAG_POP_PW_CW; if (rpath->frp_flags & FIB_ROUTE_PATH_RESOLVE_VIA_HOST) cfg_flags |= FIB_PATH_CFG_FLAG_RESOLVE_HOST; if (rpath->frp_flags & FIB_ROUTE_PATH_RESOLVE_VIA_ATTACHED) @@ -1031,6 +1260,14 @@ fib_path_route_flags_to_cfg_flags (const fib_route_path_t *rpath) cfg_flags |= FIB_PATH_CFG_FLAG_EXCLUSIVE; if (rpath->frp_flags & FIB_ROUTE_PATH_DROP) cfg_flags |= FIB_PATH_CFG_FLAG_DROP; + if (rpath->frp_flags & FIB_ROUTE_PATH_SOURCE_LOOKUP) + cfg_flags |= FIB_PATH_CFG_FLAG_DEAG_SRC; + if (rpath->frp_flags & FIB_ROUTE_PATH_ICMP_UNREACH) + cfg_flags |= FIB_PATH_CFG_FLAG_ICMP_UNREACH; + if (rpath->frp_flags & FIB_ROUTE_PATH_ICMP_PROHIBIT) + cfg_flags |= FIB_PATH_CFG_FLAG_ICMP_PROHIBIT; + if (rpath->frp_flags & FIB_ROUTE_PATH_GLEAN) + cfg_flags |= FIB_PATH_CFG_FLAG_GLEAN; return (cfg_flags); } @@ -1048,7 +1285,7 @@ fib_path_create (fib_node_index_t pl_index, fib_path_t *path; pool_get(fib_path_pool, path); - memset(path, 0, sizeof(*path)); + clib_memset(path, 0, sizeof(*path)); fib_node_init(&path->fp_node, FIB_NODE_TYPE_PATH); @@ -1066,6 +1303,7 @@ fib_path_create (fib_node_index_t pl_index, */ path->fp_weight = 1; } + path->fp_preference = rpath->frp_preference; path->fp_cfg_flags = fib_path_route_flags_to_cfg_flags(rpath); /* @@ -1077,6 +1315,11 @@ fib_path_create (fib_node_index_t pl_index, path->receive.fp_interface = rpath->frp_sw_if_index; path->receive.fp_addr = rpath->frp_addr; } + else if (rpath->frp_flags & FIB_ROUTE_PATH_UDP_ENCAP) + { + path->fp_type = FIB_PATH_TYPE_UDP_ENCAP; + path->udp_encap.fp_udp_encap_id = rpath->frp_udp_encap_id; + } else if (path->fp_cfg_flags & FIB_PATH_CFG_FLAG_INTF_RX) { path->fp_type = FIB_PATH_TYPE_INTF_RX; @@ -1088,6 +1331,52 @@ fib_path_create (fib_node_index_t pl_index, path->deag.fp_tbl_id = rpath->frp_fib_index; path->deag.fp_rpf_id = rpath->frp_rpf_id; } + else if (rpath->frp_flags & FIB_ROUTE_PATH_BIER_FMASK) + { + path->fp_type = FIB_PATH_TYPE_BIER_FMASK; + path->bier_fmask.fp_bier_fmask = rpath->frp_bier_fmask; + } + else if (rpath->frp_flags & FIB_ROUTE_PATH_BIER_IMP) + { + path->fp_type = FIB_PATH_TYPE_BIER_IMP; + path->bier_imp.fp_bier_imp = rpath->frp_bier_imp; + } + else if (rpath->frp_flags & FIB_ROUTE_PATH_BIER_TABLE) + { + path->fp_type = FIB_PATH_TYPE_BIER_TABLE; + path->bier_table.fp_bier_tbl = rpath->frp_bier_tbl; + } + else if (rpath->frp_flags & FIB_ROUTE_PATH_DEAG) + { + path->fp_type = FIB_PATH_TYPE_DEAG; + path->deag.fp_tbl_id = rpath->frp_fib_index; + } + else if (rpath->frp_flags & FIB_ROUTE_PATH_DVR) + { + path->fp_type = FIB_PATH_TYPE_DVR; + path->dvr.fp_interface = rpath->frp_sw_if_index; + } + else if (rpath->frp_flags & FIB_ROUTE_PATH_EXCLUSIVE) + { + path->fp_type = FIB_PATH_TYPE_EXCLUSIVE; + dpo_copy(&path->exclusive.fp_ex_dpo, &rpath->dpo); + } + else if ((path->fp_cfg_flags & FIB_PATH_CFG_FLAG_ICMP_PROHIBIT) || + (path->fp_cfg_flags & FIB_PATH_CFG_FLAG_ICMP_UNREACH)) + { + path->fp_type = FIB_PATH_TYPE_SPECIAL; + } + else if ((path->fp_cfg_flags & FIB_PATH_CFG_FLAG_CLASSIFY)) + { + path->fp_type = FIB_PATH_TYPE_SPECIAL; + path->classify.fp_classify_table_id = rpath->frp_classify_table_id; + } + else if (path->fp_cfg_flags & FIB_PATH_CFG_FLAG_GLEAN) + { + path->fp_type = FIB_PATH_TYPE_ATTACHED; + path->attached.fp_interface = rpath->frp_sw_if_index; + path->attached.fp_connected = rpath->frp_connected; + } else if (~0 != rpath->frp_sw_if_index) { if (ip46_address_is_zero(&rpath->frp_addr)) @@ -1114,12 +1403,13 @@ fib_path_create (fib_node_index_t pl_index, { path->fp_type = FIB_PATH_TYPE_DEAG; path->deag.fp_tbl_id = rpath->frp_fib_index; - } + path->deag.fp_rpf_id = ~0; + } } else { path->fp_type = FIB_PATH_TYPE_RECURSIVE; - if (FIB_PROTOCOL_MPLS == path->fp_nh_proto) + if (DPO_PROTO_MPLS == path->fp_nh_proto) { path->recursive.fp_nh.fp_local_label = rpath->frp_local_label; path->recursive.fp_nh.fp_eos = rpath->frp_eos; @@ -1128,7 +1418,7 @@ fib_path_create (fib_node_index_t pl_index, { path->recursive.fp_nh.fp_ip = rpath->frp_addr; } - path->recursive.fp_tbl_id = rpath->frp_fib_index; + path->recursive.fp_tbl_id = rpath->frp_fib_index; } } @@ -1145,14 +1435,14 @@ fib_path_create (fib_node_index_t pl_index, */ fib_node_index_t fib_path_create_special (fib_node_index_t pl_index, - fib_protocol_t nh_proto, + dpo_proto_t nh_proto, fib_path_cfg_flags_t flags, const dpo_id_t *dpo) { fib_path_t *path; pool_get(fib_path_pool, path); - memset(path, 0, sizeof(*path)); + clib_memset(path, 0, sizeof(*path)); fib_node_init(&path->fp_node, FIB_NODE_TYPE_PATH); @@ -1160,6 +1450,7 @@ fib_path_create_special (fib_node_index_t pl_index, path->fp_pl_index = pl_index; path->fp_weight = 1; + path->fp_preference = 0; path->fp_nh_proto = nh_proto; path->fp_via_fib = FIB_NODE_INDEX_INVALID; path->fp_cfg_flags = flags; @@ -1199,7 +1490,7 @@ fib_path_copy (fib_node_index_t path_index, orig_path = fib_path_get(path_index); ASSERT(NULL != orig_path); - memcpy(path, orig_path, sizeof(*path)); + clib_memcpy(path, orig_path, sizeof(*path)); FIB_PATH_DBG(path, "create-copy:%d", path_index); @@ -1210,7 +1501,7 @@ fib_path_copy (fib_node_index_t path_index, path->fp_oper_flags = FIB_PATH_OPER_FLAG_NONE; path->fp_pl_index = path_list_index; path->fp_via_fib = FIB_NODE_INDEX_INVALID; - memset(&path->fp_dpo, 0, sizeof(path->fp_dpo)); + clib_memset(&path->fp_dpo, 0, sizeof(path->fp_dpo)); dpo_reset(&path->fp_dpo); return (fib_path_get_index(path)); @@ -1270,7 +1561,7 @@ fib_path_cmp_i (const fib_path_t *path1, /* * paths of different types and protocol are not equal. - * different weights only are the same path. + * different weights and/or preference only are the same path. */ if (path1->fp_type != path2->fp_type) { @@ -1301,14 +1592,26 @@ fib_path_cmp_i (const fib_path_t *path1, path2->attached.fp_interface); break; case FIB_PATH_TYPE_RECURSIVE: - res = ip46_address_cmp(&path1->recursive.fp_nh, - &path2->recursive.fp_nh); + res = ip46_address_cmp(&path1->recursive.fp_nh.fp_ip, + &path2->recursive.fp_nh.fp_ip); if (0 == res) { res = (path1->recursive.fp_tbl_id - path2->recursive.fp_tbl_id); } break; + case FIB_PATH_TYPE_BIER_FMASK: + res = (path1->bier_fmask.fp_bier_fmask - + path2->bier_fmask.fp_bier_fmask); + break; + case FIB_PATH_TYPE_BIER_IMP: + res = (path1->bier_imp.fp_bier_imp - + path2->bier_imp.fp_bier_imp); + break; + case FIB_PATH_TYPE_BIER_TABLE: + res = bier_table_id_cmp(&path1->bier_table.fp_bier_tbl, + &path2->bier_table.fp_bier_tbl); + break; case FIB_PATH_TYPE_DEAG: res = (path1->deag.fp_tbl_id - path2->deag.fp_tbl_id); if (0 == res) @@ -1319,9 +1622,17 @@ fib_path_cmp_i (const fib_path_t *path1, case FIB_PATH_TYPE_INTF_RX: res = (path1->intf_rx.fp_interface - path2->intf_rx.fp_interface); break; + case FIB_PATH_TYPE_UDP_ENCAP: + res = (path1->udp_encap.fp_udp_encap_id - path2->udp_encap.fp_udp_encap_id); + break; + case FIB_PATH_TYPE_DVR: + res = (path1->dvr.fp_interface - path2->dvr.fp_interface); + break; + case FIB_PATH_TYPE_EXCLUSIVE: + res = dpo_cmp(&path1->exclusive.fp_ex_dpo, &path2->exclusive.fp_ex_dpo); + break; case FIB_PATH_TYPE_SPECIAL: case FIB_PATH_TYPE_RECEIVE: - case FIB_PATH_TYPE_EXCLUSIVE: res = 0; break; } @@ -1345,6 +1656,15 @@ fib_path_cmp_for_sort (void * v1, path1 = fib_path_get(*pi1); path2 = fib_path_get(*pi2); + /* + * when sorting paths we want the highest preference paths + * first, so that the choices set built is in prefernce order + */ + if (path1->fp_preference != path2->fp_preference) + { + return (path1->fp_preference - path2->fp_preference); + } + return (fib_path_cmp_i(path1, path2)); } @@ -1401,7 +1721,7 @@ fib_path_cmp_w_route_path (fib_node_index_t path_index, res = (path->attached.fp_interface - rpath->frp_sw_if_index); break; case FIB_PATH_TYPE_RECURSIVE: - if (FIB_PROTOCOL_MPLS == path->fp_nh_proto) + if (DPO_PROTO_MPLS == path->fp_nh_proto) { res = path->recursive.fp_nh.fp_local_label - rpath->frp_local_label; @@ -1421,9 +1741,22 @@ fib_path_cmp_w_route_path (fib_node_index_t path_index, res = (path->recursive.fp_tbl_id - rpath->frp_fib_index); } break; + case FIB_PATH_TYPE_BIER_FMASK: + res = (path->bier_fmask.fp_bier_fmask - rpath->frp_bier_fmask); + break; + case FIB_PATH_TYPE_BIER_IMP: + res = (path->bier_imp.fp_bier_imp - rpath->frp_bier_imp); + break; + case FIB_PATH_TYPE_BIER_TABLE: + res = bier_table_id_cmp(&path->bier_table.fp_bier_tbl, + &rpath->frp_bier_tbl); + break; case FIB_PATH_TYPE_INTF_RX: res = (path->intf_rx.fp_interface - rpath->frp_sw_if_index); break; + case FIB_PATH_TYPE_UDP_ENCAP: + res = (path->udp_encap.fp_udp_encap_id - rpath->frp_udp_encap_id); + break; case FIB_PATH_TYPE_DEAG: res = (path->deag.fp_tbl_id - rpath->frp_fib_index); if (0 == res) @@ -1431,9 +1764,23 @@ fib_path_cmp_w_route_path (fib_node_index_t path_index, res = (path->deag.fp_rpf_id - rpath->frp_rpf_id); } break; - case FIB_PATH_TYPE_SPECIAL: - case FIB_PATH_TYPE_RECEIVE: + case FIB_PATH_TYPE_DVR: + res = (path->dvr.fp_interface - rpath->frp_sw_if_index); + break; case FIB_PATH_TYPE_EXCLUSIVE: + res = dpo_cmp(&path->exclusive.fp_ex_dpo, &rpath->dpo); + break; + case FIB_PATH_TYPE_RECEIVE: + if (rpath->frp_flags & FIB_ROUTE_PATH_LOCAL) + { + res = 0; + } + else + { + res = 1; + } + break; + case FIB_PATH_TYPE_SPECIAL: res = 0; break; } @@ -1503,14 +1850,13 @@ fib_path_recursive_loop_detect (fib_node_index_t path_index, FIB_PATH_DBG(path, "recursive loop formed"); path->fp_oper_flags |= FIB_PATH_OPER_FLAG_RECURSIVE_LOOP; - dpo_copy(&path->fp_dpo, - drop_dpo_get(fib_proto_to_dpo(path->fp_nh_proto))); + dpo_copy(&path->fp_dpo, drop_dpo_get(path->fp_nh_proto)); } else { /* * no loop here yet. keep forward walking the graph. - */ + */ if (fib_entry_recursive_loop_detect(path->fp_via_fib, entry_indicies)) { FIB_PATH_DBG(path, "recursive loop formed"); @@ -1526,11 +1872,29 @@ fib_path_recursive_loop_detect (fib_node_index_t path_index, } case FIB_PATH_TYPE_ATTACHED_NEXT_HOP: case FIB_PATH_TYPE_ATTACHED: + if (dpo_is_adj(&path->fp_dpo) && + adj_recursive_loop_detect(path->fp_dpo.dpoi_index, + entry_indicies)) + { + FIB_PATH_DBG(path, "recursive loop formed"); + path->fp_oper_flags |= FIB_PATH_OPER_FLAG_RECURSIVE_LOOP; + } + else + { + FIB_PATH_DBG(path, "recursive loop cleared"); + path->fp_oper_flags &= ~FIB_PATH_OPER_FLAG_RECURSIVE_LOOP; + } + break; case FIB_PATH_TYPE_SPECIAL: case FIB_PATH_TYPE_DEAG: + case FIB_PATH_TYPE_DVR: case FIB_PATH_TYPE_RECEIVE: case FIB_PATH_TYPE_INTF_RX: + case FIB_PATH_TYPE_UDP_ENCAP: case FIB_PATH_TYPE_EXCLUSIVE: + case FIB_PATH_TYPE_BIER_FMASK: + case FIB_PATH_TYPE_BIER_TABLE: + case FIB_PATH_TYPE_BIER_IMP: /* * these path types cannot be part of a loop, since they are the leaves * of the graph. @@ -1558,8 +1922,7 @@ fib_path_resolve (fib_node_index_t path_index) */ if (fib_path_is_permanent_drop(path)) { - dpo_copy(&path->fp_dpo, - drop_dpo_get(fib_proto_to_dpo(path->fp_nh_proto))); + dpo_copy(&path->fp_dpo, drop_dpo_get(path->fp_nh_proto)); path->fp_oper_flags &= ~FIB_PATH_OPER_FLAG_RESOLVED; return (fib_path_is_resolved(path_index)); } @@ -1570,48 +1933,40 @@ fib_path_resolve (fib_node_index_t path_index) fib_path_attached_next_hop_set(path); break; case FIB_PATH_TYPE_ATTACHED: - /* - * path->attached.fp_interface - */ - if (!vnet_sw_interface_is_admin_up(vnet_get_main(), - path->attached.fp_interface)) - { - path->fp_oper_flags &= ~FIB_PATH_OPER_FLAG_RESOLVED; - } - if (vnet_sw_interface_is_p2p(vnet_get_main(), - path->attached.fp_interface)) - { - /* - * point-2-point interfaces do not require a glean, since - * there is nothing to ARP. Install a rewrite/nbr adj instead - */ - dpo_set(&path->fp_dpo, - DPO_ADJACENCY, - fib_proto_to_dpo(path->fp_nh_proto), - adj_nbr_add_or_lock( - path->fp_nh_proto, - fib_proto_to_link(path->fp_nh_proto), - &zero_addr, - path->attached.fp_interface)); - } - else - { - dpo_set(&path->fp_dpo, - DPO_ADJACENCY_GLEAN, - fib_proto_to_dpo(path->fp_nh_proto), - adj_glean_add_or_lock(path->fp_nh_proto, - path->attached.fp_interface, - NULL)); - } - /* - * become a child of the adjacency so we receive updates - * when the interface state changes - */ - path->fp_sibling = adj_child_add(path->fp_dpo.dpoi_index, - FIB_NODE_TYPE_PATH, - fib_path_get_index(path)); + { + dpo_id_t tmp = DPO_INVALID; + /* + * path->attached.fp_interface + */ + if (!vnet_sw_interface_is_up(vnet_get_main(), + path->attached.fp_interface)) + { + path->fp_oper_flags &= ~FIB_PATH_OPER_FLAG_RESOLVED; + } + fib_path_attached_get_adj(path, + dpo_proto_to_link(path->fp_nh_proto), + &tmp); + + /* + * re-fetch after possible mem realloc + */ + path = fib_path_get(path_index); + dpo_copy(&path->fp_dpo, &tmp); + + /* + * become a child of the adjacency so we receive updates + * when the interface state changes + */ + if (dpo_is_adj(&path->fp_dpo)) + { + path->fp_sibling = adj_child_add(path->fp_dpo.dpoi_index, + FIB_NODE_TYPE_PATH, + fib_path_get_index(path)); + } + dpo_reset(&tmp); break; + } case FIB_PATH_TYPE_RECURSIVE: { /* @@ -1626,7 +1981,7 @@ fib_path_resolve (fib_node_index_t path_index) ASSERT(FIB_NODE_INDEX_INVALID == path->fp_via_fib); - if (FIB_PROTOCOL_MPLS == path->fp_nh_proto) + if (DPO_PROTO_MPLS == path->fp_nh_proto) { fib_prefix_from_mpls_label(path->recursive.fp_nh.fp_local_label, path->recursive.fp_nh.fp_eos, @@ -1637,11 +1992,13 @@ fib_path_resolve (fib_node_index_t path_index) fib_prefix_from_ip46_addr(&path->recursive.fp_nh.fp_ip, &pfx); } + fib_table_lock(path->recursive.fp_tbl_id, + dpo_proto_to_fib(path->fp_nh_proto), + FIB_SOURCE_RR); fei = fib_table_entry_special_add(path->recursive.fp_tbl_id, &pfx, FIB_SOURCE_RR, - FIB_ENTRY_FLAG_NONE, - ADJ_INDEX_INVALID); + FIB_ENTRY_FLAG_NONE); path = fib_path_get(path_index); path->fp_via_fib = fei; @@ -1664,49 +2021,128 @@ fib_path_resolve (fib_node_index_t path_index) break; } + case FIB_PATH_TYPE_BIER_FMASK: + { + /* + * become a dependent child of the entry so the path is + * informed when the forwarding for the entry changes. + */ + path->fp_sibling = bier_fmask_child_add(path->bier_fmask.fp_bier_fmask, + FIB_NODE_TYPE_PATH, + fib_path_get_index(path)); + + path->fp_via_bier_fmask = path->bier_fmask.fp_bier_fmask; + fib_path_bier_fmask_update(path, &path->fp_dpo); + + break; + } + case FIB_PATH_TYPE_BIER_IMP: + bier_imp_lock(path->bier_imp.fp_bier_imp); + bier_imp_contribute_forwarding(path->bier_imp.fp_bier_imp, + DPO_PROTO_IP4, + &path->fp_dpo); + break; + case FIB_PATH_TYPE_BIER_TABLE: + { + /* + * Find/create the BIER table to link to + */ + ASSERT(FIB_NODE_INDEX_INVALID == path->fp_via_bier_tbl); + + path->fp_via_bier_tbl = + bier_table_ecmp_create_and_lock(&path->bier_table.fp_bier_tbl); + + bier_table_contribute_forwarding(path->fp_via_bier_tbl, + &path->fp_dpo); + break; + } case FIB_PATH_TYPE_SPECIAL: - /* - * Resolve via the drop - */ - dpo_copy(&path->fp_dpo, - drop_dpo_get(fib_proto_to_dpo(path->fp_nh_proto))); - break; + if (path->fp_cfg_flags & FIB_PATH_CFG_FLAG_ICMP_PROHIBIT) + { + ip_null_dpo_add_and_lock (path->fp_nh_proto, + IP_NULL_ACTION_SEND_ICMP_PROHIBIT, + &path->fp_dpo); + } + else if (path->fp_cfg_flags & FIB_PATH_CFG_FLAG_ICMP_UNREACH) + { + ip_null_dpo_add_and_lock (path->fp_nh_proto, + IP_NULL_ACTION_SEND_ICMP_UNREACH, + &path->fp_dpo); + } + else if (path->fp_cfg_flags & FIB_PATH_CFG_FLAG_CLASSIFY) + { + dpo_set (&path->fp_dpo, DPO_CLASSIFY, + path->fp_nh_proto, + classify_dpo_create (path->fp_nh_proto, + path->classify.fp_classify_table_id)); + } + else + { + /* + * Resolve via the drop + */ + dpo_copy(&path->fp_dpo, drop_dpo_get(path->fp_nh_proto)); + } + break; case FIB_PATH_TYPE_DEAG: { - /* - * Resolve via a lookup DPO. - * FIXME. control plane should add routes with a table ID - */ - lookup_cast_t cast; - - cast = (path->fp_cfg_flags & FIB_PATH_CFG_FLAG_RPF_ID ? - LOOKUP_MULTICAST : - LOOKUP_UNICAST); - - lookup_dpo_add_or_lock_w_fib_index(path->deag.fp_tbl_id, - fib_proto_to_dpo(path->fp_nh_proto), - cast, - LOOKUP_INPUT_DST_ADDR, - LOOKUP_TABLE_FROM_CONFIG, - &path->fp_dpo); - break; + if (DPO_PROTO_BIER == path->fp_nh_proto) + { + bier_disp_table_contribute_forwarding(path->deag.fp_tbl_id, + &path->fp_dpo); + } + else + { + /* + * Resolve via a lookup DPO. + * FIXME. control plane should add routes with a table ID + */ + lookup_input_t input; + lookup_cast_t cast; + + cast = (path->fp_cfg_flags & FIB_PATH_CFG_FLAG_RPF_ID ? + LOOKUP_MULTICAST : + LOOKUP_UNICAST); + input = (path->fp_cfg_flags & FIB_PATH_CFG_FLAG_DEAG_SRC ? + LOOKUP_INPUT_SRC_ADDR : + LOOKUP_INPUT_DST_ADDR); + + lookup_dpo_add_or_lock_w_fib_index(path->deag.fp_tbl_id, + path->fp_nh_proto, + cast, + input, + LOOKUP_TABLE_FROM_CONFIG, + &path->fp_dpo); + } + break; } + case FIB_PATH_TYPE_DVR: + dvr_dpo_add_or_lock(path->dvr.fp_interface, + path->fp_nh_proto, + &path->fp_dpo); + break; case FIB_PATH_TYPE_RECEIVE: /* * Resolve via a receive DPO. */ - receive_dpo_add_or_lock(fib_proto_to_dpo(path->fp_nh_proto), + receive_dpo_add_or_lock(path->fp_nh_proto, path->receive.fp_interface, &path->receive.fp_addr, &path->fp_dpo); break; + case FIB_PATH_TYPE_UDP_ENCAP: + udp_encap_lock(path->udp_encap.fp_udp_encap_id); + udp_encap_contribute_forwarding(path->udp_encap.fp_udp_encap_id, + path->fp_nh_proto, + &path->fp_dpo); + break; case FIB_PATH_TYPE_INTF_RX: { /* * Resolve via a receive DPO. */ - interface_dpo_add_or_lock(fib_proto_to_dpo(path->fp_nh_proto), - path->intf_rx.fp_interface, - &path->fp_dpo); + interface_rx_dpo_add_or_lock(path->fp_nh_proto, + path->intf_rx.fp_interface, + &path->fp_dpo); break; } case FIB_PATH_TYPE_EXCLUSIVE: @@ -1736,13 +2172,55 @@ fib_path_get_resolving_interface (fib_node_index_t path_index) case FIB_PATH_TYPE_RECEIVE: return (path->receive.fp_interface); case FIB_PATH_TYPE_RECURSIVE: - return (fib_entry_get_resolving_interface(path->fp_via_fib)); + if (fib_path_is_resolved(path_index)) + { + return (fib_entry_get_resolving_interface(path->fp_via_fib)); + } + break; + case FIB_PATH_TYPE_DVR: + return (path->dvr.fp_interface); case FIB_PATH_TYPE_INTF_RX: + case FIB_PATH_TYPE_UDP_ENCAP: case FIB_PATH_TYPE_SPECIAL: case FIB_PATH_TYPE_DEAG: case FIB_PATH_TYPE_EXCLUSIVE: + case FIB_PATH_TYPE_BIER_FMASK: + case FIB_PATH_TYPE_BIER_TABLE: + case FIB_PATH_TYPE_BIER_IMP: break; } + return (dpo_get_urpf(&path->fp_dpo)); +} + +index_t +fib_path_get_resolving_index (fib_node_index_t path_index) +{ + fib_path_t *path; + + path = fib_path_get(path_index); + + switch (path->fp_type) + { + case FIB_PATH_TYPE_ATTACHED_NEXT_HOP: + case FIB_PATH_TYPE_ATTACHED: + case FIB_PATH_TYPE_RECEIVE: + case FIB_PATH_TYPE_INTF_RX: + case FIB_PATH_TYPE_SPECIAL: + case FIB_PATH_TYPE_DEAG: + case FIB_PATH_TYPE_DVR: + case FIB_PATH_TYPE_EXCLUSIVE: + break; + case FIB_PATH_TYPE_UDP_ENCAP: + return (path->udp_encap.fp_udp_encap_id); + case FIB_PATH_TYPE_RECURSIVE: + return (path->fp_via_fib); + case FIB_PATH_TYPE_BIER_FMASK: + return (path->bier_fmask.fp_bier_fmask); + case FIB_PATH_TYPE_BIER_TABLE: + return (path->fp_via_bier_tbl); + case FIB_PATH_TYPE_BIER_IMP: + return (path->bier_imp.fp_bier_imp); + } return (~0); } @@ -1753,7 +2231,6 @@ fib_path_get_adj (fib_node_index_t path_index) path = fib_path_get(path_index); - ASSERT(dpo_is_adj(&path->fp_dpo)); if (dpo_is_adj(&path->fp_dpo)) { return (path->fp_dpo.dpoi_index); @@ -1761,7 +2238,7 @@ fib_path_get_adj (fib_node_index_t path_index) return (ADJ_INDEX_INVALID); } -int +u16 fib_path_get_weight (fib_node_index_t path_index) { fib_path_t *path; @@ -1773,6 +2250,35 @@ fib_path_get_weight (fib_node_index_t path_index) return (path->fp_weight); } +u16 +fib_path_get_preference (fib_node_index_t path_index) +{ + fib_path_t *path; + + path = fib_path_get(path_index); + + ASSERT(path); + + return (path->fp_preference); +} + +u32 +fib_path_get_rpf_id (fib_node_index_t path_index) +{ + fib_path_t *path; + + path = fib_path_get(path_index); + + ASSERT(path); + + if (FIB_PATH_CFG_FLAG_RPF_ID & path->fp_cfg_flags) + { + return (path->deag.fp_rpf_id); + } + + return (~0); +} + /** * @brief Contribute the path's adjacency to the list passed. * By calling this function over all paths, recursively, a child @@ -1801,11 +2307,12 @@ fib_path_contribute_urpf (fib_node_index_t path_index, break; case FIB_PATH_TYPE_RECURSIVE: - if (FIB_NODE_INDEX_INVALID != path->fp_via_fib) + if (FIB_NODE_INDEX_INVALID != path->fp_via_fib && + !fib_path_is_looped(path_index)) { /* * there's unresolved due to constraints, and there's unresolved - * due to ain't go no via. can't do nowt w'out via. + * due to ain't got no via. can't do nowt w'out via. */ fib_entry_contribute_urpf(path->fp_via_fib, urpf); } @@ -1813,23 +2320,31 @@ fib_path_contribute_urpf (fib_node_index_t path_index, case FIB_PATH_TYPE_EXCLUSIVE: case FIB_PATH_TYPE_SPECIAL: - /* + { + /* * these path types may link to an adj, if that's what * the clinet gave */ - if (dpo_is_adj(&path->fp_dpo)) - { - ip_adjacency_t *adj; + u32 rpf_sw_if_index; - adj = adj_get(path->fp_dpo.dpoi_index); + rpf_sw_if_index = dpo_get_urpf(&path->fp_dpo); - fib_urpf_list_append(urpf, adj->rewrite_header.sw_if_index); + if (~0 != rpf_sw_if_index) + { + fib_urpf_list_append(urpf, rpf_sw_if_index); } break; - + } + case FIB_PATH_TYPE_DVR: + fib_urpf_list_append(urpf, path->dvr.fp_interface); + break; case FIB_PATH_TYPE_DEAG: case FIB_PATH_TYPE_RECEIVE: case FIB_PATH_TYPE_INTF_RX: + case FIB_PATH_TYPE_UDP_ENCAP: + case FIB_PATH_TYPE_BIER_FMASK: + case FIB_PATH_TYPE_BIER_TABLE: + case FIB_PATH_TYPE_BIER_IMP: /* * these path types don't link to an adj */ @@ -1840,6 +2355,7 @@ fib_path_contribute_urpf (fib_node_index_t path_index, void fib_path_stack_mpls_disp (fib_node_index_t path_index, dpo_proto_t payload_proto, + fib_mpls_lsp_mode_t mode, dpo_id_t *dpo) { fib_path_t *path; @@ -1850,29 +2366,51 @@ fib_path_stack_mpls_disp (fib_node_index_t path_index, switch (path->fp_type) { - case FIB_PATH_TYPE_DEAG: + case FIB_PATH_TYPE_ATTACHED_NEXT_HOP: { dpo_id_t tmp = DPO_INVALID; dpo_copy(&tmp, dpo); - dpo_set(dpo, - DPO_MPLS_DISPOSITION, - payload_proto, - mpls_disp_dpo_create(payload_proto, - path->deag.fp_rpf_id, - &tmp)); + + mpls_disp_dpo_create(payload_proto, ~0, mode, &tmp, dpo); dpo_reset(&tmp); break; } + case FIB_PATH_TYPE_DEAG: + { + dpo_id_t tmp = DPO_INVALID; + + dpo_copy(&tmp, dpo); + + mpls_disp_dpo_create(payload_proto, + path->deag.fp_rpf_id, + mode, &tmp, dpo); + dpo_reset(&tmp); + break; + } case FIB_PATH_TYPE_RECEIVE: case FIB_PATH_TYPE_ATTACHED: - case FIB_PATH_TYPE_ATTACHED_NEXT_HOP: case FIB_PATH_TYPE_RECURSIVE: case FIB_PATH_TYPE_INTF_RX: + case FIB_PATH_TYPE_UDP_ENCAP: case FIB_PATH_TYPE_EXCLUSIVE: case FIB_PATH_TYPE_SPECIAL: + case FIB_PATH_TYPE_BIER_FMASK: + case FIB_PATH_TYPE_BIER_TABLE: + case FIB_PATH_TYPE_BIER_IMP: + case FIB_PATH_TYPE_DVR: break; } + + if (path->fp_cfg_flags & FIB_PATH_CFG_FLAG_POP_PW_CW) + { + dpo_id_t tmp = DPO_INVALID; + + dpo_copy(&tmp, dpo); + + pw_cw_dpo_create(&tmp, dpo); + dpo_reset(&tmp); + } } void @@ -1887,8 +2425,6 @@ fib_path_contribute_forwarding (fib_node_index_t path_index, ASSERT(path); ASSERT(FIB_FORW_CHAIN_TYPE_MPLS_EOS != fct); - FIB_PATH_DBG(path, "contribute"); - /* * The DPO stored in the path was created when the path was resolved. * This then represents the path's 'native' protocol; IP. @@ -1911,43 +2447,76 @@ fib_path_contribute_forwarding (fib_node_index_t path_index, case FIB_FORW_CHAIN_TYPE_MPLS_NON_EOS: case FIB_FORW_CHAIN_TYPE_ETHERNET: case FIB_FORW_CHAIN_TYPE_NSH: - { - adj_index_t ai; - - /* - * get a appropriate link type adj. - */ - ai = fib_path_attached_next_hop_get_adj( - path, - fib_forw_chain_type_to_link_type(fct)); - dpo_set(dpo, DPO_ADJACENCY, - fib_forw_chain_type_to_dpo_proto(fct), ai); - adj_unlock(ai); - + case FIB_FORW_CHAIN_TYPE_MCAST_IP4: + case FIB_FORW_CHAIN_TYPE_MCAST_IP6: + path = fib_path_attached_next_hop_get_adj( + path, + fib_forw_chain_type_to_link_type(fct), + dpo); + break; + case FIB_FORW_CHAIN_TYPE_BIER: break; } + break; + case FIB_PATH_TYPE_RECURSIVE: + switch (fct) + { + case FIB_FORW_CHAIN_TYPE_MPLS_EOS: + case FIB_FORW_CHAIN_TYPE_UNICAST_IP4: + case FIB_FORW_CHAIN_TYPE_UNICAST_IP6: + case FIB_FORW_CHAIN_TYPE_MPLS_NON_EOS: case FIB_FORW_CHAIN_TYPE_MCAST_IP4: case FIB_FORW_CHAIN_TYPE_MCAST_IP6: + case FIB_FORW_CHAIN_TYPE_BIER: + fib_path_recursive_adj_update(path, fct, dpo); + break; + case FIB_FORW_CHAIN_TYPE_ETHERNET: + case FIB_FORW_CHAIN_TYPE_NSH: + ASSERT(0); + break; + } break; + case FIB_PATH_TYPE_BIER_TABLE: + switch (fct) + { + case FIB_FORW_CHAIN_TYPE_BIER: + bier_table_contribute_forwarding(path->fp_via_bier_tbl, dpo); + break; + case FIB_FORW_CHAIN_TYPE_MPLS_EOS: + case FIB_FORW_CHAIN_TYPE_UNICAST_IP4: + case FIB_FORW_CHAIN_TYPE_UNICAST_IP6: + case FIB_FORW_CHAIN_TYPE_MPLS_NON_EOS: + case FIB_FORW_CHAIN_TYPE_MCAST_IP4: + case FIB_FORW_CHAIN_TYPE_MCAST_IP6: + case FIB_FORW_CHAIN_TYPE_ETHERNET: + case FIB_FORW_CHAIN_TYPE_NSH: + ASSERT(0); + break; } break; - case FIB_PATH_TYPE_RECURSIVE: + case FIB_PATH_TYPE_BIER_FMASK: switch (fct) { + case FIB_FORW_CHAIN_TYPE_BIER: + fib_path_bier_fmask_update(path, dpo); + break; case FIB_FORW_CHAIN_TYPE_MPLS_EOS: case FIB_FORW_CHAIN_TYPE_UNICAST_IP4: case FIB_FORW_CHAIN_TYPE_UNICAST_IP6: case FIB_FORW_CHAIN_TYPE_MPLS_NON_EOS: case FIB_FORW_CHAIN_TYPE_MCAST_IP4: case FIB_FORW_CHAIN_TYPE_MCAST_IP6: - fib_path_recursive_adj_update(path, fct, dpo); - break; case FIB_FORW_CHAIN_TYPE_ETHERNET: case FIB_FORW_CHAIN_TYPE_NSH: ASSERT(0); break; } break; + case FIB_PATH_TYPE_BIER_IMP: + bier_imp_contribute_forwarding(path->bier_imp.fp_bier_imp, + fib_forw_chain_type_to_dpo_proto(fct), + dpo); + break; case FIB_PATH_TYPE_DEAG: switch (fct) { @@ -1962,10 +2531,12 @@ fib_path_contribute_forwarding (fib_node_index_t path_index, case FIB_FORW_CHAIN_TYPE_MPLS_EOS: case FIB_FORW_CHAIN_TYPE_UNICAST_IP4: case FIB_FORW_CHAIN_TYPE_UNICAST_IP6: - dpo_copy(dpo, &path->fp_dpo); - break; case FIB_FORW_CHAIN_TYPE_MCAST_IP4: case FIB_FORW_CHAIN_TYPE_MCAST_IP6: + dpo_copy(dpo, &path->fp_dpo); + break; + case FIB_FORW_CHAIN_TYPE_BIER: + break; case FIB_FORW_CHAIN_TYPE_ETHERNET: case FIB_FORW_CHAIN_TYPE_NSH: ASSERT(0); @@ -1984,6 +2555,10 @@ fib_path_contribute_forwarding (fib_node_index_t path_index, case FIB_FORW_CHAIN_TYPE_MPLS_EOS: case FIB_FORW_CHAIN_TYPE_ETHERNET: case FIB_FORW_CHAIN_TYPE_NSH: + case FIB_FORW_CHAIN_TYPE_BIER: + fib_path_attached_get_adj(path, + fib_forw_chain_type_to_link_type(fct), + dpo); break; case FIB_FORW_CHAIN_TYPE_MCAST_IP4: case FIB_FORW_CHAIN_TYPE_MCAST_IP6: @@ -1993,9 +2568,24 @@ fib_path_contribute_forwarding (fib_node_index_t path_index, /* * Create the adj needed for sending IP multicast traffic */ - ai = adj_mcast_add_or_lock(path->fp_nh_proto, - fib_forw_chain_type_to_link_type(fct), - path->attached.fp_interface); + if (vnet_sw_interface_is_p2p(vnet_get_main(), + path->attached.fp_interface)) + { + /* + * point-2-point interfaces do not require a glean, since + * there is nothing to ARP. Install a rewrite/nbr adj instead + */ + ai = adj_nbr_add_or_lock(dpo_proto_to_fib(path->fp_nh_proto), + fib_forw_chain_type_to_link_type(fct), + &zero_addr, + path->attached.fp_interface); + } + else + { + ai = adj_mcast_add_or_lock(dpo_proto_to_fib(path->fp_nh_proto), + fib_forw_chain_type_to_link_type(fct), + path->attached.fp_interface); + } dpo_set(dpo, DPO_ADJACENCY, fib_forw_chain_type_to_dpo_proto(fct), ai); @@ -2008,12 +2598,18 @@ fib_path_contribute_forwarding (fib_node_index_t path_index, /* * Create the adj needed for sending IP multicast traffic */ - interface_dpo_add_or_lock(fib_forw_chain_type_to_dpo_proto(fct), - path->attached.fp_interface, - dpo); + interface_rx_dpo_add_or_lock(fib_forw_chain_type_to_dpo_proto(fct), + path->attached.fp_interface, + dpo); + break; + case FIB_PATH_TYPE_UDP_ENCAP: + udp_encap_contribute_forwarding(path->udp_encap.fp_udp_encap_id, + path->fp_nh_proto, + dpo); break; case FIB_PATH_TYPE_RECEIVE: case FIB_PATH_TYPE_SPECIAL: + case FIB_PATH_TYPE_DVR: dpo_copy(dpo, &path->fp_dpo); break; } @@ -2032,26 +2628,33 @@ fib_path_append_nh_for_multipath_hash (fib_node_index_t path_index, ASSERT(path); + vec_add2(hash_key, mnh, 1); + + mnh->path_weight = path->fp_weight; + mnh->path_index = path_index; + if (fib_path_is_resolved(path_index)) { - vec_add2(hash_key, mnh, 1); - - mnh->path_weight = path->fp_weight; - mnh->path_index = path_index; - fib_path_contribute_forwarding(path_index, fct, &mnh->path_dpo); + fib_path_contribute_forwarding(path_index, fct, &mnh->path_dpo); + } + else + { + dpo_copy(&mnh->path_dpo, + drop_dpo_get(fib_forw_chain_type_to_dpo_proto(fct))); } - return (hash_key); } int -fib_path_is_recursive (fib_node_index_t path_index) +fib_path_is_recursive_constrained (fib_node_index_t path_index) { fib_path_t *path; path = fib_path_get(path_index); - return (FIB_PATH_TYPE_RECURSIVE == path->fp_type); + return ((FIB_PATH_TYPE_RECURSIVE == path->fp_type) && + ((path->fp_cfg_flags & FIB_PATH_CFG_FLAG_RESOLVE_ATTACHED) || + (path->fp_cfg_flags & FIB_PATH_CFG_FLAG_RESOLVE_HOST))); } int @@ -2097,50 +2700,91 @@ fib_path_is_looped (fib_node_index_t path_index) return (path->fp_oper_flags & FIB_PATH_OPER_FLAG_RECURSIVE_LOOP); } -int +fib_path_list_walk_rc_t fib_path_encode (fib_node_index_t path_list_index, - fib_node_index_t path_index, - void *ctx) + fib_node_index_t path_index, + const fib_path_ext_t *path_ext, + void *args) { - fib_route_path_encode_t **api_rpaths = ctx; - fib_route_path_encode_t *api_rpath; + fib_path_encode_ctx_t *ctx = args; + fib_route_path_t *rpath; fib_path_t *path; path = fib_path_get(path_index); if (!path) - return (0); - vec_add2(*api_rpaths, api_rpath, 1); - api_rpath->rpath.frp_weight = path->fp_weight; - api_rpath->rpath.frp_proto = path->fp_nh_proto; - api_rpath->rpath.frp_sw_if_index = ~0; - api_rpath->dpo = path->exclusive.fp_ex_dpo; + return (FIB_PATH_LIST_WALK_CONTINUE); + + vec_add2(ctx->rpaths, rpath, 1); + rpath->frp_weight = path->fp_weight; + rpath->frp_preference = path->fp_preference; + rpath->frp_proto = path->fp_nh_proto; + rpath->frp_sw_if_index = ~0; + rpath->frp_fib_index = 0; + switch (path->fp_type) - { + { case FIB_PATH_TYPE_RECEIVE: - api_rpath->rpath.frp_addr = path->receive.fp_addr; - api_rpath->rpath.frp_sw_if_index = path->receive.fp_interface; + rpath->frp_addr = path->receive.fp_addr; + rpath->frp_sw_if_index = path->receive.fp_interface; + rpath->frp_flags |= FIB_ROUTE_PATH_LOCAL; break; case FIB_PATH_TYPE_ATTACHED: - api_rpath->rpath.frp_sw_if_index = path->attached.fp_interface; + rpath->frp_sw_if_index = path->attached.fp_interface; break; case FIB_PATH_TYPE_ATTACHED_NEXT_HOP: - api_rpath->rpath.frp_sw_if_index = path->attached_next_hop.fp_interface; - api_rpath->rpath.frp_addr = path->attached_next_hop.fp_nh; + rpath->frp_sw_if_index = path->attached_next_hop.fp_interface; + rpath->frp_addr = path->attached_next_hop.fp_nh; + break; + case FIB_PATH_TYPE_BIER_FMASK: + rpath->frp_bier_fmask = path->bier_fmask.fp_bier_fmask; break; case FIB_PATH_TYPE_SPECIAL: break; case FIB_PATH_TYPE_DEAG: + rpath->frp_fib_index = path->deag.fp_tbl_id; + if (path->fp_cfg_flags & FIB_PATH_CFG_FLAG_RPF_ID) + { + rpath->frp_flags |= FIB_ROUTE_PATH_RPF_ID; + } break; case FIB_PATH_TYPE_RECURSIVE: - api_rpath->rpath.frp_addr = path->recursive.fp_nh.fp_ip; + rpath->frp_addr = path->recursive.fp_nh.fp_ip; + rpath->frp_fib_index = path->recursive.fp_tbl_id; break; + case FIB_PATH_TYPE_DVR: + rpath->frp_sw_if_index = path->dvr.fp_interface; + rpath->frp_flags |= FIB_ROUTE_PATH_DVR; + break; + case FIB_PATH_TYPE_UDP_ENCAP: + rpath->frp_udp_encap_id = path->udp_encap.fp_udp_encap_id; + rpath->frp_flags |= FIB_ROUTE_PATH_UDP_ENCAP; + break; + case FIB_PATH_TYPE_INTF_RX: + rpath->frp_sw_if_index = path->receive.fp_interface; + rpath->frp_flags |= FIB_ROUTE_PATH_INTF_RX; + break; + case FIB_PATH_TYPE_EXCLUSIVE: + rpath->frp_flags |= FIB_ROUTE_PATH_EXCLUSIVE; default: break; - } - return (1); + } + + if (path_ext && path_ext->fpe_type == FIB_PATH_EXT_MPLS) + { + rpath->frp_label_stack = path_ext->fpe_path.frp_label_stack; + } + + if (path->fp_cfg_flags & FIB_PATH_CFG_FLAG_DROP) + rpath->frp_flags |= FIB_ROUTE_PATH_DROP; + if (path->fp_cfg_flags & FIB_PATH_CFG_FLAG_ICMP_UNREACH) + rpath->frp_flags |= FIB_ROUTE_PATH_ICMP_UNREACH; + if (path->fp_cfg_flags & FIB_PATH_CFG_FLAG_ICMP_PROHIBIT) + rpath->frp_flags |= FIB_ROUTE_PATH_ICMP_PROHIBIT; + + return (FIB_PATH_LIST_WALK_CONTINUE); } -fib_protocol_t +dpo_proto_t fib_path_get_proto (fib_node_index_t path_index) { fib_path_t *path; @@ -2154,6 +2798,7 @@ void fib_path_module_init (void) { fib_node_register_type (FIB_NODE_TYPE_PATH, &fib_path_vft); + fib_path_logger = vlib_log_register_class ("fib", "path"); } static clib_error_t * @@ -2172,10 +2817,11 @@ show_fib_path_command (vlib_main_t * vm, if (!pool_is_free_index(fib_path_pool, pi)) { path = fib_path_get(pi); - u8 *s = fib_path_format(pi, NULL); - s = format(s, "children:"); + u8 *s = format(NULL, "%U", format_fib_path, pi, 1, + FIB_PATH_FORMAT_FLAGS_NONE); + s = format(s, "\n children:"); s = fib_node_children_format(path->fp_node.fn_children, s); - vlib_cli_output (vm, "%s", s); + vlib_cli_output (vm, "%v", s); vec_free(s); } else @@ -2186,9 +2832,10 @@ show_fib_path_command (vlib_main_t * vm, else { vlib_cli_output (vm, "FIB Paths"); - pool_foreach(path, fib_path_pool, + pool_foreach_index (pi, fib_path_pool, ({ - vlib_cli_output (vm, "%U", format_fib_path, path); + vlib_cli_output (vm, "%U", format_fib_path, pi, 0, + FIB_PATH_FORMAT_FLAGS_NONE); })); }