From: khemendra kumar Date: Mon, 27 Nov 2017 09:45:53 +0000 (+0530) Subject: VPP-246 Coding standards cleanup - vnet/vnet/classify X-Git-Tag: v18.04-rc0~158 X-Git-Url: https://gerrit.fd.io/r/gitweb?a=commitdiff_plain;h=d7bfa0e4fedb896f9c2808053c6c6620d3a41a43;p=vpp.git VPP-246 Coding standards cleanup - vnet/vnet/classify Signed-off-by: khemendra kumar Change-Id: I0d94ef604d34da6981e7c2d2b4da5ec3ec5fb19a --- diff --git a/src/vnet/classify/flow_classify.c b/src/vnet/classify/flow_classify.c index 6c067f55759..941a030f640 100644 --- a/src/vnet/classify/flow_classify.c +++ b/src/vnet/classify/flow_classify.c @@ -18,10 +18,10 @@ flow_classify_main_t flow_classify_main; static void vnet_flow_classify_feature_enable (vlib_main_t * vnm, - flow_classify_main_t * fcm, - u32 sw_if_index, - flow_classify_table_id_t tid, - int feature_enable) + flow_classify_main_t * fcm, + u32 sw_if_index, + flow_classify_table_id_t tid, + int feature_enable) { vnet_feature_config_main_t *vfcm; u8 arc; @@ -43,13 +43,14 @@ vnet_flow_classify_feature_enable (vlib_main_t * vnm, fcm->vnet_config_main[tid] = &vfcm->config_main; } -int vnet_set_flow_classify_intfc (vlib_main_t * vm, u32 sw_if_index, - u32 ip4_table_index, u32 ip6_table_index, - u32 is_add) +int +vnet_set_flow_classify_intfc (vlib_main_t * vm, u32 sw_if_index, + u32 ip4_table_index, u32 ip6_table_index, + u32 is_add) { - flow_classify_main_t * fcm = &flow_classify_main; - vnet_classify_main_t * vcm = fcm->vnet_classify_main; - u32 pct[FLOW_CLASSIFY_N_TABLES] = {ip4_table_index, ip6_table_index}; + flow_classify_main_t *fcm = &flow_classify_main; + vnet_classify_main_t *vcm = fcm->vnet_classify_main; + u32 pct[FLOW_CLASSIFY_N_TABLES] = { ip4_table_index, ip6_table_index }; u32 ti; /* Assume that we've validated sw_if_index in the API layer */ @@ -57,34 +58,36 @@ int vnet_set_flow_classify_intfc (vlib_main_t * vm, u32 sw_if_index, for (ti = 0; ti < FLOW_CLASSIFY_N_TABLES; ti++) { if (pct[ti] == ~0) - continue; + continue; if (pool_is_free_index (vcm->tables, pct[ti])) - return VNET_API_ERROR_NO_SUCH_TABLE; + return VNET_API_ERROR_NO_SUCH_TABLE; vec_validate_init_empty - (fcm->classify_table_index_by_sw_if_index[ti], sw_if_index, ~0); + (fcm->classify_table_index_by_sw_if_index[ti], sw_if_index, ~0); /* Reject any DEL operation with wrong sw_if_index */ if (!is_add && - (pct[ti] != fcm->classify_table_index_by_sw_if_index[ti][sw_if_index])) - { - clib_warning ("Non-existent intf_idx=%d with table_index=%d for delete", - sw_if_index, pct[ti]); - return VNET_API_ERROR_NO_SUCH_TABLE; - } + (pct[ti] != + fcm->classify_table_index_by_sw_if_index[ti][sw_if_index])) + { + clib_warning + ("Non-existent intf_idx=%d with table_index=%d for delete", + sw_if_index, pct[ti]); + return VNET_API_ERROR_NO_SUCH_TABLE; + } /* Return ok on ADD operaton if feature is already enabled */ if (is_add && - fcm->classify_table_index_by_sw_if_index[ti][sw_if_index] != ~0) - return 0; + fcm->classify_table_index_by_sw_if_index[ti][sw_if_index] != ~0) + return 0; vnet_flow_classify_feature_enable (vm, fcm, sw_if_index, ti, is_add); if (is_add) - fcm->classify_table_index_by_sw_if_index[ti][sw_if_index] = pct[ti]; + fcm->classify_table_index_by_sw_if_index[ti][sw_if_index] = pct[ti]; else - fcm->classify_table_index_by_sw_if_index[ti][sw_if_index] = ~0; + fcm->classify_table_index_by_sw_if_index[ti][sw_if_index] = ~0; } @@ -93,10 +96,10 @@ int vnet_set_flow_classify_intfc (vlib_main_t * vm, u32 sw_if_index, static clib_error_t * set_flow_classify_command_fn (vlib_main_t * vm, - unformat_input_t * input, - vlib_cli_command_t * cmd) + unformat_input_t * input, + vlib_cli_command_t * cmd) { - vnet_main_t * vnm = vnet_get_main(); + vnet_main_t *vnm = vnet_get_main (); u32 sw_if_index = ~0; u32 ip4_table_index = ~0; u32 ip6_table_index = ~0; @@ -107,16 +110,16 @@ set_flow_classify_command_fn (vlib_main_t * vm, while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) { if (unformat (input, "interface %U", unformat_vnet_sw_interface, - vnm, &sw_if_index)) - ; + vnm, &sw_if_index)) + ; else if (unformat (input, "ip4-table %d", &ip4_table_index)) - idx_cnt++; + idx_cnt++; else if (unformat (input, "ip6-table %d", &ip6_table_index)) - idx_cnt++; + idx_cnt++; else if (unformat (input, "del")) - is_add = 0; + is_add = 0; else - break; + break; } if (sw_if_index == ~0) @@ -128,8 +131,8 @@ set_flow_classify_command_fn (vlib_main_t * vm, if (idx_cnt > 1) return clib_error_return (0, "Only one table index per API is allowed."); - rv = vnet_set_flow_classify_intfc(vm, sw_if_index, ip4_table_index, - ip6_table_index, is_add); + rv = vnet_set_flow_classify_intfc (vm, sw_if_index, ip4_table_index, + ip6_table_index, is_add); switch (rv) { @@ -145,6 +148,7 @@ set_flow_classify_command_fn (vlib_main_t * vm, return 0; } +/* *INDENT-OFF* */ VLIB_CLI_COMMAND (set_input_acl_command, static) = { .path = "set flow classify", .short_help = @@ -152,11 +156,12 @@ VLIB_CLI_COMMAND (set_input_acl_command, static) = { " [ip6-table ] [del]", .function = set_flow_classify_command_fn, }; +/* *INDENT-ON* */ static uword unformat_table_type (unformat_input_t * input, va_list * va) { - u32 * r = va_arg (*va, u32 *); + u32 *r = va_arg (*va, u32 *); u32 tid; if (unformat (input, "ip4")) @@ -169,14 +174,15 @@ unformat_table_type (unformat_input_t * input, va_list * va) *r = tid; return 1; } + static clib_error_t * show_flow_classify_command_fn (vlib_main_t * vm, - unformat_input_t * input, - vlib_cli_command_t * cmd) + unformat_input_t * input, + vlib_cli_command_t * cmd) { - flow_classify_main_t * fcm = &flow_classify_main; + flow_classify_main_t *fcm = &flow_classify_main; u32 type = FLOW_CLASSIFY_N_TABLES; - u32 * vec_tbl; + u32 *vec_tbl; int i; if (unformat (input, "type %U", unformat_table_type, &type)) @@ -189,26 +195,36 @@ show_flow_classify_command_fn (vlib_main_t * vm, vec_tbl = fcm->classify_table_index_by_sw_if_index[type]; - if (vec_len(vec_tbl)) - vlib_cli_output (vm, "%10s%20s\t\t%s", "Intfc idx", "Classify table", - "Interface name"); + if (vec_len (vec_tbl)) + vlib_cli_output (vm, "%10s%20s\t\t%s", "Intfc idx", "Classify table", + "Interface name"); else vlib_cli_output (vm, "No tables configured."); for (i = 0; i < vec_len (vec_tbl); i++) { - if (vec_elt(vec_tbl, i) == ~0) - continue; + if (vec_elt (vec_tbl, i) == ~0) + continue; - vlib_cli_output (vm, "%10d%20d\t\t%U", i, vec_elt(vec_tbl, i), - format_vnet_sw_if_index_name, fcm->vnet_main, i); + vlib_cli_output (vm, "%10d%20d\t\t%U", i, vec_elt (vec_tbl, i), + format_vnet_sw_if_index_name, fcm->vnet_main, i); } return 0; } +/* *INDENT-OFF* */ VLIB_CLI_COMMAND (show_flow_classify_command, static) = { .path = "show classify flow", .short_help = "show classify flow type [ip4|ip6]", .function = show_flow_classify_command_fn, }; +/* *INDENT-ON* */ + +/* + * fd.io coding-style-patch-verification: ON + * + * Local Variables: + * eval: (c-set-style "gnu") + * End: + */ diff --git a/src/vnet/classify/flow_classify.h b/src/vnet/classify/flow_classify.h index 14430df392f..e3308f3a302 100644 --- a/src/vnet/classify/flow_classify.h +++ b/src/vnet/classify/flow_classify.h @@ -20,32 +20,43 @@ #include #include -typedef enum { +typedef enum +{ FLOW_CLASSIFY_TABLE_IP4, FLOW_CLASSIFY_TABLE_IP6, FLOW_CLASSIFY_N_TABLES, } flow_classify_table_id_t; -typedef enum { +typedef enum +{ FLOW_CLASSIFY_NEXT_INDEX_DROP, FLOW_CLASSIFY_NEXT_INDEX_N_NEXT, } flow_classify_next_index_t; -typedef struct { +typedef struct +{ /* Classifier table vectors */ - u32 * classify_table_index_by_sw_if_index [FLOW_CLASSIFY_N_TABLES]; + u32 *classify_table_index_by_sw_if_index[FLOW_CLASSIFY_N_TABLES]; /* Convenience variables */ - vlib_main_t * vlib_main; - vnet_main_t * vnet_main; - vnet_classify_main_t * vnet_classify_main; - vnet_config_main_t * vnet_config_main [FLOW_CLASSIFY_N_TABLES]; + vlib_main_t *vlib_main; + vnet_main_t *vnet_main; + vnet_classify_main_t *vnet_classify_main; + vnet_config_main_t *vnet_config_main[FLOW_CLASSIFY_N_TABLES]; } flow_classify_main_t; extern flow_classify_main_t flow_classify_main; int vnet_set_flow_classify_intfc (vlib_main_t * vm, u32 sw_if_index, - u32 ip4_table_index, u32 ip6_table_index, - u32 is_add); + u32 ip4_table_index, u32 ip6_table_index, + u32 is_add); #endif /* __included_vnet_flow_classify_h__ */ + +/* + * fd.io coding-style-patch-verification: ON + * + * Local Variables: + * eval: (c-set-style "gnu") + * End: + */ diff --git a/src/vnet/classify/flow_classify_node.c b/src/vnet/classify/flow_classify_node.c index d3261d33bca..5941b6dd9b3 100644 --- a/src/vnet/classify/flow_classify_node.c +++ b/src/vnet/classify/flow_classify_node.c @@ -21,7 +21,8 @@ #include #include -typedef struct { +typedef struct +{ u32 sw_if_index; u32 next_index; u32 table_index; @@ -33,10 +34,10 @@ format_flow_classify_trace (u8 * s, va_list * args) { CLIB_UNUSED (vlib_main_t * vm) = va_arg (*args, vlib_main_t *); CLIB_UNUSED (vlib_node_t * node) = va_arg (*args, vlib_node_t *); - flow_classify_trace_t * t = va_arg (*args, flow_classify_trace_t *); + flow_classify_trace_t *t = va_arg (*args, flow_classify_trace_t *); s = format (s, "FLOW_CLASSIFY: sw_if_index %d next %d table %d offset %d", - t->sw_if_index, t->next_index, t->table_index, t->offset); + t->sw_if_index, t->next_index, t->table_index, t->offset); return s; } @@ -46,14 +47,15 @@ _(HIT, "Flow classify hits") \ _(CHAIN_HIT, "Flow classify hits after chain walk") \ _(DROP, "Flow classify action drop") -typedef enum { +typedef enum +{ #define _(sym,str) FLOW_CLASSIFY_ERROR_##sym, foreach_flow_classify_error #undef _ - FLOW_CLASSIFY_N_ERROR, + FLOW_CLASSIFY_N_ERROR, } flow_classify_error_t; -static char * flow_classify_error_strings[] = { +static char *flow_classify_error_strings[] = { #define _(sym,string) string, foreach_flow_classify_error #undef _ @@ -61,14 +63,13 @@ static char * flow_classify_error_strings[] = { static inline uword flow_classify_inline (vlib_main_t * vm, - vlib_node_runtime_t * node, - vlib_frame_t * frame, - flow_classify_table_id_t tid) + vlib_node_runtime_t * node, + vlib_frame_t * frame, flow_classify_table_id_t tid) { - u32 n_left_from, * from, * to_next; + u32 n_left_from, *from, *to_next; flow_classify_next_index_t next_index; - flow_classify_main_t * fcm = &flow_classify_main; - vnet_classify_main_t * vcm = fcm->vnet_classify_main; + flow_classify_main_t *fcm = &flow_classify_main; + vnet_classify_main_t *vcm = fcm->vnet_classify_main; f64 now = vlib_time_now (vm); u32 hits = 0; u32 misses = 0; @@ -81,24 +82,24 @@ flow_classify_inline (vlib_main_t * vm, /* First pass: compute hashes */ while (n_left_from > 2) { - vlib_buffer_t * b0, * b1; + vlib_buffer_t *b0, *b1; u32 bi0, bi1; - u8 * h0, * h1; + u8 *h0, *h1; u32 sw_if_index0, sw_if_index1; u32 table_index0, table_index1; - vnet_classify_table_t * t0, * t1; + vnet_classify_table_t *t0, *t1; /* Prefetch next iteration */ { - vlib_buffer_t * p1, * p2; + vlib_buffer_t *p1, *p2; - p1 = vlib_get_buffer (vm, from[1]); - p2 = vlib_get_buffer (vm, from[2]); + p1 = vlib_get_buffer (vm, from[1]); + p2 = vlib_get_buffer (vm, from[2]); - vlib_prefetch_buffer_header (p1, STORE); - CLIB_PREFETCH (p1->data, CLIB_CACHE_LINE_BYTES, STORE); - vlib_prefetch_buffer_header (p2, STORE); - CLIB_PREFETCH (p2->data, CLIB_CACHE_LINE_BYTES, STORE); + vlib_prefetch_buffer_header (p1, STORE); + CLIB_PREFETCH (p1->data, CLIB_CACHE_LINE_BYTES, STORE); + vlib_prefetch_buffer_header (p2, STORE); + CLIB_PREFETCH (p2->data, CLIB_CACHE_LINE_BYTES, STORE); } bi0 = from[0]; @@ -110,28 +111,30 @@ flow_classify_inline (vlib_main_t * vm, h1 = b1->data; sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX]; - table_index0 = fcm->classify_table_index_by_sw_if_index[tid][sw_if_index0]; + table_index0 = + fcm->classify_table_index_by_sw_if_index[tid][sw_if_index0]; sw_if_index1 = vnet_buffer (b1)->sw_if_index[VLIB_RX]; - table_index1 = fcm->classify_table_index_by_sw_if_index[tid][sw_if_index1]; + table_index1 = + fcm->classify_table_index_by_sw_if_index[tid][sw_if_index1]; t0 = pool_elt_at_index (vcm->tables, table_index0); t1 = pool_elt_at_index (vcm->tables, table_index1); - vnet_buffer(b0)->l2_classify.hash = - vnet_classify_hash_packet (t0, (u8 *) h0); + vnet_buffer (b0)->l2_classify.hash = + vnet_classify_hash_packet (t0, (u8 *) h0); - vnet_classify_prefetch_bucket (t0, vnet_buffer(b0)->l2_classify.hash); + vnet_classify_prefetch_bucket (t0, vnet_buffer (b0)->l2_classify.hash); - vnet_buffer(b1)->l2_classify.hash = - vnet_classify_hash_packet (t1, (u8 *) h1); + vnet_buffer (b1)->l2_classify.hash = + vnet_classify_hash_packet (t1, (u8 *) h1); - vnet_classify_prefetch_bucket (t1, vnet_buffer(b1)->l2_classify.hash); + vnet_classify_prefetch_bucket (t1, vnet_buffer (b1)->l2_classify.hash); - vnet_buffer(b0)->l2_classify.table_index = table_index0; + vnet_buffer (b0)->l2_classify.table_index = table_index0; - vnet_buffer(b1)->l2_classify.table_index = table_index1; + vnet_buffer (b1)->l2_classify.table_index = table_index1; from += 2; n_left_from -= 2; @@ -139,26 +142,27 @@ flow_classify_inline (vlib_main_t * vm, while (n_left_from > 0) { - vlib_buffer_t * b0; + vlib_buffer_t *b0; u32 bi0; - u8 * h0; + u8 *h0; u32 sw_if_index0; u32 table_index0; - vnet_classify_table_t * t0; + vnet_classify_table_t *t0; bi0 = from[0]; b0 = vlib_get_buffer (vm, bi0); h0 = b0->data; sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX]; - table_index0 = fcm->classify_table_index_by_sw_if_index[tid][sw_if_index0]; + table_index0 = + fcm->classify_table_index_by_sw_if_index[tid][sw_if_index0]; t0 = pool_elt_at_index (vcm->tables, table_index0); - vnet_buffer(b0)->l2_classify.hash = - vnet_classify_hash_packet (t0, (u8 *) h0); + vnet_buffer (b0)->l2_classify.hash = + vnet_classify_hash_packet (t0, (u8 *) h0); - vnet_buffer(b0)->l2_classify.table_index = table_index0; - vnet_classify_prefetch_bucket (t0, vnet_buffer(b0)->l2_classify.hash); + vnet_buffer (b0)->l2_classify.table_index = table_index0; + vnet_classify_prefetch_bucket (t0, vnet_buffer (b0)->l2_classify.hash); from++; n_left_from--; @@ -176,114 +180,109 @@ flow_classify_inline (vlib_main_t * vm, /* Not enough load/store slots to dual loop... */ while (n_left_from > 0 && n_left_to_next > 0) - { - u32 bi0; - vlib_buffer_t * b0; - u32 next0 = FLOW_CLASSIFY_NEXT_INDEX_DROP; - u32 table_index0; - vnet_classify_table_t * t0; - vnet_classify_entry_t * e0; - u64 hash0; - u8 * h0; - - /* Stride 3 seems to work best */ - if (PREDICT_TRUE (n_left_from > 3)) - { - vlib_buffer_t * p1 = vlib_get_buffer(vm, from[3]); - vnet_classify_table_t * tp1; - u32 table_index1; - u64 phash1; - - table_index1 = vnet_buffer(p1)->l2_classify.table_index; - - if (PREDICT_TRUE (table_index1 != ~0)) - { - tp1 = pool_elt_at_index (vcm->tables, table_index1); - phash1 = vnet_buffer(p1)->l2_classify.hash; - vnet_classify_prefetch_entry (tp1, phash1); - } - } - - /* Speculatively enqueue b0 to the current next frame */ - bi0 = from[0]; - to_next[0] = bi0; - from += 1; - to_next += 1; - n_left_from -= 1; - n_left_to_next -= 1; - - b0 = vlib_get_buffer (vm, bi0); - h0 = b0->data; - table_index0 = vnet_buffer(b0)->l2_classify.table_index; - e0 = 0; - t0 = 0; - - vnet_get_config_data (fcm->vnet_config_main[tid], - &b0->current_config_index, - &next0, - /* # bytes of config data */ 0); - - if (PREDICT_TRUE(table_index0 != ~0)) - { - hash0 = vnet_buffer(b0)->l2_classify.hash; - t0 = pool_elt_at_index (vcm->tables, table_index0); - e0 = vnet_classify_find_entry (t0, (u8 *) h0, hash0, now); - if (e0) - { - hits++; - } - else - { - misses++; - vnet_classify_add_del_session (vcm, table_index0, - h0, ~0, 0, 0, 0, 0, 1); - /* increment counter */ - vnet_classify_find_entry (t0, (u8 *) h0, hash0, now); - } - } - if (PREDICT_FALSE((node->flags & VLIB_NODE_FLAG_TRACE) - && (b0->flags & VLIB_BUFFER_IS_TRACED))) - { - flow_classify_trace_t * t = - vlib_add_trace (vm, node, b0, sizeof (*t)); - t->sw_if_index = vnet_buffer(b0)->sw_if_index[VLIB_RX]; - t->next_index = next0; - t->table_index = t0 ? t0 - vcm->tables : ~0; - t->offset = (t0 && e0) ? vnet_classify_get_offset (t0, e0): ~0; - } - - /* Verify speculative enqueue, maybe switch current next frame */ - vlib_validate_buffer_enqueue_x1 (vm, node, next_index, to_next, - n_left_to_next, bi0, next0); - } + { + u32 bi0; + vlib_buffer_t *b0; + u32 next0 = FLOW_CLASSIFY_NEXT_INDEX_DROP; + u32 table_index0; + vnet_classify_table_t *t0; + vnet_classify_entry_t *e0; + u64 hash0; + u8 *h0; + + /* Stride 3 seems to work best */ + if (PREDICT_TRUE (n_left_from > 3)) + { + vlib_buffer_t *p1 = vlib_get_buffer (vm, from[3]); + vnet_classify_table_t *tp1; + u32 table_index1; + u64 phash1; + + table_index1 = vnet_buffer (p1)->l2_classify.table_index; + + if (PREDICT_TRUE (table_index1 != ~0)) + { + tp1 = pool_elt_at_index (vcm->tables, table_index1); + phash1 = vnet_buffer (p1)->l2_classify.hash; + vnet_classify_prefetch_entry (tp1, phash1); + } + } + + /* Speculatively enqueue b0 to the current next frame */ + bi0 = from[0]; + to_next[0] = bi0; + from += 1; + to_next += 1; + n_left_from -= 1; + n_left_to_next -= 1; + + b0 = vlib_get_buffer (vm, bi0); + h0 = b0->data; + table_index0 = vnet_buffer (b0)->l2_classify.table_index; + e0 = 0; + t0 = 0; + + vnet_get_config_data (fcm->vnet_config_main[tid], + &b0->current_config_index, &next0, + /* # bytes of config data */ 0); + + if (PREDICT_TRUE (table_index0 != ~0)) + { + hash0 = vnet_buffer (b0)->l2_classify.hash; + t0 = pool_elt_at_index (vcm->tables, table_index0); + e0 = vnet_classify_find_entry (t0, (u8 *) h0, hash0, now); + if (e0) + { + hits++; + } + else + { + misses++; + vnet_classify_add_del_session (vcm, table_index0, + h0, ~0, 0, 0, 0, 0, 1); + /* increment counter */ + vnet_classify_find_entry (t0, (u8 *) h0, hash0, now); + } + } + if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE) + && (b0->flags & VLIB_BUFFER_IS_TRACED))) + { + flow_classify_trace_t *t = + vlib_add_trace (vm, node, b0, sizeof (*t)); + t->sw_if_index = vnet_buffer (b0)->sw_if_index[VLIB_RX]; + t->next_index = next0; + t->table_index = t0 ? t0 - vcm->tables : ~0; + t->offset = (t0 && e0) ? vnet_classify_get_offset (t0, e0) : ~0; + } + + /* Verify speculative enqueue, maybe switch current next frame */ + vlib_validate_buffer_enqueue_x1 (vm, node, next_index, to_next, + n_left_to_next, bi0, next0); + } vlib_put_next_frame (vm, node, next_index, n_left_to_next); } vlib_node_increment_counter (vm, node->node_index, - FLOW_CLASSIFY_ERROR_MISS, - misses); + FLOW_CLASSIFY_ERROR_MISS, misses); vlib_node_increment_counter (vm, node->node_index, - FLOW_CLASSIFY_ERROR_HIT, - hits); + FLOW_CLASSIFY_ERROR_HIT, hits); vlib_node_increment_counter (vm, node->node_index, - FLOW_CLASSIFY_ERROR_CHAIN_HIT, - chain_hits); + FLOW_CLASSIFY_ERROR_CHAIN_HIT, chain_hits); vlib_node_increment_counter (vm, node->node_index, - FLOW_CLASSIFY_ERROR_DROP, - drop); + FLOW_CLASSIFY_ERROR_DROP, drop); return frame->n_vectors; } static uword ip4_flow_classify (vlib_main_t * vm, - vlib_node_runtime_t * node, - vlib_frame_t * frame) + vlib_node_runtime_t * node, vlib_frame_t * frame) { - return flow_classify_inline(vm, node, frame, FLOW_CLASSIFY_TABLE_IP4); + return flow_classify_inline (vm, node, frame, FLOW_CLASSIFY_TABLE_IP4); } +/* *INDENT-OFF* */ VLIB_REGISTER_NODE (ip4_flow_classify_node) = { .function = ip4_flow_classify, .name = "ip4-flow-classify", @@ -296,17 +295,18 @@ VLIB_REGISTER_NODE (ip4_flow_classify_node) = { [FLOW_CLASSIFY_NEXT_INDEX_DROP] = "error-drop", }, }; +/* *INDENT-ON* */ VLIB_NODE_FUNCTION_MULTIARCH (ip4_flow_classify_node, ip4_flow_classify); static uword ip6_flow_classify (vlib_main_t * vm, - vlib_node_runtime_t * node, - vlib_frame_t * frame) + vlib_node_runtime_t * node, vlib_frame_t * frame) { - return flow_classify_inline(vm, node, frame, FLOW_CLASSIFY_TABLE_IP6); + return flow_classify_inline (vm, node, frame, FLOW_CLASSIFY_TABLE_IP6); } +/* *INDENT-OFF* */ VLIB_REGISTER_NODE (ip6_flow_classify_node) = { .function = ip6_flow_classify, .name = "ip6-flow-classify", @@ -320,19 +320,29 @@ VLIB_REGISTER_NODE (ip6_flow_classify_node) = { }, }; +/* *INDENT-ON* */ + VLIB_NODE_FUNCTION_MULTIARCH (ip6_flow_classify_node, ip6_flow_classify); static clib_error_t * -flow_classify_init (vlib_main_t *vm) +flow_classify_init (vlib_main_t * vm) { - flow_classify_main_t * fcm = &flow_classify_main; + flow_classify_main_t *fcm = &flow_classify_main; fcm->vlib_main = vm; - fcm->vnet_main = vnet_get_main(); + fcm->vnet_main = vnet_get_main (); fcm->vnet_classify_main = &vnet_classify_main; return 0; } VLIB_INIT_FUNCTION (flow_classify_init); + +/* + * fd.io coding-style-patch-verification: ON + * + * Local Variables: + * eval: (c-set-style "gnu") + * End: + */ diff --git a/src/vnet/classify/input_acl.c b/src/vnet/classify/input_acl.c index c446f2d687c..cf5053ef3ac 100644 --- a/src/vnet/classify/input_acl.c +++ b/src/vnet/classify/input_acl.c @@ -20,19 +20,18 @@ input_acl_main_t input_acl_main; static int vnet_inacl_ip_feature_enable (vlib_main_t * vnm, - input_acl_main_t *am, - u32 sw_if_index, - input_acl_table_id_t tid, - int feature_enable) + input_acl_main_t * am, + u32 sw_if_index, + input_acl_table_id_t tid, int feature_enable) { if (tid == INPUT_ACL_TABLE_L2) { l2input_intf_bitmap_enable (sw_if_index, L2INPUT_FEAT_ACL, - feature_enable); + feature_enable); } else - { /* IP[46] */ + { /* IP[46] */ vnet_feature_config_main_t *fcm; u8 arc; @@ -56,15 +55,16 @@ vnet_inacl_ip_feature_enable (vlib_main_t * vnm, return 0; } -int vnet_set_input_acl_intfc (vlib_main_t * vm, u32 sw_if_index, - u32 ip4_table_index, - u32 ip6_table_index, - u32 l2_table_index, u32 is_add) +int +vnet_set_input_acl_intfc (vlib_main_t * vm, u32 sw_if_index, + u32 ip4_table_index, + u32 ip6_table_index, u32 l2_table_index, u32 is_add) { - input_acl_main_t * am = &input_acl_main; - vnet_classify_main_t * vcm = am->vnet_classify_main; - u32 acl[INPUT_ACL_N_TABLES] = {ip4_table_index, ip6_table_index, - l2_table_index}; + input_acl_main_t *am = &input_acl_main; + vnet_classify_main_t *vcm = am->vnet_classify_main; + u32 acl[INPUT_ACL_N_TABLES] = { ip4_table_index, ip6_table_index, + l2_table_index + }; u32 ti; /* Assume that we've validated sw_if_index in the API layer */ @@ -72,34 +72,36 @@ int vnet_set_input_acl_intfc (vlib_main_t * vm, u32 sw_if_index, for (ti = 0; ti < INPUT_ACL_N_TABLES; ti++) { if (acl[ti] == ~0) - continue; + continue; if (pool_is_free_index (vcm->tables, acl[ti])) - return VNET_API_ERROR_NO_SUCH_TABLE; + return VNET_API_ERROR_NO_SUCH_TABLE; vec_validate_init_empty - (am->classify_table_index_by_sw_if_index[ti], sw_if_index, ~0); + (am->classify_table_index_by_sw_if_index[ti], sw_if_index, ~0); /* Reject any DEL operation with wrong sw_if_index */ if (!is_add && - (acl[ti] != am->classify_table_index_by_sw_if_index[ti][sw_if_index])) - { - clib_warning ("Non-existent intf_idx=%d with table_index=%d for delete", - sw_if_index, acl[ti]); - return VNET_API_ERROR_NO_SUCH_TABLE; - } + (acl[ti] != + am->classify_table_index_by_sw_if_index[ti][sw_if_index])) + { + clib_warning + ("Non-existent intf_idx=%d with table_index=%d for delete", + sw_if_index, acl[ti]); + return VNET_API_ERROR_NO_SUCH_TABLE; + } /* Return ok on ADD operaton if feature is already enabled */ if (is_add && - am->classify_table_index_by_sw_if_index[ti][sw_if_index] != ~0) - return 0; + am->classify_table_index_by_sw_if_index[ti][sw_if_index] != ~0) + return 0; vnet_inacl_ip_feature_enable (vm, am, sw_if_index, ti, is_add); if (is_add) - am->classify_table_index_by_sw_if_index[ti][sw_if_index] = acl[ti]; + am->classify_table_index_by_sw_if_index[ti][sw_if_index] = acl[ti]; else - am->classify_table_index_by_sw_if_index[ti][sw_if_index] = ~0; + am->classify_table_index_by_sw_if_index[ti][sw_if_index] = ~0; } return 0; @@ -107,10 +109,9 @@ int vnet_set_input_acl_intfc (vlib_main_t * vm, u32 sw_if_index, static clib_error_t * set_input_acl_command_fn (vlib_main_t * vm, - unformat_input_t * input, - vlib_cli_command_t * cmd) + unformat_input_t * input, vlib_cli_command_t * cmd) { - vnet_main_t * vnm = vnet_get_main(); + vnet_main_t *vnm = vnet_get_main (); u32 sw_if_index = ~0; u32 ip4_table_index = ~0; u32 ip6_table_index = ~0; @@ -122,18 +123,18 @@ set_input_acl_command_fn (vlib_main_t * vm, while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) { if (unformat (input, "intfc %U", unformat_vnet_sw_interface, - vnm, &sw_if_index)) - ; + vnm, &sw_if_index)) + ; else if (unformat (input, "ip4-table %d", &ip4_table_index)) - idx_cnt++; + idx_cnt++; else if (unformat (input, "ip6-table %d", &ip6_table_index)) - idx_cnt++; + idx_cnt++; else if (unformat (input, "l2-table %d", &l2_table_index)) - idx_cnt++; + idx_cnt++; else if (unformat (input, "del")) - is_add = 0; + is_add = 0; else - break; + break; } if (sw_if_index == ~0) @@ -146,7 +147,7 @@ set_input_acl_command_fn (vlib_main_t * vm, return clib_error_return (0, "Only one table index per API is allowed."); rv = vnet_set_input_acl_intfc (vm, sw_if_index, ip4_table_index, - ip6_table_index, l2_table_index, is_add); + ip6_table_index, l2_table_index, is_add); switch (rv) { @@ -172,6 +173,7 @@ set_input_acl_command_fn (vlib_main_t * vm, * Note: Only one table index per API call is allowed. * */ +/* *INDENT-OFF* */ VLIB_CLI_COMMAND (set_input_acl_command, static) = { .path = "set interface input acl", .short_help = @@ -179,17 +181,19 @@ VLIB_CLI_COMMAND (set_input_acl_command, static) = { " [ip6-table ] [l2-table ] [del]", .function = set_input_acl_command_fn, }; +/* *INDENT-ON* */ -clib_error_t *input_acl_init (vlib_main_t *vm) +clib_error_t * +input_acl_init (vlib_main_t * vm) { - input_acl_main_t * am = &input_acl_main; - clib_error_t * error = 0; + input_acl_main_t *am = &input_acl_main; + clib_error_t *error = 0; if ((error = vlib_call_init_function (vm, ip_inacl_init))) return error; am->vlib_main = vm; - am->vnet_main = vnet_get_main(); + am->vnet_main = vnet_get_main (); am->vnet_classify_main = &vnet_classify_main; return 0; @@ -197,61 +201,63 @@ clib_error_t *input_acl_init (vlib_main_t *vm) VLIB_INIT_FUNCTION (input_acl_init); -uword unformat_acl_type (unformat_input_t * input, va_list * args) +uword +unformat_acl_type (unformat_input_t * input, va_list * args) { - u32 * acl_type = va_arg (*args, u32 *); + u32 *acl_type = va_arg (*args, u32 *); u32 tid = INPUT_ACL_N_TABLES; - while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) { - if (unformat (input, "ip4")) - tid = INPUT_ACL_TABLE_IP4; - else if (unformat (input, "ip6")) - tid = INPUT_ACL_TABLE_IP6; - else if (unformat (input, "l2")) - tid = INPUT_ACL_TABLE_L2; - else - break; - } + while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) + { + if (unformat (input, "ip4")) + tid = INPUT_ACL_TABLE_IP4; + else if (unformat (input, "ip6")) + tid = INPUT_ACL_TABLE_IP6; + else if (unformat (input, "l2")) + tid = INPUT_ACL_TABLE_L2; + else + break; + } *acl_type = tid; return 1; } -u8 * format_vnet_inacl_info (u8 * s, va_list * va) +u8 * +format_vnet_inacl_info (u8 * s, va_list * va) { - input_acl_main_t * am = va_arg (*va, input_acl_main_t *); + input_acl_main_t *am = va_arg (*va, input_acl_main_t *); int sw_if_idx = va_arg (*va, int); u32 tid = va_arg (*va, u32); if (tid == ~0) { s = format (s, "%10s%20s\t\t%s", "Intfc idx", "Classify table", - "Interface name"); + "Interface name"); return s; } s = format (s, "%10d%20d\t\t%U", sw_if_idx, tid, - format_vnet_sw_if_index_name, am->vnet_main, sw_if_idx); + format_vnet_sw_if_index_name, am->vnet_main, sw_if_idx); return s; } static clib_error_t * show_inacl_command_fn (vlib_main_t * vm, - unformat_input_t * input, - vlib_cli_command_t * cmd) + unformat_input_t * input, vlib_cli_command_t * cmd) { - input_acl_main_t * am = &input_acl_main; + input_acl_main_t *am = &input_acl_main; u32 type = INPUT_ACL_N_TABLES; int i; - u32 * vec_tbl; + u32 *vec_tbl; while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) { if (unformat (input, "type %U", unformat_acl_type, &type)) - ; + ; else - break; + break; } if (type == INPUT_ACL_N_TABLES) @@ -259,25 +265,35 @@ show_inacl_command_fn (vlib_main_t * vm, vec_tbl = am->classify_table_index_by_sw_if_index[type]; - if (vec_len(vec_tbl)) - vlib_cli_output (vm, "%U", format_vnet_inacl_info, am, ~0 /* hdr */, ~0); + if (vec_len (vec_tbl)) + vlib_cli_output (vm, "%U", format_vnet_inacl_info, am, ~0 /* hdr */ , ~0); else vlib_cli_output (vm, "No input ACL tables configured"); for (i = 0; i < vec_len (vec_tbl); i++) { - if (vec_elt(vec_tbl, i) == ~0) - continue; + if (vec_elt (vec_tbl, i) == ~0) + continue; vlib_cli_output (vm, "%U", format_vnet_inacl_info, - am, i, vec_elt(vec_tbl, i)); + am, i, vec_elt (vec_tbl, i)); } return 0; } +/* *INDENT-OFF* */ VLIB_CLI_COMMAND (show_inacl_command, static) = { .path = "show inacl", .short_help = "show inacl type [ip4|ip6|l2]", .function = show_inacl_command_fn, }; +/* *INDENT-ON* */ + +/* + * fd.io coding-style-patch-verification: ON + * + * Local Variables: + * eval: (c-set-style "gnu") + * End: + */ diff --git a/src/vnet/classify/input_acl.h b/src/vnet/classify/input_acl.h index 7ffc189f053..a5f3bac6b99 100644 --- a/src/vnet/classify/input_acl.h +++ b/src/vnet/classify/input_acl.h @@ -20,35 +20,46 @@ #include #include -typedef enum { +typedef enum +{ INPUT_ACL_TABLE_IP4, INPUT_ACL_TABLE_IP6, INPUT_ACL_TABLE_L2, INPUT_ACL_N_TABLES, } input_acl_table_id_t; -typedef enum { +typedef enum +{ ACL_NEXT_INDEX_DENY, ACL_NEXT_INDEX_N_NEXT, } acl_next_index_t; -typedef struct { +typedef struct +{ /* classifier table vectors */ - u32 * classify_table_index_by_sw_if_index [INPUT_ACL_N_TABLES]; + u32 *classify_table_index_by_sw_if_index[INPUT_ACL_N_TABLES]; /* convenience variables */ - vlib_main_t * vlib_main; - vnet_main_t * vnet_main; - vnet_classify_main_t * vnet_classify_main; - vnet_config_main_t * vnet_config_main [INPUT_ACL_N_TABLES]; + vlib_main_t *vlib_main; + vnet_main_t *vnet_main; + vnet_classify_main_t *vnet_classify_main; + vnet_config_main_t *vnet_config_main[INPUT_ACL_N_TABLES]; } input_acl_main_t; extern input_acl_main_t input_acl_main; int vnet_set_input_acl_intfc (vlib_main_t * vm, u32 sw_if_index, - u32 ip4_table_index, - u32 ip6_table_index, - u32 l2_table_index, u32 is_add); + u32 ip4_table_index, + u32 ip6_table_index, + u32 l2_table_index, u32 is_add); #endif /* __included_vnet_input_acl_h__ */ + +/* + * fd.io coding-style-patch-verification: ON + * + * Local Variables: + * eval: (c-set-style "gnu") + * End: + */ diff --git a/src/vnet/classify/ip_classify.c b/src/vnet/classify/ip_classify.c index 44973ae5e99..1a8e1dd41e5 100644 --- a/src/vnet/classify/ip_classify.c +++ b/src/vnet/classify/ip_classify.c @@ -17,21 +17,23 @@ #include #include -typedef struct { +typedef struct +{ u32 next_index; u32 table_index; u32 entry_index; } ip_classify_trace_t; /* packet trace format function */ -static u8 * format_ip_classify_trace (u8 * s, va_list * args) +static u8 * +format_ip_classify_trace (u8 * s, va_list * args) { CLIB_UNUSED (vlib_main_t * vm) = va_arg (*args, vlib_main_t *); CLIB_UNUSED (vlib_node_t * node) = va_arg (*args, vlib_node_t *); - ip_classify_trace_t * t = va_arg (*args, ip_classify_trace_t *); - + ip_classify_trace_t *t = va_arg (*args, ip_classify_trace_t *); + s = format (s, "IP_CLASSIFY: next_index %d, table %d, entry %d", - t->next_index, t->table_index, t->entry_index); + t->next_index, t->table_index, t->entry_index); return s; } @@ -43,14 +45,15 @@ _(MISS, "Classify misses") \ _(HIT, "Classify hits") \ _(CHAIN_HIT, "Classify hits after chain walk") -typedef enum { +typedef enum +{ #define _(sym,str) IP_CLASSIFY_ERROR_##sym, foreach_ip_classify_error #undef _ - IP_CLASSIFY_N_ERROR, + IP_CLASSIFY_N_ERROR, } ip_classify_error_t; -static char * ip_classify_error_strings[] = { +static char *ip_classify_error_strings[] = { #define _(sym,string) string, foreach_ip_classify_error #undef _ @@ -58,23 +61,26 @@ static char * ip_classify_error_strings[] = { static inline uword ip_classify_inline (vlib_main_t * vm, - vlib_node_runtime_t * node, - vlib_frame_t * frame, int is_ip4) + vlib_node_runtime_t * node, + vlib_frame_t * frame, int is_ip4) { - u32 n_left_from, * from, * to_next; + u32 n_left_from, *from, *to_next; ip_lookup_next_t next_index; - vnet_classify_main_t * vcm = &vnet_classify_main; + vnet_classify_main_t *vcm = &vnet_classify_main; f64 now = vlib_time_now (vm); u32 hits = 0; u32 misses = 0; u32 chain_hits = 0; u32 n_next; - if (is_ip4) { - n_next = IP4_LOOKUP_N_NEXT; - } else { - n_next = IP6_LOOKUP_N_NEXT; - } + if (is_ip4) + { + n_next = IP4_LOOKUP_N_NEXT; + } + else + { + n_next = IP6_LOOKUP_N_NEXT; + } from = vlib_frame_vector_args (frame); n_left_from = frame->n_vectors; @@ -83,62 +89,62 @@ ip_classify_inline (vlib_main_t * vm, while (n_left_from > 2) { - vlib_buffer_t * b0, * b1; + vlib_buffer_t *b0, *b1; u32 bi0, bi1; - u8 * h0, * h1; + u8 *h0, *h1; u32 cd_index0, cd_index1; - classify_dpo_t *cd0, * cd1; + classify_dpo_t *cd0, *cd1; u32 table_index0, table_index1; - vnet_classify_table_t * t0, * t1; + vnet_classify_table_t *t0, *t1; /* prefetch next iteration */ - { - vlib_buffer_t * p1, * p2; - - p1 = vlib_get_buffer (vm, from[1]); - p2 = vlib_get_buffer (vm, from[2]); - - vlib_prefetch_buffer_header (p1, STORE); - CLIB_PREFETCH (p1->data, CLIB_CACHE_LINE_BYTES, STORE); - vlib_prefetch_buffer_header (p2, STORE); - CLIB_PREFETCH (p2->data, CLIB_CACHE_LINE_BYTES, STORE); - } - + { + vlib_buffer_t *p1, *p2; + + p1 = vlib_get_buffer (vm, from[1]); + p2 = vlib_get_buffer (vm, from[2]); + + vlib_prefetch_buffer_header (p1, STORE); + CLIB_PREFETCH (p1->data, CLIB_CACHE_LINE_BYTES, STORE); + vlib_prefetch_buffer_header (p2, STORE); + CLIB_PREFETCH (p2->data, CLIB_CACHE_LINE_BYTES, STORE); + } + bi0 = from[0]; b0 = vlib_get_buffer (vm, bi0); - h0 = (void *)vlib_buffer_get_current(b0) - - ethernet_buffer_header_size(b0); + h0 = (void *) vlib_buffer_get_current (b0) - + ethernet_buffer_header_size (b0); bi1 = from[1]; b1 = vlib_get_buffer (vm, bi1); - h1 = (void *)vlib_buffer_get_current(b1) - - ethernet_buffer_header_size(b1); - + h1 = (void *) vlib_buffer_get_current (b1) - + ethernet_buffer_header_size (b1); + cd_index0 = vnet_buffer (b0)->ip.adj_index[VLIB_TX]; - cd0 = classify_dpo_get(cd_index0); + cd0 = classify_dpo_get (cd_index0); table_index0 = cd0->cd_table_index; cd_index1 = vnet_buffer (b1)->ip.adj_index[VLIB_TX]; - cd1 = classify_dpo_get(cd_index1); + cd1 = classify_dpo_get (cd_index1); table_index1 = cd1->cd_table_index; t0 = pool_elt_at_index (vcm->tables, table_index0); t1 = pool_elt_at_index (vcm->tables, table_index1); - - vnet_buffer(b0)->l2_classify.hash = - vnet_classify_hash_packet (t0, (u8 *) h0); - vnet_classify_prefetch_bucket (t0, vnet_buffer(b0)->l2_classify.hash); + vnet_buffer (b0)->l2_classify.hash = + vnet_classify_hash_packet (t0, (u8 *) h0); - vnet_buffer(b1)->l2_classify.hash = - vnet_classify_hash_packet (t1, (u8 *) h1); + vnet_classify_prefetch_bucket (t0, vnet_buffer (b0)->l2_classify.hash); - vnet_classify_prefetch_bucket (t1, vnet_buffer(b1)->l2_classify.hash); + vnet_buffer (b1)->l2_classify.hash = + vnet_classify_hash_packet (t1, (u8 *) h1); - vnet_buffer(b0)->l2_classify.table_index = table_index0; + vnet_classify_prefetch_bucket (t1, vnet_buffer (b1)->l2_classify.hash); - vnet_buffer(b1)->l2_classify.table_index = table_index1; + vnet_buffer (b0)->l2_classify.table_index = table_index0; + + vnet_buffer (b1)->l2_classify.table_index = table_index1; from += 2; n_left_from -= 2; @@ -146,34 +152,34 @@ ip_classify_inline (vlib_main_t * vm, while (n_left_from > 0) { - vlib_buffer_t * b0; + vlib_buffer_t *b0; u32 bi0; - u8 * h0; + u8 *h0; u32 cd_index0; classify_dpo_t *cd0; u32 table_index0; - vnet_classify_table_t * t0; + vnet_classify_table_t *t0; bi0 = from[0]; b0 = vlib_get_buffer (vm, bi0); - h0 = (void *)vlib_buffer_get_current(b0) - - ethernet_buffer_header_size(b0); - + h0 = (void *) vlib_buffer_get_current (b0) - + ethernet_buffer_header_size (b0); + cd_index0 = vnet_buffer (b0)->ip.adj_index[VLIB_TX]; - cd0 = classify_dpo_get(cd_index0); + cd0 = classify_dpo_get (cd_index0); table_index0 = cd0->cd_table_index; t0 = pool_elt_at_index (vcm->tables, table_index0); - vnet_buffer(b0)->l2_classify.hash = - vnet_classify_hash_packet (t0, (u8 *) h0); + vnet_buffer (b0)->l2_classify.hash = + vnet_classify_hash_packet (t0, (u8 *) h0); - vnet_buffer(b0)->l2_classify.table_index = table_index0; - vnet_classify_prefetch_bucket (t0, vnet_buffer(b0)->l2_classify.hash); + vnet_buffer (b0)->l2_classify.table_index = table_index0; + vnet_classify_prefetch_bucket (t0, vnet_buffer (b0)->l2_classify.hash); from++; n_left_from--; } - + next_index = node->cached_next_index; from = vlib_frame_vector_args (frame); n_left_from = frame->n_vectors; @@ -182,40 +188,39 @@ ip_classify_inline (vlib_main_t * vm, { u32 n_left_to_next; - vlib_get_next_frame (vm, node, next_index, - to_next, n_left_to_next); + vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next); /* Not enough load/store slots to dual loop... */ while (n_left_from > 0 && n_left_to_next > 0) { - u32 bi0; - vlib_buffer_t * b0; - u32 next0 = IP_LOOKUP_NEXT_DROP; - u32 table_index0; - vnet_classify_table_t * t0; - vnet_classify_entry_t * e0; - u64 hash0; - u8 * h0; - - /* Stride 3 seems to work best */ - if (PREDICT_TRUE (n_left_from > 3)) - { - vlib_buffer_t * p1 = vlib_get_buffer(vm, from[3]); - vnet_classify_table_t * tp1; - u32 table_index1; - u64 phash1; - - table_index1 = vnet_buffer(p1)->l2_classify.table_index; - - if (PREDICT_TRUE (table_index1 != ~0)) - { - tp1 = pool_elt_at_index (vcm->tables, table_index1); - phash1 = vnet_buffer(p1)->l2_classify.hash; - vnet_classify_prefetch_entry (tp1, phash1); - } - } - - /* speculatively enqueue b0 to the current next frame */ + u32 bi0; + vlib_buffer_t *b0; + u32 next0 = IP_LOOKUP_NEXT_DROP; + u32 table_index0; + vnet_classify_table_t *t0; + vnet_classify_entry_t *e0; + u64 hash0; + u8 *h0; + + /* Stride 3 seems to work best */ + if (PREDICT_TRUE (n_left_from > 3)) + { + vlib_buffer_t *p1 = vlib_get_buffer (vm, from[3]); + vnet_classify_table_t *tp1; + u32 table_index1; + u64 phash1; + + table_index1 = vnet_buffer (p1)->l2_classify.table_index; + + if (PREDICT_TRUE (table_index1 != ~0)) + { + tp1 = pool_elt_at_index (vcm->tables, table_index1); + phash1 = vnet_buffer (p1)->l2_classify.hash; + vnet_classify_prefetch_entry (tp1, phash1); + } + } + + /* speculatively enqueue b0 to the current next frame */ bi0 = from[0]; to_next[0] = bi0; from += 1; @@ -224,72 +229,71 @@ ip_classify_inline (vlib_main_t * vm, n_left_to_next -= 1; b0 = vlib_get_buffer (vm, bi0); - h0 = b0->data; - table_index0 = vnet_buffer(b0)->l2_classify.table_index; - e0 = 0; - t0 = 0; - vnet_buffer(b0)->l2_classify.opaque_index = ~0; - - if (PREDICT_TRUE(table_index0 != ~0)) - { - hash0 = vnet_buffer(b0)->l2_classify.hash; - t0 = pool_elt_at_index (vcm->tables, table_index0); - - e0 = vnet_classify_find_entry (t0, (u8 *) h0, hash0, - now); - if (e0) - { - vnet_buffer(b0)->l2_classify.opaque_index - = e0->opaque_index; - vlib_buffer_advance (b0, e0->advance); - next0 = (e0->next_index < node->n_next_nodes)? - e0->next_index:next0; - hits++; - } - else - { - while (1) - { - if (t0->next_table_index != ~0) - t0 = pool_elt_at_index (vcm->tables, - t0->next_table_index); - else - { - next0 = (t0->miss_next_index < n_next) ? - t0->miss_next_index : next0; - misses++; - break; - } - - hash0 = vnet_classify_hash_packet (t0, (u8 *) h0); - e0 = vnet_classify_find_entry - (t0, (u8 *) h0, hash0, now); - if (e0) - { - vnet_buffer(b0)->l2_classify.opaque_index - = e0->opaque_index; - vlib_buffer_advance (b0, e0->advance); - next0 = (e0->next_index < node->n_next_nodes)? - e0->next_index:next0; - hits++; - chain_hits++; - break; - } - } - } - } - - if (PREDICT_FALSE((node->flags & VLIB_NODE_FLAG_TRACE) - && (b0->flags & VLIB_BUFFER_IS_TRACED))) - { - ip_classify_trace_t *t = - vlib_add_trace (vm, node, b0, sizeof (*t)); - t->next_index = next0; - t->table_index = t0 ? t0 - vcm->tables : ~0; - t->entry_index = e0 ? e0 - t0->entries : ~0; - } - - /* verify speculative enqueue, maybe switch current next frame */ + h0 = b0->data; + table_index0 = vnet_buffer (b0)->l2_classify.table_index; + e0 = 0; + t0 = 0; + vnet_buffer (b0)->l2_classify.opaque_index = ~0; + + if (PREDICT_TRUE (table_index0 != ~0)) + { + hash0 = vnet_buffer (b0)->l2_classify.hash; + t0 = pool_elt_at_index (vcm->tables, table_index0); + + e0 = vnet_classify_find_entry (t0, (u8 *) h0, hash0, now); + if (e0) + { + vnet_buffer (b0)->l2_classify.opaque_index + = e0->opaque_index; + vlib_buffer_advance (b0, e0->advance); + next0 = (e0->next_index < node->n_next_nodes) ? + e0->next_index : next0; + hits++; + } + else + { + while (1) + { + if (t0->next_table_index != ~0) + t0 = pool_elt_at_index (vcm->tables, + t0->next_table_index); + else + { + next0 = (t0->miss_next_index < n_next) ? + t0->miss_next_index : next0; + misses++; + break; + } + + hash0 = vnet_classify_hash_packet (t0, (u8 *) h0); + e0 = vnet_classify_find_entry + (t0, (u8 *) h0, hash0, now); + if (e0) + { + vnet_buffer (b0)->l2_classify.opaque_index + = e0->opaque_index; + vlib_buffer_advance (b0, e0->advance); + next0 = (e0->next_index < node->n_next_nodes) ? + e0->next_index : next0; + hits++; + chain_hits++; + break; + } + } + } + } + + if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE) + && (b0->flags & VLIB_BUFFER_IS_TRACED))) + { + ip_classify_trace_t *t = + vlib_add_trace (vm, node, b0, sizeof (*t)); + t->next_index = next0; + t->table_index = t0 ? t0 - vcm->tables : ~0; + t->entry_index = e0 ? e0 - t0->entries : ~0; + } + + /* verify speculative enqueue, maybe switch current next frame */ vlib_validate_buffer_enqueue_x1 (vm, node, next_index, to_next, n_left_to_next, bi0, next0); @@ -298,27 +302,24 @@ ip_classify_inline (vlib_main_t * vm, vlib_put_next_frame (vm, node, next_index, n_left_to_next); } - vlib_node_increment_counter (vm, node->node_index, - IP_CLASSIFY_ERROR_MISS, - misses); - vlib_node_increment_counter (vm, node->node_index, - IP_CLASSIFY_ERROR_HIT, - hits); - vlib_node_increment_counter (vm, node->node_index, - IP_CLASSIFY_ERROR_CHAIN_HIT, - chain_hits); + vlib_node_increment_counter (vm, node->node_index, + IP_CLASSIFY_ERROR_MISS, misses); + vlib_node_increment_counter (vm, node->node_index, + IP_CLASSIFY_ERROR_HIT, hits); + vlib_node_increment_counter (vm, node->node_index, + IP_CLASSIFY_ERROR_CHAIN_HIT, chain_hits); return frame->n_vectors; } static uword ip4_classify (vlib_main_t * vm, - vlib_node_runtime_t * node, - vlib_frame_t * frame) + vlib_node_runtime_t * node, vlib_frame_t * frame) { - return ip_classify_inline (vm, node, frame, 1 /* is_ip4 */); + return ip_classify_inline (vm, node, frame, 1 /* is_ip4 */ ); } +/* *INDENT-OFF* */ VLIB_REGISTER_NODE (ip4_classify_node) = { .function = ip4_classify, .name = "ip4-classify", @@ -330,18 +331,18 @@ VLIB_REGISTER_NODE (ip4_classify_node) = { .n_next_nodes = 0, }; +/* *INDENT-ON* */ VLIB_NODE_FUNCTION_MULTIARCH (ip4_classify_node, ip4_classify) - -static uword -ip6_classify (vlib_main_t * vm, - vlib_node_runtime_t * node, - vlib_frame_t * frame) + static uword + ip6_classify (vlib_main_t * vm, + vlib_node_runtime_t * node, vlib_frame_t * frame) { - return ip_classify_inline (vm, node, frame, 0 /* is_ip4 */); + return ip_classify_inline (vm, node, frame, 0 /* is_ip4 */ ); } +/* *INDENT-OFF* */ VLIB_REGISTER_NODE (ip6_classify_node) = { .function = ip6_classify, .name = "ip6-classify", @@ -353,13 +354,20 @@ VLIB_REGISTER_NODE (ip6_classify_node) = { .n_next_nodes = 0, }; +/* *INDENT-ON* */ VLIB_NODE_FUNCTION_MULTIARCH (ip6_classify_node, ip6_classify) - -static clib_error_t * -ip_classify_init (vlib_main_t * vm) + static clib_error_t *ip_classify_init (vlib_main_t * vm) { return 0; } VLIB_INIT_FUNCTION (ip_classify_init); + +/* + * fd.io coding-style-patch-verification: ON + * + * Local Variables: + * eval: (c-set-style "gnu") + * End: + */ diff --git a/src/vnet/classify/policer_classify.c b/src/vnet/classify/policer_classify.c index 2a4fa62c023..542ba1015ed 100644 --- a/src/vnet/classify/policer_classify.c +++ b/src/vnet/classify/policer_classify.c @@ -18,19 +18,19 @@ policer_classify_main_t policer_classify_main; static void vnet_policer_classify_feature_enable (vlib_main_t * vnm, - policer_classify_main_t * pcm, - u32 sw_if_index, - policer_classify_table_id_t tid, - int feature_enable) + policer_classify_main_t * pcm, + u32 sw_if_index, + policer_classify_table_id_t tid, + int feature_enable) { if (tid == POLICER_CLASSIFY_TABLE_L2) { l2input_intf_bitmap_enable (sw_if_index, L2INPUT_FEAT_POLICER_CLAS, - feature_enable); + feature_enable); } else { - vnet_feature_config_main_t * fcm; + vnet_feature_config_main_t *fcm; u8 arc; if (tid == POLICER_CLASSIFY_TABLE_IP4) @@ -39,7 +39,7 @@ vnet_policer_classify_feature_enable (vlib_main_t * vnm, sw_if_index, feature_enable, 0, 0); arc = vnet_get_feature_arc_index ("ip4-unicast"); } - + else { vnet_feature_enable_disable ("ip6-unicast", "ip6-policer-classify", @@ -52,14 +52,16 @@ vnet_policer_classify_feature_enable (vlib_main_t * vnm, } } -int vnet_set_policer_classify_intfc (vlib_main_t * vm, u32 sw_if_index, - u32 ip4_table_index, u32 ip6_table_index, - u32 l2_table_index, u32 is_add) +int +vnet_set_policer_classify_intfc (vlib_main_t * vm, u32 sw_if_index, + u32 ip4_table_index, u32 ip6_table_index, + u32 l2_table_index, u32 is_add) { - policer_classify_main_t * pcm = &policer_classify_main; - vnet_classify_main_t * vcm = pcm->vnet_classify_main; - u32 pct[POLICER_CLASSIFY_N_TABLES] = {ip4_table_index, ip6_table_index, - l2_table_index}; + policer_classify_main_t *pcm = &policer_classify_main; + vnet_classify_main_t *vcm = pcm->vnet_classify_main; + u32 pct[POLICER_CLASSIFY_N_TABLES] = { ip4_table_index, ip6_table_index, + l2_table_index + }; u32 ti; /* Assume that we've validated sw_if_index in the API layer */ @@ -67,34 +69,36 @@ int vnet_set_policer_classify_intfc (vlib_main_t * vm, u32 sw_if_index, for (ti = 0; ti < POLICER_CLASSIFY_N_TABLES; ti++) { if (pct[ti] == ~0) - continue; + continue; if (pool_is_free_index (vcm->tables, pct[ti])) - return VNET_API_ERROR_NO_SUCH_TABLE; + return VNET_API_ERROR_NO_SUCH_TABLE; vec_validate_init_empty - (pcm->classify_table_index_by_sw_if_index[ti], sw_if_index, ~0); + (pcm->classify_table_index_by_sw_if_index[ti], sw_if_index, ~0); /* Reject any DEL operation with wrong sw_if_index */ if (!is_add && - (pct[ti] != pcm->classify_table_index_by_sw_if_index[ti][sw_if_index])) - { - clib_warning ("Non-existent intf_idx=%d with table_index=%d for delete", - sw_if_index, pct[ti]); - return VNET_API_ERROR_NO_SUCH_TABLE; - } + (pct[ti] != + pcm->classify_table_index_by_sw_if_index[ti][sw_if_index])) + { + clib_warning + ("Non-existent intf_idx=%d with table_index=%d for delete", + sw_if_index, pct[ti]); + return VNET_API_ERROR_NO_SUCH_TABLE; + } /* Return ok on ADD operaton if feature is already enabled */ if (is_add && - pcm->classify_table_index_by_sw_if_index[ti][sw_if_index] != ~0) - return 0; + pcm->classify_table_index_by_sw_if_index[ti][sw_if_index] != ~0) + return 0; vnet_policer_classify_feature_enable (vm, pcm, sw_if_index, ti, is_add); if (is_add) - pcm->classify_table_index_by_sw_if_index[ti][sw_if_index] = pct[ti]; + pcm->classify_table_index_by_sw_if_index[ti][sw_if_index] = pct[ti]; else - pcm->classify_table_index_by_sw_if_index[ti][sw_if_index] = ~0; + pcm->classify_table_index_by_sw_if_index[ti][sw_if_index] = ~0; } @@ -103,10 +107,10 @@ int vnet_set_policer_classify_intfc (vlib_main_t * vm, u32 sw_if_index, static clib_error_t * set_policer_classify_command_fn (vlib_main_t * vm, - unformat_input_t * input, - vlib_cli_command_t * cmd) + unformat_input_t * input, + vlib_cli_command_t * cmd) { - vnet_main_t * vnm = vnet_get_main(); + vnet_main_t *vnm = vnet_get_main (); u32 sw_if_index = ~0; u32 ip4_table_index = ~0; u32 ip6_table_index = ~0; @@ -118,18 +122,18 @@ set_policer_classify_command_fn (vlib_main_t * vm, while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) { if (unformat (input, "interface %U", unformat_vnet_sw_interface, - vnm, &sw_if_index)) - ; + vnm, &sw_if_index)) + ; else if (unformat (input, "ip4-table %d", &ip4_table_index)) - idx_cnt++; + idx_cnt++; else if (unformat (input, "ip6-table %d", &ip6_table_index)) - idx_cnt++; + idx_cnt++; else if (unformat (input, "l2-table %d", &l2_table_index)) - idx_cnt++; + idx_cnt++; else if (unformat (input, "del")) - is_add = 0; + is_add = 0; else - break; + break; } if (sw_if_index == ~0) @@ -141,8 +145,9 @@ set_policer_classify_command_fn (vlib_main_t * vm, if (idx_cnt > 1) return clib_error_return (0, "Only one table index per API is allowed."); - rv = vnet_set_policer_classify_intfc(vm, sw_if_index, ip4_table_index, - ip6_table_index, l2_table_index, is_add); + rv = vnet_set_policer_classify_intfc (vm, sw_if_index, ip4_table_index, + ip6_table_index, l2_table_index, + is_add); switch (rv) { @@ -158,6 +163,7 @@ set_policer_classify_command_fn (vlib_main_t * vm, return 0; } +/* *INDENT-OFF* */ VLIB_CLI_COMMAND (set_policer_classify_command, static) = { .path = "set policer classify", .short_help = @@ -165,11 +171,12 @@ VLIB_CLI_COMMAND (set_policer_classify_command, static) = { " [ip6-table ] [l2-table ] [del]", .function = set_policer_classify_command_fn, }; +/* *INDENT-ON* */ static uword unformat_table_type (unformat_input_t * input, va_list * va) { - u32 * r = va_arg (*va, u32 *); + u32 *r = va_arg (*va, u32 *); u32 tid; if (unformat (input, "ip4")) @@ -184,14 +191,15 @@ unformat_table_type (unformat_input_t * input, va_list * va) *r = tid; return 1; } + static clib_error_t * show_policer_classify_command_fn (vlib_main_t * vm, - unformat_input_t * input, - vlib_cli_command_t * cmd) + unformat_input_t * input, + vlib_cli_command_t * cmd) { - policer_classify_main_t * pcm = &policer_classify_main; + policer_classify_main_t *pcm = &policer_classify_main; u32 type = POLICER_CLASSIFY_N_TABLES; - u32 * vec_tbl; + u32 *vec_tbl; int i; if (unformat (input, "type %U", unformat_table_type, &type)) @@ -204,26 +212,36 @@ show_policer_classify_command_fn (vlib_main_t * vm, vec_tbl = pcm->classify_table_index_by_sw_if_index[type]; - if (vec_len(vec_tbl)) - vlib_cli_output (vm, "%10s%20s\t\t%s", "Intfc idx", "Classify table", - "Interface name"); + if (vec_len (vec_tbl)) + vlib_cli_output (vm, "%10s%20s\t\t%s", "Intfc idx", "Classify table", + "Interface name"); else vlib_cli_output (vm, "No tables configured."); for (i = 0; i < vec_len (vec_tbl); i++) { - if (vec_elt(vec_tbl, i) == ~0) - continue; + if (vec_elt (vec_tbl, i) == ~0) + continue; - vlib_cli_output (vm, "%10d%20d\t\t%U", i, vec_elt(vec_tbl, i), - format_vnet_sw_if_index_name, pcm->vnet_main, i); + vlib_cli_output (vm, "%10d%20d\t\t%U", i, vec_elt (vec_tbl, i), + format_vnet_sw_if_index_name, pcm->vnet_main, i); } return 0; } +/* *INDENT-OFF* */ VLIB_CLI_COMMAND (show_policer_classify_command, static) = { .path = "show classify policer", .short_help = "show classify policer type [ip4|ip6|l2]", .function = show_policer_classify_command_fn, }; +/* *INDENT-ON* */ + +/* + * fd.io coding-style-patch-verification: ON + * + * Local Variables: + * eval: (c-set-style "gnu") + * End: + */ diff --git a/src/vnet/classify/policer_classify.h b/src/vnet/classify/policer_classify.h index 6ab424fbf9c..a1bc9c54162 100644 --- a/src/vnet/classify/policer_classify.h +++ b/src/vnet/classify/policer_classify.h @@ -20,36 +20,47 @@ #include #include -typedef enum { +typedef enum +{ POLICER_CLASSIFY_TABLE_IP4, POLICER_CLASSIFY_TABLE_IP6, POLICER_CLASSIFY_TABLE_L2, POLICER_CLASSIFY_N_TABLES, } policer_classify_table_id_t; -typedef enum { +typedef enum +{ POLICER_CLASSIFY_NEXT_INDEX_DROP, POLICER_CLASSIFY_NEXT_INDEX_N_NEXT, } policer_classify_next_index_t; -typedef struct { +typedef struct +{ /* Classifier table vectors */ - u32 * classify_table_index_by_sw_if_index [POLICER_CLASSIFY_N_TABLES]; + u32 *classify_table_index_by_sw_if_index[POLICER_CLASSIFY_N_TABLES]; /* L2 next nodes for each feature */ u32 feat_next_node_index[32]; /* Convenience variables */ - vlib_main_t * vlib_main; - vnet_main_t * vnet_main; - vnet_classify_main_t * vnet_classify_main; - vnet_config_main_t * vnet_config_main [POLICER_CLASSIFY_N_TABLES]; + vlib_main_t *vlib_main; + vnet_main_t *vnet_main; + vnet_classify_main_t *vnet_classify_main; + vnet_config_main_t *vnet_config_main[POLICER_CLASSIFY_N_TABLES]; } policer_classify_main_t; extern policer_classify_main_t policer_classify_main; int vnet_set_policer_classify_intfc (vlib_main_t * vm, u32 sw_if_index, - u32 ip4_table_index, u32 ip6_table_index, - u32 l2_table_index, u32 is_add); + u32 ip4_table_index, u32 ip6_table_index, + u32 l2_table_index, u32 is_add); #endif /* __included_vnet_policer_classify_h__ */ + +/* + * fd.io coding-style-patch-verification: ON + * + * Local Variables: + * eval: (c-set-style "gnu") + * End: + */ diff --git a/src/vnet/classify/vnet_classify.c b/src/vnet/classify/vnet_classify.c index a70782339d4..ef45d1a4167 100644 --- a/src/vnet/classify/vnet_classify.c +++ b/src/vnet/classify/vnet_classify.c @@ -15,110 +15,122 @@ #include #include #include -#include /* for API error numbers */ -#include /* for L2_INPUT_CLASSIFY_NEXT_xxx */ +#include /* for API error numbers */ +#include /* for L2_INPUT_CLASSIFY_NEXT_xxx */ #include vnet_classify_main_t vnet_classify_main; #if VALIDATION_SCAFFOLDING /* Validation scaffolding */ -void mv (vnet_classify_table_t * t) +void +mv (vnet_classify_table_t * t) { - void * oldheap; + void *oldheap; oldheap = clib_mem_set_heap (t->mheap); - clib_mem_validate(); + clib_mem_validate (); clib_mem_set_heap (oldheap); } -void rogue (vnet_classify_table_t * t) +void +rogue (vnet_classify_table_t * t) { int i, j, k; - vnet_classify_entry_t * v, * save_v; + vnet_classify_entry_t *v, *save_v; u32 active_elements = 0; - vnet_classify_bucket_t * b; - + vnet_classify_bucket_t *b; + for (i = 0; i < t->nbuckets; i++) { - b = &t->buckets [i]; + b = &t->buckets[i]; if (b->offset == 0) - continue; + continue; save_v = vnet_classify_get_entry (t, b->offset); - for (j = 0; j < (1<log2_pages); j++) - { - for (k = 0; k < t->entries_per_page; k++) - { - v = vnet_classify_entry_at_index - (t, save_v, j*t->entries_per_page + k); - - if (vnet_classify_entry_is_busy (v)) - active_elements++; - } - } + for (j = 0; j < (1 << b->log2_pages); j++) + { + for (k = 0; k < t->entries_per_page; k++) + { + v = vnet_classify_entry_at_index + (t, save_v, j * t->entries_per_page + k); + + if (vnet_classify_entry_is_busy (v)) + active_elements++; + } + } } if (active_elements != t->active_elements) - clib_warning ("found %u expected %u elts", active_elements, - t->active_elements); + clib_warning ("found %u expected %u elts", active_elements, + t->active_elements); } #else -void mv (vnet_classify_table_t * t) { } -void rogue (vnet_classify_table_t * t) { } +void +mv (vnet_classify_table_t * t) +{ +} + +void +rogue (vnet_classify_table_t * t) +{ +} #endif -void vnet_classify_register_unformat_l2_next_index_fn (unformat_function_t * fn) +void +vnet_classify_register_unformat_l2_next_index_fn (unformat_function_t * fn) { - vnet_classify_main_t * cm = &vnet_classify_main; + vnet_classify_main_t *cm = &vnet_classify_main; vec_add1 (cm->unformat_l2_next_index_fns, fn); } -void vnet_classify_register_unformat_ip_next_index_fn (unformat_function_t * fn) +void +vnet_classify_register_unformat_ip_next_index_fn (unformat_function_t * fn) { - vnet_classify_main_t * cm = &vnet_classify_main; + vnet_classify_main_t *cm = &vnet_classify_main; vec_add1 (cm->unformat_ip_next_index_fns, fn); } -void +void vnet_classify_register_unformat_acl_next_index_fn (unformat_function_t * fn) { - vnet_classify_main_t * cm = &vnet_classify_main; + vnet_classify_main_t *cm = &vnet_classify_main; vec_add1 (cm->unformat_acl_next_index_fns, fn); } void -vnet_classify_register_unformat_policer_next_index_fn (unformat_function_t * fn) +vnet_classify_register_unformat_policer_next_index_fn (unformat_function_t * + fn) { - vnet_classify_main_t * cm = &vnet_classify_main; + vnet_classify_main_t *cm = &vnet_classify_main; vec_add1 (cm->unformat_policer_next_index_fns, fn); } -void vnet_classify_register_unformat_opaque_index_fn (unformat_function_t * fn) +void +vnet_classify_register_unformat_opaque_index_fn (unformat_function_t * fn) { - vnet_classify_main_t * cm = &vnet_classify_main; + vnet_classify_main_t *cm = &vnet_classify_main; vec_add1 (cm->unformat_opaque_index_fns, fn); } -vnet_classify_table_t * -vnet_classify_new_table (vnet_classify_main_t *cm, - u8 * mask, u32 nbuckets, u32 memory_size, - u32 skip_n_vectors, - u32 match_n_vectors) +vnet_classify_table_t * +vnet_classify_new_table (vnet_classify_main_t * cm, + u8 * mask, u32 nbuckets, u32 memory_size, + u32 skip_n_vectors, u32 match_n_vectors) { - vnet_classify_table_t * t; - void * oldheap; - + vnet_classify_table_t *t; + void *oldheap; + nbuckets = 1 << (max_log2 (nbuckets)); pool_get_aligned (cm->tables, t, CLIB_CACHE_LINE_BYTES); - memset(t, 0, sizeof (*t)); - - vec_validate_aligned (t->mask, match_n_vectors - 1, sizeof(u32x4)); + memset (t, 0, sizeof (*t)); + + vec_validate_aligned (t->mask, match_n_vectors - 1, sizeof (u32x4)); clib_memcpy (t->mask, mask, match_n_vectors * sizeof (u32x4)); t->next_table_index = ~0; @@ -128,23 +140,24 @@ vnet_classify_new_table (vnet_classify_main_t *cm, t->skip_n_vectors = skip_n_vectors; t->entries_per_page = 2; - t->mheap = mheap_alloc (0 /* use VM */, memory_size); + t->mheap = mheap_alloc (0 /* use VM */ , memory_size); vec_validate_aligned (t->buckets, nbuckets - 1, CLIB_CACHE_LINE_BYTES); oldheap = clib_mem_set_heap (t->mheap); - t->writer_lock = clib_mem_alloc_aligned (CLIB_CACHE_LINE_BYTES, - CLIB_CACHE_LINE_BYTES); + t->writer_lock = clib_mem_alloc_aligned (CLIB_CACHE_LINE_BYTES, + CLIB_CACHE_LINE_BYTES); t->writer_lock[0] = 0; clib_mem_set_heap (oldheap); return (t); } -void vnet_classify_delete_table_index (vnet_classify_main_t *cm, - u32 table_index, int del_chain) +void +vnet_classify_delete_table_index (vnet_classify_main_t * cm, + u32 table_index, int del_chain) { - vnet_classify_table_t * t; + vnet_classify_table_t *t; /* Tolerate multiple frees, up to a point */ if (pool_is_free_index (cm->tables, table_index)) @@ -158,26 +171,26 @@ void vnet_classify_delete_table_index (vnet_classify_main_t *cm, vec_free (t->mask); vec_free (t->buckets); mheap_free (t->mheap); - + pool_put (cm->tables, t); } static vnet_classify_entry_t * vnet_classify_entry_alloc (vnet_classify_table_t * t, u32 log2_pages) { - vnet_classify_entry_t * rv = 0; + vnet_classify_entry_t *rv = 0; u32 required_length; - void * oldheap; + void *oldheap; ASSERT (t->writer_lock[0]); - required_length = - (sizeof(vnet_classify_entry_t) + (t->match_n_vectors*sizeof(u32x4))) - * t->entries_per_page * (1<match_n_vectors * sizeof (u32x4))) + * t->entries_per_page * (1 << log2_pages); - if (log2_pages >= vec_len (t->freelists) || t->freelists [log2_pages] == 0) + if (log2_pages >= vec_len (t->freelists) || t->freelists[log2_pages] == 0) { oldheap = clib_mem_set_heap (t->mheap); - + vec_validate (t->freelists, log2_pages); rv = clib_mem_alloc_aligned (required_length, CLIB_CACHE_LINE_BYTES); @@ -186,9 +199,9 @@ vnet_classify_entry_alloc (vnet_classify_table_t * t, u32 log2_pages) } rv = t->freelists[log2_pages]; t->freelists[log2_pages] = rv->next_free; - + initialize: - ASSERT(rv); + ASSERT (rv); memset (rv, 0xff, required_length); return rv; @@ -196,24 +209,24 @@ initialize: static void vnet_classify_entry_free (vnet_classify_table_t * t, - vnet_classify_entry_t * v, u32 log2_pages) + vnet_classify_entry_t * v, u32 log2_pages) { - ASSERT (t->writer_lock[0]); + ASSERT (t->writer_lock[0]); - ASSERT(vec_len (t->freelists) > log2_pages); + ASSERT (vec_len (t->freelists) > log2_pages); - v->next_free = t->freelists[log2_pages]; - t->freelists[log2_pages] = v; + v->next_free = t->freelists[log2_pages]; + t->freelists[log2_pages] = v; } static inline void make_working_copy -(vnet_classify_table_t * t, vnet_classify_bucket_t * b) + (vnet_classify_table_t * t, vnet_classify_bucket_t * b) { - vnet_classify_entry_t * v; - vnet_classify_bucket_t working_bucket __attribute__((aligned (8))); - void * oldheap; - vnet_classify_entry_t * working_copy; - u32 thread_index = vlib_get_thread_index(); + vnet_classify_entry_t *v; + vnet_classify_bucket_t working_bucket __attribute__ ((aligned (8))); + void *oldheap; + vnet_classify_entry_t *working_copy; + u32 thread_index = vlib_get_thread_index (); int working_copy_length, required_length; if (thread_index >= vec_len (t->working_copies)) @@ -225,16 +238,16 @@ static inline void make_working_copy clib_mem_set_heap (oldheap); } - /* + /* * working_copies are per-cpu so that near-simultaneous * updates from multiple threads will not result in sporadic, spurious - * lookup failures. + * lookup failures. */ working_copy = t->working_copies[thread_index]; working_copy_length = t->working_copy_lengths[thread_index]; - required_length = - (sizeof(vnet_classify_entry_t) + (t->match_n_vectors*sizeof(u32x4))) - * t->entries_per_page * (1<log2_pages); + required_length = + (sizeof (vnet_classify_entry_t) + (t->match_n_vectors * sizeof (u32x4))) + * t->entries_per_page * (1 << b->log2_pages); t->saved_bucket.as_u64 = b->as_u64; oldheap = clib_mem_set_heap (t->mheap); @@ -242,119 +255,118 @@ static inline void make_working_copy if (required_length > working_copy_length) { if (working_copy) - clib_mem_free (working_copy); + clib_mem_free (working_copy); working_copy = - clib_mem_alloc_aligned (required_length, CLIB_CACHE_LINE_BYTES); + clib_mem_alloc_aligned (required_length, CLIB_CACHE_LINE_BYTES); t->working_copies[thread_index] = working_copy; } clib_mem_set_heap (oldheap); v = vnet_classify_get_entry (t, b->offset); - + clib_memcpy (working_copy, v, required_length); - + working_bucket.as_u64 = b->as_u64; working_bucket.offset = vnet_classify_get_offset (t, working_copy); - CLIB_MEMORY_BARRIER(); + CLIB_MEMORY_BARRIER (); b->as_u64 = working_bucket.as_u64; t->working_copies[thread_index] = working_copy; } static vnet_classify_entry_t * split_and_rehash (vnet_classify_table_t * t, - vnet_classify_entry_t * old_values, u32 old_log2_pages, - u32 new_log2_pages) + vnet_classify_entry_t * old_values, u32 old_log2_pages, + u32 new_log2_pages) { - vnet_classify_entry_t * new_values, * v, * new_v; + vnet_classify_entry_t *new_values, *v, *new_v; int i, j, length_in_entries; - + new_values = vnet_classify_entry_alloc (t, new_log2_pages); - length_in_entries = (1<entries_per_page; - + length_in_entries = (1 << old_log2_pages) * t->entries_per_page; + for (i = 0; i < length_in_entries; i++) { u64 new_hash; - + v = vnet_classify_entry_at_index (t, old_values, i); - + if (vnet_classify_entry_is_busy (v)) - { - /* Hack so we can use the packet hash routine */ - u8 * key_minus_skip; - key_minus_skip = (u8 *) v->key; - key_minus_skip -= t->skip_n_vectors * sizeof (u32x4); - - new_hash = vnet_classify_hash_packet (t, key_minus_skip); - new_hash >>= t->log2_nbuckets; - new_hash &= (1<entries_per_page; j++) - { - new_v = vnet_classify_entry_at_index (t, new_values, - new_hash + j); - - if (vnet_classify_entry_is_free (new_v)) - { - clib_memcpy (new_v, v, sizeof (vnet_classify_entry_t) - + (t->match_n_vectors * sizeof (u32x4))); - new_v->flags &= ~(VNET_CLASSIFY_ENTRY_FREE); - goto doublebreak; - } - } - /* Crap. Tell caller to try again */ - vnet_classify_entry_free (t, new_values, new_log2_pages); - return 0; - doublebreak: - ; - } + { + /* Hack so we can use the packet hash routine */ + u8 *key_minus_skip; + key_minus_skip = (u8 *) v->key; + key_minus_skip -= t->skip_n_vectors * sizeof (u32x4); + + new_hash = vnet_classify_hash_packet (t, key_minus_skip); + new_hash >>= t->log2_nbuckets; + new_hash &= (1 << new_log2_pages) - 1; + + for (j = 0; j < t->entries_per_page; j++) + { + new_v = vnet_classify_entry_at_index (t, new_values, + new_hash + j); + + if (vnet_classify_entry_is_free (new_v)) + { + clib_memcpy (new_v, v, sizeof (vnet_classify_entry_t) + + (t->match_n_vectors * sizeof (u32x4))); + new_v->flags &= ~(VNET_CLASSIFY_ENTRY_FREE); + goto doublebreak; + } + } + /* Crap. Tell caller to try again */ + vnet_classify_entry_free (t, new_values, new_log2_pages); + return 0; + doublebreak: + ; + } } return new_values; } static vnet_classify_entry_t * split_and_rehash_linear (vnet_classify_table_t * t, - vnet_classify_entry_t * old_values, - u32 old_log2_pages, - u32 new_log2_pages) + vnet_classify_entry_t * old_values, + u32 old_log2_pages, u32 new_log2_pages) { - vnet_classify_entry_t * new_values, * v, * new_v; + vnet_classify_entry_t *new_values, *v, *new_v; int i, j, new_length_in_entries, old_length_in_entries; - + new_values = vnet_classify_entry_alloc (t, new_log2_pages); - new_length_in_entries = (1<entries_per_page; - old_length_in_entries = (1<entries_per_page; - + new_length_in_entries = (1 << new_log2_pages) * t->entries_per_page; + old_length_in_entries = (1 << old_log2_pages) * t->entries_per_page; + j = 0; for (i = 0; i < old_length_in_entries; i++) { v = vnet_classify_entry_at_index (t, old_values, i); - + if (vnet_classify_entry_is_busy (v)) - { - for (; j < new_length_in_entries; j++) - { - new_v = vnet_classify_entry_at_index (t, new_values, j); - - if (vnet_classify_entry_is_busy (new_v)) - { - clib_warning ("BUG: linear rehash new entry not free!"); - continue; - } - clib_memcpy (new_v, v, sizeof (vnet_classify_entry_t) - + (t->match_n_vectors * sizeof (u32x4))); - new_v->flags &= ~(VNET_CLASSIFY_ENTRY_FREE); - j++; - goto doublebreak; - } - /* - * Crap. Tell caller to try again. - * This should never happen... - */ - clib_warning ("BUG: linear rehash failed!"); - vnet_classify_entry_free (t, new_values, new_log2_pages); - return 0; - } + { + for (; j < new_length_in_entries; j++) + { + new_v = vnet_classify_entry_at_index (t, new_values, j); + + if (vnet_classify_entry_is_busy (new_v)) + { + clib_warning ("BUG: linear rehash new entry not free!"); + continue; + } + clib_memcpy (new_v, v, sizeof (vnet_classify_entry_t) + + (t->match_n_vectors * sizeof (u32x4))); + new_v->flags &= ~(VNET_CLASSIFY_ENTRY_FREE); + j++; + goto doublebreak; + } + /* + * Crap. Tell caller to try again. + * This should never happen... + */ + clib_warning ("BUG: linear rehash failed!"); + vnet_classify_entry_free (t, new_values, new_log2_pages); + return 0; + } doublebreak: ; } @@ -363,52 +375,52 @@ split_and_rehash_linear (vnet_classify_table_t * t, } static void -vnet_classify_entry_claim_resource (vnet_classify_entry_t *e) +vnet_classify_entry_claim_resource (vnet_classify_entry_t * e) { - switch (e->action) + switch (e->action) { case CLASSIFY_ACTION_SET_IP4_FIB_INDEX: - fib_table_lock (e->metadata, FIB_PROTOCOL_IP4, FIB_SOURCE_CLASSIFY); - break; + fib_table_lock (e->metadata, FIB_PROTOCOL_IP4, FIB_SOURCE_CLASSIFY); + break; case CLASSIFY_ACTION_SET_IP6_FIB_INDEX: - fib_table_lock (e->metadata, FIB_PROTOCOL_IP6, FIB_SOURCE_CLASSIFY); - break; + fib_table_lock (e->metadata, FIB_PROTOCOL_IP6, FIB_SOURCE_CLASSIFY); + break; case CLASSIFY_ACTION_SET_METADATA: - break; + break; } } static void -vnet_classify_entry_release_resource (vnet_classify_entry_t *e) +vnet_classify_entry_release_resource (vnet_classify_entry_t * e) { - switch (e->action) + switch (e->action) { case CLASSIFY_ACTION_SET_IP4_FIB_INDEX: - fib_table_unlock (e->metadata, FIB_PROTOCOL_IP4, FIB_SOURCE_CLASSIFY); - break; + fib_table_unlock (e->metadata, FIB_PROTOCOL_IP4, FIB_SOURCE_CLASSIFY); + break; case CLASSIFY_ACTION_SET_IP6_FIB_INDEX: - fib_table_unlock (e->metadata, FIB_PROTOCOL_IP6, FIB_SOURCE_CLASSIFY); - break; + fib_table_unlock (e->metadata, FIB_PROTOCOL_IP6, FIB_SOURCE_CLASSIFY); + break; case CLASSIFY_ACTION_SET_METADATA: - break; + break; } } -int vnet_classify_add_del (vnet_classify_table_t * t, - vnet_classify_entry_t * add_v, - int is_add) +int +vnet_classify_add_del (vnet_classify_table_t * t, + vnet_classify_entry_t * add_v, int is_add) { u32 bucket_index; - vnet_classify_bucket_t * b, tmp_b; - vnet_classify_entry_t * v, * new_v, * save_new_v, * working_copy, * save_v; + vnet_classify_bucket_t *b, tmp_b; + vnet_classify_entry_t *v, *new_v, *save_new_v, *working_copy, *save_v; u32 value_index; int rv = 0; int i; u64 hash, new_hash; u32 limit; u32 old_log2_pages, new_log2_pages; - u32 thread_index = vlib_get_thread_index(); - u8 * key_minus_skip; + u32 thread_index = vlib_get_thread_index (); + u8 *key_minus_skip; int resplit_once = 0; int mark_bucket_linear; @@ -419,26 +431,26 @@ int vnet_classify_add_del (vnet_classify_table_t * t, hash = vnet_classify_hash_packet (t, key_minus_skip); - bucket_index = hash & (t->nbuckets-1); + bucket_index = hash & (t->nbuckets - 1); b = &t->buckets[bucket_index]; hash >>= t->log2_nbuckets; while (__sync_lock_test_and_set (t->writer_lock, 1)) - ; + ; /* First elt in the bucket? */ if (b->offset == 0) { if (is_add == 0) - { - rv = -1; - goto unlock; - } + { + rv = -1; + goto unlock; + } - v = vnet_classify_entry_alloc (t, 0 /* new_log2_pages */); + v = vnet_classify_entry_alloc (t, 0 /* new_log2_pages */ ); clib_memcpy (v, add_v, sizeof (vnet_classify_entry_t) + - t->match_n_vectors * sizeof (u32x4)); + t->match_n_vectors * sizeof (u32x4)); v->flags &= ~(VNET_CLASSIFY_ENTRY_FREE); vnet_classify_entry_claim_resource (v); @@ -446,84 +458,86 @@ int vnet_classify_add_del (vnet_classify_table_t * t, tmp_b.offset = vnet_classify_get_offset (t, v); b->as_u64 = tmp_b.as_u64; - t->active_elements ++; + t->active_elements++; goto unlock; } - + make_working_copy (t, b); - + save_v = vnet_classify_get_entry (t, t->saved_bucket.offset); - value_index = hash & ((1<saved_bucket.log2_pages)-1); + value_index = hash & ((1 << t->saved_bucket.log2_pages) - 1); limit = t->entries_per_page; if (PREDICT_FALSE (b->linear_search)) { value_index = 0; - limit *= (1<log2_pages); + limit *= (1 << b->log2_pages); } - + if (is_add) { - /* + /* * For obvious (in hindsight) reasons, see if we're supposed to * replace an existing key, then look for an empty slot. */ for (i = 0; i < limit; i++) - { - v = vnet_classify_entry_at_index (t, save_v, value_index + i); - - if (!memcmp (v->key, add_v->key, t->match_n_vectors * sizeof (u32x4))) - { - clib_memcpy (v, add_v, sizeof (vnet_classify_entry_t) + - t->match_n_vectors * sizeof(u32x4)); - v->flags &= ~(VNET_CLASSIFY_ENTRY_FREE); - vnet_classify_entry_claim_resource (v); - - CLIB_MEMORY_BARRIER(); - /* Restore the previous (k,v) pairs */ - b->as_u64 = t->saved_bucket.as_u64; - goto unlock; - } - } + { + v = vnet_classify_entry_at_index (t, save_v, value_index + i); + + if (!memcmp + (v->key, add_v->key, t->match_n_vectors * sizeof (u32x4))) + { + clib_memcpy (v, add_v, sizeof (vnet_classify_entry_t) + + t->match_n_vectors * sizeof (u32x4)); + v->flags &= ~(VNET_CLASSIFY_ENTRY_FREE); + vnet_classify_entry_claim_resource (v); + + CLIB_MEMORY_BARRIER (); + /* Restore the previous (k,v) pairs */ + b->as_u64 = t->saved_bucket.as_u64; + goto unlock; + } + } for (i = 0; i < limit; i++) - { - v = vnet_classify_entry_at_index (t, save_v, value_index + i); - - if (vnet_classify_entry_is_free (v)) - { - clib_memcpy (v, add_v, sizeof (vnet_classify_entry_t) + - t->match_n_vectors * sizeof(u32x4)); - v->flags &= ~(VNET_CLASSIFY_ENTRY_FREE); - vnet_classify_entry_claim_resource (v); - - CLIB_MEMORY_BARRIER(); - b->as_u64 = t->saved_bucket.as_u64; - t->active_elements ++; - goto unlock; - } - } + { + v = vnet_classify_entry_at_index (t, save_v, value_index + i); + + if (vnet_classify_entry_is_free (v)) + { + clib_memcpy (v, add_v, sizeof (vnet_classify_entry_t) + + t->match_n_vectors * sizeof (u32x4)); + v->flags &= ~(VNET_CLASSIFY_ENTRY_FREE); + vnet_classify_entry_claim_resource (v); + + CLIB_MEMORY_BARRIER (); + b->as_u64 = t->saved_bucket.as_u64; + t->active_elements++; + goto unlock; + } + } /* no room at the inn... split case... */ } else { for (i = 0; i < limit; i++) - { - v = vnet_classify_entry_at_index (t, save_v, value_index + i); - - if (!memcmp (v->key, add_v->key, t->match_n_vectors * sizeof (u32x4))) - { - vnet_classify_entry_release_resource (v); - memset (v, 0xff, sizeof (vnet_classify_entry_t) + - t->match_n_vectors * sizeof(u32x4)); - v->flags |= VNET_CLASSIFY_ENTRY_FREE; - - CLIB_MEMORY_BARRIER(); - b->as_u64 = t->saved_bucket.as_u64; - t->active_elements --; - goto unlock; - } - } + { + v = vnet_classify_entry_at_index (t, save_v, value_index + i); + + if (!memcmp + (v->key, add_v->key, t->match_n_vectors * sizeof (u32x4))) + { + vnet_classify_entry_release_resource (v); + memset (v, 0xff, sizeof (vnet_classify_entry_t) + + t->match_n_vectors * sizeof (u32x4)); + v->flags |= VNET_CLASSIFY_ENTRY_FREE; + + CLIB_MEMORY_BARRIER (); + b->as_u64 = t->saved_bucket.as_u64; + t->active_elements--; + goto unlock; + } + } rv = -3; b->as_u64 = t->saved_bucket.as_u64; goto unlock; @@ -546,22 +560,22 @@ int vnet_classify_add_del (vnet_classify_table_t * t, resplit_once = 1; new_log2_pages++; - new_v = split_and_rehash (t, working_copy, old_log2_pages, - new_log2_pages); + new_v = split_and_rehash (t, working_copy, old_log2_pages, + new_log2_pages); if (new_v == 0) - { - mark_linear: - new_log2_pages--; + { + mark_linear: + new_log2_pages--; - linear_resplit: + linear_resplit: /* pinned collisions, use linear search */ - new_v = split_and_rehash_linear (t, working_copy, old_log2_pages, - new_log2_pages); - /* A new linear-search bucket? */ - if (!t->saved_bucket.linear_search) - t->linear_buckets ++; - mark_bucket_linear = 1; - } + new_v = split_and_rehash_linear (t, working_copy, old_log2_pages, + new_log2_pages); + /* A new linear-search bucket? */ + if (!t->saved_bucket.linear_search) + t->linear_buckets++; + mark_bucket_linear = 1; + } } /* Try to add the new entry */ @@ -572,28 +586,28 @@ int vnet_classify_add_del (vnet_classify_table_t * t, new_hash = vnet_classify_hash_packet_inline (t, key_minus_skip); new_hash >>= t->log2_nbuckets; - new_hash &= (1<entries_per_page; if (mark_bucket_linear) { - limit *= (1<match_n_vectors * sizeof(u32x4)); - new_v->flags &= ~(VNET_CLASSIFY_ENTRY_FREE); - vnet_classify_entry_claim_resource (new_v); - - goto expand_ok; - } + { + clib_memcpy (new_v, add_v, sizeof (vnet_classify_entry_t) + + t->match_n_vectors * sizeof (u32x4)); + new_v->flags &= ~(VNET_CLASSIFY_ENTRY_FREE); + vnet_classify_entry_claim_resource (new_v); + + goto expand_ok; + } } /* Crap. Try again */ vnet_classify_entry_free (t, save_new_v, new_log2_pages); @@ -601,116 +615,121 @@ int vnet_classify_add_del (vnet_classify_table_t * t, if (resplit_once) goto mark_linear; - else + else goto try_resplit; - expand_ok: +expand_ok: tmp_b.log2_pages = new_log2_pages; tmp_b.offset = vnet_classify_get_offset (t, save_new_v); tmp_b.linear_search = mark_bucket_linear; - CLIB_MEMORY_BARRIER(); + CLIB_MEMORY_BARRIER (); b->as_u64 = tmp_b.as_u64; - t->active_elements ++; + t->active_elements++; v = vnet_classify_get_entry (t, t->saved_bucket.offset); vnet_classify_entry_free (t, v, old_log2_pages); - unlock: - CLIB_MEMORY_BARRIER(); +unlock: + CLIB_MEMORY_BARRIER (); t->writer_lock[0] = 0; return rv; } +/* *INDENT-OFF* */ typedef CLIB_PACKED(struct { ethernet_header_t eh; ip4_header_t ip; }) classify_data_or_mask_t; +/* *INDENT-ON* */ -u64 vnet_classify_hash_packet (vnet_classify_table_t * t, u8 * h) +u64 +vnet_classify_hash_packet (vnet_classify_table_t * t, u8 * h) { - return vnet_classify_hash_packet_inline (t, h); + return vnet_classify_hash_packet_inline (t, h); } -vnet_classify_entry_t * +vnet_classify_entry_t * vnet_classify_find_entry (vnet_classify_table_t * t, - u8 * h, u64 hash, f64 now) + u8 * h, u64 hash, f64 now) { return vnet_classify_find_entry_inline (t, h, hash, now); } -static u8 * format_classify_entry (u8 * s, va_list * args) - { - vnet_classify_table_t * t = va_arg (*args, vnet_classify_table_t *); - vnet_classify_entry_t * e = va_arg (*args, vnet_classify_entry_t *); +static u8 * +format_classify_entry (u8 * s, va_list * args) +{ + vnet_classify_table_t *t = va_arg (*args, vnet_classify_table_t *); + vnet_classify_entry_t *e = va_arg (*args, vnet_classify_entry_t *); s = format (s, "[%u]: next_index %d advance %d opaque %d action %d metadata %d\n", - vnet_classify_get_offset (t, e), e->next_index, e->advance, + vnet_classify_get_offset (t, e), e->next_index, e->advance, e->opaque_index, e->action, e->metadata); s = format (s, " k: %U\n", format_hex_bytes, e->key, - t->match_n_vectors * sizeof(u32x4)); - + t->match_n_vectors * sizeof (u32x4)); + if (vnet_classify_entry_is_busy (e)) s = format (s, " hits %lld, last_heard %.2f\n", - e->hits, e->last_heard); + e->hits, e->last_heard); else s = format (s, " entry is free\n"); return s; - } - -u8 * format_classify_table (u8 * s, va_list * args) +} + +u8 * +format_classify_table (u8 * s, va_list * args) { - vnet_classify_table_t * t = va_arg (*args, vnet_classify_table_t *); + vnet_classify_table_t *t = va_arg (*args, vnet_classify_table_t *); int verbose = va_arg (*args, int); - vnet_classify_bucket_t * b; - vnet_classify_entry_t * v, * save_v; + vnet_classify_bucket_t *b; + vnet_classify_entry_t *v, *save_v; int i, j, k; u64 active_elements = 0; - + for (i = 0; i < t->nbuckets; i++) { - b = &t->buckets [i]; + b = &t->buckets[i]; if (b->offset == 0) - { - if (verbose > 1) - s = format (s, "[%d]: empty\n", i); - continue; - } + { + if (verbose > 1) + s = format (s, "[%d]: empty\n", i); + continue; + } if (verbose) - { - s = format (s, "[%d]: heap offset %d, elts %d, %s\n", i, - b->offset, (1<log2_pages)*t->entries_per_page, - b->linear_search ? "LINEAR" : "normal"); - } + { + s = format (s, "[%d]: heap offset %d, elts %d, %s\n", i, + b->offset, (1 << b->log2_pages) * t->entries_per_page, + b->linear_search ? "LINEAR" : "normal"); + } save_v = vnet_classify_get_entry (t, b->offset); - for (j = 0; j < (1<log2_pages); j++) - { - for (k = 0; k < t->entries_per_page; k++) - { - - v = vnet_classify_entry_at_index (t, save_v, - j*t->entries_per_page + k); - - if (vnet_classify_entry_is_free (v)) - { - if (verbose > 1) - s = format (s, " %d: empty\n", - j * t->entries_per_page + k); - continue; - } - if (verbose) - { - s = format (s, " %d: %U\n", - j * t->entries_per_page + k, - format_classify_entry, t, v); - } - active_elements++; - } - } + for (j = 0; j < (1 << b->log2_pages); j++) + { + for (k = 0; k < t->entries_per_page; k++) + { + + v = vnet_classify_entry_at_index (t, save_v, + j * t->entries_per_page + k); + + if (vnet_classify_entry_is_free (v)) + { + if (verbose > 1) + s = format (s, " %d: empty\n", + j * t->entries_per_page + k); + continue; + } + if (verbose) + { + s = format (s, " %d: %U\n", + j * t->entries_per_page + k, + format_classify_entry, t, v); + } + active_elements++; + } + } } s = format (s, " %lld active elements\n", active_elements); @@ -719,50 +738,50 @@ u8 * format_classify_table (u8 * s, va_list * args) return s; } -int vnet_classify_add_del_table (vnet_classify_main_t * cm, - u8 * mask, - u32 nbuckets, - u32 memory_size, - u32 skip, - u32 match, - u32 next_table_index, - u32 miss_next_index, - u32 * table_index, - u8 current_data_flag, - i16 current_data_offset, - int is_add, - int del_chain) +int +vnet_classify_add_del_table (vnet_classify_main_t * cm, + u8 * mask, + u32 nbuckets, + u32 memory_size, + u32 skip, + u32 match, + u32 next_table_index, + u32 miss_next_index, + u32 * table_index, + u8 current_data_flag, + i16 current_data_offset, + int is_add, int del_chain) { - vnet_classify_table_t * t; + vnet_classify_table_t *t; if (is_add) { - if (*table_index == ~0) /* add */ - { - if (memory_size == 0) - return VNET_API_ERROR_INVALID_MEMORY_SIZE; - - if (nbuckets == 0) - return VNET_API_ERROR_INVALID_VALUE; - - t = vnet_classify_new_table (cm, mask, nbuckets, memory_size, - skip, match); - t->next_table_index = next_table_index; - t->miss_next_index = miss_next_index; - t->current_data_flag = current_data_flag; - t->current_data_offset = current_data_offset; - *table_index = t - cm->tables; - } - else /* update */ - { - vnet_classify_main_t *cm = &vnet_classify_main; - t = pool_elt_at_index (cm->tables, *table_index); - - t->next_table_index = next_table_index; - } + if (*table_index == ~0) /* add */ + { + if (memory_size == 0) + return VNET_API_ERROR_INVALID_MEMORY_SIZE; + + if (nbuckets == 0) + return VNET_API_ERROR_INVALID_VALUE; + + t = vnet_classify_new_table (cm, mask, nbuckets, memory_size, + skip, match); + t->next_table_index = next_table_index; + t->miss_next_index = miss_next_index; + t->current_data_flag = current_data_flag; + t->current_data_offset = current_data_offset; + *table_index = t - cm->tables; + } + else /* update */ + { + vnet_classify_main_t *cm = &vnet_classify_main; + t = pool_elt_at_index (cm->tables, *table_index); + + t->next_table_index = next_table_index; + } return 0; } - + vnet_classify_delete_table_index (cm, *table_index, del_chain); return 0; } @@ -785,12 +804,13 @@ _(ttl) \ _(protocol) \ _(checksum) -uword unformat_tcp_mask (unformat_input_t * input, va_list * args) +uword +unformat_tcp_mask (unformat_input_t * input, va_list * args) { - u8 ** maskp = va_arg (*args, u8 **); - u8 * mask = 0; + u8 **maskp = va_arg (*args, u8 **); + u8 *mask = 0; u8 found_something = 0; - tcp_header_t * tcp; + tcp_header_t *tcp; #define _(a) u8 a=0; foreach_tcp_proto_field; @@ -798,12 +818,12 @@ uword unformat_tcp_mask (unformat_input_t * input, va_list * args) while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) { - if (0) ; + if (0); #define _(a) else if (unformat (input, #a)) a=1; foreach_tcp_proto_field #undef _ - else - break; + else + break; } #define _(a) found_something += a; @@ -825,12 +845,13 @@ uword unformat_tcp_mask (unformat_input_t * input, va_list * args) return 1; } -uword unformat_udp_mask (unformat_input_t * input, va_list * args) +uword +unformat_udp_mask (unformat_input_t * input, va_list * args) { - u8 ** maskp = va_arg (*args, u8 **); - u8 * mask = 0; + u8 **maskp = va_arg (*args, u8 **); + u8 *mask = 0; u8 found_something = 0; - udp_header_t * udp; + udp_header_t *udp; #define _(a) u8 a=0; foreach_udp_proto_field; @@ -838,12 +859,12 @@ uword unformat_udp_mask (unformat_input_t * input, va_list * args) while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) { - if (0) ; + if (0); #define _(a) else if (unformat (input, #a)) a=1; foreach_udp_proto_field #undef _ - else - break; + else + break; } #define _(a) found_something += a; @@ -865,34 +886,36 @@ uword unformat_udp_mask (unformat_input_t * input, va_list * args) return 1; } -typedef struct { +typedef struct +{ u16 src_port, dst_port; } tcpudp_header_t; -uword unformat_l4_mask (unformat_input_t * input, va_list * args) +uword +unformat_l4_mask (unformat_input_t * input, va_list * args) { - u8 ** maskp = va_arg (*args, u8 **); + u8 **maskp = va_arg (*args, u8 **); u16 src_port = 0, dst_port = 0; - tcpudp_header_t * tcpudp; + tcpudp_header_t *tcpudp; while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) { if (unformat (input, "tcp %U", unformat_tcp_mask, maskp)) - return 1; + return 1; else if (unformat (input, "udp %U", unformat_udp_mask, maskp)) - return 1; + return 1; else if (unformat (input, "src_port")) - src_port = 0xFFFF; + src_port = 0xFFFF; else if (unformat (input, "dst_port")) - dst_port = 0xFFFF; + dst_port = 0xFFFF; else - return 0; + return 0; } if (!src_port && !dst_port) return 0; - u8 * mask = 0; + u8 *mask = 0; vec_validate (mask, sizeof (tcpudp_header_t) - 1); tcpudp = (tcpudp_header_t *) mask; @@ -904,63 +927,64 @@ uword unformat_l4_mask (unformat_input_t * input, va_list * args) return 1; } -uword unformat_ip4_mask (unformat_input_t * input, va_list * args) +uword +unformat_ip4_mask (unformat_input_t * input, va_list * args) { - u8 ** maskp = va_arg (*args, u8 **); - u8 * mask = 0; + u8 **maskp = va_arg (*args, u8 **); + u8 *mask = 0; u8 found_something = 0; - ip4_header_t * ip; - + ip4_header_t *ip; + #define _(a) u8 a=0; foreach_ip4_proto_field; #undef _ u8 version = 0; u8 hdr_length = 0; - - - while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) + + + while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) { - if (unformat (input, "version")) - version = 1; + if (unformat (input, "version")) + version = 1; else if (unformat (input, "hdr_length")) - hdr_length = 1; + hdr_length = 1; else if (unformat (input, "src")) - src_address = 1; + src_address = 1; else if (unformat (input, "dst")) - dst_address = 1; + dst_address = 1; else if (unformat (input, "proto")) - protocol = 1; - + protocol = 1; + #define _(a) else if (unformat (input, #a)) a=1; foreach_ip4_proto_field #undef _ - else - break; + else + break; } - + #define _(a) found_something += a; foreach_ip4_proto_field; #undef _ - + if (found_something == 0) return 0; - + vec_validate (mask, sizeof (*ip) - 1); - + ip = (ip4_header_t *) mask; - + #define _(a) if (a) memset (&ip->a, 0xff, sizeof (ip->a)); foreach_ip4_proto_field; #undef _ - + ip->ip_version_and_header_length = 0; - + if (version) ip->ip_version_and_header_length |= 0xF0; - + if (hdr_length) ip->ip_version_and_header_length |= 0x0F; - + *maskp = mask; return 1; } @@ -972,60 +996,61 @@ _(payload_length) \ _(hop_limit) \ _(protocol) -uword unformat_ip6_mask (unformat_input_t * input, va_list * args) +uword +unformat_ip6_mask (unformat_input_t * input, va_list * args) { - u8 ** maskp = va_arg (*args, u8 **); - u8 * mask = 0; + u8 **maskp = va_arg (*args, u8 **); + u8 *mask = 0; u8 found_something = 0; - ip6_header_t * ip; + ip6_header_t *ip; u32 ip_version_traffic_class_and_flow_label; - + #define _(a) u8 a=0; foreach_ip6_proto_field; #undef _ u8 version = 0; u8 traffic_class = 0; u8 flow_label = 0; - - while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) + + while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) { - if (unformat (input, "version")) - version = 1; + if (unformat (input, "version")) + version = 1; else if (unformat (input, "traffic-class")) - traffic_class = 1; + traffic_class = 1; else if (unformat (input, "flow-label")) - flow_label = 1; + flow_label = 1; else if (unformat (input, "src")) - src_address = 1; + src_address = 1; else if (unformat (input, "dst")) - dst_address = 1; + dst_address = 1; else if (unformat (input, "proto")) - protocol = 1; - + protocol = 1; + #define _(a) else if (unformat (input, #a)) a=1; foreach_ip6_proto_field #undef _ - else - break; + else + break; } - + #define _(a) found_something += a; foreach_ip6_proto_field; #undef _ - + if (found_something == 0) return 0; - + vec_validate (mask, sizeof (*ip) - 1); - + ip = (ip6_header_t *) mask; - + #define _(a) if (a) memset (&ip->a, 0xff, sizeof (ip->a)); foreach_ip6_proto_field; #undef _ - + ip_version_traffic_class_and_flow_label = 0; - + if (version) ip_version_traffic_class_and_flow_label |= 0xF0000000; @@ -1035,32 +1060,35 @@ uword unformat_ip6_mask (unformat_input_t * input, va_list * args) if (flow_label) ip_version_traffic_class_and_flow_label |= 0x000FFFFF; - ip->ip_version_traffic_class_and_flow_label = + ip->ip_version_traffic_class_and_flow_label = clib_host_to_net_u32 (ip_version_traffic_class_and_flow_label); - + *maskp = mask; return 1; } -uword unformat_l3_mask (unformat_input_t * input, va_list * args) +uword +unformat_l3_mask (unformat_input_t * input, va_list * args) { - u8 ** maskp = va_arg (*args, u8 **); + u8 **maskp = va_arg (*args, u8 **); - while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) { - if (unformat (input, "ip4 %U", unformat_ip4_mask, maskp)) - return 1; - else if (unformat (input, "ip6 %U", unformat_ip6_mask, maskp)) - return 1; - else - break; - } + while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) + { + if (unformat (input, "ip4 %U", unformat_ip4_mask, maskp)) + return 1; + else if (unformat (input, "ip6 %U", unformat_ip6_mask, maskp)) + return 1; + else + break; + } return 0; } -uword unformat_l2_mask (unformat_input_t * input, va_list * args) +uword +unformat_l2_mask (unformat_input_t * input, va_list * args) { - u8 ** maskp = va_arg (*args, u8 **); - u8 * mask = 0; + u8 **maskp = va_arg (*args, u8 **); + u8 *mask = 0; u8 src = 0; u8 dst = 0; u8 proto = 0; @@ -1074,34 +1102,35 @@ uword unformat_l2_mask (unformat_input_t * input, va_list * args) u8 dot1ad = 0; int len = 14; - while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) { - if (unformat (input, "src")) - src = 1; - else if (unformat (input, "dst")) - dst = 1; - else if (unformat (input, "proto")) - proto = 1; - else if (unformat (input, "tag1")) - tag1 = 1; - else if (unformat (input, "tag2")) - tag2 = 1; - else if (unformat (input, "ignore-tag1")) - ignore_tag1 = 1; - else if (unformat (input, "ignore-tag2")) - ignore_tag2 = 1; - else if (unformat (input, "cos1")) - cos1 = 1; - else if (unformat (input, "cos2")) - cos2 = 1; - else if (unformat (input, "dot1q")) - dot1q = 1; - else if (unformat (input, "dot1ad")) - dot1ad = 1; - else - break; - } + while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) + { + if (unformat (input, "src")) + src = 1; + else if (unformat (input, "dst")) + dst = 1; + else if (unformat (input, "proto")) + proto = 1; + else if (unformat (input, "tag1")) + tag1 = 1; + else if (unformat (input, "tag2")) + tag2 = 1; + else if (unformat (input, "ignore-tag1")) + ignore_tag1 = 1; + else if (unformat (input, "ignore-tag2")) + ignore_tag2 = 1; + else if (unformat (input, "cos1")) + cos1 = 1; + else if (unformat (input, "cos2")) + cos2 = 1; + else if (unformat (input, "dot1q")) + dot1q = 1; + else if (unformat (input, "dot1ad")) + dot1ad = 1; + else + break; + } if ((src + dst + proto + tag1 + tag2 + dot1q + dot1ad + - ignore_tag1 + ignore_tag2 + cos1 + cos2) == 0) + ignore_tag1 + ignore_tag2 + cos1 + cos2) == 0) return 0; if (tag1 || ignore_tag1 || cos1 || dot1q) @@ -1109,47 +1138,47 @@ uword unformat_l2_mask (unformat_input_t * input, va_list * args) if (tag2 || ignore_tag2 || cos2 || dot1ad) len = 22; - vec_validate (mask, len-1); + vec_validate (mask, len - 1); if (dst) memset (mask, 0xff, 6); if (src) memset (mask + 6, 0xff, 6); - + if (tag2 || dot1ad) { /* inner vlan tag */ if (tag2) - { - mask[19] = 0xff; - mask[18] = 0x0f; - } + { + mask[19] = 0xff; + mask[18] = 0x0f; + } if (cos2) - mask[18] |= 0xe0; + mask[18] |= 0xe0; if (proto) - mask[21] = mask [20] = 0xff; + mask[21] = mask[20] = 0xff; if (tag1) - { - mask [15] = 0xff; - mask [14] = 0x0f; - } + { + mask[15] = 0xff; + mask[14] = 0x0f; + } if (cos1) - mask[14] |= 0xe0; + mask[14] |= 0xe0; *maskp = mask; return 1; } if (tag1 | dot1q) { if (tag1) - { - mask [15] = 0xff; - mask [14] = 0x0f; - } + { + mask[15] = 0xff; + mask[14] = 0x0f; + } if (cos1) - mask[14] |= 0xe0; + mask[14] |= 0xe0; if (proto) - mask[16] = mask [17] = 0xff; + mask[16] = mask[17] = 0xff; *maskp = mask; return 1; } @@ -1158,90 +1187,93 @@ uword unformat_l2_mask (unformat_input_t * input, va_list * args) if (cos1) mask[14] |= 0xe0; if (proto) - mask[12] = mask [13] = 0xff; - + mask[12] = mask[13] = 0xff; + *maskp = mask; return 1; } -uword unformat_classify_mask (unformat_input_t * input, va_list * args) +uword +unformat_classify_mask (unformat_input_t * input, va_list * args) { - u8 ** maskp = va_arg (*args, u8 **); - u32 * skipp = va_arg (*args, u32 *); - u32 * matchp = va_arg (*args, u32 *); + u8 **maskp = va_arg (*args, u8 **); + u32 *skipp = va_arg (*args, u32 *); + u32 *matchp = va_arg (*args, u32 *); u32 match; - u8 * mask = 0; - u8 * l2 = 0; - u8 * l3 = 0; - u8 * l4 = 0; + u8 *mask = 0; + u8 *l2 = 0; + u8 *l3 = 0; + u8 *l4 = 0; int i; - - while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) { - if (unformat (input, "hex %U", unformat_hex_string, &mask)) - ; - else if (unformat (input, "l2 %U", unformat_l2_mask, &l2)) - ; - else if (unformat (input, "l3 %U", unformat_l3_mask, &l3)) - ; - else if (unformat (input, "l4 %U", unformat_l4_mask, &l4)) - ; - else - break; - } - if (l4 && !l3) { - vec_free (mask); - vec_free (l2); - vec_free (l4); - return 0; - } + while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) + { + if (unformat (input, "hex %U", unformat_hex_string, &mask)) + ; + else if (unformat (input, "l2 %U", unformat_l2_mask, &l2)) + ; + else if (unformat (input, "l3 %U", unformat_l3_mask, &l3)) + ; + else if (unformat (input, "l4 %U", unformat_l4_mask, &l4)) + ; + else + break; + } + + if (l4 && !l3) + { + vec_free (mask); + vec_free (l2); + vec_free (l4); + return 0; + } if (mask || l2 || l3 || l4) { if (l2 || l3 || l4) - { - /* "With a free Ethernet header in every package" */ - if (l2 == 0) - vec_validate (l2, 13); - mask = l2; - if (l3) - { - vec_append (mask, l3); - vec_free (l3); - } - if (l4) - { - vec_append (mask, l4); - vec_free (l4); - } - } + { + /* "With a free Ethernet header in every package" */ + if (l2 == 0) + vec_validate (l2, 13); + mask = l2; + if (l3) + { + vec_append (mask, l3); + vec_free (l3); + } + if (l4) + { + vec_append (mask, l4); + vec_free (l4); + } + } /* Scan forward looking for the first significant mask octet */ for (i = 0; i < vec_len (mask); i++) - if (mask[i]) - break; + if (mask[i]) + break; /* compute (skip, match) params */ - *skipp = i / sizeof(u32x4); - vec_delete (mask, *skipp * sizeof(u32x4), 0); + *skipp = i / sizeof (u32x4); + vec_delete (mask, *skipp * sizeof (u32x4), 0); /* Pad mask to an even multiple of the vector size */ while (vec_len (mask) % sizeof (u32x4)) - vec_add1 (mask, 0); + vec_add1 (mask, 0); match = vec_len (mask) / sizeof (u32x4); - for (i = match*sizeof(u32x4); i > 0; i-= sizeof(u32x4)) - { - u64 *tmp = (u64 *)(mask + (i-sizeof(u32x4))); - if (*tmp || *(tmp+1)) - break; - match--; - } + for (i = match * sizeof (u32x4); i > 0; i -= sizeof (u32x4)) + { + u64 *tmp = (u64 *) (mask + (i - sizeof (u32x4))); + if (*tmp || *(tmp + 1)) + break; + match--; + } if (match == 0) - clib_warning ("BUG: match 0"); + clib_warning ("BUG: match 0"); - _vec_len (mask) = match * sizeof(u32x4); + _vec_len (mask) = match * sizeof (u32x4); *matchp = match; *maskp = mask; @@ -1259,38 +1291,39 @@ _(ip4, IP4_INPUT) \ _(ip6, IP6_INPUT) \ _(li, LI) -uword unformat_l2_input_next_index (unformat_input_t * input, va_list * args) +uword +unformat_l2_input_next_index (unformat_input_t * input, va_list * args) { - vnet_classify_main_t * cm = &vnet_classify_main; - u32 * miss_next_indexp = va_arg (*args, u32 *); + vnet_classify_main_t *cm = &vnet_classify_main; + u32 *miss_next_indexp = va_arg (*args, u32 *); u32 next_index = 0; u32 tmp; int i; - + /* First try registered unformat fns, allowing override... */ for (i = 0; i < vec_len (cm->unformat_l2_next_index_fns); i++) { if (unformat (input, "%U", cm->unformat_l2_next_index_fns[i], &tmp)) - { - next_index = tmp; - goto out; - } + { + next_index = tmp; + goto out; + } } #define _(n,N) \ if (unformat (input, #n)) { next_index = L2_INPUT_CLASSIFY_NEXT_##N; goto out;} foreach_l2_input_next; #undef _ - + if (unformat (input, "%d", &tmp)) - { - next_index = tmp; - goto out; + { + next_index = tmp; + goto out; } - + return 0; - out: +out: *miss_next_indexp = next_index; return 1; } @@ -1298,38 +1331,39 @@ uword unformat_l2_input_next_index (unformat_input_t * input, va_list * args) #define foreach_l2_output_next \ _(drop, DROP) -uword unformat_l2_output_next_index (unformat_input_t * input, va_list * args) +uword +unformat_l2_output_next_index (unformat_input_t * input, va_list * args) { - vnet_classify_main_t * cm = &vnet_classify_main; - u32 * miss_next_indexp = va_arg (*args, u32 *); + vnet_classify_main_t *cm = &vnet_classify_main; + u32 *miss_next_indexp = va_arg (*args, u32 *); u32 next_index = 0; u32 tmp; int i; - + /* First try registered unformat fns, allowing override... */ for (i = 0; i < vec_len (cm->unformat_l2_next_index_fns); i++) { if (unformat (input, "%U", cm->unformat_l2_next_index_fns[i], &tmp)) - { - next_index = tmp; - goto out; - } + { + next_index = tmp; + goto out; + } } #define _(n,N) \ if (unformat (input, #n)) { next_index = L2_OUTPUT_CLASSIFY_NEXT_##N; goto out;} foreach_l2_output_next; #undef _ - + if (unformat (input, "%d", &tmp)) - { - next_index = tmp; - goto out; + { + next_index = tmp; + goto out; } - + return 0; - out: +out: *miss_next_indexp = next_index; return 1; } @@ -1338,38 +1372,39 @@ uword unformat_l2_output_next_index (unformat_input_t * input, va_list * args) _(drop, DROP) \ _(rewrite, REWRITE) -uword unformat_ip_next_index (unformat_input_t * input, va_list * args) +uword +unformat_ip_next_index (unformat_input_t * input, va_list * args) { - u32 * miss_next_indexp = va_arg (*args, u32 *); - vnet_classify_main_t * cm = &vnet_classify_main; + u32 *miss_next_indexp = va_arg (*args, u32 *); + vnet_classify_main_t *cm = &vnet_classify_main; u32 next_index = 0; u32 tmp; int i; - + /* First try registered unformat fns, allowing override... */ for (i = 0; i < vec_len (cm->unformat_ip_next_index_fns); i++) { if (unformat (input, "%U", cm->unformat_ip_next_index_fns[i], &tmp)) - { - next_index = tmp; - goto out; - } + { + next_index = tmp; + goto out; + } } #define _(n,N) \ if (unformat (input, #n)) { next_index = IP_LOOKUP_NEXT_##N; goto out;} foreach_ip_next; #undef _ - + if (unformat (input, "%d", &tmp)) - { - next_index = tmp; - goto out; + { + next_index = tmp; + goto out; } - + return 0; - out: +out: *miss_next_indexp = next_index; return 1; } @@ -1377,10 +1412,11 @@ uword unformat_ip_next_index (unformat_input_t * input, va_list * args) #define foreach_acl_next \ _(deny, DENY) -uword unformat_acl_next_index (unformat_input_t * input, va_list * args) +uword +unformat_acl_next_index (unformat_input_t * input, va_list * args) { - u32 * next_indexp = va_arg (*args, u32 *); - vnet_classify_main_t * cm = &vnet_classify_main; + u32 *next_indexp = va_arg (*args, u32 *); + vnet_classify_main_t *cm = &vnet_classify_main; u32 next_index = 0; u32 tmp; int i; @@ -1389,10 +1425,10 @@ uword unformat_acl_next_index (unformat_input_t * input, va_list * args) for (i = 0; i < vec_len (cm->unformat_acl_next_index_fns); i++) { if (unformat (input, "%U", cm->unformat_acl_next_index_fns[i], &tmp)) - { - next_index = tmp; - goto out; - } + { + next_index = tmp; + goto out; + } } #define _(n,N) \ @@ -1413,15 +1449,16 @@ uword unformat_acl_next_index (unformat_input_t * input, va_list * args) return 0; - out: +out: *next_indexp = next_index; return 1; } -uword unformat_policer_next_index (unformat_input_t * input, va_list * args) +uword +unformat_policer_next_index (unformat_input_t * input, va_list * args) { - u32 * next_indexp = va_arg (*args, u32 *); - vnet_classify_main_t * cm = &vnet_classify_main; + u32 *next_indexp = va_arg (*args, u32 *); + vnet_classify_main_t *cm = &vnet_classify_main; u32 next_index = 0; u32 tmp; int i; @@ -1429,11 +1466,12 @@ uword unformat_policer_next_index (unformat_input_t * input, va_list * args) /* First try registered unformat fns, allowing override... */ for (i = 0; i < vec_len (cm->unformat_policer_next_index_fns); i++) { - if (unformat (input, "%U", cm->unformat_policer_next_index_fns[i], &tmp)) - { - next_index = tmp; - goto out; - } + if (unformat + (input, "%U", cm->unformat_policer_next_index_fns[i], &tmp)) + { + next_index = tmp; + goto out; + } } if (unformat (input, "%d", &tmp)) @@ -1444,15 +1482,14 @@ uword unformat_policer_next_index (unformat_input_t * input, va_list * args) return 0; - out: +out: *next_indexp = next_index; return 1; } static clib_error_t * classify_table_command_fn (vlib_main_t * vm, - unformat_input_t * input, - vlib_cli_command_t * cmd) + unformat_input_t * input, vlib_cli_command_t * cmd) { u32 nbuckets = 2; u32 skip = ~0; @@ -1462,61 +1499,67 @@ classify_table_command_fn (vlib_main_t * vm, u32 table_index = ~0; u32 next_table_index = ~0; u32 miss_next_index = ~0; - u32 memory_size = 2<<20; + u32 memory_size = 2 << 20; u32 tmp; u32 current_data_flag = 0; int current_data_offset = 0; - u8 * mask = 0; - vnet_classify_main_t * cm = &vnet_classify_main; + u8 *mask = 0; + vnet_classify_main_t *cm = &vnet_classify_main; int rv; - while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) { - if (unformat (input, "del")) - is_add = 0; - else if (unformat (input, "del-chain")) - { + while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) + { + if (unformat (input, "del")) is_add = 0; - del_chain = 1; - } - else if (unformat (input, "buckets %d", &nbuckets)) - ; - else if (unformat (input, "skip %d", &skip)) - ; - else if (unformat (input, "match %d", &match)) - ; - else if (unformat (input, "table %d", &table_index)) - ; - else if (unformat (input, "mask %U", unformat_classify_mask, - &mask, &skip, &match)) - ; - else if (unformat (input, "memory-size %uM", &tmp)) - memory_size = tmp<<20; - else if (unformat (input, "memory-size %uG", &tmp)) - memory_size = tmp<<30; - else if (unformat (input, "next-table %d", &next_table_index)) - ; - else if (unformat (input, "miss-next %U", unformat_ip_next_index, - &miss_next_index)) - ; - else if (unformat (input, "l2-input-miss-next %U", unformat_l2_input_next_index, - &miss_next_index)) - ; - else if (unformat (input, "l2-output-miss-next %U", unformat_l2_output_next_index, - &miss_next_index)) - ; - else if (unformat (input, "acl-miss-next %U", unformat_acl_next_index, - &miss_next_index)) - ; - else if (unformat (input, "current-data-flag %d", ¤t_data_flag)) - ; - else if (unformat (input, "current-data-offset %d", ¤t_data_offset)) - ; + else if (unformat (input, "del-chain")) + { + is_add = 0; + del_chain = 1; + } + else if (unformat (input, "buckets %d", &nbuckets)) + ; + else if (unformat (input, "skip %d", &skip)) + ; + else if (unformat (input, "match %d", &match)) + ; + else if (unformat (input, "table %d", &table_index)) + ; + else if (unformat (input, "mask %U", unformat_classify_mask, + &mask, &skip, &match)) + ; + else if (unformat (input, "memory-size %uM", &tmp)) + memory_size = tmp << 20; + else if (unformat (input, "memory-size %uG", &tmp)) + memory_size = tmp << 30; + else if (unformat (input, "next-table %d", &next_table_index)) + ; + else if (unformat (input, "miss-next %U", unformat_ip_next_index, + &miss_next_index)) + ; + else + if (unformat + (input, "l2-input-miss-next %U", unformat_l2_input_next_index, + &miss_next_index)) + ; + else + if (unformat + (input, "l2-output-miss-next %U", unformat_l2_output_next_index, + &miss_next_index)) + ; + else if (unformat (input, "acl-miss-next %U", unformat_acl_next_index, + &miss_next_index)) + ; + else if (unformat (input, "current-data-flag %d", ¤t_data_flag)) + ; + else + if (unformat (input, "current-data-offset %d", ¤t_data_offset)) + ; + + else + break; + } - else - break; - } - if (is_add && mask == 0 && table_index == ~0) return clib_error_return (0, "Mask required"); @@ -1525,13 +1568,15 @@ classify_table_command_fn (vlib_main_t * vm, if (is_add && match == ~0 && table_index == ~0) return clib_error_return (0, "match count required"); - + if (!is_add && table_index == ~0) return clib_error_return (0, "table index required for delete"); rv = vnet_classify_add_del_table (cm, mask, nbuckets, memory_size, - skip, match, next_table_index, miss_next_index, &table_index, - current_data_flag, current_data_offset, is_add, del_chain); + skip, match, next_table_index, + miss_next_index, &table_index, + current_data_flag, current_data_offset, + is_add, del_chain); switch (rv) { case 0: @@ -1539,14 +1584,15 @@ classify_table_command_fn (vlib_main_t * vm, default: return clib_error_return (0, "vnet_classify_add_del_table returned %d", - rv); + rv); } return 0; } +/* *INDENT-OFF* */ VLIB_CLI_COMMAND (classify_table, static) = { .path = "classify table", - .short_help = + .short_help = "classify table [miss-next|l2-miss_next|acl-miss-next ]" "\n mask buckets [skip ] [match ]" "\n [current-data-flag ] [current-data-offset ] [table ]" @@ -1554,79 +1600,83 @@ VLIB_CLI_COMMAND (classify_table, static) = { "\n [del] [del-chain]", .function = classify_table_command_fn, }; +/* *INDENT-ON* */ -static u8 * format_vnet_classify_table (u8 * s, va_list * args) +static u8 * +format_vnet_classify_table (u8 * s, va_list * args) { - vnet_classify_main_t * cm = va_arg (*args, vnet_classify_main_t *); + vnet_classify_main_t *cm = va_arg (*args, vnet_classify_main_t *); int verbose = va_arg (*args, int); u32 index = va_arg (*args, u32); - vnet_classify_table_t * t; + vnet_classify_table_t *t; if (index == ~0) { s = format (s, "%10s%10s%10s%10s", "TableIdx", "Sessions", "NextTbl", - "NextNode", verbose ? "Details" : ""); + "NextNode", verbose ? "Details" : ""); return s; } t = pool_elt_at_index (cm->tables, index); s = format (s, "%10u%10d%10d%10d", index, t->active_elements, - t->next_table_index, t->miss_next_index); + t->next_table_index, t->miss_next_index); - s = format (s, "\n Heap: %U", format_mheap, t->mheap, 0 /*verbose*/); + s = format (s, "\n Heap: %U", format_mheap, t->mheap, 0 /*verbose */ ); s = format (s, "\n nbuckets %d, skip %d match %d flag %d offset %d", - t->nbuckets, t->skip_n_vectors, t->match_n_vectors, - t->current_data_flag, t->current_data_offset); - s = format (s, "\n mask %U", format_hex_bytes, t->mask, - t->match_n_vectors * sizeof (u32x4)); + t->nbuckets, t->skip_n_vectors, t->match_n_vectors, + t->current_data_flag, t->current_data_offset); + s = format (s, "\n mask %U", format_hex_bytes, t->mask, + t->match_n_vectors * sizeof (u32x4)); s = format (s, "\n linear-search buckets %d\n", t->linear_buckets); if (verbose == 0) return s; s = format (s, "\n%U", format_classify_table, t, verbose); - + return s; } static clib_error_t * show_classify_tables_command_fn (vlib_main_t * vm, - unformat_input_t * input, - vlib_cli_command_t * cmd) + unformat_input_t * input, + vlib_cli_command_t * cmd) { - vnet_classify_main_t * cm = &vnet_classify_main; - vnet_classify_table_t * t; + vnet_classify_main_t *cm = &vnet_classify_main; + vnet_classify_table_t *t; u32 match_index = ~0; - u32 * indices = 0; + u32 *indices = 0; int verbose = 0; int i; - while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) + while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) { if (unformat (input, "index %d", &match_index)) - ; + ; else if (unformat (input, "verbose %d", &verbose)) - ; + ; else if (unformat (input, "verbose")) - verbose = 1; - else - break; + verbose = 1; + else + break; } - - pool_foreach (t, cm->tables, + + /* *INDENT-OFF* */ + pool_foreach (t, cm->tables, ({ if (match_index == ~0 || (match_index == t - cm->tables)) vec_add1 (indices, t - cm->tables); })); + /* *INDENT-ON* */ - if (vec_len(indices)) + if (vec_len (indices)) { vlib_cli_output (vm, "%U", format_vnet_classify_table, cm, verbose, - ~0 /* hdr */); + ~0 /* hdr */ ); for (i = 0; i < vec_len (indices); i++) - vlib_cli_output (vm, "%U", format_vnet_classify_table, cm, - verbose, indices[i]); + vlib_cli_output (vm, "%U", format_vnet_classify_table, cm, + verbose, indices[i]); } else vlib_cli_output (vm, "No classifier tables configured"); @@ -1636,17 +1686,20 @@ show_classify_tables_command_fn (vlib_main_t * vm, return 0; } +/* *INDENT-OFF* */ VLIB_CLI_COMMAND (show_classify_table_command, static) = { .path = "show classify tables", .short_help = "show classify tables [index ]", .function = show_classify_tables_command_fn, }; +/* *INDENT-ON* */ -uword unformat_l4_match (unformat_input_t * input, va_list * args) +uword +unformat_l4_match (unformat_input_t * input, va_list * args) { - u8 ** matchp = va_arg (*args, u8 **); + u8 **matchp = va_arg (*args, u8 **); - u8 * proto_header = 0; + u8 *proto_header = 0; int src_port = 0; int dst_port = 0; @@ -1655,28 +1708,29 @@ uword unformat_l4_match (unformat_input_t * input, va_list * args) while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) { if (unformat (input, "src_port %d", &src_port)) - ; + ; else if (unformat (input, "dst_port %d", &dst_port)) - ; + ; else - return 0; + return 0; } - h.src_port = clib_host_to_net_u16(src_port); - h.dst_port = clib_host_to_net_u16(dst_port); - vec_validate(proto_header, sizeof(h)-1); - memcpy(proto_header, &h, sizeof(h)); + h.src_port = clib_host_to_net_u16 (src_port); + h.dst_port = clib_host_to_net_u16 (dst_port); + vec_validate (proto_header, sizeof (h) - 1); + memcpy (proto_header, &h, sizeof (h)); *matchp = proto_header; return 1; } -uword unformat_ip4_match (unformat_input_t * input, va_list * args) +uword +unformat_ip4_match (unformat_input_t * input, va_list * args) { - u8 ** matchp = va_arg (*args, u8 **); - u8 * match = 0; - ip4_header_t * ip; + u8 **matchp = va_arg (*args, u8 **); + u8 *match = 0; + ip4_header_t *ip; int version = 0; u32 version_val; int hdr_length = 0; @@ -1696,67 +1750,67 @@ uword unformat_ip4_match (unformat_input_t * input, va_list * args) int checksum = 0; u32 checksum_val; - while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) + while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) { - if (unformat (input, "version %d", &version_val)) - version = 1; + if (unformat (input, "version %d", &version_val)) + version = 1; else if (unformat (input, "hdr_length %d", &hdr_length_val)) - hdr_length = 1; + hdr_length = 1; else if (unformat (input, "src %U", unformat_ip4_address, &src_val)) - src = 1; + src = 1; else if (unformat (input, "dst %U", unformat_ip4_address, &dst_val)) - dst = 1; + dst = 1; else if (unformat (input, "proto %d", &proto_val)) - proto = 1; + proto = 1; else if (unformat (input, "tos %d", &tos_val)) - tos = 1; + tos = 1; else if (unformat (input, "length %d", &length_val)) - length = 1; + length = 1; else if (unformat (input, "fragment_id %d", &fragment_id_val)) - fragment_id = 1; + fragment_id = 1; else if (unformat (input, "ttl %d", &ttl_val)) - ttl = 1; + ttl = 1; else if (unformat (input, "checksum %d", &checksum_val)) - checksum = 1; + checksum = 1; else - break; + break; } - + if (version + hdr_length + src + dst + proto + tos + length + fragment_id + ttl + checksum == 0) return 0; - /* + /* * Aligned because we use the real comparison functions */ - vec_validate_aligned (match, sizeof (*ip) - 1, sizeof(u32x4)); - + vec_validate_aligned (match, sizeof (*ip) - 1, sizeof (u32x4)); + ip = (ip4_header_t *) match; - + /* These are realistically matched in practice */ if (src) ip->src_address.as_u32 = src_val.as_u32; if (dst) ip->dst_address.as_u32 = dst_val.as_u32; - + if (proto) ip->protocol = proto_val; - + /* These are not, but they're included for completeness */ if (version) - ip->ip_version_and_header_length |= (version_val & 0xF)<<4; + ip->ip_version_and_header_length |= (version_val & 0xF) << 4; if (hdr_length) ip->ip_version_and_header_length |= (hdr_length_val & 0xF); - + if (tos) ip->tos = tos_val; - + if (length) ip->length = clib_host_to_net_u16 (length_val); - + if (ttl) ip->ttl = ttl_val; @@ -1767,17 +1821,18 @@ uword unformat_ip4_match (unformat_input_t * input, va_list * args) return 1; } -uword unformat_ip6_match (unformat_input_t * input, va_list * args) +uword +unformat_ip6_match (unformat_input_t * input, va_list * args) { - u8 ** matchp = va_arg (*args, u8 **); - u8 * match = 0; - ip6_header_t * ip; + u8 **matchp = va_arg (*args, u8 **); + u8 *match = 0; + ip6_header_t *ip; int version = 0; u32 version_val; - u8 traffic_class = 0; + u8 traffic_class = 0; u32 traffic_class_val; - u8 flow_label = 0; - u8 flow_label_val; + u8 flow_label = 0; + u8 flow_label_val; int src = 0, dst = 0; ip6_address_t src_val, dst_val; int proto = 0; @@ -1788,65 +1843,66 @@ uword unformat_ip6_match (unformat_input_t * input, va_list * args) int hop_limit_val; u32 ip_version_traffic_class_and_flow_label; - while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) + while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) { - if (unformat (input, "version %d", &version_val)) - version = 1; + if (unformat (input, "version %d", &version_val)) + version = 1; else if (unformat (input, "traffic_class %d", &traffic_class_val)) - traffic_class = 1; + traffic_class = 1; else if (unformat (input, "flow_label %d", &flow_label_val)) - flow_label = 1; + flow_label = 1; else if (unformat (input, "src %U", unformat_ip6_address, &src_val)) - src = 1; + src = 1; else if (unformat (input, "dst %U", unformat_ip6_address, &dst_val)) - dst = 1; + dst = 1; else if (unformat (input, "proto %d", &proto_val)) - proto = 1; + proto = 1; else if (unformat (input, "payload_length %d", &payload_length_val)) - payload_length = 1; + payload_length = 1; else if (unformat (input, "hop_limit %d", &hop_limit_val)) - hop_limit = 1; + hop_limit = 1; else - break; + break; } - + if (version + traffic_class + flow_label + src + dst + proto + payload_length + hop_limit == 0) return 0; - /* + /* * Aligned because we use the real comparison functions */ - vec_validate_aligned (match, sizeof (*ip) - 1, sizeof(u32x4)); - + vec_validate_aligned (match, sizeof (*ip) - 1, sizeof (u32x4)); + ip = (ip6_header_t *) match; - + if (src) clib_memcpy (&ip->src_address, &src_val, sizeof (ip->src_address)); if (dst) clib_memcpy (&ip->dst_address, &dst_val, sizeof (ip->dst_address)); - + if (proto) ip->protocol = proto_val; - + ip_version_traffic_class_and_flow_label = 0; if (version) ip_version_traffic_class_and_flow_label |= (version_val & 0xF) << 28; if (traffic_class) - ip_version_traffic_class_and_flow_label |= (traffic_class_val & 0xFF) << 20; + ip_version_traffic_class_and_flow_label |= + (traffic_class_val & 0xFF) << 20; if (flow_label) ip_version_traffic_class_and_flow_label |= (flow_label_val & 0xFFFFF); - - ip->ip_version_traffic_class_and_flow_label = + + ip->ip_version_traffic_class_and_flow_label = clib_host_to_net_u32 (ip_version_traffic_class_and_flow_label); if (payload_length) ip->payload_length = clib_host_to_net_u16 (payload_length_val); - + if (hop_limit) ip->hop_limit = hop_limit_val; @@ -1854,30 +1910,33 @@ uword unformat_ip6_match (unformat_input_t * input, va_list * args) return 1; } -uword unformat_l3_match (unformat_input_t * input, va_list * args) +uword +unformat_l3_match (unformat_input_t * input, va_list * args) { - u8 ** matchp = va_arg (*args, u8 **); + u8 **matchp = va_arg (*args, u8 **); - while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) { - if (unformat (input, "ip4 %U", unformat_ip4_match, matchp)) - return 1; - else if (unformat (input, "ip6 %U", unformat_ip6_match, matchp)) - return 1; - /* $$$$ add mpls */ - else - break; - } + while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) + { + if (unformat (input, "ip4 %U", unformat_ip4_match, matchp)) + return 1; + else if (unformat (input, "ip6 %U", unformat_ip6_match, matchp)) + return 1; + /* $$$$ add mpls */ + else + break; + } return 0; } -uword unformat_vlan_tag (unformat_input_t * input, va_list * args) +uword +unformat_vlan_tag (unformat_input_t * input, va_list * args) { - u8 * tagp = va_arg (*args, u8 *); + u8 *tagp = va_arg (*args, u8 *); u32 tag; - if (unformat(input, "%d", &tag)) + if (unformat (input, "%d", &tag)) { - tagp[0] = (tag>>8) & 0x0F; + tagp[0] = (tag >> 8) & 0x0F; tagp[1] = tag & 0xFF; return 1; } @@ -1885,10 +1944,11 @@ uword unformat_vlan_tag (unformat_input_t * input, va_list * args) return 0; } -uword unformat_l2_match (unformat_input_t * input, va_list * args) +uword +unformat_l2_match (unformat_input_t * input, va_list * args) { - u8 ** matchp = va_arg (*args, u8 **); - u8 * match = 0; + u8 **matchp = va_arg (*args, u8 **); + u8 *match = 0; u8 src = 0; u8 src_val[6]; u8 dst = 0; @@ -1896,9 +1956,9 @@ uword unformat_l2_match (unformat_input_t * input, va_list * args) u8 proto = 0; u16 proto_val; u8 tag1 = 0; - u8 tag1_val [2]; + u8 tag1_val[2]; u8 tag2 = 0; - u8 tag2_val [2]; + u8 tag2_val[2]; int len = 14; u8 ignore_tag1 = 0; u8 ignore_tag2 = 0; @@ -1907,31 +1967,33 @@ uword unformat_l2_match (unformat_input_t * input, va_list * args) u32 cos1_val = 0; u32 cos2_val = 0; - while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) { - if (unformat (input, "src %U", unformat_ethernet_address, &src_val)) - src = 1; - else if (unformat (input, "dst %U", unformat_ethernet_address, &dst_val)) - dst = 1; - else if (unformat (input, "proto %U", - unformat_ethernet_type_host_byte_order, &proto_val)) - proto = 1; - else if (unformat (input, "tag1 %U", unformat_vlan_tag, tag1_val)) - tag1 = 1; - else if (unformat (input, "tag2 %U", unformat_vlan_tag, tag2_val)) - tag2 = 1; - else if (unformat (input, "ignore-tag1")) - ignore_tag1 = 1; - else if (unformat (input, "ignore-tag2")) - ignore_tag2 = 1; - else if (unformat (input, "cos1 %d", &cos1_val)) - cos1 = 1; - else if (unformat (input, "cos2 %d", &cos2_val)) - cos2 = 1; - else - break; - } + while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) + { + if (unformat (input, "src %U", unformat_ethernet_address, &src_val)) + src = 1; + else + if (unformat (input, "dst %U", unformat_ethernet_address, &dst_val)) + dst = 1; + else if (unformat (input, "proto %U", + unformat_ethernet_type_host_byte_order, &proto_val)) + proto = 1; + else if (unformat (input, "tag1 %U", unformat_vlan_tag, tag1_val)) + tag1 = 1; + else if (unformat (input, "tag2 %U", unformat_vlan_tag, tag2_val)) + tag2 = 1; + else if (unformat (input, "ignore-tag1")) + ignore_tag1 = 1; + else if (unformat (input, "ignore-tag2")) + ignore_tag2 = 1; + else if (unformat (input, "cos1 %d", &cos1_val)) + cos1 = 1; + else if (unformat (input, "cos2 %d", &cos2_val)) + cos2 = 1; + else + break; + } if ((src + dst + proto + tag1 + tag2 + - ignore_tag1 + ignore_tag2 + cos1 + cos2) == 0) + ignore_tag1 + ignore_tag2 + cos1 + cos2) == 0) return 0; if (tag1 || ignore_tag1 || cos1) @@ -1939,130 +2001,135 @@ uword unformat_l2_match (unformat_input_t * input, va_list * args) if (tag2 || ignore_tag2 || cos2) len = 22; - vec_validate_aligned (match, len-1, sizeof(u32x4)); + vec_validate_aligned (match, len - 1, sizeof (u32x4)); if (dst) clib_memcpy (match, dst_val, 6); if (src) clib_memcpy (match + 6, src_val, 6); - + if (tag2) { /* inner vlan tag */ match[19] = tag2_val[1]; match[18] = tag2_val[0]; if (cos2) - match [18] |= (cos2_val & 0x7) << 5; + match[18] |= (cos2_val & 0x7) << 5; if (proto) - { - match[21] = proto_val & 0xff; - match[20] = proto_val >> 8; - } + { + match[21] = proto_val & 0xff; + match[20] = proto_val >> 8; + } if (tag1) - { - match [15] = tag1_val[1]; - match [14] = tag1_val[0]; - } + { + match[15] = tag1_val[1]; + match[14] = tag1_val[0]; + } if (cos1) - match [14] |= (cos1_val & 0x7) << 5; + match[14] |= (cos1_val & 0x7) << 5; *matchp = match; return 1; } if (tag1) { - match [15] = tag1_val[1]; - match [14] = tag1_val[0]; + match[15] = tag1_val[1]; + match[14] = tag1_val[0]; if (proto) - { - match[17] = proto_val & 0xff; - match[16] = proto_val >> 8; - } + { + match[17] = proto_val & 0xff; + match[16] = proto_val >> 8; + } if (cos1) - match [14] |= (cos1_val & 0x7) << 5; + match[14] |= (cos1_val & 0x7) << 5; *matchp = match; return 1; } if (cos2) - match [18] |= (cos2_val & 0x7) << 5; + match[18] |= (cos2_val & 0x7) << 5; if (cos1) - match [14] |= (cos1_val & 0x7) << 5; + match[14] |= (cos1_val & 0x7) << 5; if (proto) { match[13] = proto_val & 0xff; match[12] = proto_val >> 8; } - + *matchp = match; return 1; } -uword unformat_classify_match (unformat_input_t * input, va_list * args) +uword +unformat_classify_match (unformat_input_t * input, va_list * args) { - vnet_classify_main_t * cm = va_arg (*args, vnet_classify_main_t *); - u8 ** matchp = va_arg (*args, u8 **); + vnet_classify_main_t *cm = va_arg (*args, vnet_classify_main_t *); + u8 **matchp = va_arg (*args, u8 **); u32 table_index = va_arg (*args, u32); - vnet_classify_table_t * t; - - u8 * match = 0; - u8 * l2 = 0; - u8 * l3 = 0; - u8 * l4 = 0; + vnet_classify_table_t *t; + + u8 *match = 0; + u8 *l2 = 0; + u8 *l3 = 0; + u8 *l4 = 0; if (pool_is_free_index (cm->tables, table_index)) return 0; t = pool_elt_at_index (cm->tables, table_index); - while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) { - if (unformat (input, "hex %U", unformat_hex_string, &match)) - ; - else if (unformat (input, "l2 %U", unformat_l2_match, &l2)) - ; - else if (unformat (input, "l3 %U", unformat_l3_match, &l3)) - ; - else if (unformat (input, "l4 %U", unformat_l4_match, &l4)) - ; - else - break; - } + while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) + { + if (unformat (input, "hex %U", unformat_hex_string, &match)) + ; + else if (unformat (input, "l2 %U", unformat_l2_match, &l2)) + ; + else if (unformat (input, "l3 %U", unformat_l3_match, &l3)) + ; + else if (unformat (input, "l4 %U", unformat_l4_match, &l4)) + ; + else + break; + } - if (l4 && !l3) { - vec_free (match); - vec_free (l2); - vec_free (l4); - return 0; - } + if (l4 && !l3) + { + vec_free (match); + vec_free (l2); + vec_free (l4); + return 0; + } if (match || l2 || l3 || l4) { if (l2 || l3 || l4) - { - /* "Win a free Ethernet header in every packet" */ - if (l2 == 0) - vec_validate_aligned (l2, 13, sizeof(u32x4)); - match = l2; - if (l3) - { - vec_append_aligned (match, l3, sizeof(u32x4)); - vec_free (l3); - } - if (l4) - { - vec_append_aligned (match, l4, sizeof(u32x4)); - vec_free (l4); - } - } + { + /* "Win a free Ethernet header in every packet" */ + if (l2 == 0) + vec_validate_aligned (l2, 13, sizeof (u32x4)); + match = l2; + if (l3) + { + vec_append_aligned (match, l3, sizeof (u32x4)); + vec_free (l3); + } + if (l4) + { + vec_append_aligned (match, l4, sizeof (u32x4)); + vec_free (l4); + } + } /* Make sure the vector is big enough even if key is all 0's */ - vec_validate_aligned - (match, ((t->match_n_vectors + t->skip_n_vectors) * sizeof(u32x4)) - 1, - sizeof(u32x4)); - - /* Set size, include skipped vectors*/ - _vec_len (match) = (t->match_n_vectors+t->skip_n_vectors) * sizeof(u32x4); + vec_validate_aligned + (match, + ((t->match_n_vectors + t->skip_n_vectors) * sizeof (u32x4)) - 1, + sizeof (u32x4)); + + /* Set size, include skipped vectors */ + _vec_len (match) = + (t->match_n_vectors + t->skip_n_vectors) * sizeof (u32x4); *matchp = match; @@ -2072,27 +2139,26 @@ uword unformat_classify_match (unformat_input_t * input, va_list * args) return 0; } -int vnet_classify_add_del_session (vnet_classify_main_t * cm, - u32 table_index, - u8 * match, - u32 hit_next_index, - u32 opaque_index, - i32 advance, - u8 action, - u32 metadata, - int is_add) +int +vnet_classify_add_del_session (vnet_classify_main_t * cm, + u32 table_index, + u8 * match, + u32 hit_next_index, + u32 opaque_index, + i32 advance, + u8 action, u32 metadata, int is_add) { - vnet_classify_table_t * t; - vnet_classify_entry_5_t _max_e __attribute__((aligned (16))); - vnet_classify_entry_t * e; + vnet_classify_table_t *t; + vnet_classify_entry_5_t _max_e __attribute__ ((aligned (16))); + vnet_classify_entry_t *e; int i, rv; if (pool_is_free_index (cm->tables, table_index)) return VNET_API_ERROR_NO_SUCH_TABLE; - + t = pool_elt_at_index (cm->tables, table_index); - - e = (vnet_classify_entry_t *)&_max_e; + + e = (vnet_classify_entry_t *) & _max_e; e->next_index = hit_next_index; e->opaque_index = opaque_index; e->advance = advance; @@ -2102,12 +2168,12 @@ int vnet_classify_add_del_session (vnet_classify_main_t * cm, e->action = action; if (e->action == CLASSIFY_ACTION_SET_IP4_FIB_INDEX) e->metadata = fib_table_find_or_create_and_lock (FIB_PROTOCOL_IP4, - metadata, - FIB_SOURCE_CLASSIFY); + metadata, + FIB_SOURCE_CLASSIFY); else if (e->action == CLASSIFY_ACTION_SET_IP6_FIB_INDEX) e->metadata = fib_table_find_or_create_and_lock (FIB_PROTOCOL_IP6, - metadata, - FIB_SOURCE_CLASSIFY); + metadata, + FIB_SOURCE_CLASSIFY); else if (e->action == CLASSIFY_ACTION_SET_METADATA) e->metadata = metadata; else @@ -2115,7 +2181,7 @@ int vnet_classify_add_del_session (vnet_classify_main_t * cm, /* Copy key data, honoring skip_n_vectors */ clib_memcpy (&e->key, match + t->skip_n_vectors * sizeof (u32x4), - t->match_n_vectors * sizeof (u32x4)); + t->match_n_vectors * sizeof (u32x4)); /* Clear don't-care bits; likely when dynamically creating sessions */ for (i = 0; i < t->match_n_vectors; i++) @@ -2123,7 +2189,7 @@ int vnet_classify_add_del_session (vnet_classify_main_t * cm, rv = vnet_classify_add_del (t, e, is_add); - vnet_classify_entry_release_resource(e); + vnet_classify_entry_release_resource (e); if (rv) return VNET_API_ERROR_NO_SUCH_ENTRY; @@ -2132,65 +2198,69 @@ int vnet_classify_add_del_session (vnet_classify_main_t * cm, static clib_error_t * classify_session_command_fn (vlib_main_t * vm, - unformat_input_t * input, - vlib_cli_command_t * cmd) + unformat_input_t * input, + vlib_cli_command_t * cmd) { - vnet_classify_main_t * cm = &vnet_classify_main; + vnet_classify_main_t *cm = &vnet_classify_main; int is_add = 1; u32 table_index = ~0; u32 hit_next_index = ~0; u64 opaque_index = ~0; - u8 * match = 0; + u8 *match = 0; i32 advance = 0; u32 action = 0; u32 metadata = 0; int i, rv; - while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) + while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) { if (unformat (input, "del")) - is_add = 0; + is_add = 0; else if (unformat (input, "hit-next %U", unformat_ip_next_index, - &hit_next_index)) - ; - else if (unformat (input, "l2-input-hit-next %U", unformat_l2_input_next_index, - &hit_next_index)) - ; - else if (unformat (input, "l2-output-hit-next %U", unformat_l2_output_next_index, - &hit_next_index)) - ; + &hit_next_index)) + ; + else + if (unformat + (input, "l2-input-hit-next %U", unformat_l2_input_next_index, + &hit_next_index)) + ; + else + if (unformat + (input, "l2-output-hit-next %U", unformat_l2_output_next_index, + &hit_next_index)) + ; else if (unformat (input, "acl-hit-next %U", unformat_acl_next_index, - &hit_next_index)) - ; + &hit_next_index)) + ; else if (unformat (input, "policer-hit-next %U", - unformat_policer_next_index, &hit_next_index)) - ; + unformat_policer_next_index, &hit_next_index)) + ; else if (unformat (input, "opaque-index %lld", &opaque_index)) - ; + ; else if (unformat (input, "match %U", unformat_classify_match, - cm, &match, table_index)) - ; + cm, &match, table_index)) + ; else if (unformat (input, "advance %d", &advance)) - ; + ; else if (unformat (input, "table-index %d", &table_index)) - ; + ; else if (unformat (input, "action set-ip4-fib-id %d", &metadata)) - action = 1; + action = 1; else if (unformat (input, "action set-ip6-fib-id %d", &metadata)) - action = 2; + action = 2; else if (unformat (input, "action set-sr-policy-index %d", &metadata)) - action = 3; + action = 3; else - { - /* Try registered opaque-index unformat fns */ - for (i = 0; i < vec_len (cm->unformat_opaque_index_fns); i++) - { - if (unformat (input, "%U", cm->unformat_opaque_index_fns[i], - &opaque_index)) - goto found_opaque; - } - break; - } + { + /* Try registered opaque-index unformat fns */ + for (i = 0; i < vec_len (cm->unformat_opaque_index_fns); i++) + { + if (unformat (input, "%U", cm->unformat_opaque_index_fns[i], + &opaque_index)) + goto found_opaque; + } + break; + } found_opaque: ; } @@ -2201,24 +2271,26 @@ classify_session_command_fn (vlib_main_t * vm, if (is_add && match == 0) return clib_error_return (0, "Match value required"); - rv = vnet_classify_add_del_session (cm, table_index, match, - hit_next_index, - opaque_index, advance, - action, metadata, is_add); + rv = vnet_classify_add_del_session (cm, table_index, match, + hit_next_index, + opaque_index, advance, + action, metadata, is_add); - switch(rv) + switch (rv) { case 0: break; default: - return clib_error_return (0, "vnet_classify_add_del_session returned %d", - rv); + return clib_error_return (0, + "vnet_classify_add_del_session returned %d", + rv); } return 0; } +/* *INDENT-OFF* */ VLIB_CLI_COMMAND (classify_session_command, static) = { .path = "classify session", .short_help = @@ -2228,15 +2300,16 @@ VLIB_CLI_COMMAND (classify_session_command, static) = { "\n [action set-ip4-fib-id|set-ip6-fib-id|set-sr-policy-index ] [del]", .function = classify_session_command_fn, }; +/* *INDENT-ON* */ -static uword +static uword unformat_opaque_sw_if_index (unformat_input_t * input, va_list * args) { - u64 * opaquep = va_arg (*args, u64 *); + u64 *opaquep = va_arg (*args, u64 *); u32 sw_if_index; if (unformat (input, "opaque-sw_if_index %U", unformat_vnet_sw_interface, - vnet_get_main(), &sw_if_index)) + vnet_get_main (), &sw_if_index)) { *opaquep = sw_if_index; return 1; @@ -2247,13 +2320,13 @@ unformat_opaque_sw_if_index (unformat_input_t * input, va_list * args) static uword unformat_ip_next_node (unformat_input_t * input, va_list * args) { - vnet_classify_main_t * cm = &vnet_classify_main; - u32 * next_indexp = va_arg (*args, u32 *); + vnet_classify_main_t *cm = &vnet_classify_main; + u32 *next_indexp = va_arg (*args, u32 *); u32 node_index; u32 next_index = ~0; if (unformat (input, "ip6-node %U", unformat_vlib_node, - cm->vlib_main, &node_index)) + cm->vlib_main, &node_index)) { next_index = vlib_node_add_next (cm->vlib_main, ip6_classify_node.index, node_index); @@ -2271,16 +2344,16 @@ unformat_ip_next_node (unformat_input_t * input, va_list * args) return 1; } -static uword +static uword unformat_acl_next_node (unformat_input_t * input, va_list * args) { - vnet_classify_main_t * cm = &vnet_classify_main; - u32 * next_indexp = va_arg (*args, u32 *); + vnet_classify_main_t *cm = &vnet_classify_main; + u32 *next_indexp = va_arg (*args, u32 *); u32 node_index; u32 next_index; if (unformat (input, "ip6-node %U", unformat_vlib_node, - cm->vlib_main, &node_index)) + cm->vlib_main, &node_index)) { next_index = vlib_node_add_next (cm->vlib_main, ip6_inacl_node.index, node_index); @@ -2298,19 +2371,19 @@ unformat_acl_next_node (unformat_input_t * input, va_list * args) return 1; } -static uword +static uword unformat_l2_input_next_node (unformat_input_t * input, va_list * args) { - vnet_classify_main_t * cm = &vnet_classify_main; - u32 * next_indexp = va_arg (*args, u32 *); + vnet_classify_main_t *cm = &vnet_classify_main; + u32 *next_indexp = va_arg (*args, u32 *); u32 node_index; u32 next_index; if (unformat (input, "input-node %U", unformat_vlib_node, - cm->vlib_main, &node_index)) + cm->vlib_main, &node_index)) { - next_index = vlib_node_add_next - (cm->vlib_main, l2_input_classify_node.index, node_index); + next_index = vlib_node_add_next + (cm->vlib_main, l2_input_classify_node.index, node_index); *next_indexp = next_index; return 1; @@ -2318,19 +2391,19 @@ unformat_l2_input_next_node (unformat_input_t * input, va_list * args) return 0; } -static uword +static uword unformat_l2_output_next_node (unformat_input_t * input, va_list * args) { - vnet_classify_main_t * cm = &vnet_classify_main; - u32 * next_indexp = va_arg (*args, u32 *); + vnet_classify_main_t *cm = &vnet_classify_main; + u32 *next_indexp = va_arg (*args, u32 *); u32 node_index; u32 next_index; if (unformat (input, "output-node %U", unformat_vlib_node, - cm->vlib_main, &node_index)) + cm->vlib_main, &node_index)) { - next_index = vlib_node_add_next - (cm->vlib_main, l2_output_classify_node.index, node_index); + next_index = vlib_node_add_next + (cm->vlib_main, l2_output_classify_node.index, node_index); *next_indexp = next_index; return 1; @@ -2338,19 +2411,18 @@ unformat_l2_output_next_node (unformat_input_t * input, va_list * args) return 0; } -static clib_error_t * +static clib_error_t * vnet_classify_init (vlib_main_t * vm) { - vnet_classify_main_t * cm = &vnet_classify_main; + vnet_classify_main_t *cm = &vnet_classify_main; cm->vlib_main = vm; - cm->vnet_main = vnet_get_main(); + cm->vnet_main = vnet_get_main (); - vnet_classify_register_unformat_opaque_index_fn + vnet_classify_register_unformat_opaque_index_fn (unformat_opaque_sw_if_index); - vnet_classify_register_unformat_ip_next_index_fn - (unformat_ip_next_node); + vnet_classify_register_unformat_ip_next_index_fn (unformat_ip_next_node); vnet_classify_register_unformat_l2_next_index_fn (unformat_l2_input_next_node); @@ -2358,8 +2430,7 @@ vnet_classify_init (vlib_main_t * vm) vnet_classify_register_unformat_l2_next_index_fn (unformat_l2_output_next_node); - vnet_classify_register_unformat_acl_next_index_fn - (unformat_acl_next_node); + vnet_classify_register_unformat_acl_next_index_fn (unformat_acl_next_node); return 0; } @@ -2370,7 +2441,7 @@ VLIB_INIT_FUNCTION (vnet_classify_init); #if TEST_CODE > 0 -typedef struct +typedef struct { ip4_address_t addr; int in_table; @@ -2394,8 +2465,8 @@ typedef struct u32 seed; /* Test data */ - classify_data_or_mask_t * mask; - classify_data_or_mask_t * data; + classify_data_or_mask_t *mask; + classify_data_or_mask_t *data; /* convenience */ vnet_classify_main_t *classify_main; @@ -2406,7 +2477,7 @@ typedef struct static test_classify_main_t test_classify_main; static clib_error_t * -test_classify_churn (test_classify_main_t *tm) +test_classify_churn (test_classify_main_t * tm) { classify_data_or_mask_t *mask, *data; vlib_main_t *vm = tm->vlib_main; @@ -2415,8 +2486,8 @@ test_classify_churn (test_classify_main_t *tm) u32 tmp; int i, rv; - vec_validate_aligned (mp, 3 * sizeof(u32x4), sizeof(u32x4)); - vec_validate_aligned (dp, 3 * sizeof(u32x4), sizeof(u32x4)); + vec_validate_aligned (mp, 3 * sizeof (u32x4), sizeof (u32x4)); + vec_validate_aligned (dp, 3 * sizeof (u32x4), sizeof (u32x4)); mask = (classify_data_or_mask_t *) mp; data = (classify_data_or_mask_t *) dp; @@ -2434,47 +2505,45 @@ test_classify_churn (test_classify_main_t *tm) tmp++; } - tm->table = vnet_classify_new_table (tm->classify_main, - (u8 *)mask, - tm->buckets, - tm->memory_size, - 0 /* skip */, - 3 /* vectors to match */); + tm->table = vnet_classify_new_table (tm->classify_main, + (u8 *) mask, + tm->buckets, + tm->memory_size, 0 /* skip */ , + 3 /* vectors to match */ ); tm->table->miss_next_index = IP_LOOKUP_NEXT_DROP; tm->table_index = tm->table - tm->classify_main->tables; - vlib_cli_output (vm, "Created table %d, buckets %d", - tm->table_index, tm->buckets); + vlib_cli_output (vm, "Created table %d, buckets %d", + tm->table_index, tm->buckets); + + vlib_cli_output (vm, "Initialize: add %d (approx. half of %d sessions)...", + tm->sessions / 2, tm->sessions); - vlib_cli_output (vm, "Initialize: add %d (approx. half of %d sessions)...", - tm->sessions/2, tm->sessions); - - for (i = 0; i < tm->sessions/2; i++) + for (i = 0; i < tm->sessions / 2; i++) { ep = vec_elt_at_index (tm->entries, i); data->ip.src_address.as_u32 = ep->addr.as_u32; ep->in_table = 1; - rv = vnet_classify_add_del_session (tm->classify_main, - tm->table_index, - (u8 *) data, - IP_LOOKUP_NEXT_DROP, - i /* opaque_index */, - 0 /* advance */, - 0 /* action*/, - 0 /* metadata */, - 1 /* is_add */); - + rv = vnet_classify_add_del_session (tm->classify_main, + tm->table_index, + (u8 *) data, + IP_LOOKUP_NEXT_DROP, + i /* opaque_index */ , + 0 /* advance */ , + 0 /* action */ , + 0 /* metadata */ , + 1 /* is_add */ ); + if (rv != 0) - clib_warning ("add: returned %d", rv); + clib_warning ("add: returned %d", rv); if (tm->verbose) - vlib_cli_output (vm, "add: %U", format_ip4_address, - &ep->addr.as_u32); + vlib_cli_output (vm, "add: %U", format_ip4_address, &ep->addr.as_u32); } vlib_cli_output (vm, "Execute %d random add/delete operations", - tm->iterations); + tm->iterations); for (i = 0; i < tm->iterations; i++) { @@ -2482,7 +2551,7 @@ test_classify_churn (test_classify_main_t *tm) /* Pick a random entry */ index = random_u32 (&tm->seed) % tm->sessions; - + ep = vec_elt_at_index (tm->entries, index); data->ip.src_address.as_u32 = ep->addr.as_u32; @@ -2491,109 +2560,102 @@ test_classify_churn (test_classify_main_t *tm) is_add = !ep->in_table; if (tm->verbose) - vlib_cli_output (vm, "%s: %U", - is_add ? "add" : "del", - format_ip4_address, - &ep->addr.as_u32); - - rv = vnet_classify_add_del_session (tm->classify_main, - tm->table_index, - (u8 *) data, - IP_LOOKUP_NEXT_DROP, - i /* opaque_index */, - 0 /* advance */, - 0 /* action*/, - 0 /* metadata */, - is_add); + vlib_cli_output (vm, "%s: %U", + is_add ? "add" : "del", + format_ip4_address, &ep->addr.as_u32); + + rv = vnet_classify_add_del_session (tm->classify_main, + tm->table_index, + (u8 *) data, + IP_LOOKUP_NEXT_DROP, + i /* opaque_index */ , + 0 /* advance */ , + 0 /* action */ , + 0 /* metadata */ , + is_add); if (rv != 0) - vlib_cli_output (vm, - "%s[%d]: %U returned %d", is_add ? "add" : "del", - index, - format_ip4_address, - &ep->addr.as_u32, rv); + vlib_cli_output (vm, + "%s[%d]: %U returned %d", is_add ? "add" : "del", + index, format_ip4_address, &ep->addr.as_u32, rv); else - ep->in_table = is_add; + ep->in_table = is_add; } vlib_cli_output (vm, "Remove remaining %d entries from the table", - tm->table->active_elements); + tm->table->active_elements); for (i = 0; i < tm->sessions; i++) { - u8 * key_minus_skip; + u8 *key_minus_skip; u64 hash; - vnet_classify_entry_t * e; - + vnet_classify_entry_t *e; + ep = tm->entries + i; if (ep->in_table == 0) - continue; + continue; data->ip.src_address.as_u32 = ep->addr.as_u32; hash = vnet_classify_hash_packet (tm->table, (u8 *) data); - e = vnet_classify_find_entry (tm->table, - (u8 *) data, hash, 0 /* time_now */); + e = vnet_classify_find_entry (tm->table, + (u8 *) data, hash, 0 /* time_now */ ); if (e == 0) - { - clib_warning ("Couldn't find %U index %d which should be present", - format_ip4_address, ep->addr, i); - continue; - } + { + clib_warning ("Couldn't find %U index %d which should be present", + format_ip4_address, ep->addr, i); + continue; + } - key_minus_skip = (u8 *)e->key; + key_minus_skip = (u8 *) e->key; key_minus_skip -= tm->table->skip_n_vectors * sizeof (u32x4); - rv = vnet_classify_add_del_session - (tm->classify_main, - tm->table_index, - key_minus_skip, - IP_LOOKUP_NEXT_DROP, - i /* opaque_index */, - 0 /* advance */, 0, 0, - 0 /* is_add */); + rv = vnet_classify_add_del_session + (tm->classify_main, + tm->table_index, + key_minus_skip, IP_LOOKUP_NEXT_DROP, i /* opaque_index */ , + 0 /* advance */ , 0, 0, + 0 /* is_add */ ); if (rv != 0) - clib_warning ("del: returned %d", rv); + clib_warning ("del: returned %d", rv); if (tm->verbose) - vlib_cli_output (vm, "del: %U", format_ip4_address, - &ep->addr.as_u32); + vlib_cli_output (vm, "del: %U", format_ip4_address, &ep->addr.as_u32); } vlib_cli_output (vm, "%d entries remain, MUST be zero", - tm->table->active_elements); + tm->table->active_elements); vlib_cli_output (vm, "Table after cleanup: \n%U\n", - format_classify_table, tm->table, 0 /* verbose */); + format_classify_table, tm->table, 0 /* verbose */ ); vec_free (mp); vec_free (dp); vnet_classify_delete_table_index (tm->classify_main, - tm->table_index, 1 /* del_chain */); + tm->table_index, 1 /* del_chain */ ); tm->table = 0; tm->table_index = ~0; - vec_free(tm->entries); + vec_free (tm->entries); return 0; } static clib_error_t * test_classify_command_fn (vlib_main_t * vm, - unformat_input_t * input, - vlib_cli_command_t * cmd) + unformat_input_t * input, vlib_cli_command_t * cmd) { test_classify_main_t *tm = &test_classify_main; - vnet_classify_main_t * cm = &vnet_classify_main; + vnet_classify_main_t *cm = &vnet_classify_main; u32 tmp; int which = 0; - clib_error_t * error = 0; - + clib_error_t *error = 0; + tm->buckets = 1024; tm->sessions = 8192; tm->iterations = 8192; - tm->memory_size = 64<<20; + tm->memory_size = 64 << 20; tm->src.as_u32 = clib_net_to_host_u32 (0x0100000A); tm->table = 0; tm->seed = 0xDEADDABE; @@ -2603,28 +2665,30 @@ test_classify_command_fn (vlib_main_t * vm, /* Default starting address 1.0.0.10 */ - while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) { - if (unformat (input, "sessions %d", &tmp)) - tm->sessions = tmp; - else if (unformat (input, "src %U", unformat_ip4_address, &tm->src.as_u32)) - ; - else if (unformat (input, "buckets %d", &tm->buckets)) - ; - else if (unformat (input, "memory-size %uM", &tmp)) - tm->memory_size = tmp<<20; - else if (unformat (input, "memory-size %uG", &tmp)) - tm->memory_size = tmp<<30; - else if (unformat (input, "seed %d", &tm->seed)) - ; - else if (unformat (input, "verbose")) - tm->verbose = 1; + while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) + { + if (unformat (input, "sessions %d", &tmp)) + tm->sessions = tmp; + else + if (unformat (input, "src %U", unformat_ip4_address, &tm->src.as_u32)) + ; + else if (unformat (input, "buckets %d", &tm->buckets)) + ; + else if (unformat (input, "memory-size %uM", &tmp)) + tm->memory_size = tmp << 20; + else if (unformat (input, "memory-size %uG", &tmp)) + tm->memory_size = tmp << 30; + else if (unformat (input, "seed %d", &tm->seed)) + ; + else if (unformat (input, "verbose")) + tm->verbose = 1; - else if (unformat (input, "iterations %d", &tm->iterations)) - ; - else if (unformat (input, "churn-test")) - which = 0; - else - break; + else if (unformat (input, "iterations %d", &tm->iterations)) + ; + else if (unformat (input, "churn-test")) + which = 0; + else + break; } switch (which) @@ -2640,12 +2704,22 @@ test_classify_command_fn (vlib_main_t * vm, return error; } +/* *INDENT-OFF* */ VLIB_CLI_COMMAND (test_classify_command, static) = { .path = "test classify", - .short_help = + .short_help = "test classify [src ] [sessions ] [buckets ] [seed ]\n" " [memory-size [M|G]]\n" " [churn-test]", .function = test_classify_command_fn, }; +/* *INDENT-ON* */ #endif /* TEST_CODE */ + +/* + * fd.io coding-style-patch-verification: ON + * + * Local Variables: + * eval: (c-set-style "gnu") + * End: + */ diff --git a/src/vnet/classify/vnet_classify.h b/src/vnet/classify/vnet_classify.h index 4304d33bae8..40628015501 100644 --- a/src/vnet/classify/vnet_classify.h +++ b/src/vnet/classify/vnet_classify.h @@ -28,7 +28,7 @@ #include #include #include -#include /* for API error numbers */ +#include /* for API error numbers */ #include #include @@ -41,7 +41,7 @@ extern vlib_node_registration_t ip6_classify_node; #define CLASSIFY_TRACE 0 #if !defined( __aarch64__) && !defined(__arm__) -#define CLASSIFY_USE_SSE //Allow usage of SSE operations +#define CLASSIFY_USE_SSE //Allow usage of SSE operations #endif #define U32X4_ALIGNED(p) PREDICT_TRUE((((intptr_t)p) & 0xf) == 0) @@ -79,6 +79,7 @@ _(3) \ _(4) \ _(5) +/* *INDENT-OFF* */ typedef CLIB_PACKED(struct _vnet_classify_entry { /* Graph node next index */ u32 next_index; @@ -105,24 +106,28 @@ typedef CLIB_PACKED(struct _vnet_classify_entry { u64 hits; struct _vnet_classify_entry * next_free; }; - + f64 last_heard; /* Must be aligned to a 16-octet boundary */ u32x4 key[0]; }) vnet_classify_entry_t; +/* *INDENT-ON* */ -static inline int vnet_classify_entry_is_free (vnet_classify_entry_t * e) +static inline int +vnet_classify_entry_is_free (vnet_classify_entry_t * e) { return e->flags & VNET_CLASSIFY_ENTRY_FREE; } -static inline int vnet_classify_entry_is_busy (vnet_classify_entry_t * e) +static inline int +vnet_classify_entry_is_busy (vnet_classify_entry_t * e) { return ((e->flags & VNET_CLASSIFY_ENTRY_FREE) == 0); } /* Need these to con the vector allocator */ +/* *INDENT-OFF* */ #define _(size) \ typedef CLIB_PACKED(struct { \ u32 pad0[4]; \ @@ -130,11 +135,15 @@ typedef CLIB_PACKED(struct { \ u32x4 key[size]; \ }) vnet_classify_entry_##size##_t; foreach_size_in_u32x4; +/* *INDENT-ON* */ #undef _ -typedef struct { - union { - struct { +typedef struct +{ + union + { + struct + { u32 offset; u8 linear_search; u8 pad[2]; @@ -144,13 +153,14 @@ typedef struct { }; } vnet_classify_bucket_t; -typedef struct { +typedef struct +{ /* Mask to apply after skipping N vectors */ u32x4 *mask; /* Buckets and entries */ - vnet_classify_bucket_t * buckets; - vnet_classify_entry_t * entries; - + vnet_classify_bucket_t *buckets; + vnet_classify_entry_t *entries; + /* Config parameters */ u32 match_n_vectors; u32 skip_n_vectors; @@ -164,217 +174,219 @@ typedef struct { u32 data_offset; /* Index of next table to try */ u32 next_table_index; - + /* Miss next index, return if next_table_index = 0 */ u32 miss_next_index; - + /* Per-bucket working copies, one per thread */ - vnet_classify_entry_t ** working_copies; + vnet_classify_entry_t **working_copies; int *working_copy_lengths; vnet_classify_bucket_t saved_bucket; - + /* Free entry freelists */ vnet_classify_entry_t **freelists; - u8 * name; - + u8 *name; + /* Private allocation arena, protected by the writer lock */ - void * mheap; - + void *mheap; + /* Writer (only) lock for this table */ - volatile u32 * writer_lock; - + volatile u32 *writer_lock; + } vnet_classify_table_t; -struct _vnet_classify_main { +struct _vnet_classify_main +{ /* Table pool */ - vnet_classify_table_t * tables; - + vnet_classify_table_t *tables; + /* Registered next-index, opaque unformat fcns */ - unformat_function_t ** unformat_l2_next_index_fns; - unformat_function_t ** unformat_ip_next_index_fns; - unformat_function_t ** unformat_acl_next_index_fns; - unformat_function_t ** unformat_policer_next_index_fns; - unformat_function_t ** unformat_opaque_index_fns; + unformat_function_t **unformat_l2_next_index_fns; + unformat_function_t **unformat_ip_next_index_fns; + unformat_function_t **unformat_acl_next_index_fns; + unformat_function_t **unformat_policer_next_index_fns; + unformat_function_t **unformat_opaque_index_fns; /* convenience variables */ - vlib_main_t * vlib_main; - vnet_main_t * vnet_main; + vlib_main_t *vlib_main; + vnet_main_t *vnet_main; }; extern vnet_classify_main_t vnet_classify_main; -u8 * format_classify_table (u8 * s, va_list * args); +u8 *format_classify_table (u8 * s, va_list * args); u64 vnet_classify_hash_packet (vnet_classify_table_t * t, u8 * h); -static inline u64 -vnet_classify_hash_packet_inline (vnet_classify_table_t * t, - u8 * h) +static inline u64 +vnet_classify_hash_packet_inline (vnet_classify_table_t * t, u8 * h) { u32x4 *mask; - union { + union + { u32x4 as_u32x4; u64 as_u64[2]; - } xor_sum __attribute__((aligned(sizeof(u32x4)))); + } xor_sum __attribute__ ((aligned (sizeof (u32x4)))); - ASSERT(t); + ASSERT (t); mask = t->mask; #ifdef CLASSIFY_USE_SSE - if (U32X4_ALIGNED(h)) { //SSE can't handle unaligned data - u32x4 *data = (u32x4 *)h; - xor_sum.as_u32x4 = data[0 + t->skip_n_vectors] & mask[0]; - switch (t->match_n_vectors) - { - case 5: - xor_sum.as_u32x4 ^= data[4 + t->skip_n_vectors] & mask[4]; - /* FALLTHROUGH */ - case 4: - xor_sum.as_u32x4 ^= data[3 + t->skip_n_vectors] & mask[3]; - /* FALLTHROUGH */ - case 3: - xor_sum.as_u32x4 ^= data[2 + t->skip_n_vectors] & mask[2]; - /* FALLTHROUGH */ - case 2: - xor_sum.as_u32x4 ^= data[1 + t->skip_n_vectors] & mask[1]; - /* FALLTHROUGH */ - case 1: - break; - default: - abort(); + if (U32X4_ALIGNED (h)) + { //SSE can't handle unaligned data + u32x4 *data = (u32x4 *) h; + xor_sum.as_u32x4 = data[0 + t->skip_n_vectors] & mask[0]; + switch (t->match_n_vectors) + { + case 5: + xor_sum.as_u32x4 ^= data[4 + t->skip_n_vectors] & mask[4]; + /* FALLTHROUGH */ + case 4: + xor_sum.as_u32x4 ^= data[3 + t->skip_n_vectors] & mask[3]; + /* FALLTHROUGH */ + case 3: + xor_sum.as_u32x4 ^= data[2 + t->skip_n_vectors] & mask[2]; + /* FALLTHROUGH */ + case 2: + xor_sum.as_u32x4 ^= data[1 + t->skip_n_vectors] & mask[1]; + /* FALLTHROUGH */ + case 1: + break; + default: + abort (); + } } - } else + else #endif /* CLASSIFY_USE_SSE */ - { - u32 skip_u64 = t->skip_n_vectors * 2; - u64 *data64 = (u64 *)h; - xor_sum.as_u64[0] = data64[0 + skip_u64] & ((u64 *)mask)[0]; - xor_sum.as_u64[1] = data64[1 + skip_u64] & ((u64 *)mask)[1]; - switch (t->match_n_vectors) { - case 5: - xor_sum.as_u64[0] ^= data64[8 + skip_u64] & ((u64 *)mask)[8]; - xor_sum.as_u64[1] ^= data64[9 + skip_u64] & ((u64 *)mask)[9]; - /* FALLTHROUGH */ - case 4: - xor_sum.as_u64[0] ^= data64[6 + skip_u64] & ((u64 *)mask)[6]; - xor_sum.as_u64[1] ^= data64[7 + skip_u64] & ((u64 *)mask)[7]; - /* FALLTHROUGH */ - case 3: - xor_sum.as_u64[0] ^= data64[4 + skip_u64] & ((u64 *)mask)[4]; - xor_sum.as_u64[1] ^= data64[5 + skip_u64] & ((u64 *)mask)[5]; - /* FALLTHROUGH */ - case 2: - xor_sum.as_u64[0] ^= data64[2 + skip_u64] & ((u64 *)mask)[2]; - xor_sum.as_u64[1] ^= data64[3 + skip_u64] & ((u64 *)mask)[3]; - /* FALLTHROUGH */ - case 1: - break; - - default: - abort(); + u32 skip_u64 = t->skip_n_vectors * 2; + u64 *data64 = (u64 *) h; + xor_sum.as_u64[0] = data64[0 + skip_u64] & ((u64 *) mask)[0]; + xor_sum.as_u64[1] = data64[1 + skip_u64] & ((u64 *) mask)[1]; + switch (t->match_n_vectors) + { + case 5: + xor_sum.as_u64[0] ^= data64[8 + skip_u64] & ((u64 *) mask)[8]; + xor_sum.as_u64[1] ^= data64[9 + skip_u64] & ((u64 *) mask)[9]; + /* FALLTHROUGH */ + case 4: + xor_sum.as_u64[0] ^= data64[6 + skip_u64] & ((u64 *) mask)[6]; + xor_sum.as_u64[1] ^= data64[7 + skip_u64] & ((u64 *) mask)[7]; + /* FALLTHROUGH */ + case 3: + xor_sum.as_u64[0] ^= data64[4 + skip_u64] & ((u64 *) mask)[4]; + xor_sum.as_u64[1] ^= data64[5 + skip_u64] & ((u64 *) mask)[5]; + /* FALLTHROUGH */ + case 2: + xor_sum.as_u64[0] ^= data64[2 + skip_u64] & ((u64 *) mask)[2]; + xor_sum.as_u64[1] ^= data64[3 + skip_u64] & ((u64 *) mask)[3]; + /* FALLTHROUGH */ + case 1: + break; + + default: + abort (); + } } - } - + return clib_xxhash (xor_sum.as_u64[0] ^ xor_sum.as_u64[1]); } -static inline void +static inline void vnet_classify_prefetch_bucket (vnet_classify_table_t * t, u64 hash) { u32 bucket_index; - - ASSERT (is_pow2(t->nbuckets)); - + + ASSERT (is_pow2 (t->nbuckets)); + bucket_index = hash & (t->nbuckets - 1); - - CLIB_PREFETCH(&t->buckets[bucket_index], CLIB_CACHE_LINE_BYTES, LOAD); + + CLIB_PREFETCH (&t->buckets[bucket_index], CLIB_CACHE_LINE_BYTES, LOAD); } -static inline vnet_classify_entry_t * +static inline vnet_classify_entry_t * vnet_classify_get_entry (vnet_classify_table_t * t, uword offset) { - u8 * hp = t->mheap; - u8 * vp = hp + offset; - + u8 *hp = t->mheap; + u8 *vp = hp + offset; + return (void *) vp; } -static inline uword vnet_classify_get_offset (vnet_classify_table_t * t, - vnet_classify_entry_t * v) +static inline uword +vnet_classify_get_offset (vnet_classify_table_t * t, + vnet_classify_entry_t * v) { - u8 * hp, * vp; + u8 *hp, *vp; hp = (u8 *) t->mheap; vp = (u8 *) v; - ASSERT((vp - hp) < 0x100000000ULL); + ASSERT ((vp - hp) < 0x100000000ULL); return vp - hp; } static inline vnet_classify_entry_t * -vnet_classify_entry_at_index (vnet_classify_table_t * t, - vnet_classify_entry_t * e, - u32 index) +vnet_classify_entry_at_index (vnet_classify_table_t * t, + vnet_classify_entry_t * e, u32 index) { - u8 * eu8; + u8 *eu8; - eu8 = (u8 *)e; + eu8 = (u8 *) e; eu8 += index * (sizeof (vnet_classify_entry_t) + - (t->match_n_vectors * sizeof (u32x4))); + (t->match_n_vectors * sizeof (u32x4))); return (vnet_classify_entry_t *) eu8; } static inline void -vnet_classify_prefetch_entry (vnet_classify_table_t * t, - u64 hash) +vnet_classify_prefetch_entry (vnet_classify_table_t * t, u64 hash) { u32 bucket_index; u32 value_index; - vnet_classify_bucket_t * b; - vnet_classify_entry_t * e; + vnet_classify_bucket_t *b; + vnet_classify_entry_t *e; bucket_index = hash & (t->nbuckets - 1); b = &t->buckets[bucket_index]; - + if (b->offset == 0) return; hash >>= t->log2_nbuckets; e = vnet_classify_get_entry (t, b->offset); - value_index = hash & ((1<log2_pages)-1); + value_index = hash & ((1 << b->log2_pages) - 1); e = vnet_classify_entry_at_index (t, e, value_index); - CLIB_PREFETCH(e, CLIB_CACHE_LINE_BYTES, LOAD); + CLIB_PREFETCH (e, CLIB_CACHE_LINE_BYTES, LOAD); } -vnet_classify_entry_t * -vnet_classify_find_entry (vnet_classify_table_t * t, - u8 * h, u64 hash, f64 now); +vnet_classify_entry_t *vnet_classify_find_entry (vnet_classify_table_t * t, + u8 * h, u64 hash, f64 now); static inline vnet_classify_entry_t * vnet_classify_find_entry_inline (vnet_classify_table_t * t, - u8 * h, u64 hash, f64 now) + u8 * h, u64 hash, f64 now) { - vnet_classify_entry_t * v; + vnet_classify_entry_t *v; u32x4 *mask, *key; - union { + union + { u32x4 as_u32x4; u64 as_u64[2]; - } result __attribute__((aligned(sizeof(u32x4)))); - vnet_classify_bucket_t * b; + } result __attribute__ ((aligned (sizeof (u32x4)))); + vnet_classify_bucket_t *b; u32 value_index; u32 bucket_index; u32 limit; int i; - bucket_index = hash & (t->nbuckets-1); + bucket_index = hash & (t->nbuckets - 1); b = &t->buckets[bucket_index]; mask = t->mask; @@ -384,128 +396,147 @@ vnet_classify_find_entry_inline (vnet_classify_table_t * t, hash >>= t->log2_nbuckets; v = vnet_classify_get_entry (t, b->offset); - value_index = hash & ((1<log2_pages)-1); + value_index = hash & ((1 << b->log2_pages) - 1); limit = t->entries_per_page; if (PREDICT_FALSE (b->linear_search)) { value_index = 0; - limit *= (1<log2_pages); + limit *= (1 << b->log2_pages); } v = vnet_classify_entry_at_index (t, v, value_index); #ifdef CLASSIFY_USE_SSE - if (U32X4_ALIGNED(h)) { - u32x4 *data = (u32x4 *) h; - for (i = 0; i < limit; i++) { - key = v->key; - result.as_u32x4 = (data[0 + t->skip_n_vectors] & mask[0]) ^ key[0]; - switch (t->match_n_vectors) - { - case 5: - result.as_u32x4 |= (data[4 + t->skip_n_vectors] & mask[4]) ^ key[4]; - /* FALLTHROUGH */ - case 4: - result.as_u32x4 |= (data[3 + t->skip_n_vectors] & mask[3]) ^ key[3]; - /* FALLTHROUGH */ - case 3: - result.as_u32x4 |= (data[2 + t->skip_n_vectors] & mask[2]) ^ key[2]; - /* FALLTHROUGH */ - case 2: - result.as_u32x4 |= (data[1 + t->skip_n_vectors] & mask[1]) ^ key[1]; - /* FALLTHROUGH */ - case 1: - break; - default: - abort(); - } - - if (u32x4_zero_byte_mask (result.as_u32x4) == 0xffff) { - if (PREDICT_TRUE(now)) { - v->hits++; - v->last_heard = now; - } - return (v); - } - v = vnet_classify_entry_at_index (t, v, 1); + if (U32X4_ALIGNED (h)) + { + u32x4 *data = (u32x4 *) h; + for (i = 0; i < limit; i++) + { + key = v->key; + result.as_u32x4 = (data[0 + t->skip_n_vectors] & mask[0]) ^ key[0]; + switch (t->match_n_vectors) + { + case 5: + result.as_u32x4 |= + (data[4 + t->skip_n_vectors] & mask[4]) ^ key[4]; + /* FALLTHROUGH */ + case 4: + result.as_u32x4 |= + (data[3 + t->skip_n_vectors] & mask[3]) ^ key[3]; + /* FALLTHROUGH */ + case 3: + result.as_u32x4 |= + (data[2 + t->skip_n_vectors] & mask[2]) ^ key[2]; + /* FALLTHROUGH */ + case 2: + result.as_u32x4 |= + (data[1 + t->skip_n_vectors] & mask[1]) ^ key[1]; + /* FALLTHROUGH */ + case 1: + break; + default: + abort (); + } + + if (u32x4_zero_byte_mask (result.as_u32x4) == 0xffff) + { + if (PREDICT_TRUE (now)) + { + v->hits++; + v->last_heard = now; + } + return (v); + } + v = vnet_classify_entry_at_index (t, v, 1); + } } - } else + else #endif /* CLASSIFY_USE_SSE */ { u32 skip_u64 = t->skip_n_vectors * 2; - u64 *data64 = (u64 *)h; - for (i = 0; i < limit; i++) { - key = v->key; - - result.as_u64[0] = (data64[0 + skip_u64] & ((u64 *)mask)[0]) ^ ((u64 *)key)[0]; - result.as_u64[1] = (data64[1 + skip_u64] & ((u64 *)mask)[1]) ^ ((u64 *)key)[1]; - switch (t->match_n_vectors) - { - case 5: - result.as_u64[0] |= (data64[8 + skip_u64] & ((u64 *)mask)[8]) ^ ((u64 *)key)[8]; - result.as_u64[1] |= (data64[9 + skip_u64] & ((u64 *)mask)[9]) ^ ((u64 *)key)[9]; - /* FALLTHROUGH */ - case 4: - result.as_u64[0] |= (data64[6 + skip_u64] & ((u64 *)mask)[6]) ^ ((u64 *)key)[6]; - result.as_u64[1] |= (data64[7 + skip_u64] & ((u64 *)mask)[7]) ^ ((u64 *)key)[7]; - /* FALLTHROUGH */ - case 3: - result.as_u64[0] |= (data64[4 + skip_u64] & ((u64 *)mask)[4]) ^ ((u64 *)key)[4]; - result.as_u64[1] |= (data64[5 + skip_u64] & ((u64 *)mask)[5]) ^ ((u64 *)key)[5]; - /* FALLTHROUGH */ - case 2: - result.as_u64[0] |= (data64[2 + skip_u64] & ((u64 *)mask)[2]) ^ ((u64 *)key)[2]; - result.as_u64[1] |= (data64[3 + skip_u64] & ((u64 *)mask)[3]) ^ ((u64 *)key)[3]; - /* FALLTHROUGH */ - case 1: - break; - default: - abort(); - } - - if (result.as_u64[0] == 0 && result.as_u64[1] == 0) { - if (PREDICT_TRUE(now)) { - v->hits++; - v->last_heard = now; - } - return (v); - } - - v = vnet_classify_entry_at_index (t, v, 1); - } + u64 *data64 = (u64 *) h; + for (i = 0; i < limit; i++) + { + key = v->key; + + result.as_u64[0] = + (data64[0 + skip_u64] & ((u64 *) mask)[0]) ^ ((u64 *) key)[0]; + result.as_u64[1] = + (data64[1 + skip_u64] & ((u64 *) mask)[1]) ^ ((u64 *) key)[1]; + switch (t->match_n_vectors) + { + case 5: + result.as_u64[0] |= + (data64[8 + skip_u64] & ((u64 *) mask)[8]) ^ ((u64 *) key)[8]; + result.as_u64[1] |= + (data64[9 + skip_u64] & ((u64 *) mask)[9]) ^ ((u64 *) key)[9]; + /* FALLTHROUGH */ + case 4: + result.as_u64[0] |= + (data64[6 + skip_u64] & ((u64 *) mask)[6]) ^ ((u64 *) key)[6]; + result.as_u64[1] |= + (data64[7 + skip_u64] & ((u64 *) mask)[7]) ^ ((u64 *) key)[7]; + /* FALLTHROUGH */ + case 3: + result.as_u64[0] |= + (data64[4 + skip_u64] & ((u64 *) mask)[4]) ^ ((u64 *) key)[4]; + result.as_u64[1] |= + (data64[5 + skip_u64] & ((u64 *) mask)[5]) ^ ((u64 *) key)[5]; + /* FALLTHROUGH */ + case 2: + result.as_u64[0] |= + (data64[2 + skip_u64] & ((u64 *) mask)[2]) ^ ((u64 *) key)[2]; + result.as_u64[1] |= + (data64[3 + skip_u64] & ((u64 *) mask)[3]) ^ ((u64 *) key)[3]; + /* FALLTHROUGH */ + case 1: + break; + default: + abort (); + } + + if (result.as_u64[0] == 0 && result.as_u64[1] == 0) + { + if (PREDICT_TRUE (now)) + { + v->hits++; + v->last_heard = now; + } + return (v); + } + + v = vnet_classify_entry_at_index (t, v, 1); + } } return 0; } -vnet_classify_table_t * -vnet_classify_new_table (vnet_classify_main_t *cm, - u8 * mask, u32 nbuckets, u32 memory_size, - u32 skip_n_vectors, - u32 match_n_vectors); - -int vnet_classify_add_del_session (vnet_classify_main_t * cm, - u32 table_index, - u8 * match, - u32 hit_next_index, - u32 opaque_index, - i32 advance, - u8 action, - u32 metadata, - int is_add); +vnet_classify_table_t *vnet_classify_new_table (vnet_classify_main_t * cm, + u8 * mask, u32 nbuckets, + u32 memory_size, + u32 skip_n_vectors, + u32 match_n_vectors); + +int vnet_classify_add_del_session (vnet_classify_main_t * cm, + u32 table_index, + u8 * match, + u32 hit_next_index, + u32 opaque_index, + i32 advance, + u8 action, u32 metadata, int is_add); int vnet_classify_add_del_table (vnet_classify_main_t * cm, - u8 * mask, - u32 nbuckets, - u32 memory_size, - u32 skip, - u32 match, - u32 next_table_index, - u32 miss_next_index, - u32 * table_index, - u8 current_data_flag, - i16 current_data_offset, - int is_add, - int del_chain); + u8 * mask, + u32 nbuckets, + u32 memory_size, + u32 skip, + u32 match, + u32 next_table_index, + u32 miss_next_index, + u32 * table_index, + u8 current_data_flag, + i16 current_data_offset, + int is_add, int del_chain); unformat_function_t unformat_ip4_mask; unformat_function_t unformat_ip6_mask; @@ -522,18 +553,27 @@ unformat_function_t unformat_vlan_tag; unformat_function_t unformat_l2_match; unformat_function_t unformat_classify_match; -void vnet_classify_register_unformat_ip_next_index_fn -(unformat_function_t * fn); +void vnet_classify_register_unformat_ip_next_index_fn + (unformat_function_t * fn); -void vnet_classify_register_unformat_l2_next_index_fn -(unformat_function_t * fn); +void vnet_classify_register_unformat_l2_next_index_fn + (unformat_function_t * fn); -void vnet_classify_register_unformat_acl_next_index_fn -(unformat_function_t * fn); +void vnet_classify_register_unformat_acl_next_index_fn + (unformat_function_t * fn); -void vnet_classify_register_unformat_policer_next_index_fn -(unformat_function_t * fn); +void vnet_classify_register_unformat_policer_next_index_fn + (unformat_function_t * fn); -void vnet_classify_register_unformat_opaque_index_fn (unformat_function_t * fn); +void vnet_classify_register_unformat_opaque_index_fn (unformat_function_t * + fn); #endif /* __included_vnet_classify_h__ */ + +/* + * fd.io coding-style-patch-verification: ON + * + * Local Variables: + * eval: (c-set-style "gnu") + * End: + */