/* * Copyright (c) 2017 Cisco and/or its affiliates. * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at: * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #ifndef __included_ioam_cache_h__ #define __included_ioam_cache_h__ #include #include #include #include #include #include #include #include #include #include #include #include #include /* * ioam_cache.h * This header contains routines for caching of ioam header and * buffer: * 1 - On application facing node: to cache ioam header recvd * in request and reattach in response to provide round * trip path visibility. Since request response matching * is needed works with TCP and relies on (5 tuples,seq no) * 2 - On M-Anycast server node: This node replicates requests * towards multiple anycast service nodes serving anycast * IP6 address. It evaluates response and forwards the best * response towards the client of requesting the service. * Again since request-response matching is needed, works * with TCP and relies on (5 tuples,seq no) for matching. * To do this it caches SYN-ACK responses for a short time to * evaluate multiple responses received before the selected * SYN-ACK response is forwared and others dropped. * * M-Anycast server cache: * - There is a pool of cache entries per worker thread. * - Cache entry is created when SYN is received expected * number of responses are marked based on number of * SR tunnels for the anycast destination address * - The pool/thread id and pool index are attached in the * message as an ioam option for quick look up. * - When is received SYN-ACK the ioam option containing * thread id + pool index of the cache entry is used to * look up cache entry. * - Cache synchronization: * - This is achieved by cache entry add/del/update all handled * by the same worker/main thread * - Packets from client to threads - syn packets, can be disctributed * based on incoming interface affinity to the cpu core pinned to * the thread or a simple sequence number based distribution * if thread per interface is not scaling * - Response packets from server towards clients - syn-acks, are * forced to the same thread that created the cache entry * using SR and the destination of SR v6 address assigned * to the core/thread. This adderss is sent as an ioam option * in the syn that can be then used on the other side to * populate v6 dst address in the response * - Timeout: timer wheel per thread is used to track the syn-ack wait * time. The timer wheel tick is updated via an input node per thread. * * Application facing node/Service side cache: * - Single pool of cache entries. * - Cache entry is created when SYN is received. Caches the ioam * header. Hash table entry is created based on 5 tuple and * TCP seq no to pool index * - Response SYN-ACK processed by looking up pool index in hash table * and cache entry in the pool is used to get the ioam header rewrite * string. Entry is freed from pool and hash table after use. * - Locking/Synchronization: Currently this functionality is deployed * with main/single thread only. Hence no locking is used. * - Deployment: A VPP node per application server servicing anycast * address is expected. Locking/synchronization needed when the server * /application facing node is started with multiple worker threads. * */ /* * Application facing server side caching: * Cache entry for ioam header * Currently caters to TCP and relies on * TCP - 5 tuples + seqno to cache and reinsert * ioam header b/n TCP request response */ typedef struct { ip6_address_t src_address; ip6_address_t dst_address; u16 src_port; u16 dst_port; u8 protocol; u32 seq_no; ip6_address_t next_hop; u16 my_address_offset; u8 *ioam_rewrite_string; } ioam_cache_entry_t; /* * Cache entry for anycast server selection * Works for TCP as 5 tuple + sequence number * is required for request response matching * max_responses expected is set based on number * of SR tunnels for the dst_address * Timeout or all response_received = max_responses * will clear the entry * buffer_index index of the response msg vlib buffer * that is currently the best response */ typedef struct { u32 pool_id; u32 pool_index; ip6_address_t src_address; ip6_address_t dst_address; u16 src_port; u16 dst_port; u8 protocol; u32 seq_no; u32 buffer_index; ip6_hop_by_hop_header_t *hbh; //pointer to hbh header in the buffer u64 created_at; u8 response_received; u8 max_responses; u32 stop_timer_handle; /** Handle returned from tw_start_timer */ u32 timer_handle; /** entry should expire at this clock tick */ u32 expected_to_expire; } ioam_cache_ts_entry_t; /* * Per thread tunnel selection cache stats */ typedef struct { u64 inuse; u64 add_failed; } ioam_cache_ts_pool_stats_t; /* Server side: iOAM header caching */ #define MAX_CACHE_ENTRIES 4096 /* M-Anycast: Cache for SR tunnel selection */ #define MAX_CACHE_TS_ENTRIES 1048576 #define IOAM_CACHE_TABLE_DEFAULT_HASH_NUM_BUCKETS (4 * 1024) #define IOAM_CACHE_TABLE_DEFAULT_HASH_MEMORY_SIZE (2<<20) typedef struct { /* API message ID base */ u16 msg_id_base; /* Pool of ioam_cache_buffer_t */ ioam_cache_entry_t *ioam_rewrite_pool; /* For steering packets ioam cache entry is followed by * SR header. This is the SR rewrite template */ u8 *sr_rewrite_template; /* The current rewrite string being used */ u8 *rewrite; u8 rewrite_pool_index_offset; ip6_address_t sr_localsid_cache; u64 lookup_table_nbuckets; u64 lookup_table_size; clib_bihash_8_8_t ioam_rewrite_cache_table; /* M-Anycast: Pool of ioam_cache_ts_entry_t per thread */ ioam_cache_ts_entry_t **ioam_ts_pool; ioam_cache_ts_pool_stats_t *ts_stats; /** per thread single-wheel */ tw_timer_wheel_16t_2w_512sl_t *timer_wheels; /* * Selection criteria: oneway delay: Server to M-Anycast * or RTT */ bool criteria_oneway; u8 wait_for_responses; ip6_address_t sr_localsid_ts; /* convenience */ vlib_main_t *vlib_main; uword cache_hbh_slot; uword ts_hbh_slot; u32 ip6_hbh_pop_node_index; u32 error_node_index; u32 cleanup_process_node_index; } ioam_cache_main_t; ioam_cache_main_t ioam_cache_main; extern vlib_node_registration_t ioam_cache_node; extern vlib_node_registration_t ioam_cache_ts_node; /* Compute flow hash. We'll use it to select which Sponge to use for this * flow. And other things. * ip6_compute_flow_hash in ip6.h doesnt locate tcp/udp when * ext headers are present. While it could be made to it will be a * performance hit for ECMP flows. * HEnce this function here, with L4 information directly input * Useful when tcp/udp headers are already located in presence of * ext headers */ always_inline u32 ip6_compute_flow_hash_ext (const ip6_header_t * ip, u8 protocol, u16 src_port, u16 dst_port, flow_hash_config_t flow_hash_config) { u64 a, b, c; u64 t1, t2; t1 = (ip->src_address.as_u64[0] ^ ip->src_address.as_u64[1]); t1 = (flow_hash_config & IP_FLOW_HASH_SRC_ADDR) ? t1 : 0; t2 = (ip->dst_address.as_u64[0] ^ ip->dst_address.as_u64[1]); t2 = (flow_hash_config & IP_FLOW_HASH_DST_ADDR) ? t2 : 0; a = (flow_hash_config & IP_FLOW_HASH_REVERSE_SRC_DST) ? t2 : t1; b = (flow_hash_config & IP_FLOW_HASH_REVERSE_SRC_DST) ? t1 : t2; b ^= (flow_hash_config & IP_FLOW_HASH_PROTO) ? protocol : 0; t1 = src_port; t2 = dst_port; t1 = (flow_hash_config & IP_FLOW_HASH_SRC_PORT) ? t1 : 0; t2 = (flow_hash_config & IP_FLOW_HASH_DST_PORT) ? t2 : 0; c = (flow_hash_config & IP_FLOW_HASH_REVERSE_SRC_DST) ? ((t1 << 16) | t2) : ((t2 << 16) | t1); hash_mix64 (a, b, c); return (u32) c; } /* 2 new ioam E2E options : * 1. HBH_OPTION_TYPE_IOAM_EDGE_TO_EDGE_ID: IP6 address * of ioam node that inserted ioam header * 2. HBH_OPTION_TYPE_IOAM_E2E_CACHE_ID: Pool id and index * to look up tunnel select cache entry */ #define HBH_OPTION_TYPE_IOAM_EDGE_TO_EDGE_ID 30 #define HBH_OPTION_TYPE_IOAM_E2E_CACHE_ID 31 typedef CLIB_PACKED (struct { ip6_hop_by_hop_option_t hdr; u8 e2e_type; u8 reserved[5]; ip6_address_t id; }) ioam_e2e_id_option_t; typedef CLIB_PACKED (struct { ip6_hop_by_hop_option_t hdr; u8 e2e_type; u8 pool_id; u32 pool_index; }) ioam_e2e_cache_option_t; #define IOAM_E2E_ID_OPTION_RND ((sizeof(ioam_e2e_id_option_t) + 7) & ~7) #define IOAM_E2E_ID_HBH_EXT_LEN (IOAM_E2E_ID_OPTION_RND >> 3) #define IOAM_E2E_CACHE_OPTION_RND ((sizeof(ioam_e2e_cache_option_t) + 7) & ~7) #define IOAM_E2E_CACHE_HBH_EXT_LEN (IOAM_E2E_CACHE_OPTION_RND >> 3) static inline void ioam_e2e_id_rewrite_handler (ioam_e2e_id_option_t * e2e_option, ip6_address_t * address) { e2e_option->id.as_u64[0] = address->as_u64[0]; e2e_option->id.as_u64[1] = address->as_u64[1]; } /* Following functions are for the caching of ioam header * to enable reattaching it for a complete request-response * message exchange */ inline static void ioam_cache_entry_free (ioam_cache_entry_t * entry) { ioam_cache_main_t *cm = &ioam_cache_main; if (entry) { vec_free (entry->ioam_rewrite_string); memset (entry, 0, sizeof (*entry)); pool_put (cm->ioam_rewrite_pool, entry); } } inline static ioam_cache_entry_t * ioam_cache_entry_cleanup (u32 pool_index) { ioam_cache_main_t *cm = &ioam_cache_main; ioam_cache_entry_t *entry = 0; entry = pool_elt_at_index (cm->ioam_rewrite_pool, pool_index); ioam_cache_entry_free (entry); return (0); } inline static ioam_cache_entry_t * ioam_cache_lookup (ip6_header_t * ip0, u16 src_port, u16 dst_port, u32 seq_no) { ioam_cache_main_t *cm = &ioam_cache_main; u32 flow_hash = ip6_compute_flow_hash_ext (ip0, ip0->protocol, src_port, dst_port, IP_FLOW_HASH_DEFAULT | IP_FLOW_HASH_REVERSE_SRC_DST); clib_bihash_kv_8_8_t kv, value; kv.key = (u64) flow_hash << 32 | seq_no; kv.value = 0; value.key = 0; value.value = 0; if (clib_bihash_search_8_8 (&cm->ioam_rewrite_cache_table, &kv, &value) >= 0) { ioam_cache_entry_t *entry = 0; entry = pool_elt_at_index (cm->ioam_rewrite_pool, value.value); /* match */ if (ip6_address_compare (&ip0->src_address, &entry->dst_address) == 0 && ip6_address_compare (&ip0->dst_address, &entry->src_address) == 0 && entry->src_port == dst_port && entry->dst_port == src_port && entry->seq_no == seq_no) { /* If lookup is successful remove it from the hash */ clib_bihash_add_del_8_8 (&cm->ioam_rewrite_cache_table, &kv, 0); return (entry); } else return (0); } return (0); } /* * Caches ioam hbh header * Extends the hbh header with option to contain IP6 address of the node * that caches it */ inline static int ioam_cache_add (vlib_buffer_t * b0, ip6_header_t * ip0, u16 src_port, u16 dst_port, ip6_hop_by_hop_header_t * hbh0, u32 seq_no) { ioam_cache_main_t *cm = &ioam_cache_main; ioam_cache_entry_t *entry = 0; u32 rewrite_len = 0, e2e_id_offset = 0; u32 pool_index = 0; ioam_e2e_id_option_t *e2e = 0; pool_get_aligned (cm->ioam_rewrite_pool, entry, CLIB_CACHE_LINE_BYTES); memset (entry, 0, sizeof (*entry)); pool_index = entry - cm->ioam_rewrite_pool; clib_memcpy (entry->dst_address.as_u64, ip0->dst_address.as_u64, sizeof (ip6_address_t)); clib_memcpy (entry->src_address.as_u64, ip0->src_address.as_u64, sizeof (ip6_address_t)); entry->src_port = src_port; entry->dst_port = dst_port; entry->seq_no = seq_no; rewrite_len = ((hbh0->length + 1) << 3); vec_validate (entry->ioam_rewrite_string, rewrite_len - 1); e2e = ip6_ioam_find_hbh_option (hbh0, HBH_OPTION_TYPE_IOAM_EDGE_TO_EDGE_ID); if (e2e) { entry->next_hop.as_u64[0] = e2e->id.as_u64[0]; entry->next_hop.as_u64[1] = e2e->id.as_u64[1]; } else { return (-1); } e2e_id_offset = (u8 *) e2e - (u8 *) hbh0; /* setup e2e id option to insert v6 address of the node caching it */ clib_memcpy (entry->ioam_rewrite_string, hbh0, rewrite_len); hbh0 = (ip6_hop_by_hop_header_t *) entry->ioam_rewrite_string; /* suffix rewrite string with e2e ID option */ e2e = (ioam_e2e_id_option_t *) (entry->ioam_rewrite_string + e2e_id_offset); ioam_e2e_id_rewrite_handler (e2e, &cm->sr_localsid_cache); entry->my_address_offset = (u8 *) (&e2e->id) - (u8 *) hbh0; /* add it to hash, replacing and freeing any collision for now */ u32 flow_hash = ip6_compute_flow_hash_ext (ip0, hbh0->protocol, src_port, dst_port, IP_FLOW_HASH_DEFAULT); clib_bihash_kv_8_8_t kv, value; kv.key = (u64) flow_hash << 32 | seq_no; kv.value = 0; if (clib_bihash_search_8_8 (&cm->ioam_rewrite_cache_table, &kv, &value) >= 0) { /* replace */ ioam_cache_entry_cleanup (value.value); } kv.value = pool_index; clib_bihash_add_del_8_8 (&cm->ioam_rewrite_cache_table, &kv, 1); return (0); } /* Creates SR rewrite string * This is appended with ioam header on the server facing * node. * This SR header is necessary to attract packets towards * selected Anycast server. */ inline static void ioam_cache_sr_rewrite_template_create (void) { ioam_cache_main_t *cm = &ioam_cache_main; ip6_address_t *segments = 0; ip6_address_t *this_seg = 0; /* This nodes address and the original dest will be * filled when the packet is processed */ vec_add2 (segments, this_seg, 1); memset (this_seg, 0xfe, sizeof (ip6_address_t)); cm->sr_rewrite_template = ip6_sr_compute_rewrite_string_insert (segments); vec_free (segments); } inline static int ioam_cache_table_init (vlib_main_t * vm) { ioam_cache_main_t *cm = &ioam_cache_main; pool_alloc_aligned (cm->ioam_rewrite_pool, MAX_CACHE_ENTRIES, CLIB_CACHE_LINE_BYTES); cm->lookup_table_nbuckets = IOAM_CACHE_TABLE_DEFAULT_HASH_NUM_BUCKETS; cm->lookup_table_nbuckets = 1 << max_log2 (cm->lookup_table_nbuckets); cm->lookup_table_size = IOAM_CACHE_TABLE_DEFAULT_HASH_MEMORY_SIZE; clib_bihash_init_8_8 (&cm->ioam_rewrite_cache_table, "ioam rewrite cache table", cm->lookup_table_nbuckets, cm->lookup_table_size); /* Create SR rewrite template */ ioam_cache_sr_rewrite_template_create (); return (1); } inline static int ioam_cache_table_destroy (vlib_main_t * vm) { ioam_cache_main_t *cm = &ioam_cache_main; ioam_cache_entry_t *entry = 0; /* free pool and hash table */ clib_bihash_free_8_8 (&cm->ioam_rewrite_cache_table); pool_foreach (entry, cm->ioam_rewrite_pool, ( { ioam_cache_entry_free (entry); })); pool_free (cm->ioam_rewrite_pool); cm->ioam_rewrite_pool = 0; vec_free (cm->sr_rewrite_template); cm->sr_rewrite_template = 0; return (0); } inline static u8 * format_ioam_cache_entry (u8 * s, va_list * args) { ioam_cache_entry_t *e = va_arg (*args, ioam_cache_entry_t *); ioam_cache_main_t *cm = &ioam_cache_main; int rewrite_len = vec_len (e->ioam_rewrite_string); s = format (s, "%d: %U:%d to %U:%d seq_no %lu\n", (e - cm->ioam_rewrite_pool), format_ip6_address, &e->src_address, e->src_port, format_ip6_address, &e->dst_address, e->dst_port, e->seq_no); if (rewrite_len) { s = format (s, " %U", format_ip6_hop_by_hop_ext_hdr, (ip6_hop_by_hop_header_t *) e->ioam_rewrite_string, rewrite_len - 1); } return s; } void ioam_cache_ts_timer_node_enable (vlib_main_t * vm, u8 enable); #define IOAM_CACHE_TS_TIMEOUT 1.0 //SYN timeout 1 sec #define IOAM_CACHE_TS_TICK 100e-3 /* Timer delays as multiples of 100ms */ #define IOAM_CACHE_TS_TIMEOUT_TICKS IOAM_CACHE_TS_TICK*9 #define TIMER_HANDLE_INVALID ((u32) ~0) void expired_cache_ts_timer_callback (u32 * expired_timers); /* * Following functions are to manage M-Anycast server selection * cache * There is a per worker thread pool to create a cache entry * for a TCP SYN received. TCP SYN-ACK contians ioam header * with HBH_OPTION_TYPE_IOAM_E2E_CACHE_ID option to point to the * entry. */ inline static int ioam_cache_ts_table_init (vlib_main_t * vm) { ioam_cache_main_t *cm = &ioam_cache_main; int no_of_threads = vec_len (vlib_worker_threads); int i; vec_validate_aligned (cm->ioam_ts_pool, no_of_threads - 1, CLIB_CACHE_LINE_BYTES); vec_validate_aligned (cm->ts_stats, no_of_threads - 1, CLIB_CACHE_LINE_BYTES); vec_validate_aligned (cm->timer_wheels, no_of_threads - 1, CLIB_CACHE_LINE_BYTES); cm->lookup_table_nbuckets = IOAM_CACHE_TABLE_DEFAULT_HASH_NUM_BUCKETS; cm->lookup_table_nbuckets = 1 << max_log2 (cm->lookup_table_nbuckets); cm->lookup_table_size = IOAM_CACHE_TABLE_DEFAULT_HASH_MEMORY_SIZE; for (i = 0; i < no_of_threads; i++) { pool_alloc_aligned (cm->ioam_ts_pool[i], MAX_CACHE_TS_ENTRIES, CLIB_CACHE_LINE_BYTES); memset (&cm->ts_stats[i], 0, sizeof (ioam_cache_ts_pool_stats_t)); tw_timer_wheel_init_16t_2w_512sl (&cm->timer_wheels[i], expired_cache_ts_timer_callback, IOAM_CACHE_TS_TICK /* timer period 100ms */ , 10e4); cm->timer_wheels[i].last_run_time = vlib_time_now (vm); } ioam_cache_ts_timer_node_enable (vm, 1); return (1); } always_inline void ioam_cache_ts_timer_set (ioam_cache_main_t * cm, ioam_cache_ts_entry_t * entry, u32 interval) { entry->timer_handle = tw_timer_start_16t_2w_512sl (&cm->timer_wheels[entry->pool_id], entry->pool_index, 1, interval); } always_inline void ioam_cache_ts_timer_reset (ioam_cache_main_t * cm, ioam_cache_ts_entry_t * entry) { tw_timer_stop_16t_2w_512sl (&cm->timer_wheels[entry->pool_id], entry->timer_handle); entry->timer_handle = TIMER_HANDLE_INVALID; } inline static void ioam_cache_ts_entry_free (u32 thread_id, ioam_cache_ts_entry_t * entry, u32 node_index) { ioam_cache_main_t *cm = &ioam_cache_main; vlib_main_t *vm = cm->vlib_main; vlib_frame_t *nf = 0; u32 *to_next; if (entry) { if (entry->hbh != 0) { nf = vlib_get_frame_to_node (vm, node_index); nf->n_vectors = 0; to_next = vlib_frame_vector_args (nf); nf->n_vectors = 1; to_next[0] = entry->buffer_index; vlib_put_frame_to_node (vm, node_index, nf); } pool_put (cm->ioam_ts_pool[thread_id], entry); cm->ts_stats[thread_id].inuse--; memset (entry, 0, sizeof (*entry)); } } inline static int ioam_cache_ts_table_destroy (vlib_main_t * vm) { ioam_cache_main_t *cm = &ioam_cache_main; ioam_cache_ts_entry_t *entry = 0; int no_of_threads = vec_len (vlib_worker_threads); int i; /* free pool and hash table */ for (i = 0; i < no_of_threads; i++) { pool_foreach (entry, cm->ioam_ts_pool[i], ( { ioam_cache_ts_entry_free (i, entry, cm->error_node_index); } )); pool_free (cm->ioam_ts_pool[i]); cm->ioam_ts_pool = 0; tw_timer_wheel_free_16t_2w_512sl (&cm->timer_wheels[i]); } vec_free (cm->ioam_ts_pool); return (0); } inline static int ioam_cache_ts_entry_cleanup (u32 thread_id, u32 pool_index) { ioam_cache_main_t *cm = &ioam_cache_main; ioam_cache_ts_entry_t *entry = 0; entry = pool_elt_at_index (cm->ioam_ts_pool[thread_id], pool_index); ioam_cache_ts_entry_free (thread_id, entry, cm->error_node_index); return (0); } /* * Caches buffer for ioam SR tunnel select for Anycast service */ inline static int ioam_cache_ts_add (ip6_header_t * ip0, u16 src_port, u16 dst_port, u32 seq_no, u8 max_responses, u64 now, u32 thread_id, u32 * pool_index) { ioam_cache_main_t *cm = &ioam_cache_main; ioam_cache_ts_entry_t *entry = 0; if (cm->ts_stats[thread_id].inuse == MAX_CACHE_TS_ENTRIES) { cm->ts_stats[thread_id].add_failed++; return (-1); } pool_get_aligned (cm->ioam_ts_pool[thread_id], entry, CLIB_CACHE_LINE_BYTES); memset (entry, 0, sizeof (*entry)); *pool_index = entry - cm->ioam_ts_pool[thread_id]; clib_memcpy (entry->dst_address.as_u64, ip0->dst_address.as_u64, sizeof (ip6_address_t)); clib_memcpy (entry->src_address.as_u64, ip0->src_address.as_u64, sizeof (ip6_address_t)); entry->src_port = src_port; entry->dst_port = dst_port; entry->seq_no = seq_no; entry->response_received = 0; entry->max_responses = max_responses; entry->created_at = now; entry->hbh = 0; entry->buffer_index = 0; entry->pool_id = thread_id; entry->pool_index = *pool_index; ioam_cache_ts_timer_set (cm, entry, IOAM_CACHE_TS_TIMEOUT); cm->ts_stats[thread_id].inuse++; return (0); } inline static void ioam_cache_ts_send (u32 thread_id, i32 pool_index) { ioam_cache_main_t *cm = &ioam_cache_main; ioam_cache_ts_entry_t *entry = 0; entry = pool_elt_at_index (cm->ioam_ts_pool[thread_id], pool_index); if (!pool_is_free (cm->ioam_ts_pool[thread_id], entry) && entry) { /* send and free pool entry */ ioam_cache_ts_entry_free (thread_id, entry, cm->ip6_hbh_pop_node_index); } } inline static void ioam_cache_ts_check_and_send (u32 thread_id, i32 pool_index) { ioam_cache_main_t *cm = &ioam_cache_main; ioam_cache_ts_entry_t *entry = 0; entry = pool_elt_at_index (cm->ioam_ts_pool[thread_id], pool_index); if (entry && entry->hbh) { if (entry->response_received == entry->max_responses || entry->created_at + IOAM_CACHE_TS_TIMEOUT <= vlib_time_now (cm->vlib_main)) { ioam_cache_ts_timer_reset (cm, entry); ioam_cache_ts_send (thread_id, pool_index); } } } inline static int ioam_cache_ts_update (u32 thread_id, i32 pool_index, u32 buffer_index, ip6_hop_by_hop_header_t * hbh) { ioam_cache_main_t *cm = &ioam_cache_main; ioam_cache_ts_entry_t *entry = 0; vlib_main_t *vm = cm->vlib_main; vlib_frame_t *nf = 0; u32 *to_next; entry = pool_elt_at_index (cm->ioam_ts_pool[thread_id], pool_index); if (!pool_is_free (cm->ioam_ts_pool[thread_id], entry) && entry) { /* drop existing buffer */ if (entry->hbh != 0) { nf = vlib_get_frame_to_node (vm, cm->error_node_index); nf->n_vectors = 0; to_next = vlib_frame_vector_args (nf); nf->n_vectors = 1; to_next[0] = entry->buffer_index; vlib_put_frame_to_node (vm, cm->error_node_index, nf); } /* update */ entry->buffer_index = buffer_index; entry->hbh = hbh; /* check and send */ ioam_cache_ts_check_and_send (thread_id, pool_index); return (0); } return (-1); } /* * looks up the entry based on the e2e option pool index * result = 0 found the entry * result < 0 indicates failture to find an entry */ inline static int ioam_cache_ts_lookup (ip6_header_t * ip0, u8 protocol, u16 src_port, u16 dst_port, u32 seq_no, ip6_hop_by_hop_header_t ** hbh, u32 * pool_index, u8 * thread_id, u8 response_seen) { ioam_cache_main_t *cm = &ioam_cache_main; ip6_hop_by_hop_header_t *hbh0 = 0; ioam_e2e_cache_option_t *e2e = 0; hbh0 = (ip6_hop_by_hop_header_t *) (ip0 + 1); e2e = (ioam_e2e_cache_option_t *) ((u8 *) hbh0 + cm->rewrite_pool_index_offset); if ((u8 *) e2e < ((u8 *) hbh0 + ((hbh0->length + 1) << 3)) && e2e->hdr.type == HBH_OPTION_TYPE_IOAM_E2E_CACHE_ID) { ioam_cache_ts_entry_t *entry = 0; *pool_index = e2e->pool_index; *thread_id = e2e->pool_id; entry = pool_elt_at_index (cm->ioam_ts_pool[*thread_id], *pool_index); /* match */ if (entry && ip6_address_compare (&ip0->src_address, &entry->dst_address) == 0 && ip6_address_compare (&ip0->dst_address, &entry->src_address) == 0 && entry->src_port == dst_port && entry->dst_port == src_port && entry->seq_no == seq_no) { *hbh = entry->hbh; entry->response_received += response_seen; return (0); } else if (entry) { return (-1); } } return (-1); } inline static u8 * format_ioam_cache_ts_entry (u8 * s, va_list * args) { ioam_cache_ts_entry_t *e = va_arg (*args, ioam_cache_ts_entry_t *); u32 thread_id = va_arg (*args, u32); ioam_cache_main_t *cm = &ioam_cache_main; ioam_e2e_id_option_t *e2e = 0; vlib_main_t *vm = cm->vlib_main; clib_time_t *ct = &vm->clib_time; if (!e) goto end; if (e->hbh) { e2e = ip6_ioam_find_hbh_option (e->hbh, HBH_OPTION_TYPE_IOAM_EDGE_TO_EDGE_ID); s = format (s, "%d: %U:%d to %U:%d seq_no %u buffer %u %U \n\t\tCreated at %U Received %d\n", (e - cm->ioam_ts_pool[thread_id]), format_ip6_address, &e->src_address, e->src_port, format_ip6_address, &e->dst_address, e->dst_port, e->seq_no, e->buffer_index, format_ip6_address, e2e ? &e2e->id : 0, format_time_interval, "h:m:s:u", (e->created_at - vm->cpu_time_main_loop_start) * ct->seconds_per_clock, e->response_received); } else { s = format (s, "%d: %U:%d to %U:%d seq_no %u Buffer %u \n\t\tCreated at %U Received %d\n", (e - cm->ioam_ts_pool[thread_id]), format_ip6_address, &e->src_address, e->src_port, format_ip6_address, &e->dst_address, e->dst_port, e->seq_no, e->buffer_index, format_time_interval, "h:m:s:u", (e->created_at - vm->cpu_time_main_loop_start) * ct->seconds_per_clock, e->response_received); } end: return s; } /* * Get extended rewrite string for iOAM data in v6 * This makes space for an e2e options to carry cache pool info * and manycast server address. * It set the rewrite string per configs in ioam ip6 + new option * for cache along with offset to the option to populate cache * pool id and index */ static inline int ip6_ioam_ts_cache_set_rewrite (void) { ip6_hop_by_hop_ioam_main_t *hm = &ip6_hop_by_hop_ioam_main; ioam_cache_main_t *cm = &ioam_cache_main; ip6_hop_by_hop_header_t *hbh; u32 rewrite_len = 0; ioam_e2e_cache_option_t *e2e = 0; ioam_e2e_id_option_t *e2e_id = 0; vec_free (cm->rewrite); ip6_ioam_set_rewrite (&(cm->rewrite), hm->has_trace_option, hm->has_pot_option, hm->has_seqno_option); hbh = (ip6_hop_by_hop_header_t *) cm->rewrite; rewrite_len = ((hbh->length + 1) << 3); vec_validate (cm->rewrite, rewrite_len - 1 + IOAM_E2E_CACHE_OPTION_RND + IOAM_E2E_ID_OPTION_RND); hbh = (ip6_hop_by_hop_header_t *) cm->rewrite; /* setup e2e id option to insert pool id and index of the node caching it */ hbh->length += IOAM_E2E_CACHE_HBH_EXT_LEN + IOAM_E2E_ID_HBH_EXT_LEN; cm->rewrite_pool_index_offset = rewrite_len; e2e = (ioam_e2e_cache_option_t *) (cm->rewrite + rewrite_len); e2e->hdr.type = HBH_OPTION_TYPE_IOAM_E2E_CACHE_ID | HBH_OPTION_TYPE_SKIP_UNKNOWN; e2e->hdr.length = sizeof (ioam_e2e_cache_option_t) - sizeof (ip6_hop_by_hop_option_t); e2e->e2e_type = 2; e2e_id = (ioam_e2e_id_option_t *) ((u8 *) e2e + sizeof (ioam_e2e_cache_option_t)); e2e_id->hdr.type = HBH_OPTION_TYPE_IOAM_EDGE_TO_EDGE_ID | HBH_OPTION_TYPE_SKIP_UNKNOWN; e2e_id->hdr.length = sizeof (ioam_e2e_id_option_t) - sizeof (ip6_hop_by_hop_option_t); e2e_id->e2e_type = 1; return (0); } static inline int ip6_ioam_ts_cache_cleanup_rewrite (void) { ioam_cache_main_t *cm = &ioam_cache_main; vec_free (cm->rewrite); cm->rewrite = 0; cm->rewrite_pool_index_offset = 0; return (0); } #endif /* __included_ioam_cache_h__ */ /* * fd.io coding-style-patch-verification: ON * * Local Variables: * eval: (c-set-style "gnu") * End: */