/* * Copyright (c) 2015 Cisco and/or its affiliates. * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at: * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #include #include #include #include #include #include #include #include #include #include #define MAP_SKIP_IP6_LOOKUP 1 #define MAP_ERR_GOOD 0 #define MAP_ERR_BAD_POOL_SIZE -1 #define MAP_ERR_BAD_HT_RATIO -2 #define MAP_ERR_BAD_LIFETIME -3 #define MAP_ERR_BAD_BUFFERS -4 #define MAP_ERR_BAD_BUFFERS_TOO_LARGE -5 int map_create_domain (ip4_address_t * ip4_prefix, u8 ip4_prefix_len, ip6_address_t * ip6_prefix, u8 ip6_prefix_len, ip6_address_t * ip6_src, u8 ip6_src_len, u8 ea_bits_len, u8 psid_offset, u8 psid_length, u32 * map_domain_index, u16 mtu, u8 flags); int map_delete_domain (u32 map_domain_index); int map_add_del_psid (u32 map_domain_index, u16 psid, ip6_address_t * tep, u8 is_add); int map_if_enable_disable (bool is_enable, u32 sw_if_index, bool is_translation); u8 *format_map_trace (u8 * s, va_list * args); int map_param_set_fragmentation (bool inner, bool ignore_df); int map_param_set_icmp (ip4_address_t * ip4_err_relay_src); int map_param_set_icmp6 (u8 enable_unreachable); void map_pre_resolve (ip4_address_t * ip4, ip6_address_t * ip6, int is_del); int map_param_set_reassembly (bool is_ipv6, u16 lifetime_ms, u16 pool_size, u32 buffers, f64 ht_ratio, u32 * reass, u32 * packets); int map_param_set_security_check (bool enable, bool fragments); int map_param_set_traffic_class (bool copy, u8 tc); int map_param_set_tcp (u16 tcp_mss); typedef enum { MAP_DOMAIN_PREFIX = 1 << 0, MAP_DOMAIN_TRANSLATION = 1 << 1, // The domain uses MAP-T MAP_DOMAIN_RFC6052 = 1 << 2, } __attribute__ ((__packed__)) map_domain_flags_e; /** * IP4 reassembly logic: * One virtually reassembled flow requires a map_ip4_reass_t structure in order * to keep the first-fragment port number and, optionally, cache out of sequence * packets. * There are up to MAP_IP4_REASS_MAX_REASSEMBLY such structures. * When in use, those structures are stored in a hash table of MAP_IP4_REASS_BUCKETS buckets. * When a new structure needs to be used, it is allocated from available ones. * If there is no structure available, the oldest in use is selected and used if and * only if it was first allocated more than MAP_IP4_REASS_LIFETIME seconds ago. * In case no structure can be allocated, the fragment is dropped. */ #define MAP_IP4_REASS_LIFETIME_DEFAULT (100) /* ms */ #define MAP_IP4_REASS_HT_RATIO_DEFAULT (1.0) #define MAP_IP4_REASS_POOL_SIZE_DEFAULT 1024 // Number of reassembly structures #define MAP_IP4_REASS_BUFFERS_DEFAULT 2048 #define MAP_IP4_REASS_MAX_FRAGMENTS_PER_REASSEMBLY 5 // Number of fragment per reassembly #define MAP_IP6_REASS_LIFETIME_DEFAULT (100) /* ms */ #define MAP_IP6_REASS_HT_RATIO_DEFAULT (1.0) #define MAP_IP6_REASS_POOL_SIZE_DEFAULT 1024 // Number of reassembly structures #define MAP_IP6_REASS_BUFFERS_DEFAULT 2048 #define MAP_IP6_REASS_MAX_FRAGMENTS_PER_REASSEMBLY 5 #define MAP_IP6_REASS_COUNT_BYTES #define MAP_IP4_REASS_COUNT_BYTES //#define IP6_MAP_T_OVERRIDE_TOS 0 /* * This structure _MUST_ be no larger than a single cache line (64 bytes). * If more space is needed make a union of ip6_prefix and *rules, those are mutually exclusive. */ typedef struct { /* Required for pool_get_aligned */ CLIB_CACHE_LINE_ALIGN_MARK (cacheline0); ip6_address_t ip6_src; ip6_address_t ip6_prefix; ip6_address_t *rules; u32 suffix_mask; ip4_address_t ip4_prefix; u16 psid_mask; u16 mtu; map_domain_flags_e flags; u8 ip6_prefix_len; u8 ip6_src_len; u8 ea_bits_len; u8 psid_offset; u8 psid_length; /* helpers */ u8 psid_shift; u8 suffix_shift; u8 ea_shift; /* not used by forwarding */ u8 ip4_prefix_len; } map_domain_t; STATIC_ASSERT ((sizeof (map_domain_t) <= CLIB_CACHE_LINE_BYTES), "MAP domain fits in one cacheline"); #define MAP_REASS_INDEX_NONE ((u16)0xffff) /* * Hash key, padded out to 16 bytes for fast compare */ /* *INDENT-OFF* */ typedef union { CLIB_PACKED (struct { ip4_address_t src; ip4_address_t dst; u16 fragment_id; u8 protocol; }); u64 as_u64[2]; u32 as_u32[4]; } map_ip4_reass_key_t; /* *INDENT-ON* */ typedef struct { map_ip4_reass_key_t key; f64 ts; #ifdef MAP_IP4_REASS_COUNT_BYTES u16 expected_total; u16 forwarded; #endif i32 port; u16 bucket; u16 bucket_next; u16 fifo_prev; u16 fifo_next; u32 fragments[MAP_IP4_REASS_MAX_FRAGMENTS_PER_REASSEMBLY]; } map_ip4_reass_t; /* * MAP domain counters */ typedef enum { /* Simple counters */ MAP_DOMAIN_IPV4_FRAGMENT = 0, /* Combined counters */ MAP_DOMAIN_COUNTER_RX = 0, MAP_DOMAIN_COUNTER_TX, MAP_N_DOMAIN_COUNTER } map_domain_counter_t; /* * main_main_t */ /* *INDENT-OFF* */ typedef union { CLIB_PACKED (struct { ip6_address_t src; ip6_address_t dst; u32 fragment_id; u8 protocol; }); u64 as_u64[5]; u32 as_u32[10]; } map_ip6_reass_key_t; /* *INDENT-OFF* */ typedef struct { u32 pi; //Cached packet or ~0 u16 next_data_offset; //The data offset of the additional 20 bytes or ~0 u8 next_data_len; //Number of bytes ready to be copied (20 if not last fragment) u8 next_data[20]; //The 20 additional bytes } map_ip6_fragment_t; typedef struct { map_ip6_reass_key_t key; f64 ts; #ifdef MAP_IP6_REASS_COUNT_BYTES u16 expected_total; u16 forwarded; #endif u16 bucket; //What hash bucket this element is linked in u16 bucket_next; u16 fifo_prev; u16 fifo_next; ip4_header_t ip4_header; map_ip6_fragment_t fragments[MAP_IP6_REASS_MAX_FRAGMENTS_PER_REASSEMBLY]; } map_ip6_reass_t; #ifdef MAP_SKIP_IP6_LOOKUP /** * A pre-resolved next-hop */ typedef struct map_main_pre_resolved_t_ { /** * Linkage into the FIB graph */ fib_node_t node; /** * The FIB entry index of the next-hop */ fib_node_index_t fei; /** * This object sibling index on the FIB entry's child dependency list */ u32 sibling; /** * The Load-balance object index to use to forward */ dpo_id_t dpo; } map_main_pre_resolved_t; /** * Pre-resolved next hops for v4 and v6. Why these are global and not * per-domain is beyond me. */ extern map_main_pre_resolved_t pre_resolved[FIB_PROTOCOL_MAX]; #endif typedef struct { /* pool of MAP domains */ map_domain_t *domains; /* MAP Domain packet/byte counters indexed by map domain index */ vlib_simple_counter_main_t *simple_domain_counters; vlib_combined_counter_main_t *domain_counters; volatile u32 *counter_lock; /* API message id base */ u16 msg_id_base; /* Traffic class: zero, copy (~0) or fixed value */ u8 tc; bool tc_copy; bool sec_check; /* Inbound security check */ bool sec_check_frag; /* Inbound security check for (subsequent) fragments */ bool icmp6_enabled; /* Send destination unreachable for security check failure */ u16 tcp_mss; /* TCP MSS clamp value */ /* ICMPv6 -> ICMPv4 relay parameters */ ip4_address_t icmp4_src_address; vlib_simple_counter_main_t icmp_relayed; /* convenience */ vlib_main_t *vlib_main; vnet_main_t *vnet_main; /* * IPv4 encap and decap reassembly */ /* Configuration */ f32 ip4_reass_conf_ht_ratio; //Size of ht is 2^ceil(log2(ratio*pool_size)) u16 ip4_reass_conf_pool_size; //Max number of allocated reass structures u16 ip4_reass_conf_lifetime_ms; //Time a reassembly struct is considered valid in ms u32 ip4_reass_conf_buffers; //Maximum number of buffers used by ip4 reassembly /* Runtime */ map_ip4_reass_t *ip4_reass_pool; u8 ip4_reass_ht_log2len; //Hash table size is 2^log2len u16 ip4_reass_allocated; u16 *ip4_reass_hash_table; u16 ip4_reass_fifo_last; volatile u32 *ip4_reass_lock; /* Counters */ u32 ip4_reass_buffered_counter; bool frag_inner; /* Inner or outer fragmentation */ bool frag_ignore_df; /* Fragment (outer) packet even if DF is set */ /* * IPv6 decap reassembly */ /* Configuration */ f32 ip6_reass_conf_ht_ratio; //Size of ht is 2^ceil(log2(ratio*pool_size)) u16 ip6_reass_conf_pool_size; //Max number of allocated reass structures u16 ip6_reass_conf_lifetime_ms; //Time a reassembly struct is considered valid in ms u32 ip6_reass_conf_buffers; //Maximum number of buffers used by ip6 reassembly /* Runtime */ map_ip6_reass_t *ip6_reass_pool; u8 ip6_reass_ht_log2len; //Hash table size is 2^log2len u16 ip6_reass_allocated; u16 *ip6_reass_hash_table; u16 ip6_reass_fifo_last; volatile u32 *ip6_reass_lock; /* Counters */ u32 ip6_reass_buffered_counter; } map_main_t; /* * MAP Error counters/messages */ #define foreach_map_error \ /* Must be first. */ \ _(NONE, "valid MAP packets") \ _(BAD_PROTOCOL, "bad protocol") \ _(SEC_CHECK, "security check failed") \ _(ENCAP_SEC_CHECK, "encap security check failed") \ _(DECAP_SEC_CHECK, "decap security check failed") \ _(ICMP, "unable to translate ICMP") \ _(ICMP_RELAY, "unable to relay ICMP") \ _(UNKNOWN, "unknown") \ _(NO_BINDING, "no binding") \ _(NO_DOMAIN, "no domain") \ _(FRAGMENTED, "packet is a fragment") \ _(FRAGMENT_MEMORY, "could not cache fragment") \ _(FRAGMENT_MALFORMED, "fragment has unexpected format")\ _(FRAGMENT_DROPPED, "dropped cached fragment") \ _(MALFORMED, "malformed packet") \ _(DF_SET, "can't fragment, DF set") typedef enum { #define _(sym,str) MAP_ERROR_##sym, foreach_map_error #undef _ MAP_N_ERROR, } map_error_t; u64 map_error_counter_get(u32 node_index, map_error_t map_error); typedef struct { u32 map_domain_index; u16 port; } map_trace_t; extern map_main_t map_main; extern vlib_node_registration_t ip4_map_node; extern vlib_node_registration_t ip6_map_node; extern vlib_node_registration_t ip4_map_t_node; extern vlib_node_registration_t ip4_map_t_fragmented_node; extern vlib_node_registration_t ip4_map_t_tcp_udp_node; extern vlib_node_registration_t ip4_map_t_icmp_node; extern vlib_node_registration_t ip6_map_t_node; extern vlib_node_registration_t ip6_map_t_fragmented_node; extern vlib_node_registration_t ip6_map_t_tcp_udp_node; extern vlib_node_registration_t ip6_map_t_icmp_node; /* * map_get_pfx */ static_always_inline u64 map_get_pfx (map_domain_t *d, u32 addr, u16 port) { u16 psid = (port >> d->psid_shift) & d->psid_mask; if (d->ea_bits_len == 0 && d->rules) return clib_net_to_host_u64(d->rules[psid].as_u64[0]); u32 suffix = (addr >> d->suffix_shift) & d->suffix_mask; u64 ea = d->ea_bits_len == 0 ? 0 : (((u64) suffix << d->psid_length)) | psid; return clib_net_to_host_u64(d->ip6_prefix.as_u64[0]) | ea << d->ea_shift; } static_always_inline u64 map_get_pfx_net (map_domain_t *d, u32 addr, u16 port) { return clib_host_to_net_u64(map_get_pfx(d, clib_net_to_host_u32(addr), clib_net_to_host_u16(port))); } /* * map_get_sfx */ static_always_inline u64 map_get_sfx (map_domain_t *d, u32 addr, u16 port) { u16 psid = (port >> d->psid_shift) & d->psid_mask; /* Shared 1:1 mode. */ if (d->ea_bits_len == 0 && d->rules) return clib_net_to_host_u64(d->rules[psid].as_u64[1]); if (d->ip6_prefix_len == 128) return clib_net_to_host_u64(d->ip6_prefix.as_u64[1]); if (d->flags & MAP_DOMAIN_RFC6052) return (clib_net_to_host_u64(d->ip6_prefix.as_u64[1]) | addr); /* IPv4 prefix */ if (d->flags & MAP_DOMAIN_PREFIX) return (u64) (addr & (0xFFFFFFFF << d->suffix_shift)) << 16; /* Shared or full IPv4 address */ return ((u64) addr << 16) | psid; } static_always_inline u64 map_get_sfx_net (map_domain_t *d, u32 addr, u16 port) { return clib_host_to_net_u64(map_get_sfx(d, clib_net_to_host_u32(addr), clib_net_to_host_u16(port))); } static_always_inline u32 map_get_ip4 (ip6_address_t *addr, map_domain_flags_e flags) { if (flags & MAP_DOMAIN_RFC6052) return clib_host_to_net_u32(clib_net_to_host_u64(addr->as_u64[1])); else return clib_host_to_net_u32(clib_net_to_host_u64(addr->as_u64[1]) >> 16); } /* * Get the MAP domain from an IPv4 lookup adjacency. */ static_always_inline map_domain_t * ip4_map_get_domain (u32 mdi) { map_main_t *mm = &map_main; return pool_elt_at_index(mm->domains, mdi); } /* * Get the MAP domain from an IPv6 lookup adjacency. * If the IPv6 address or prefix is not shared, no lookup is required. * The IPv4 address is used otherwise. */ static_always_inline map_domain_t * ip6_map_get_domain (u32 mdi, ip4_address_t *addr, u32 *map_domain_index, u8 *error) { map_main_t *mm = &map_main; #ifdef TODO /* * Disable direct MAP domain lookup on decap, until the security check is updated to verify IPv4 SA. * (That's done implicitly when MAP domain is looked up in the IPv4 FIB) */ //#ifdef MAP_NONSHARED_DOMAIN_ENABLED //#error "How can you be sure this domain is not shared?" #endif *map_domain_index = mdi; return pool_elt_at_index(mm->domains, mdi); #ifdef TODO u32 lbi = ip4_fib_forwarding_lookup(0, addr); const dpo_id_t *dpo = load_balance_get_bucket(lbi, 0); if (PREDICT_TRUE(dpo->dpoi_type == map_dpo_type || dpo->dpoi_type == map_t_dpo_type)) { *map_domain_index = dpo->dpoi_index; return pool_elt_at_index(mm->domains, *map_domain_index); } *error = MAP_ERROR_NO_DOMAIN; return NULL; #endif } map_ip4_reass_t * map_ip4_reass_get(u32 src, u32 dst, u16 fragment_id, u8 protocol, u32 **pi_to_drop); void map_ip4_reass_free(map_ip4_reass_t *r, u32 **pi_to_drop); #define map_ip4_reass_lock() while (clib_atomic_test_and_set (map_main.ip4_reass_lock)) {} #define map_ip4_reass_unlock() do {CLIB_MEMORY_BARRIER(); *map_main.ip4_reass_lock = 0;} while(0) static_always_inline void map_ip4_reass_get_fragments(map_ip4_reass_t *r, u32 **pi) { int i; for (i=0; ifragments[i] != ~0) { vec_add1(*pi, r->fragments[i]); r->fragments[i] = ~0; map_main.ip4_reass_buffered_counter--; } } clib_error_t * map_plugin_api_hookup (vlib_main_t * vm); int map_ip4_reass_add_fragment(map_ip4_reass_t *r, u32 pi); map_ip6_reass_t * map_ip6_reass_get(ip6_address_t *src, ip6_address_t *dst, u32 fragment_id, u8 protocol, u32 **pi_to_drop); void map_ip6_reass_free(map_ip6_reass_t *r, u32 **pi_to_drop); #define map_ip6_reass_lock() while (clib_atomic_test_and_set (map_main.ip6_reass_lock)) {} #define map_ip6_reass_unlock() do {CLIB_MEMORY_BARRIER(); *map_main.ip6_reass_lock = 0;} while(0) int map_ip6_reass_add_fragment(map_ip6_reass_t *r, u32 pi, u16 data_offset, u16 next_data_offset, u8 *data_start, u16 data_len); void map_ip4_drop_pi(u32 pi); int map_ip4_reass_conf_ht_ratio(f32 ht_ratio, u32 *trashed_reass, u32 *dropped_packets); #define MAP_IP4_REASS_CONF_HT_RATIO_MAX 100 int map_ip4_reass_conf_pool_size(u16 pool_size, u32 *trashed_reass, u32 *dropped_packets); #define MAP_IP4_REASS_CONF_POOL_SIZE_MAX (0xfeff) int map_ip4_reass_conf_lifetime(u16 lifetime_ms); #define MAP_IP4_REASS_CONF_LIFETIME_MAX 0xffff int map_ip4_reass_conf_buffers(u32 buffers); #define MAP_IP4_REASS_CONF_BUFFERS_MAX (0xffffffff) void map_ip6_drop_pi(u32 pi); int map_ip6_reass_conf_ht_ratio(f32 ht_ratio, u32 *trashed_reass, u32 *dropped_packets); #define MAP_IP6_REASS_CONF_HT_RATIO_MAX 100 int map_ip6_reass_conf_pool_size(u16 pool_size, u32 *trashed_reass, u32 *dropped_packets); #define MAP_IP6_REASS_CONF_POOL_SIZE_MAX (0xfeff) int map_ip6_reass_conf_lifetime(u16 lifetime_ms); #define MAP_IP6_REASS_CONF_LIFETIME_MAX 0xffff int map_ip6_reass_conf_buffers(u32 buffers); #define MAP_IP6_REASS_CONF_BUFFERS_MAX (0xffffffff) static_always_inline void ip4_map_t_embedded_address (map_domain_t *d, ip6_address_t *ip6, const ip4_address_t *ip4) { ASSERT(d->ip6_src_len == 96 || d->ip6_src_len == 64); //No support for other lengths for now u8 offset = d->ip6_src_len == 64 ? 9 : 12; ip6->as_u64[0] = d->ip6_src.as_u64[0]; ip6->as_u64[1] = d->ip6_src.as_u64[1]; clib_memcpy_fast(&ip6->as_u8[offset], ip4, 4); } static_always_inline u32 ip6_map_t_embedded_address (map_domain_t *d, ip6_address_t *addr) { ASSERT(d->ip6_src_len == 64 || d->ip6_src_len == 96); u32 x; u8 offset = d->ip6_src_len == 64 ? 9 : 12; clib_memcpy(&x, &addr->as_u8[offset], 4); return x; } static inline void map_domain_counter_lock (map_main_t *mm) { if (mm->counter_lock) while (clib_atomic_test_and_set (mm->counter_lock)) /* zzzz */ ; } static inline void map_domain_counter_unlock (map_main_t *mm) { if (mm->counter_lock) clib_atomic_release (mm->counter_lock); } static_always_inline void map_send_all_to_node(vlib_main_t *vm, u32 *pi_vector, vlib_node_runtime_t *node, vlib_error_t *error, u32 next) { u32 n_left_from, *from, next_index, *to_next, n_left_to_next; //Deal with fragments that are ready from = pi_vector; n_left_from = vec_len(pi_vector); next_index = node->cached_next_index; while (n_left_from > 0) { vlib_get_next_frame(vm, node, next_index, to_next, n_left_to_next); while (n_left_from > 0 && n_left_to_next > 0) { u32 pi0 = to_next[0] = from[0]; from += 1; n_left_from -= 1; to_next += 1; n_left_to_next -= 1; vlib_buffer_t *p0 = vlib_get_buffer(vm, pi0); p0->error = *error; vlib_validate_buffer_enqueue_x1(vm, node, next_index, to_next, n_left_to_next, pi0, next); } vlib_put_next_frame(vm, node, next_index, n_left_to_next); } } /* * fd.io coding-style-patch-verification: ON * * Local Variables: * eval: (c-set-style "gnu") * End: */