X-Git-Url: https://gerrit.fd.io/r/gitweb?a=blobdiff_plain;ds=sidebyside;f=vnet%2Fvnet%2Freplication.h;h=5dc554c97b34f47984f896b1e7e49891234fa1fb;hb=a290d7c4f50270387e90091524e307c4dea88bd5;hp=9de5717f4d34e96a6b139c8427238eb20a7ec9ce;hpb=cb9cadad578297ffd78fa8a33670bdf1ab669e7e;p=vpp.git diff --git a/vnet/vnet/replication.h b/vnet/vnet/replication.h index 9de5717f4d3..5dc554c97b3 100644 --- a/vnet/vnet/replication.h +++ b/vnet/vnet/replication.h @@ -24,43 +24,46 @@ #include -typedef struct { - - // The entire vnet buffer header restored for each replica - u8 vnet_buffer[32]; // 16B aligned to allow vector unit copy - u8 reserved[32]; // space for future expansion of vnet buffer header +typedef struct +{ + /* The entire vnet buffer header restored for each replica */ + u8 vnet_buffer[32]; /* 16B aligned to allow vector unit copy */ + u8 reserved[32]; /* space for future expansion of vnet buffer header */ - // feature state used during this replication - u64 feature_replicas; // feature's id for its set of replicas - u32 feature_counter; // feature's current index into set of replicas - u32 recycle_node_index; // feature's recycle node index + /* feature state used during this replication */ + u64 feature_replicas; /* feature's id for its set of replicas */ + u32 feature_counter; /* feature's current index into set of replicas */ + u32 recycle_node_index; /* feature's recycle node index */ - // data saved from the start of replication and restored at the end of replication - u32 saved_clone_count; // from vlib buffer - u32 saved_free_list_index; // from vlib buffer + /* + * data saved from the start of replication and restored + * at the end of replication + */ + u32 saved_free_list_index; /* from vlib buffer */ - // data saved from the original packet and restored for each replica - u64 l2_header[3]; // 24B (must be at least 22B for l2 packets) - u16 ip_tos; // v4 and v6 - u16 ip4_checksum; // needed for v4 only + /* data saved from the original packet and restored for each replica */ + u64 l2_header[3]; /* 24B (must be at least 22B for l2 packets) */ + u16 ip_tos; /* v4 and v6 */ + u16 ip4_checksum; /* needed for v4 only */ - // data saved from the vlib buffer header and restored for each replica - i16 current_data; // offset of first byte of packet in packet data - u8 pad[6]; // to 64B - u8 l2_packet; // flag for l2 vs l3 packet data + /* data saved from the vlib buffer header and restored for each replica */ + i16 current_data; /* offset of first byte of packet in packet data */ + u8 pad[6]; /* to 64B */ + u8 l2_packet; /* flag for l2 vs l3 packet data */ -} replication_context_t; // 128B +} replication_context_t; /* 128B */ -typedef struct { +typedef struct +{ u32 recycle_list_index; - // per-thread pools of replication contexts - replication_context_t ** contexts; + /* per-thread pools of replication contexts */ + replication_context_t **contexts; - vlib_main_t * vlib_main; - vnet_main_t * vnet_main; + vlib_main_t *vlib_main; + vnet_main_t *vnet_main; } replication_main_t; @@ -68,56 +71,66 @@ typedef struct { extern replication_main_t replication_main; -// Return 1 if this buffer just came from the replication recycle handler. +/* Return 1 if this buffer just came from the replication recycle handler. */ always_inline u32 replication_is_recycled (vlib_buffer_t * b0) { return b0->flags & VLIB_BUFFER_IS_RECYCLED; } -// Clear the recycle flag. If buffer came from the replication recycle -// handler, this flag must be cleared before the packet is transmitted again. +/* + * Clear the recycle flag. If buffer came from the replication recycle + * handler, this flag must be cleared before the packet is transmitted again. + */ always_inline void replication_clear_recycled (vlib_buffer_t * b0) { b0->flags &= ~VLIB_BUFFER_IS_RECYCLED; } -// Return the active replication context if this buffer has -// been recycled, otherwise return 0. (Note that this essentially -// restricts access to the replication context to the replication -// feature's prep and recycle nodes.) +/* + * Return the active replication context if this buffer has + * been recycled, otherwise return 0. (Note that this essentially + * restricts access to the replication context to the replication + * feature's prep and recycle nodes.) + */ always_inline replication_context_t * replication_get_ctx (vlib_buffer_t * b0) { - replication_main_t * rm = &replication_main; + replication_main_t *rm = &replication_main; - return replication_is_recycled (b0) ? - pool_elt_at_index (rm->contexts[os_get_cpu_number()], b0->clone_count) : - 0; + return replication_is_recycled (b0) ? + pool_elt_at_index (rm->contexts[os_get_cpu_number ()], + b0->recycle_count) : 0; } -// Prefetch the replication context for this buffer, if it exists +/* Prefetch the replication context for this buffer, if it exists */ always_inline void replication_prefetch_ctx (vlib_buffer_t * b0) { replication_context_t *ctx = replication_get_ctx (b0); - if (ctx) { - CLIB_PREFETCH (ctx, (2*CLIB_CACHE_LINE_BYTES), STORE); - } + if (ctx) + { + CLIB_PREFETCH (ctx, (2 * CLIB_CACHE_LINE_BYTES), STORE); + } } -replication_context_t * -replication_prep (vlib_main_t * vm, - vlib_buffer_t * b0, - u32 recycle_node_index, - u32 l2_packet); +replication_context_t *replication_prep (vlib_main_t * vm, + vlib_buffer_t * b0, + u32 recycle_node_index, + u32 l2_packet); -replication_context_t * -replication_recycle (vlib_main_t * vm, - vlib_buffer_t * b0, - u32 is_last); +replication_context_t *replication_recycle (vlib_main_t * vm, + vlib_buffer_t * b0, u32 is_last); #endif + +/* + * fd.io coding-style-patch-verification: ON + * + * Local Variables: + * eval: (c-set-style "gnu") + * End: + */