summary |
shortlog |
log |
commit | commitdiff |
review |
tree
raw |
patch |
inline | side by side (from parent 1:
2222986)
Change-Id: Ic8b193e93ce18ca82b294816aa7ee0ef31d64bc2
Signed-off-by: Pierre Pfister <ppfister@cisco.com>
Signed-off-by: Hongjun Ni <hongjun.ni@intel.com>
lb/node.c \
lb/cli.c \
lb/util.c \
lb/node.c \
lb/cli.c \
lb/util.c \
lb/api.c
BUILT_SOURCES += \
lb/api.c
BUILT_SOURCES += \
noinst_HEADERS += \
lb/lb.h \
lb/util.h \
noinst_HEADERS += \
lb/lb.h \
lb/util.h \
lb/lbhash.h \
lb/lb.api.h
lb/lbhash.h \
lb/lb.api.h
#define LB_PLUGIN_LB_LB_H_
#include <lb/util.h>
#define LB_PLUGIN_LB_LB_H_
#include <lb/util.h>
-#include <lb/refcount.h>
+#include <vnet/util/refcount.h>
#include <vnet/vnet.h>
#include <vnet/ip/ip.h>
#include <vnet/vnet.h>
#include <vnet/ip/ip.h>
vnet/vnet.h \
vnet/vnet_all_api_h.h \
vnet/vnet_msg_enum.h \
vnet/vnet.h \
vnet/vnet_all_api_h.h \
vnet/vnet_msg_enum.h \
+ vnet/util/radix.h \
+ vnet/util/refcount.h
API_FILES += vnet/interface.api
API_FILES += vnet/interface.api
########################################
libvnet_la_SOURCES += \
########################################
libvnet_la_SOURCES += \
+ vnet/util/radix.c \
+ vnet/util/refcount.c \
vnet/util/trajectory.c
########################################
vnet/util/trajectory.c
########################################
* limitations under the License.
*/
* limitations under the License.
*/
-#include <lb/refcount.h>
+#include <vnet/util/refcount.h>
void __vlib_refcount_resize(vlib_refcount_per_cpu_t *per_cpu, u32 size)
{
u32 *new_counter = 0, *old_counter;
vec_validate(new_counter, size);
void __vlib_refcount_resize(vlib_refcount_per_cpu_t *per_cpu, u32 size)
{
u32 *new_counter = 0, *old_counter;
vec_validate(new_counter, size);
- memcpy(new_counter, per_cpu->counters, per_cpu->length);
+ vlib_refcount_lock(per_cpu->counter_lock);
+ memcpy(new_counter, per_cpu->counters, vec_len(per_cpu->counters)*4);
old_counter = per_cpu->counters;
per_cpu->counters = new_counter;
old_counter = per_cpu->counters;
per_cpu->counters = new_counter;
+ vlib_refcount_unlock(per_cpu->counter_lock);
- per_cpu->length = vec_len(new_counter);
vlib_thread_main_t *tm = vlib_get_thread_main ();
u32 thread_index;
for (thread_index = 0; thread_index < tm->n_vlib_mains; thread_index++) {
vlib_thread_main_t *tm = vlib_get_thread_main ();
u32 thread_index;
for (thread_index = 0; thread_index < tm->n_vlib_mains; thread_index++) {
- if (r->per_cpu[thread_index].length > index)
- count += r->per_cpu[thread_index].counters[index];
+ vlib_refcount_lock(r->per_cpu[thread_index].counter_lock);
+ if (index < vec_len(r->per_cpu[thread_index].counters))
+ {
+ count += r->per_cpu[thread_index].counters[index];
+ }
+ vlib_refcount_unlock(r->per_cpu[thread_index].counter_lock);
+/*
+ * Reference counting
+ * A specific reference counter is used. The design is quite
+ * similar to vlib counters but:
+ * - It is possible to decrease the value
+ * - Summing will not zero the per-thread counters
+ * - Only the thread can reallocate its own counters vector (to avoid concurrency issues)
+*/
typedef struct {
u32 *counters;
typedef struct {
u32 *counters;
- u32 length;
- u32 *reader_lengths;
+ volatile u32 *counter_lock;
CLIB_CACHE_LINE_ALIGN_MARK(o);
} vlib_refcount_per_cpu_t;
CLIB_CACHE_LINE_ALIGN_MARK(o);
} vlib_refcount_per_cpu_t;
vlib_refcount_per_cpu_t *per_cpu;
} vlib_refcount_t;
vlib_refcount_per_cpu_t *per_cpu;
} vlib_refcount_t;
+static_always_inline
+void vlib_refcount_lock (volatile u32 *counter_lock)
+{
+ while (__sync_lock_test_and_set (counter_lock, 1))
+ ;
+}
+
+static_always_inline
+void vlib_refcount_unlock (volatile u32 *counter_lock)
+{
+ *counter_lock = 0;
+}
+
void __vlib_refcount_resize(vlib_refcount_per_cpu_t *per_cpu, u32 size);
static_always_inline
void vlib_refcount_add(vlib_refcount_t *r, u32 thread_index, u32 counter_index, i32 v)
{
vlib_refcount_per_cpu_t *per_cpu = &r->per_cpu[thread_index];
void __vlib_refcount_resize(vlib_refcount_per_cpu_t *per_cpu, u32 size);
static_always_inline
void vlib_refcount_add(vlib_refcount_t *r, u32 thread_index, u32 counter_index, i32 v)
{
vlib_refcount_per_cpu_t *per_cpu = &r->per_cpu[thread_index];
- if (PREDICT_FALSE(counter_index >= per_cpu->length))
- __vlib_refcount_resize(per_cpu, clib_max(counter_index + 16, per_cpu->length * 2));
+ if (PREDICT_FALSE(counter_index >= vec_len(per_cpu->counters)))
+ __vlib_refcount_resize(per_cpu, clib_max(counter_index + 16,(vec_len(per_cpu->counters)) * 2));
per_cpu->counters[counter_index] += v;
}
per_cpu->counters[counter_index] += v;
}
void vlib_refcount_init(vlib_refcount_t *r)
{
vlib_thread_main_t *tm = vlib_get_thread_main ();
void vlib_refcount_init(vlib_refcount_t *r)
{
vlib_thread_main_t *tm = vlib_get_thread_main ();
r->per_cpu = 0;
vec_validate (r->per_cpu, tm->n_vlib_mains - 1);
r->per_cpu = 0;
vec_validate (r->per_cpu, tm->n_vlib_mains - 1);
+
+ for (thread_index = 0; thread_index < tm->n_vlib_mains; thread_index++)
+ {
+ r->per_cpu[thread_index].counter_lock =
+ clib_mem_alloc_aligned(CLIB_CACHE_LINE_BYTES,CLIB_CACHE_LINE_BYTES);
+ r->per_cpu[thread_index].counter_lock[0] = 0;
+ }