map_ip4_reass_free(map_ip4_reass_t *r, u32 **pi_to_drop);
#define map_ip4_reass_lock() while (clib_atomic_test_and_set (map_main.ip4_reass_lock)) { CLIB_PAUSE (); }
-#define map_ip4_reass_unlock() do {CLIB_MEMORY_BARRIER(); *map_main.ip4_reass_lock = 0;} while(0)
+#define map_ip4_reass_unlock() clib_atomic_release (map_main.ip4_reass_lock)
static_always_inline void
map_ip4_reass_get_fragments(map_ip4_reass_t *r, u32 **pi)
map_ip6_reass_free(map_ip6_reass_t *r, u32 **pi_to_drop);
#define map_ip6_reass_lock() while (clib_atomic_test_and_set (map_main.ip6_reass_lock)) { CLIB_PAUSE (); }
-#define map_ip6_reass_unlock() do {CLIB_MEMORY_BARRIER(); *map_main.ip6_reass_lock = 0;} while(0)
+#define map_ip6_reass_unlock() clib_atomic_release (map_main.ip6_reass_lock)
int
map_ip6_reass_add_fragment(map_ip6_reass_t *r, u32 pi,
{
h->owner_pid = 0;
h->tag = 0;
- CLIB_MEMORY_BARRIER ();
- h->lock = 0;
+ clib_atomic_release (&h->lock);
}
}
ssvm_unlock_non_recursive (ssvm_shared_header_t * h)
{
h->tag = 0;
- CLIB_MEMORY_BARRIER ();
- h->lock = 0;
+ clib_atomic_release (&h->lock);
}
static inline void *
vnet_classify_entry_free (t, v, old_log2_pages);
unlock:
- CLIB_MEMORY_BARRIER ();
- t->writer_lock[0] = 0;
+ clib_atomic_release (&t->writer_lock[0]);
return rv;
}
{
if (dm->cache_lock)
{
- CLIB_MEMORY_BARRIER ();
- *dm->cache_lock = 0;
+ clib_atomic_release (dm->cache_lock);
}
}
{
if (PREDICT_FALSE (em->lock != 0))
{
- CLIB_MEMORY_BARRIER ();
- *em->lock = 0;
+ clib_atomic_release (em->lock);
}
}
clib_spinlock_unlock (clib_spinlock_t * p)
{
CLIB_LOCK_DBG_CLEAR (p);
- /* Make sure all writes are complete before releasing the lock */
- CLIB_MEMORY_BARRIER ();
- (*p)->lock = 0;
+ /* Make sure all reads/writes are complete before releasing the lock */
+ clib_atomic_release (&(*p)->lock);
}
static_always_inline void
while (clib_atomic_test_and_set (&(*p)->writer_lock))
CLIB_PAUSE ();
}
- CLIB_MEMORY_BARRIER ();
- (*p)->n_readers_lock = 0;
-
+ clib_atomic_release (&(*p)->n_readers_lock);
CLIB_LOCK_DBG (p);
}
(*p)->n_readers -= 1;
if ((*p)->n_readers == 0)
{
- CLIB_MEMORY_BARRIER ();
- (*p)->writer_lock = 0;
+ clib_atomic_release (&(*p)->writer_lock);
}
-
- CLIB_MEMORY_BARRIER ();
- (*p)->n_readers_lock = 0;
+ clib_atomic_release (&(*p)->n_readers_lock);
}
always_inline void
clib_rwlock_writer_unlock (clib_rwlock_t * p)
{
CLIB_LOCK_DBG_CLEAR (p);
- CLIB_MEMORY_BARRIER ();
- (*p)->writer_lock = 0;
+ clib_atomic_release (&(*p)->writer_lock);
}
#endif
if (--h->recursion_count == 0)
{
h->owner_cpu = ~0;
- CLIB_MEMORY_BARRIER ();
- h->lock = 0;
+ clib_atomic_release (&h->lock);
}
}
}