ioam_path_map_t *path;
u8 k, i;
- while (__sync_lock_test_and_set (data->writer_lock, 1))
+ while (clib_atomic_test_and_set (data->writer_lock))
;
trace_data = &data->trace_data;
for (k = 0; k < trace_record->num_nodes; k++)
path[k].state_up = 0;
}
- *(data->writer_lock) = 0;
+ clib_atomic_release (data->writer_lock);
}
always_inline void
u16 size_of_traceopt_per_node;
u16 size_of_all_traceopts;
- while (__sync_lock_test_and_set (data->writer_lock, 1))
+ while (clib_atomic_test_and_set (data->writer_lock))
;
trace_data = &data->trace_data;
}
}
end:
- *(data->writer_lock) = 0;
+ clib_atomic_release (data->writer_lock);
}
always_inline int
ioam_path_map_t *path = NULL;
ioam_analyse_trace_record *trace_record;
- while (__sync_lock_test_and_set (data->writer_lock, 1))
+ while (clib_atomic_test_and_set (data->writer_lock))
;
trace_data = &data->trace_data;
(u32) ((sum + delay) / (data->seqno_data.rx_packets + 1));
}
DONE:
- *(data->writer_lock) = 0;
+ clib_atomic_release (data->writer_lock);
return 0;
}
ip6_ioam_analyse_hbh_e2e (ioam_analyser_data_t * data,
ioam_e2e_packet_t * e2e, u16 len)
{
- while (__sync_lock_test_and_set (data->writer_lock, 1))
+ while (clib_atomic_test_and_set (data->writer_lock))
;
ioam_analyze_seqno (&data->seqno_data,
(u64) clib_net_to_host_u32 (e2e->e2e_data));
- *(data->writer_lock) = 0;
+ clib_atomic_release (data->writer_lock);
+
return 0;
}
data->writer_lock = clib_mem_alloc_aligned (CLIB_CACHE_LINE_BYTES,
CLIB_CACHE_LINE_BYTES);
- *(data->writer_lock) = 0;
+
+ clib_atomic_release (data->writer_lock);
trace_data = &(data->trace_data);
for (j = 0; j < IOAM_MAX_PATHS_PER_FLOW; j++)