6 #ifndef included_memcpy_h 7 #define included_memcpy_h 14 #if defined(CLIB_HAVE_VEC128) 15 u32x4_store_unaligned (u32x4_load_unaligned (src), dst);
23 #if defined(CLIB_HAVE_VEC256) 24 u32x8_store_unaligned (u32x8_load_unaligned (src), dst);
34 #if defined(CLIB_HAVE_VEC512) 35 u32x16_store_unaligned (u32x16_load_unaligned (src), dst);
45 #if defined(CLIB_HAVE_VEC128) 81 #if defined(CLIB_HAVE_VEC512) 94 #if defined(CLIB_HAVE_VEC256) 113 #if defined(CLIB_HAVE_VEC512_MASK_LOAD_STORE) 117 u32x16_mask_store (u32x16_mask_load_zero (src, mask), dst, mask);
130 #if defined(CLIB_HAVE_VEC256_MASK_LOAD_STORE) 134 u32x8_mask_store (u32x8_mask_load_zero (src, mask), dst, mask);
161 memcpy (
dst,
src, n_left *
sizeof (
u32));
#define COMPILE_TIME_CONST(x)
static_always_inline void clib_memcpy_u32_x8(u32 *dst, u32 *src)
#define static_always_inline
static uword pow2_mask(uword x)
static_always_inline void clib_memcpy_u32_x4(u32 *dst, u32 *src)
static_always_inline void clib_memcpy_u32(u32 *dst, u32 *src, u32 n_left)
static_always_inline void * clib_memcpy_fast(void *restrict dst, const void *restrict src, size_t n)
static_always_inline void clib_memcpy_u32_x16(u32 *dst, u32 *src)